Multimodal Sentiment Analysis using Multiple Neural Networks and Natural Language Processing Models
Keywords:
audio signals, BERT model, CMU-MOSI dataset, facial expression, Haar Cascade Algorithm, MobileNet model, mouth state detection, sentiment analysis.Abstract
Multimodal Sentiment Analysis (MSA) identifies emotional expressions over time using visual and audio information. MSA, an emerging field, detects and analyzes emotions across text, speech, and images, with practical applications in social media analysis, customer feedback analysis, healthcare monitoring, and investigations. In the realm of e-commerce, product reviews significantly influence consumer purchasing behaviours, with user-generated content offering diverse perspectives on products and services. Sentiment analysis systematically deciphers the emotional undertones within text, providing businesses with actionable insights. Videos and audio recordings introduce richer dimensions to these reviews, capturing nuanced expressions and vocal cues. Integrating both modalities into sentiment analysis offers a deeper comprehension of reviewers' emotional states, enhancing the accuracy of product perception and satisfaction assessments.
Our innovative approach merges the Haar cascade algorithm for facial detection and sentiment analysis in videos with the BERT model for audio sentiment analysis, synergizing visual and auditory cues. This integration improves accuracy and robustness, facilitating precise assessments of product sentiment. By fusing outputs from these modalities, we can identify the most salient emotional cues expressed by reviewers, providing businesses with a comprehensive understanding of consumer sentiment and aiding informed decision-making processes.
This paper presents a novel approach to sentiment analysis using facial expression and audio data from the CMU-MOSI dataset. The process involves extracting a focused video of the subject's face, converting the corresponding audio into text, and analyzing sentiment. The proposed methods achieved promising accuracy and hold immense potential when combined with other modalities.
Downloads
References
Qiupu Chen, Guimin Huang and Yabing Wang, "The Weighted Cross-Modal Attention Mechanism With Sentiment Prediction Auxiliary Task for Multimodal Sentiment Analysis" Published in: IEEE/ACM Transactions on Audio, Speech, and Language Processing. Date of Publication: 20 July 2022. DOI: 10.1109/TASLP.2022.3192728. Publisher: IEEE.
Sarah A. Abdu, Ahmed H. Yousef, Ashraf Salem, Multimodal Video Sentiment Analysis Using Deep Learning Approaches, a Survey, Information Fusion, Volume 76, 2021, Pages 204-226, ISSN 1566-2535, https://doi.org/10.1016/j.inffus.2021.06.003.
Yakaew, Atitaya & Dailey, Matthew & Racharak, Teeradaj. (2021). Multimodal Sentiment Analysis on Video Streams using Lightweight Deep Neural Networks. 442-451.
10.5220/0010304404420451.
Chi Sun, Xipeng Qiu, Yige Xu & Xuanjing Huang. (2019). How to Fine-Tune BERT for Text Classification? LNAI,volume 11856
https://doi.org/10.1007/978-3-030-32381-3_16
Google. (2023). Google Cloud Speech-to-Text API [Computer software]. Retrieved from https://cloud.google.com/speech-to-text.
Kaldi website, URL: https://kaldi-asr.org/. Accessed on [14th April 2023].
Gina Khayatun Nufus, Mustafid Mustafid, and Rahmat "Sentiment Analysis for Video on Demand Application User Satisfaction with Long Short Term Memory Model" E3S Web Conference Volume 317,2021. The 6th International Conference on Energy, Environment, Epidemiology, and Information System (ICENIS 2021)
Gaurav Meena, Krishna Kumar Mohbey, Sunil Kumar "Sentiment analysis on images using convolutional neural networks based Inception-V3 transfer learning approach"
Nur Hasifah A Razak, Muhammad Firdaus Mustapha. "Amazon Product Sentiment Analysis using RapidMiner"
Dorca Manuel-Ilie, Pitic Antoniu Gabriel, Crețulescu Radu George "Sentiment Analysis Using Bert Model" https://www.researchgate.net/publication/376670839_Sentiment_Analysis_Using_Bert_Model
Amira Samy Talaat. "Sentiment analysis classification system using hybrid BERT models".
https://journalofbigdata.springeropen.com/articles/10.1186/s40537-023-00781-w
Downloads
Published
How to Cite
Issue
Section
License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.