The Research and Preprocessing Results of The End-to-End Multimodal Sentiment Analysis System
DOI:
https://doi.org/10.54097/ecjqy928Keywords:
Summary of Preprocessing, End-to-End Systems, Practical Applications.Abstract
With the rapid development of artificial intelligence technology, the application of sentiment analysis in areas such as social media monitoring, customer service, and intelligent assistants is gradually increasing. Traditional sentiment analysis usually relies on unimodal analysis, which has certain limitations. Therefore, end-to-end multimodal sentiment analysis systems have emerged. This paper combines several typical multimodal sentiment analysis systems, with the Tension-Tension Experimental Engineering (TTEE) model for efficient computation and reasoning, and incorporates weighted vector fusion and cross-modal attention mechanisms to improve the accuracy of sentiment analysis. The system uses Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) to extract high-level features from text, speech, and images, enabling deep fusion of multimodal data. Additionally, the system adopts the Multi-Task Multi-View Network (MTMVN) architecture for multi-task and multi-perspective learning, thereby enhancing the robustness and accuracy of the sentiment analysis model. Compared to traditional unimodal analysis methods, multimodal sentiment analysis systems demonstrate higher accuracy and stability across multiple sentiment analysis tasks.
Downloads
References
[1] Yang L, Na J C, Yu J F. Cross-Modal Multitask Transformer for End-to-End Multimodal Aspect-Based Sentiment Analysis. Information Processing & Management. 2022, Volume 59, Issue 5, September 2022, 103038.
[2] Li Z L, Song Y Q, Lu X L, Liu M. Twin Towers End to End model for aspect-based sentiment analysis. Expert Systems with Applications. 2024, Volume 249, Part C, 1 September 2024, 123713.
[3] Tzirakis P, Chen J X, Zafeiriou S, Schuller B. End-to-end multimodal affect recognition in real-world environments. Information Fusion. 2024, Volume 249, Part C, 1 September 2024, 123713
[4] Bie Y, Yang Y. A Multitask Multiview Neural Network for End-to-End Aspect-Based Sentiment Analysis. Big Data Mining and Analytics. 2021, ISSN 2096-0654 05/06 pp195–207, Volume 4, Number 3, September 2021.
[5] Wei Q, Huang X, Zhang Y. FV2ES: A fully end2end multimodal system for fast yet effective video emotion recognition inference. IEEE Transactions on Broadcasting, 2023, 69(1): 10-20.
[6] Yin M, Qiao S, Chen W, et al. Social network rumor detection method based on group emotion stabilization. Journal of Software, 2020, 1-24.
[7] Wang D, He Y, Liang X, et al. Tmfn: A target-oriented multi-grained fusion network for end-to-end aspect-based multimodal sentiment analysis//Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024). 2024, 16187-16197.
[8] Wei H Z. Technical research on real-time multimodal interaction architecture based on AI big language model for dynamic adjustment of semantic features in intelligent customer service robots. COMBIN. MATH. COMBIN. COMPUT. 2025, 127a (2025) 3867--3886
[9] Chen C, Zhang J. Public Attitudes towards Autonomous Driving: Sentiment and Topic Analysis Based on Weibo Data. The Chinese Library Classification (CLC). 2023.
[10] Tang Q W, Zhang H, Wu Y A. Construction of a Multimodal Learning Analysis Model for Smart Classrooms. Journal of Teacher Education. 2024.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Highlights in Science, Engineering and Technology

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.







