计算机科学
情绪分析
自然语言处理
人工智能
人机交互
多媒体
语音识别
作者
Zemin Tang,Qi Xiao,Yunchuan Qin,Xu Zhou,Joey Tianyi Zhou,Kenli Li
出处
期刊:IEEE Transactions on Consumer Electronics
[Institute of Electrical and Electronics Engineers]
日期:2024-01-23
卷期号:70 (1): 4095-4107
标识
DOI:10.1109/tce.2024.3357480
摘要
Multimodal Sentiment Analysis (MSA) technology, prevalent in consumer applications and mobile edge computing (MEC), enables sentiment examination through user data collected by smart devices. Despite the focus on representation learning in MSA, current methods often prioritize recognition performance through modality interaction and fusion. However, they struggle to capture multi-view sentiment cues across different interaction states, limiting multimodal sentiment representations' expressiveness. This paper develops an innovative MSA framework, MVIR, learning multi-view interactive representations in diverse interaction states. Multilple meticulously designed sentiment tasks and an introduced self-supervised label generation algorithm (SSLGM) enable a comprehensive understanding of multi-view sentiment tendencies. The dual-view attention weighted fusion (DVAWF) module is designed to facilitate inter-modality information exchange in different interaction states. Extensive experiments on three MSA datasets affirm the efficacy and superiority of MVIR, showcasing its ability to capture sentiment information from multimodal data across various interaction states.
科研通智能强力驱动
Strongly Powered by AbleSci AI