计算机科学
情态动词
情绪分析
模式
水准点(测量)
人工智能
特征(语言学)
自然语言处理
机器学习
光学(聚焦)
语音识别
语言学
社会科学
化学
哲学
物理
大地测量学
社会学
高分子化学
光学
地理
作者
Yanping Fu,Zhiyuan Zhang,Ruidi Yang,Cuiyou Yao
出处
期刊:Neurocomputing
[Elsevier BV]
日期:2023-12-28
卷期号:571: 127201-127201
被引量:12
标识
DOI:10.1016/j.neucom.2023.127201
摘要
Multimodal sentiment analysis (MSA) predicts the sentiment polarity of an unlabeled utterance that carries multiple modalities, such as text, vision and audio, by analyzing labeled utterances. Existing fusion methods mainly focus on establishing the relationship of characteristics among different modalities to enhance their emotion recognition abilities. However, they always ignore the all-round interaction between different modalities, especially the cross-modal interaction, which is critical to the sentiment decision of multimodal data. To address these issues, we propose a novel hybrid cross-modal interaction learning (HCIL) framework for hybrid learning of intra-modal, inter-modal, interactive-modal and cross-modal interactions, with which the model can fully utilize the sentiment information of multimodalities and enhance the sentiment assistance between modalities. Specifically, we propose two core substructures to learn discriminative multimodal features. One is the comparative learning interaction structure that can track the class dynamics in the intra-modal, reduce the modal gap in the inter-modal and establish emotional communication in the interactive-modal; the other is the cross-modal prediction structure, which can build the sentiment relationship between cross-modal pairs, especially exploring the auxiliary sentiment effect of audio on the vision and text. Furthermore, we adopt a hierarchical feature fusion structure to generate the multimodal feature for the final sentiment prediction. Extensive experiments on three benchmark datasets showed that our HCIL approach can obtain significant performance on the MSA task and that the design of a cross-modal interaction structure can directly promote the improvement of sentiment classification performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI