计算机科学
情绪分析
模态(人机交互)
领域(数学)
构造(python库)
人工智能
模式
融合
特征(语言学)
编码(集合论)
自然语言处理
语言学
社会科学
哲学
数学
集合(抽象数据类型)
社会学
纯数学
程序设计语言
作者
Chuanbo Zhu,Min Chen,Sheng Zhang,Chao Sun,Han Liang,Yifan Liu,Jincai Chen
标识
DOI:10.1016/j.inffus.2023.101958
摘要
Multimodal sentiment analysis is an active research field that aims to recognize the user's sentiment information from multimodal data. The primary challenge in this field is to develop a high-quality fusion framework that effectively addresses the heterogeneity among different modalities. However, prior research has primarily concentrated on intermodal interactions while neglecting the semantic sentiment information conveyed by words in the text modality. In this paper, we propose the Sentiment Knowledge Enhanced Attention Fusion Network (SKEAFN), a novel end-to-end fusion network that enhances multimodal fusion by incorporating additional sentiment knowledge representations from an external knowledge base. Firstly, we construct an external knowledge enhancement module to acquire additional representations for the text modality. Then, we design a text-guided interaction module that facilitates the interaction between text and the visual/acoustic modality. Finally, we propose a feature-wised attention fusion module that achieves multimodal fusion by dynamically adjusting the weights of the additional and each modality's representations. We evaluate our method on three challenging multimodal sentiment analysis datasets: CMU-MOSI, CMU-MOSEI, and Twitter2019. The experiment results demonstrate that our model significantly outperforms the state-of-the-art models. The source code is publicly available at https://github.com/doubibobo/SKEAFN.
科研通智能强力驱动
Strongly Powered by AbleSci AI