计算机科学
背景(考古学)
代表(政治)
情态动词
人工智能
情绪分析
对偶(语法数字)
降噪
模式识别(心理学)
语言学
古生物学
化学
哲学
政治
政治学
高分子化学
法学
生物
作者
Zuhe Li,Zhenwei Huang,Yushan Pan,Jun Yu,Weihua Liu,Haoran Chen,Yiming Luo,Di Wu,Hao Wang
标识
DOI:10.1016/j.eswa.2024.124236
摘要
Multimodal sentiment analysis aims to extract sentiment cues from various modalities, such as textual, acoustic, and visual data, and manipulate them to determine the inherent sentiment polarity in the data. Despite significant achievements in multimodal sentiment analysis, challenges persist in addressing noise features in modal representations, eliminating substantial gaps in sentiment information among modal representations, and exploring contextual information that expresses different sentiments between modalities. To tackle these challenges, our paper proposes a new Multimodal Sentiment Analysis (MSA) framework. Firstly, we introduce the Hierarchical Denoising Representation Disentanglement module (HDRD), which employs hierarchical disentanglement techniques. This ensures the extraction of both common and private sentiment information while eliminating interference noise from modal representations. Furthermore, to address the uneven distribution of sentiment information among modalities, our Inter-Modal Representation Enhancement module (IMRE) enhances non-textual representations by extracting sentiment information related to non-textual representations from textual representations. Next, we introduce a new interaction mechanism, the Dual-Channel Cross-Modal Context Interaction module (DCCMCI). This module not only mines correlated contextual sentiment information within modalities but also explores positive and negative correlation contextual sentiment information between modalities. We conducted extensive experiments on two benchmark datasets, MOSI and MOSEI, and the results indicate that our proposed method offers state-of-the-art approaches.
科研通智能强力驱动
Strongly Powered by AbleSci AI