计算机科学
模式
人工智能
特征(语言学)
卷积神经网络
情绪分析
典型相关
模式识别(心理学)
模态(人机交互)
代表(政治)
机器学习
社会科学
语言学
哲学
社会学
政治
政治学
法学
作者
Hongju Cheng,Zizhen Yang,Xiaoqi Zhang,Yang Yang
出处
期刊:IEEE Transactions on Affective Computing
[Institute of Electrical and Electronics Engineers]
日期:2023-04-07
卷期号:14 (4): 3149-3163
被引量:5
标识
DOI:10.1109/taffc.2023.3265653
摘要
Multimodal sentiment analysis aims to extract and integrate information from different modalities to accurately identify the sentiment expressed in multimodal data. How to effectively capture the relevant information within a specific modality and how to fully exploit the complementary information among multiple modalities are two major challenges in multimodal sentiment analysis. Traditional approaches fail to obtain the global contextual information of long time-series data when extracting unimodal temporal features, and they usually fuse the features from multiple modalities with the same method and ignore the correlation between different modalities when modeling inter-modal interactions. In this paper, we first propose an Attentional Temporal Convolutional Network (ATCN) to extract unimodal temporal features for enhancing the feature representation ability, then introduce a Multi-layer Feature Fusion (MFF) model to improve the effectiveness of multimodal fusion, which fuses the different-level features by different methods according to the correlation coefficient between the features, and cross-modal multi-head attention is used to fully explore the potential relationship between the low-level features. The experimental results on SIMS and CMU-MOSI datasets show that the proposed model achieves superior performance on sentiment analysis tasks compared to state-of-the-art baselines.
科研通智能强力驱动
Strongly Powered by AbleSci AI