计算机科学
粒度
滤波器(信号处理)
人工智能
噪音(视频)
特征(语言学)
情绪分析
传感器融合
数据挖掘
模式
特征提取
任务(项目管理)
模式识别(心理学)
机器学习
计算机视觉
图像(数学)
社会科学
语言学
哲学
管理
社会学
经济
操作系统
作者
Xiaojun Xue,Chunxia Zhang,Zhendong Niu,Xindong Wu
出处
期刊:IEEE Transactions on Knowledge and Data Engineering
[Institute of Electrical and Electronics Engineers]
日期:2022-01-01
卷期号:: 1-1
被引量:32
标识
DOI:10.1109/tkde.2022.3155290
摘要
Multimodal sentiment analysis (MSA) is a very challenging task due to its complex and complementary interactions between multiple modalities, which can be widely applied into areas of product marketing, public opinion monitoring, and so on. However, previous works directly utilized the features extracted from multimodal data, in which the noise reduction within and among multiple modalities has been largely ignored before multimodal fusion. This paper proposes a multi-level attention map network (MAMN) to filter noise before multimodal fusion and capture the consistent and heterogeneous correlations among multi-granularity features for multimodal sentiment analysis. Architecturally, MAMN is comprised of three modules: multi-granularity feature extraction module, multi-level attention map generation module, and attention map fusion module. The first module is designed to sufficiently extract multi-granularity features from multimodal data. The second module is constructed to filter noise and enhance the representation ability for multi-granularity features before multimodal fusion. And the third module is built to extensibly mine the interactions among multi-level attention maps by the proposed extensible co-attention fusion method. Extensive experimental results on three public datasets show the proposed model is significantly superior to the state-of-the-art methods, and demonstrate its effectiveness on two tasks of document-based and aspect-based MSA tasks.
科研通智能强力驱动
Strongly Powered by AbleSci AI