模式
计算机科学
加权
人工智能
模态(人机交互)
多模式学习
水准点(测量)
模式治疗法
情绪分析
机器学习
自然语言处理
医学
社会科学
外科
大地测量学
社会学
放射科
地理
作者
Yifeng Wang,Jiahao He,Di Wang,Quan Wang,Bo Wan,Xuemei Luo
出处
期刊:Neurocomputing
[Elsevier]
日期:2024-03-01
卷期号:572: 127181-127181
被引量:1
标识
DOI:10.1016/j.neucom.2023.127181
摘要
Multimodal Sentiment Analysis (MSA) constitutes a pivotal technology in the realm of multimedia research. The efficacy of MSA models largely hinges on the quality of multimodal fusion. Notably, when conveying information pertinent to specific tasks or applications, not all modalities hold equal importance. Previous research, however, has either disregarded the importance of modalities altogether or solely focused on the importance of linguistic and non-linguistic modalities while neglecting the importance between non-linguistic modalities. To facilitate effective multimodal information fusion based on the relative importance of modalities, a novel multimodal fusion mode named Multimodal Transformer with Adaptive Modality Weighting (MTAMW) is proposed in this paper. Specifically, we introduce a multimodal adaptive weight matrix that allocates appropriate weights to each modality based on its contribution to sentiment analysis. Furthermore, a multimodal attention mechanism is introduced, utilizing multiple Softmax functions to compute attention weights, thereby efficiently fusion multimodal information via a single-stream Transformer. By meticulously considering the relative importance of each modality during the fusion process, more effective multimodal information fusion is achievable. Extensive experiments on benchmark datasets show that it is superior to or comparable to state-of-the-art methods on MSA tasks. The codes for our experiments are available at https://github.com/Vamos66/MTAMW.
科研通智能强力驱动
Strongly Powered by AbleSci AI