计算机科学
模式
情绪分析
杠杆(统计)
人工智能
相关性
典型相关
多模式学习
利用
机器学习
任务(项目管理)
自然语言处理
社会科学
几何学
数学
计算机安全
管理
社会学
经济
作者
Zuhe Li,Qingbing Guo,Yushan Pan,Weiping Ding,Jun Yu,Yazhou Zhang,Weihua Liu,Haoran Chen,Hao Wang,Ying Xie
标识
DOI:10.1016/j.inffus.2023.101891
摘要
Fusion and co-learning are major challenges in multimodal sentiment analysis. Most existing methods either ignore the basic relationships among modalities or fail to maximize their potential correlations. They also do not leverage the knowledge from resource-rich modalities in the analysis of resource-poor modalities. To address these challenges, we propose a multimodal sentiment analysis method based on multilevel correlation mining and self-supervised multi-task learning. First, we propose a unimodal feature fusion- and linguistics-guided Transformer-based framework, multi-level correlation mining framework, to overcome the difficulty of multimodal information fusion. The module exploits the correlation information between modalities from low to high levels. Second, we divided the multimodal sentiment analysis task into one multimodal task and three unimodal tasks (linguistic, acoustic, and visual tasks), and designed a self-supervised label generation module (SLGM) to generate sentiment labels for unimodal tasks. SLGM-based multi-task learning overcomes the lack of unimodal labels in co-learning. Through extensive experiments on the CMU-MOSI and CMU-MOSEI datasets, we demonstrated the superiority of the proposed multi-level correlation mining framework to state-of-the-art methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI