模式
模态(人机交互)
人工智能
情绪识别
水准点(测量)
不变(物理)
模式识别(心理学)
计算机科学
特征学习
情态动词
代表(政治)
特征(语言学)
深度学习
语音识别
自然语言处理
机器学习
数学
语言学
政治学
高分子化学
法学
地理
政治
大地测量学
化学
数学物理
社会学
哲学
社会科学
作者
Rui Liu,Haolin Zuo,Zheng Lian,Björn W. Schuller,Haizhou Li
标识
DOI:10.1109/taffc.2024.3378570
摘要
Multimodal emotion recognition (MER) aims to understand the way that humans express their emotions by exploring complementary information across modalities. However, it is hard to guarantee that full-modality data is always available in real-world scenarios. To deal with missing modalities, researchers focused on meaningful joint multimodal representation learning during cross-modal missing modality imagination. However, the cross-modal imagination mechanism is highly susceptible to errors due to the "modality gap" issue, which affects the imagination accuracy, thus, the final recognition performance. To this end, we introduce the concept of a modality-invariant feature into the missing modality imagination network, which contains two key modules: 1) a novel contrastive learning-based module to extract modality-invariant features under full modalities; 2) a robust imagination module based on imagined invariant features to reconstruct missing information under missing conditions. Finally, we incorporate imagined and available modalities for emotion recognition. Experimental results on benchmark datasets demonstrate that our proposed method outperforms existing state-of-the-art strategies. Compared with our previous work, our extended version is more effective on multimodal emotion recognition with missing modalities. The code is released at https://github.com/ZhuoYulang/CIF-MMIN .
科研通智能强力驱动
Strongly Powered by AbleSci AI