计算机科学
模式
判别式
人工智能
推论
稳健性(进化)
杠杆(统计)
情态动词
利用
约束(计算机辅助设计)
机器学习
共享空间
空格(标点符号)
机械工程
化学
高分子化学
工程类
社会科学
生物化学
计算机安全
社会学
基因
操作系统
作者
Chiqin Li,Liang‐Liang Xie,Xingmao Shao,Hang Pan,Zhiliang Wang
标识
DOI:10.1016/j.engappai.2024.108413
摘要
Continuous emotion recognition has been a compelling topic in affective computing because it can interpret human emotions subtly and continuously. Existing studies have achieved advanced emotion recognition performance using multimodal knowledge. However, these studies generally ignore the circumstances where some particular modalities are missing in the inference phase and thus become sensitive to the absence of modalities. To resolve this issue, we propose a novel multimodal shared network with a cross-modal distribution constraint, i.e. the DS-Net, which aims to improve the robustness of the model to missing modalities. The training process of the proposed network generally includes two components: multimodal shared space modeling and a cross-modal distribution matching constraint. The former utilizes the local and temporal information of multimodal signals for multimodal shared space modeling, while the latter further enhances the multimodal shared space via a loose constraint method. Coupled with the latter, the former can effectively exploit the complementarity between videos and peripheral physiological signals (PPSs), thus enhancing the discriminative capability of the shared space. Based on the shared space, the DS-Net works during the inference phase with only one modality input and can leverage multimodal knowledge to improve emotion recognition accuracy. Comprehensive experiments were conducted on two public datasets. Results demonstrate that the proposed method is competitive or superior to the current state-of-the-art methods. Further experiments indicate that the proposed method can be extended to handle other modalities and to deal with partially missing modalities, demonstrating its potential in real-world applications.
科研通智能强力驱动
Strongly Powered by AbleSci AI