计算机科学
模式
模态(人机交互)
人工智能
缺少数据
稳健性(进化)
机器学习
RGB颜色模型
模式识别(心理学)
面部识别系统
代表(政治)
面子(社会学概念)
情态动词
政治
基因
生物化学
社会学
社会科学
化学
高分子化学
法学
政治学
作者
Yizhe Zhu,Sun Xin,Xi Zhou
标识
DOI:10.1007/978-3-031-44210-0_23
摘要
Current RGB-D-T face recognition methods are able to alleviate the sensitivity to facial variations, posture, occlusions, and illumination by incorporating complementary information, while they rely heavily on the availability of complete modalities. Given the likelihood of missing modalities in real-world scenarios and the fact that current multi-modal recognition models perform poorly when faced with incomplete data, robust multi-modal models for face recognition that can handle missing modalities are highly desirable. To this end, we propose a multi-modal fusion framework for robustly learning face representations in the presence of missing modalities, using a combination of RGB, depth, and thermal modalities. Our approach effectively blends these modalities together while also alleviating the semantic gap among them. Specifically, we put forward a novel modality-missing loss function to learn the modality-specific features that are robust to missing-modality data conditions. To project various modality features to the same semantic space, we exploit a joint modality-invariant representation with a central moment discrepancy (CMD) based distance constraint training strategy. We conduct extensive experiments on several benchmark datasets, such as VAP RGBD-T and Lock3DFace, and the results demonstrate the effectiveness and robustness of the proposed approach under uncertain missing-modality conditions compared with all baseline algorithms.
科研通智能强力驱动
Strongly Powered by AbleSci AI