计算机科学
人工智能
互补性(分子生物学)
情态动词
RGB颜色模型
分割
融合机制
融合
特征(语言学)
模式识别(心理学)
冗余(工程)
加权
编码器
保险丝(电气)
计算机视觉
工程类
遗传学
哲学
高分子化学
生物
医学
化学
放射科
脂质双层融合
操作系统
语言学
电气工程
作者
Wei Wu,Tao Chu,Qiong Liu
标识
DOI:10.1016/j.patcog.2022.108881
摘要
RGB-T semantic segmentation has attracted growing attention because it makes a model robust towards challenging illumination. Most existing methods fuse RGB and thermal information in an equal manner along spatial dimensions, which results in feature redundancy and affects the discriminability of cross-modal features. In this paper, we propose a Complementarity-aware Cross-modal Feature Fusion Network (CCFFNet) including a Complementarity-Aware Encoder (CAE) and a Three-Path Fusion and Supervision (TPFS). The CAE, which consists of cascaded cross-modal fusion modules, can select complementary information from RGB and thermal features via a novel gate and fuse them by a channel-wise weighting mechanism. TPFS not only iteratively performs Three-Path Fusion (TPF) to further enhance cross-modal features, but also supervise the training of CCFFNet along three branches by Three-Supervision (TS). Extensive experiments are carried out and the results demonstrate that our model outperforms the state-of-the-art models by at least 1.6% mIoU on MFNet dataset and 2.9% mIoU on PST900 dataset, respectively. And a single-modality-based model can be easily applied to multi-modal semantic segmentation when plugging our CAE.
科研通智能强力驱动
Strongly Powered by AbleSci AI