计算机科学
人工智能
分割
RGB颜色模型
背景(考古学)
融合机制
计算机视觉
模态(人机交互)
模式识别(心理学)
特征(语言学)
语义学(计算机科学)
特征提取
图像分割
融合
地理
语言学
哲学
考古
脂质双层融合
程序设计语言
作者
Xunjie He,Meiling Wang,Tong Liu,Lin Zhao,Yufeng Yue
出处
期刊:IEEE Transactions on Instrumentation and Measurement
[Institute of Electrical and Electronics Engineers]
日期:2023-01-01
卷期号:72: 1-10
被引量:11
标识
DOI:10.1109/tim.2023.3267529
摘要
The fusion of RGB and thermal images has profound implications for the semantic segmentation of challenging urban scenes, such as those with poor illumination. Nevertheless, existing RGB-T fusion networks pay less attention to modality differences; i.e., RGB and thermal images are commonly fused with fixed weights. In addition, spatial context details are lost during regular extraction operations, inevitably leading to imprecise object segmentation. To improve the segmentation accuracy, a novel network named spatial feature aggregation and fusion with modality adaptation (SFAF-MA) is proposed in this paper. The modality difference adaptive fusion (MDAF) module is introduced to adaptively fuse RGB and thermal images with corresponding weights generated from an attention mechanism. In addition, the spatial semantic fusion (SSF) module is designed to tap into more information by capturing multiscale perceptive fields with dilated convolutions of different rates, and aggregate shallower-level features with rich visual information and deeper-level features with strong semantics. Compared with existing methods on the public MFNet dataset and PST900 dataset, the proposed network significantly improves the segmentation effectiveness. The code is available at https://github.com/hexunjie/SFAF-MA.
科研通智能强力驱动
Strongly Powered by AbleSci AI