多光谱图像
计算机科学
人工智能
特征(语言学)
目标检测
模式识别(心理学)
计算机视觉
滤波器(信号处理)
图像融合
融合
集合(抽象数据类型)
对象(语法)
传感器融合
像素
图像(数学)
哲学
程序设计语言
语言学
作者
Yumin Xie,Langwen Zhang,Xiaoyuan Yu,Wei Xie
出处
期刊:IEEE Transactions on Cognitive and Developmental Systems
[Institute of Electrical and Electronics Engineers]
日期:2023-01-19
卷期号:15 (4): 2132-2143
被引量:7
标识
DOI:10.1109/tcds.2023.3238181
摘要
Object detection is essential for an autonomous driving sensing system. Since the light condition is changed in unconstrained scenarios, the detection accuracy based on visible images can be greatly degraded. Although the detection accuracy can be improved by fusing visible and infrared images, existing multispectral object detection (MOD) algorithms suffer from inadequate intermodal interaction and a lack of global dependence in the fusion approach. Thus, we propose an MOD framework called YOLO-MS by designing a feature interaction and self-attention fusion network (FISAFN) as the backbone network. Within the FISAFN, correlations between two modalities are extracted by the feature interaction module (FIM) for reconstructing the information components of each modality and enhancing capability of information exchange. To filter redundant features and enhance complementary features, long-range information dependence between two modalities are established by using a self-attention feature fusion module (SAFFM). Thus, a better information richness of the fused features can be achieved. Experimental results on the FLIR-aligned data set and the M3FD data set demonstrate that the proposed YOLO-MS performs favorably against state-of-the-art approaches, including feature-level fusion and pixel-level fusion. And further, the proposed YOLO-MS possesses good detection performance under diverse scene conditions.
科研通智能强力驱动
Strongly Powered by AbleSci AI