计算机科学
RGB颜色模型
人工智能
编码器
突出
融合机制
卷积神经网络
稳健性(进化)
模式识别(心理学)
计算机视觉
模式
融合
语言学
哲学
社会科学
生物化学
化学
脂质双层融合
社会学
基因
操作系统
作者
Junyi Wu,Wujie Zhou,Xiaohong Qian,Jingsheng Lei,Lu Yu,Ting Luo
标识
DOI:10.1016/j.dsp.2022.103827
摘要
Recent progress in salient object detection (SOD) has been fueled substantially by the development of convolutional neural networks. However, several SOD methods do not fully exploit information from different modalities, consequently performing only marginally better than methods using a single modality. Therefore, we propose a multitype fusion and enhancement network (MFENet), following three steps “Encoder- Pre-decoder- Decoder” for RGB-thermal (RGB-T) SOD by completely exploiting the advantages of the RGB and thermal modalities through feature integration and enhancement. To better fuse two modalities' features, we have designed the cross-modality fusion module (CMFM) in the encoder part. As shallow features describe details and deep features provide semantic information, a multiscale interactive refinement module is designed in the pre-decoder part to complement multilevel features. Additionally, to further sharpen salient objects, we have proposed a high-level, low-level module that takes inputs from adjacent layers for gradual translation into a saliency map in the decoder part. This module provides semantic information for shallower features and the boundaries of salient objects can be gradually sharpened with subtle details. Extensive experiments show the effectiveness and robustness of the proposed MFENet and its substantial improvement over state-of-the-art RGB-T SOD methods. The codes and results will be available at: https://github.com/wujunyi1412/MFENet_DSP.
科研通智能强力驱动
Strongly Powered by AbleSci AI