模态(人机交互)
计算机科学
任务(项目管理)
人工智能
计算机视觉
图像融合
融合
比例(比率)
红外线的
图像(数学)
工程类
光学
物理
系统工程
哲学
语言学
量子力学
作者
Wei Zhou,Yingyuan Wang,Lina Zuo,Yuan Gao,Yugen Yi
出处
期刊:Measurement
[Elsevier BV]
日期:2024-05-24
卷期号:237: 114977-114977
被引量:3
标识
DOI:10.1016/j.measurement.2024.114977
摘要
Infrared-visible image fusion aims to merge data for a more accurate scene representation. Current research focuses mainly on enhancing visual appeal rather than improving performance in high-level vision tasks. To address this gap, we propose the Semantic Enhanced Multi-scale Cross-modality Interactive Image Fusion Network (SeMIFusion). Initially, Multi-scale Cross-modality Feature Fusion (MCFF) module is devised to extract shallow and deep features across different modalities. During feature extraction, Texture Enhancer (TE) and Semantify Enhancer (SE) blocks capture diverse hierarchical features across multi-scale layers, seamlessly integrating into the Semantic Feature Integration (SFI) module for profound semantic information extraction. Furthermore, an Image Scene Reconstruction (ISR) module maintains original image details in fused features, ensuring image fidelity. Additionally, incorporating a visual preservation guiding mask prioritizes retaining visual quality during reconstruction, preventing degradation. Extensive experiments demonstrate our method's superiority in preserving visual effects and texture details, especially in high-level vision tasks. Our code will be released.
科研通智能强力驱动
Strongly Powered by AbleSci AI