计算机科学
人工智能
计算机视觉
变压器
图像融合
模式识别(心理学)
融合
图像(数学)
工程类
电压
电气工程
语言学
哲学
作者
Xiao Wu,Zihan Cao,Ting‐Zhu Huang,Liang-Jian Deng,Jocelyn Chanussot,Gemine Vivone
标识
DOI:10.1109/tpami.2024.3523364
摘要
Multi-source image fusion combines the information coming from multiple images into one data, thus improving imaging quality. This topic has aroused great interest in the community. How to integrate information from different sources is still a big challenge, although the existing self-attention based transformer methods can capture spatial and channel similarities. In this paper, we first discuss the mathematical concepts behind the proposed generalized self-attention mechanism, where the existing self-attentions are considered basic forms. The proposed mechanism employs multilinear algebra to drive the development of a novel fully-connected self-attention (FCSA) method to fully exploit local and non-local domain-specific correlations among multi-source images. Moreover, we propose a multi-source image representation embedding it into the FCSA framework as a non-local prior within an optimization problem. Some different fusion problems are unfolded into the proposed fully-connected transformer fusion network (FC-Former). More specifically, the concept of generalized self-attention can promote the potential development of self-attention. Hence, the FC-Former can be viewed as a network model unifying different fusion tasks. Compared with state-of-the-art methods, the proposed FC-Former method exhibits robust and superior performance, showing its capability of faithfully preserving information.
科研通智能强力驱动
Strongly Powered by AbleSci AI