计算机科学
卷积神经网络
图像融合
融合规则
人工智能
图像(数学)
光学(聚焦)
融合
基本事实
计算机视觉
一般化
模式识别(心理学)
数学
哲学
数学分析
物理
光学
语言学
作者
Yu Zhang,Yu Liu,Peng Sun,Yan Han,Xiaolin Zhao,Li Zhang
标识
DOI:10.1016/j.inffus.2019.07.011
摘要
In this paper, we propose a general image fusion framework based on the convolutional neural network, named as IFCNN. Inspired by the transform-domain image fusion algorithms, we firstly utilize two convolutional layers to extract the salient image features from multiple input images. Afterwards, the convolutional features of multiple input images are fused by an appropriate fusion rule (elementwise-max, elementwise-min or elementwise-mean), which is selected according to the type of input images. Finally, the fused features are reconstructed by two convolutional layers to produce the informative fusion image. The proposed model is fully convolutional, so it could be trained in the end-to-end manner without any post-processing procedures. In order to fully train the model, we have generated a large-scale multi-focus image dataset based on the large-scale RGB-D dataset (i.e., NYU-D2), which owns ground-truth fusion images and contains more diverse and larger images than the existing datasets for image fusion. Without finetuning on other types of image datasets, the experimental results show that the proposed model demonstrates better generalization ability than the existing image fusion models for fusing various types of images, such as multi-focus, infrared-visual, multi-modal medical and multi-exposure images. Moreover, the results also verify that our model has achieved comparable or even better results compared to the state-of-the-art image fusion algorithms on four types of image datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI