计算机科学
人工智能
图像配准
合成孔径雷达
人工神经网络
反向传播
转化(遗传学)
特征(语言学)
计算机视觉
相似性(几何)
基本事实
模式识别(心理学)
管道(软件)
几何变换
深度学习
图像(数学)
基因
哲学
生物化学
化学
程序设计语言
语言学
作者
Yuanxin Ye,Tengfeng Tang,Bai Zhu,Chao Yang,Bo Li,Siyuan Hao
出处
期刊:IEEE Transactions on Geoscience and Remote Sensing
[Institute of Electrical and Electronics Engineers]
日期:2022-01-01
卷期号:60: 1-15
被引量:90
标识
DOI:10.1109/tgrs.2022.3167644
摘要
Registration for multisensor or multimodal image pairs with a large degree of distortions is a fundamental task for many remote sensing applications. To achieve accurate and low-cost remote sensing image registration, we propose a multiscale framework with unsupervised learning, named MU-Net. Without costly ground truth labels, MU-Net directly learns the end-to-end mapping from the image pairs to their transformation parameters. MU-Net stacks several deep neural network (DNN) models on multiple scales to generate a coarse-to-fine registration pipeline, which prevents the backpropagation from falling into a local extremum and resists significant image distortions. We design a novel loss function paradigm based on structural similarity, which makes MU-Net suitable for various types of multimodal images. MU-Net is compared with traditional feature-based and area-based methods, as well as supervised and other unsupervised learning methods on the optical-optical, optical-infrared, optical-synthetic aperture radar (SAR), and optical-map datasets. Experimental results show that MU-Net achieves more comprehensive and accurate registration performance between these image pairs with geometric and radiometric distortions. We share the code implemented by Pytorch at https://github.com/yeyuanxin110/MU-Net .
科研通智能强力驱动
Strongly Powered by AbleSci AI