计算机科学
交叉口(航空)
领域(数学分析)
域适应
一般化
人工智能
符号
适应(眼睛)
信息抽取
数据挖掘
机器学习
情报检索
地理
光学
数学分析
物理
分类器(UML)
算术
地图学
数学
作者
Shuang Wang,Qi Zang,Dong Zhao,Chaowei Fang,Dou Quan,Yutong Wan,Yanhe Guo,Licheng Jiao
标识
DOI:10.1109/tnnls.2023.3291876
摘要
Accurately extracting buildings from aerial images has essential research significance for timely understanding human intervention on the land. The distribution discrepancies between diversified unlabeled remote sensing images (changes in imaging sensor, location, and environment) and labeled historical images significantly degrade the generalization performance of deep learning algorithms. Unsupervised domain adaptation (UDA) algorithms have recently been proposed to eliminate the distribution discrepancies without re-annotating training data for new domains. Nevertheless, due to the limited information provided by a single-source domain, single-source UDA (SSUDA) is not an optimal choice when multitemporal and multiregion remote sensing images are available. We propose a multisource UDA (MSUDA) framework SPENet for building extraction, aiming at selecting, purifying, and exchanging information from multisource domains to better adapt the model to the target domain. Specifically, the framework effectively utilizes richer knowledge by extracting target-relevant information from multiple-source domains, purifying target domain information with low-level features of buildings, and exchanging target domain information in an interactive learning manner. Extensive experiments and ablation studies constructed on 12 city datasets prove the effectiveness of our method against existing state-of-the-art methods, e.g., our method achieves $59.1\%$ intersection over union (IoU) on Austin and Kitsap $\longrightarrow $ Potsdam, which surpasses the target domain supervised method by $2.2\%$ . The code is available at https://github.com/QZangXDU/SPENet.
科研通智能强力驱动
Strongly Powered by AbleSci AI