计算机科学
合成孔径雷达
人工智能
分割
多光谱图像
计算机视觉
RGB颜色模型
遥感
特征提取
模式识别(心理学)
地质学
作者
Zhe Zhao,Boya Zhao,Yuanfeng Wu,Zhonghua He,Lianru Gao
标识
DOI:10.1109/jstars.2025.3525709
摘要
Automatically extracting buildings with high precision from remote sensing images is crucial for various applications. Due to their distinct imaging modalities and complementary characteristics, optical and synthetic aperture radar (SAR) images serve as primary data sources for this task. We propose a novel Boundary-Link Multimodal Fusion Network (BLMFNet) for joint semantic segmentation to leverage the information in these images. An initial building extraction result is obtained from the multimodal fusion network, followed by refinement using building boundaries. The model achieves high-precision building delineation by leveraging building boundary and semantic information from optical and SAR images. It distinguishes buildings from the background in complex environments, such as dense urban areas or regions with mixed vegetation, particularly when small buildings lack distinct texture or color features. We conducted experiments using the MSAW dataset (RGBNIR and SAR data) and DFC track2 datasets (RGB and SAR data). The results indicate that our model significantly enhances extraction accuracy and improves building boundary delineation. The intersection over union (IoU) metric is 2.5% to 3.5% higher than that of other multimodal joint segmentation methods. The code is available at: https://github.com/tianyamokeZZ/BLMFNet
科研通智能强力驱动
Strongly Powered by AbleSci AI