计算机科学
模式
特征(语言学)
人工智能
模态(人机交互)
模式识别(心理学)
传感器融合
判别式
融合
语义学(计算机科学)
分割
语言学
哲学
社会科学
社会学
程序设计语言
作者
Shiyang Feng,Zhaowei Li,Bo Zhang,Bin Wang
出处
期刊:IEEE Transactions on Geoscience and Remote Sensing
[Institute of Electrical and Electronics Engineers]
日期:2025-01-01
卷期号:: 1-1
标识
DOI:10.1109/tgrs.2025.3526247
摘要
Although remote sensing (RS) data with multiple modalities can be used to significantly improve the accuracy of semantic segmentation in RS data, how to effectively extract multimodal information through multimodal feature fusion remains a challenging task. Specifically, existing methods for multimodal feature fusion still face two major challenges: 1) Due to the diverse imaging mechanisms of multimodal RS data, the boundaries of the same foreground may vary across different modalities, leading to the inclusion of unwanted background semantics in the fused foreground features; 2) RS data from different modalities exhibit varying discriminative abilities for different foregrounds, making it challenging to determine the proportion of semantic information for each modality in the fusion results. To address the above issues, we propose a dynamic feature fusion method based on region-wise queries, namely DF 2 RQ, for SS of multimodal RS data. This method is primarily composed of two components: the spatial reconstruction (SR) module and the dynamic fusion (DF) module. Within the SR module, we propose a spatial reconstruction scheme that samples foreground features from different modalities, achieving independent reconstruction of different unimodal features, thereby alleviating the semantic mixing between foreground and background across modalities. In the DF module, a feature fusion scheme based on unimodal feature reference positions is proposed to obtain fusion weights for each modality, thereby enabling the dynamic fusion of complementary features from multiple modalities. The performance of the proposed method has been extensively evaluated on various multimodal RS datasets for SS, and the experimental results consistently show that the proposed method achieves state-of-the-art accuracy on multiple commonly used metrics. In addition, our code is available at https://github.com/I3ab/DF2RQ.
科研通智能强力驱动
Strongly Powered by AbleSci AI