计算机科学
人工智能
分割
编码器
棱锥(几何)
模态(人机交互)
RGB颜色模型
特征(语言学)
计算机视觉
过程(计算)
模式识别(心理学)
语言学
操作系统
光学
物理
哲学
作者
Longze Zhu,Z. Kang,Mei Zhou,Xi Yang,Zhen Wang,Zhen Cao,Chenming Ye
出处
期刊:Sensors
[MDPI AG]
日期:2022-11-05
卷期号:22 (21): 8520-8520
摘要
Indoor-scene semantic segmentation is of great significance to indoor navigation, high-precision map creation, route planning, etc. However, incorporating RGB and HHA images for indoor-scene semantic segmentation is a promising yet challenging task, due to the diversity of textures and structures and the disparity of multi-modality in physical significance. In this paper, we propose a Cross-Modality Attention Network (CMANet) that facilitates the extraction of both RGB and HHA features and enhances the cross-modality feature integration. CMANet is constructed under the encoder–decoder architecture. The encoder consists of two parallel branches that successively extract the latent modality features from RGB and HHA images, respectively. Particularly, a novel self-attention mechanism-based Cross-Modality Refine Gate (CMRG) is presented, which bridges the two branches. More importantly, the CMRG achieves cross-modality feature fusion and produces certain refined aggregated features; it serves as the most crucial part of CMANet. The decoder is a multi-stage up-sampled backbone that is composed of different residual blocks at each up-sampling stage. Furthermore, bi-directional multi-step propagation and pyramid supervision are applied to assist the leaning process. To evaluate the effectiveness and efficiency of the proposed method, extensive experiments are conducted on NYUDv2 and SUN RGB-D datasets. Experimental results demonstrate that our method outperforms the existing ones for indoor semantic-segmentation tasks.
科研通智能强力驱动
Strongly Powered by AbleSci AI