计算机科学
人工智能
粒度
RGB颜色模型
保险丝(电气)
突出
情态动词
利用
计算机视觉
模式识别(心理学)
水准点(测量)
目标检测
化学
计算机安全
大地测量学
高分子化学
地理
电气工程
工程类
操作系统
作者
Zongwei Wu,Guillaume Allibert,Fabrice Mériaudeau,Chao Ma,Cédric Demonceaux
出处
期刊:IEEE transactions on image processing
[Institute of Electrical and Electronics Engineers]
日期:2023-01-01
卷期号:32: 2160-2173
被引量:37
标识
DOI:10.1109/tip.2023.3263111
摘要
RGB-D saliency detection aims to fuse multi-modal cues to accurately localize salient regions. Existing works often adopt attention modules for feature modeling, with few methods explicitly leveraging fine-grained details to merge with semantic cues. Thus, despite the auxiliary depth information, it is still challenging for existing models to distinguish objects with similar appearances but at distinct camera distances. In this paper, from a new perspective, we propose a novel Hierarchical Depth Awareness network (HiDAnet) for RGB-D saliency detection. Our motivation comes from the observation that the multi-granularity properties of geometric priors correlate well with the neural network hierarchies. To realize multi-modal and multi-level fusion, we first use a granularity-based attention scheme to strengthen the discriminatory power of RGB and depth features separately. Then we introduce a unified cross dual-attention module for multi-modal and multi-level fusion in a coarse-to-fine manner. The encoded multi-modal features are gradually aggregated into a shared decoder. Further, we exploit a multi-scale loss to take full advantage of the hierarchical information. Extensive experiments on challenging benchmark datasets demonstrate that our HiDAnet performs favorably over the state-of-the-art methods by large margins. The source code can be found in https://github.com/Zongwei97/HIDANet/.
科研通智能强力驱动
Strongly Powered by AbleSci AI