计算机科学
特征(语言学)
水准点(测量)
判别式
RGB颜色模型
人工智能
情态动词
突出
骨料(复合)
特征学习
模式识别(心理学)
对象(语法)
模式
目标检测
计算机视觉
社会学
哲学
复合材料
化学
语言学
高分子化学
材料科学
地理
社会科学
大地测量学
作者
Hongbo Bi,Ranwan Wu,Ziqi Liu,Huihui Zhu,Cong Zhang,Tian-Zhu Xiang
标识
DOI:10.1016/j.patcog.2022.109194
摘要
How to effectively exchange and aggregate the information of multiple modalities (e.g. RGB image and depth map) is a big challenge in the RGB-D salient object detection community. To address this problem, in this paper, we propose a cross-modal Hierarchical Interaction Network (HINet), which boosts the salient object detection by excavating the cross-modal feature interaction and progressively multi-level feature fusion. To achieve it, we design two modules: cross-modal information exchange (CIE) module and multi-level information progressively guided fusion (PGF) module. Specifically, the CIE module is proposed to exchange the cross-modal features for learning the shared representations, as well as the beneficial feedback to facilitate the discriminative feature learning of different modalities. Besides, the PGF module is designed to aggregate the hierarchical features progressively with the reverse guidance mechanism, which employs the high-level feature fusion to guide the low-level feature fusion and thus improve the saliency detection performance. Extensive experiments show that our proposed model significantly outperforms the existing nine state-of-the-art models on five challenging benchmark datasets. Codes and results are available at: https://github.com/RanwanWu/HINet.
科研通智能强力驱动
Strongly Powered by AbleSci AI