计算机科学
RGB颜色模型
人工智能
计算机视觉
背景(考古学)
滤波器(信号处理)
卷积(计算机科学)
突出
模态(人机交互)
模式识别(心理学)
人工神经网络
古生物学
生物
作者
Miao Zhang,Shunyu Yao,Beiqi Hu,Yongri Piao,Wei Ji
标识
DOI:10.1109/tmm.2022.3187856
摘要
The ability to deal with intra and inter-modality features has been critical to the development of RGB-D salient object detection. While many works have advanced in leaps and bounds in this field, most existing methods have not taken their way down into the inherent differences between the RGB and depth data due to widely adopted conventional convolution in which fixed parameter kernels are applied during inference. To promote intra and inter-modality interaction conditioned on various scenarios, as RGB and depth data are processed independently and later fused interactively, we develop a new insight and a better model. In this paper, we introduce a criss-cross dynamic filter network by decoupling dynamic convolution. First, we propose a Model-specific Dynamic Enhanced Module (MDEM) that dynamically enhances the intra-modality features with global context guidance. Second, we propose a Scene-aware Dynamic Fusion Module (SDFM) to realize dynamic feature selection between two modalities. As a result, our model achieves accurate predictions of salient objects. Extensive experiments demonstrate that our method achieves competitive performance over 28 state-of-the-art RGB-D methods on 7 public datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI