计算机科学
保险丝(电气)
RGB颜色模型
人工智能
深度图
水准点(测量)
情态动词
模式识别(心理学)
深度学习
图像(数学)
计算机视觉
卷积神经网络
工程类
化学
大地测量学
高分子化学
地理
电气工程
作者
Yanjie Ke,Kun Li,Wei Yang,Zhenbo Xu,Dayang Hao,Liusheng Huang,Gang Wang
出处
期刊:International Conference on Robotics and Automation
日期:2021-05-30
卷期号:: 4288-4294
被引量:9
标识
DOI:10.1109/icra48506.2021.9561490
摘要
Depth completion aims to recover the dense depth map from sparse depth data and RGB image respectively. However, due to the huge difference between the multi-modal signal input, vanilla convolutional neural network and simple fusion strategy cannot extract features from sparse data and aggregate multi-modal information effectively. To tackle this problem, we design a novel network architecture that takes full advantage of multi-modal features for depth completion. An effective Pre-completion algorithm is first put forward to increase the density of the input depth map and to provide distribution priors. Moreover, to effectively fuse the image features and the depth features, we propose a multi-modal deep aggregation block that consists of multiple connection and aggregation pathways for deeper fusion. Furthermore, based on the intuition that semantic image features are beneficial for accurate contour, we introduce the deformable guided fusion layer to guide the generation of the dense depth map. The resulting architecture, called MDANet, outperforms all the stateof-the-art methods on the popular KITTI Depth Completion Benchmark, meanwhile with fewer parameters than recent methods. The code of this work will be available at https://github.com/USTC-Keyanjie/MDANet_ICRA2021.
科研通智能强力驱动
Strongly Powered by AbleSci AI