计算机科学
人工智能
棱锥(几何)
单眼
块(置换群论)
模式识别(心理学)
GSM演进的增强数据速率
计算机视觉
联营
比例(比率)
特征(语言学)
深度图
光学(聚焦)
水准点(测量)
空间分析
图像(数学)
数学
地理
遥感
哲学
几何学
物理
光学
地图学
语言学
大地测量学
作者
Xianfa Xu,Zhe Chen,Fuliang Yin
出处
期刊:IEEE transactions on image processing
[Institute of Electrical and Electronics Engineers]
日期:2021-01-01
卷期号:30: 8811-8822
被引量:22
标识
DOI:10.1109/tip.2021.3120670
摘要
Depth estimation from single monocular image is a vital but challenging task in 3D vision and scene understanding. Previous unsupervised methods have yielded impressive results, but the predicted depth maps still have several disadvantages such as missing small objects and object edge blurring. To address these problems, a multi-scale spatial attention guided monocular depth estimation method with semantic enhancement is proposed. Specifically, we first construct a multi-scale spatial attention-guided block based on atrous spatial pyramid pooling and spatial attention. Then, the correlation between the left and right views is fully explored by mutual information to obtain a more robust feature representation. Finally, we design a double-path prediction network to simultaneously generate depth maps and semantic labels. The proposed multi-scale spatial attention-guided block can focus more on the objects, especially on small objects. Moreover, the additional semantic information also enables the objects edge in the predicted depth maps more sharper. We conduct comprehensive evaluations on public benchmark datasets, such as KITTI and Make3D. The experiment results well demonstrate the effectiveness of the proposed method and achieve better performance than other self-supervised methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI