点云
外推法
计算机视觉
人工智能
激光雷达
渲染(计算机图形)
计算机科学
单眼
像素
遥感
地理
数学
数学分析
作者
Chunlan Zhang,Chunyu Lin,Kang Liao,Shujuan Huang,Yao Zhao
出处
期刊:IEEE Transactions on Intelligent Transportation Systems
[Institute of Electrical and Electronics Engineers]
日期:2022-12-15
卷期号:24 (3): 2814-2826
被引量:1
标识
DOI:10.1109/tits.2022.3226566
摘要
Recently, a pseudo-LiDAR point cloud extrapolation algorithm equipped with stereo cameras has been introduced, bridging the gap between the expensive 3D sensor LiDAR and relatively cheap 2D sensor camera in autonomous driving. In this paper, we explore an approach to further bridge this gap using only a monocular camera and extrapolate a wide field of view 3D point cloud from a limited 2D view. However, this task is extremely challenging as it requires inferring the occluded contents in the scene. To this end, we propose a 'render-refine-iterate-fuse' framework that takes advantage of both image view synthesis and image inpainting techniques, guiding the neural network to learn the potential spatial distribution. In addition, we design a hybrid rendering scheme to ensure that the visible content moves in a geometrically correct manner and fills the pixels caused by occlusion. Benefitting from the proposed framework, our approach achieves significant improvements on the pseudo-LiDAR point cloud extrapolation task. The gap between LiDAR and cameras is further bridged, showing an economical and practical application in the environment perception module of autonomous driving. The experimental results evaluated on the KITTI dataset demonstrate that our approach achieves superior quantitative and qualitative performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI