Zaipeng Duan,Xuzhong Hu,Junfeng Ding,Pei An,Xiao Huang,Jie Ma
出处
期刊:IEEE robotics and automation letters日期:2024-01-01卷期号:9 (1): 627-634被引量:1
标识
DOI:10.1109/lra.2023.3336250
摘要
Multi-sensor collaborative perception has been a significant trend in self-driving and robot navigation. The precondition for multi-sensor fusion is the accurate calibration between sensors. Traditional LiDAR-Camera calibrations rely on laborious manual operations. Several recent studies have demonstrated the advantages of convolutional neural networks regarding feature extraction capabilities. However, the vast modality discrepancy between RGB images and point clouds makes it difficult to explore corresponding features, remaining a challenge for LiDAR-Camera calibrations. In this paper, we propose a new robust online LiDAR-Camera self-calibration network (SCNet). To reduce the search dimensionality for feature matching, we exploit self-supervised learning to align RGB images with projected depth images in 2D pixel coordinates, thereby achieving pre-alignment of the roll angle. In addition, to generate more accurate initial similarity measures for RGB image pixels and possible corresponding projected depth image pixels, we propose a novel multi-level patch matching method that concatenates cost volume constructed from multi-level feature maps. Our method achieves a mean absolute calibration error of 0.724 cm in translation and 0.055 $^{\circ }$ in rotation in a single frame analysis with miscalibration magnitudes of up to $\pm$ 1.5 m and $\pm 20^{\circ }$ on the KITTI odometry dataset, which demonstrates the superiority of our method.