计算机科学
激光雷达
校准
人工智能
过程(计算)
计算机视觉
传感器融合
特征(语言学)
摄像机切除
遥感
地理
数学
语言学
统计
操作系统
哲学
作者
Xingchen Li,Yuxuan Xiao,Beibei Wang,Haojie Ren,Yanyong Zhang,Jianmin Ji
标识
DOI:10.1007/s10462-022-10317-y
摘要
The recent trend of fusing complementary data from LiDARs and cameras for more accurate perception has made the extrinsic calibration between the two sensors critically important. Indeed, to align the sensors spatially for proper data fusion, the calibration process usually involves estimating the extrinsic parameters between them. Traditional LiDAR–camera calibration methods often depend on explicit targets or human intervention, which can be prohibitively expensive and cumbersome. Recognizing these weaknesses, recent methods usually adopt the autonomic targetless calibration approach, which can be conducted at a much lower cost. This paper presents a thorough review of these automatic targetless LiDAR–camera calibration methods. Specifically, based on how the potential cues in the environment are retrieved and utilized in the calibration process, we divide the methods into four categories: information theory based, feature based, ego-motion based, and learning based methods. For each category, we provide an in-depth overview with insights we have gathered, hoping to serve as a potential guidance for researchers in the related fields.
科研通智能强力驱动
Strongly Powered by AbleSci AI