人工智能
计算机视觉
激光雷达
计算机科学
卷积神经网络
目标检测
点云
特征(语言学)
RGB颜色模型
传感器融合
水准点(测量)
图像传感器
探测器
模式识别(心理学)
遥感
地理
电信
哲学
语言学
大地测量学
作者
Jaekyum Kim,Jaehyung Choi,Yechol Kim,Junho Koh,Chung Choo Chung,Jun Won Choi
标识
DOI:10.1109/ivs.2018.8500711
摘要
In this paper, we introduce a new deep learning architecture for camera and Lidar sensor fusion. The proposed scheme performs 2D object detection using the RGB camera image and the depth, height, and intensity images generated by projecting the 3D Lidar point cloud into camera image plane. The proposed object detector consists of two convolutional neural networks (CNNs) that process the RGB and Lidar images separately as well as the fusion network that combines the feature maps produced at the intermediate layers of the CNNs. We aim to develop a robust object detector that maintains good object detection accuracy even when the quality of the sensor signals is degraded for object detection. Towards this end, we devise the gated fusion unit (GFU) that adjusts the contribution of the feature maps generated by two CNN structures via gating mechanism. Using the GFU, the proposed object detector can fuse the high level feature maps drawn from two modalities with appropriate weights to achieve robust performance. Experiments conducted on the challenging KITTI benchmark show that the proposed camera and Lidar fusion network outperforms the conventional sensor fusion methods even when either of the camera and Lidar sensor signals is corrupted by missing data, occlusion, noise, and illumination change.
科研通智能强力驱动
Strongly Powered by AbleSci AI