激光雷达
点云
人工智能
行人检测
计算机视觉
计算机科学
RGB颜色模型
深度学习
遥感
行人
模式识别(心理学)
地质学
地理
考古
作者
Haoran Xu,Shuo Huang,Yixin Yang,Xiaodao Chen,Shiyan Hu
出处
期刊:IEEE Transactions on Industrial Informatics
[Institute of Electrical and Electronics Engineers]
日期:2024-01-31
卷期号:20 (5): 7149-7161
被引量:1
标识
DOI:10.1109/tii.2024.3353845
摘要
One of the fundamental tasks in autonomous driving is environment perception for pedestrian detection, where the fused pedestrian detection using camera and light detection and ranging (LiDAR) information imposes challenges since the data alignment, compensation, and fusion between different data modes are challenging and the simultaneous acquisition of data from two different modalities also increases the difficulty. This work addresses the above challenges from both of the hardware and software dimensions. First, a multimodal pedestrian data acquisition platform is designed and constructed using an RGB camera, sparse LiDAR, and data processing module including hardware connection and deployment, sensor distortion correction and joint calibration, and data acquisition synchronization. Pedestrian data from multiple scenes are then collected using this platform to produce and form a dedicated multimodal pedestrian detection dataset. Further, a two-branch multimodal multilevel fusion pedestrian detection network (MM-Net) is proposed, which includes a two-branch feature extraction module and a feature-level data fusion module. Extensive experiments are performed on the multimodal pedestrian detection dataset and KITTI dataset for the comparison with the existing models. The experimental results demonstrate the superior performance of MM-Net.
科研通智能强力驱动
Strongly Powered by AbleSci AI