情态动词
计算机科学
智能交通系统
感知
多智能体系统
智能代理
人机交互
人工智能
运输工程
工程类
心理学
神经科学
化学
高分子化学
作者
Zonglin Meng,Xin Xia,Zhaoliang Zheng,Letian Gao,Wei Liu,Jiaqi Zhu,Jiaqi Ma
出处
期刊:SAE International Journal of Advances and Current Practices in Mobility
日期:2024-12-12
卷期号:07 (4): 1530-1537
摘要
<div class="section abstract"><div class="htmlview paragraph">Cooperative perception has attracted wide attention given its capability to leverage shared information across connected automated vehicles (CAVs) and smart infrastructure to address the occlusion and sensing range limitation issues. To date, existing research is mainly focused on prototyping cooperative perception using only one type of sensor such as LiDAR and camera. In such cases, the performance of cooperative perception is constrained by individual sensor limitations. To exploit the multi-modality of sensors to further improve distant object detection accuracy, in this paper, we propose a unified multi-modal multi-agent cooperative perception framework that integrates camera and LiDAR data to enhance perception performance in intelligent transportation systems. By leveraging the complementary strengths of LiDAR and camera sensors, our framework utilizes the geometry information from LiDAR and the semantic information from cameras to achieve an accurate cooperative perception system. In order to fuse the multi-agent and multi-modal features, we use a bird’s-eye view (BEV) space as the consistent and unified feature representations and employ a transformer-based network for effective multi-agent multi-modal BEV feature fusion. We validate our method on the OPV2V and V2XSim benchmarks, achieving state-of-the-art performance in 3D cooperative perception tasks. The proposed framework significantly improves object detection accuracy and robustness, especially in complex traffic scenarios with occlusions such as dense intersections.</div></div>
科研通智能强力驱动
Strongly Powered by AbleSci AI