计算机科学
点云
人工智能
同时定位和映射
计算机视觉
语义映射
稳健性(进化)
解析
保险丝(电气)
分割
机器人
移动机器人
生物化学
化学
电气工程
基因
工程类
作者
Weixiang Shen,Yongxing Jia,Mingcan Li,Junchao Zhu
标识
DOI:10.1109/cisp-bmei53629.2021.9624443
摘要
Visual SLAM (V-SLAM) uses cameras for information input. In mapping, the spatial geometric information of the point cloud is used, which lacks the semantic information of the objects in the environment. This paper proposes a new semantic mapping algorithm based on improved YOLOv5. Firstly, A Pyramid Scene Parsing Network (PSPNet) segmentation head is added to YOLOv5 for performing semantic extraction of the environment. Subsequently, the robot pose is estimated with the ORB-SLAM2 framework. Finally, the semantic images, the depth images and the pose transformation matrix are sent to a mapping module to fuse a dense point cloud semantic map. Experiments show that the algorithm in this paper builds an accurate semantic map on KITTI dataset. Combined with the depth map that eliminates interference factors, it has good accuracy and robustness for semantic mapping in large-scale scenarios.
科研通智能强力驱动
Strongly Powered by AbleSci AI