兰萨克
人工智能
特征(语言学)
计算机视觉
同时定位和映射
计算机科学
束流调整
分割
单眼
单目视觉
模式识别(心理学)
图像(数学)
移动机器人
机器人
哲学
语言学
作者
Masaya Kaneko,Kazuya Iwami,Torn Ogawa,Toshihiko Yamasaki,Kiyoharu Aizawa
标识
DOI:10.1109/cvprw.2018.00063
摘要
In this paper, we propose a novel method that combines monocular visual simultaneous localization and mapping (vSLAM) and deep-learning-based semantic segmentation. For stable operation, vSLAM requires feature points on static objects. In conventional vSLAM, random sample consensus (RANSAC) [5] is used to select those feature points. However, if a major portion of the view is occupied by moving objects, many feature points become inappropriate and RANSAC does not perform well. Based on our empirical studies, feature points in the sky and on cars often cause errors in vSLAM. We propose a new framework to exclude feature points using a mask produced by semantic segmentation. Excluding feature points in masked areas enables vSLAM to stably estimate camera motion. We apply ORB-SLAM [15] in our framework, which is a state-of-the-art implementation of monocular vSLAM. For our experiments, we created vSLAM evaluation datasets by using the CARLA simulator [3] under various conditions. Compared to state-of-the-art methods, our method can achieve significantly higher accuracy.
科研通智能强力驱动
Strongly Powered by AbleSci AI