强化学习
比例(比率)
人工智能
计算机科学
机器人
钢筋
人机交互
心理学
地理
地图学
社会心理学
作者
Yuhong Cao,Rui Zhao,Yizhuo Wang,Bairan Xiang,Guillaume Sartoretti
出处
期刊:IEEE robotics and automation letters
日期:2024-03-20
卷期号:9 (5): 4631-4638
被引量:2
标识
DOI:10.1109/lra.2024.3379804
摘要
In this work, we propose a deep reinforcement learning (DRL) based reactive planner to solve large-scale Lidar-based autonomous robot exploration problems in 2D action space. Our DRL-based planner allows the agent to reactively plan its exploration path by making implicit predictions about unknown areas, based on a learned estimation of the underlying transition model of the environment. To this end, our approach relies on learned attention mechanisms for their powerful ability to capture long-term dependencies at different spatial scales to reason about the robot's entire belief over known areas. Our approach relies on ground truth information (i.e., privileged learning) to guide the environment estimation during training, as well as on a graph rarefaction algorithm, which allows models trained in small-scale environments to scale to large-scale ones. Simulation results show that our model exhibits better exploration efficiency (12% in path length, 6% in makespan) and lower planning time (60%) than the state-of-the-art planners in a 130m×100m benchmark scenario. We also validate our learned model on hardware.
科研通智能强力驱动
Strongly Powered by AbleSci AI