强化学习
互惠的
避碰
计算机科学
障碍物
碰撞
避障
新颖性
机器人
人工智能
差速器(机械装置)
移动机器人
工程类
哲学
航空航天工程
政治学
法学
语言学
计算机安全
神学
作者
Ruihua Han,Shengduo Chen,Shuaijun Wang,Zeqing Zhang,Ruxin Gao,Qi Hao,Jia Pan
出处
期刊:IEEE robotics and automation letters
日期:2022-07-01
卷期号:7 (3): 5896-5903
被引量:35
标识
DOI:10.1109/lra.2022.3161699
摘要
The challenges to solving the collision avoidance problem lie in adaptively choosing optimal robot velocities in complex scenarios full of interactive obstacles. In this letter, we propose a distributed approach for multi-robot navigation which combines the concept of reciprocal velocity obstacle (RVO) and the scheme of deep reinforcement learning (DRL) to solve the reciprocal collision avoidance problem under limited information. The novelty of this work is threefold: (1) using a set of sequential VO and RVO vectors to represent the interactive environmental states of static and dynamic obstacles, respectively; (2) developing a bidirectional recurrent module based neural network, which maps the states of a varying number of surrounding obstacles to the actions directly; (3) developing a RVO area and expected collision time based reward function to encourage reciprocal collision avoidance behaviors and trade off between collision risk and travel time. The proposed policy is trained through simulated scenarios and updated by the actor-critic based DRL algorithm. We validate the policy in complex environments with various numbers of differential drive robots and obstacles. The experiment results demonstrate that our approach outperforms the state-of-art methods and other learning based approaches in terms of the success rate, travel time, and average speed.
科研通智能强力驱动
Strongly Powered by AbleSci AI