强化学习
计算机科学
避碰
启发式
机器人
人工智能
行人
碰撞
自主代理人
增强学习
机器学习
计算机安全
操作系统
运输工程
工程类
作者
Michael Everett,Yu Fan Chen,Jonathan P. How
出处
期刊:IEEE Access
[Institute of Electrical and Electronics Engineers]
日期:2021-01-01
卷期号:9: 10357-10377
被引量:206
标识
DOI:10.1109/access.2021.3050338
摘要
Collision avoidance algorithms are essential for safe and efficient robot\noperation among pedestrians. This work proposes using deep reinforcement (RL)\nlearning as a framework to model the complex interactions and cooperation with\nnearby, decision-making agents, such as pedestrians and other robots. Existing\nRL-based works assume homogeneity of agent properties, use specific motion\nmodels over short timescales, or lack a principled method to handle a large,\npossibly varying number of agents. Therefore, this work develops an algorithm\nthat learns collision avoidance among a variety of heterogeneous,\nnon-communicating, dynamic agents without assuming they follow any particular\nbehavior rules. It extends our previous work by introducing a strategy using\nLong Short-Term Memory (LSTM) that enables the algorithm to use observations of\nan arbitrary number of other agents, instead of a small, fixed number of\nneighbors. The proposed algorithm is shown to outperform a classical collision\navoidance algorithm, another deep RL-based algorithm, and scales with the\nnumber of agents better (fewer collisions, shorter time to goal) than our\npreviously published learning-based approach. Analysis of the LSTM provides\ninsights into how observations of nearby agents affect the hidden state and\nquantifies the performance impact of various agent ordering heuristics. The\nlearned policy generalizes to several applications beyond the training\nscenarios: formation control (arrangement into letters), demonstrations on a\nfleet of four multirotors and on a fully autonomous robotic vehicle capable of\ntraveling at human walking speed among pedestrians.\n
科研通智能强力驱动
Strongly Powered by AbleSci AI