强化学习
稳健性(进化)
计算机科学
人工智能
多智能体系统
分布式计算
路径(计算)
计算机网络
生物化学
化学
基因
作者
Lin Chen,Yaonan Wang,Yang Mo,Zhiqiang Miao,Hesheng Wang,Mingtao Feng,Sifei Wang
出处
期刊:IEEE Transactions on Industrial Electronics
[Institute of Electrical and Electronics Engineers]
日期:2022-09-20
卷期号:70 (7): 7032-7040
被引量:10
标识
DOI:10.1109/tie.2022.3206745
摘要
Multiagent path finding (MAPF) is employed to find collision-free paths to guide agents traveling from an initial to a target position. The advanced decentralized approach utilizes communication between agents to improve their performance in environments with high-density obstacles. However, it dramatically reduces the robustness of multiagent systems. To overcome this difficulty, we propose a novel method for solving MAPF problems. In this method, expert data are transformed into supervised signals by proposing a hot supervised contrastive loss, which is combined with reinforcement learning to teach fully-decentralized policies. Agents reactively plan paths online in a partially observable world while exhibiting implicit coordination without communication with others. We introduce the self-attention mechanism in the policy network, which improves the ability of the policy network to extract collaborative information between agents from the observation data. By designing simulation experiments, we demonstrate that the learned policy achieved good performance without communication between agents. Furthermore, real-world application experiments demonstrate the effectiveness of our method in practical applications.
科研通智能强力驱动
Strongly Powered by AbleSci AI