强化学习
计算机科学
机器人
运动规划
变压器
人工智能
模仿
人工神经网络
移动机器人
路径(计算)
分布式计算
工程类
计算机网络
电压
社会心理学
电气工程
心理学
作者
Lin Chen,Yaonan Wang,Zhiqiang Miao,Yang Mo,Mingtao Feng,Zhen Zhou,Hesheng Wang
出处
期刊:IEEE Transactions on Industrial Informatics
[Institute of Electrical and Electronics Engineers]
日期:2023-01-30
卷期号:19 (10): 10233-10243
被引量:18
标识
DOI:10.1109/tii.2023.3240585
摘要
Multirobot path planning leads multiple robots from start positions to designated goal positions by generating efficient and collision-free paths. Multirobot systems realize coordination solutions and decentralized path planning, which is essential for large-scale systems. The state-of-the-art decentralized methods utilize imitation learning and reinforcement learning methods to teach fully decentralized policies, dramatically improving their performance. However, these methods cannot enable robots to perform tasks efficiently in relatively dense environments without communication between robots. We introduce the transformer structure into policy neural networks for the first time, dramatically enhancing the ability of policy neural networks to extract features that facilitate collaboration between robots. It mainly focuses on improving the performance of policies in relatively dense multirobot environments under conditions where robots do not communicate with each other. Furthermore, a novel imitation reinforcement learning framework is proposed by combining contrastive learning and double deep Q-network to solve the problem of difficulty training policy neural networks after introducing the transformer structure. We present results in the simulation environment and compare the resulting policy against advanced multirobot path-planning methods in terms of success rate. Simulation results show that our policy achieves state-of-the-art performance when there is no communication between robots. Finally, we experimented with a real-world case using a total of three robots in our robotic laboratory.
科研通智能强力驱动
Strongly Powered by AbleSci AI