计算机科学
运动规划
趋同(经济学)
移动机器人
算法
适应性
路径(计算)
人工神经网络
机器人
数学优化
人工智能
数学
生物
程序设计语言
经济
经济增长
生态学
作者
Yijian Tan,Yunhan Lin,Tong Liu,Huasong Min
标识
DOI:10.1109/smc53654.2022.9945119
摘要
In this paper, Prioritized Experience Replay (PER) strategy and Long Short Term Memory (LSTM) neural network are introduced to the path planning process of mobile robots, which solves the problems of slow convergence and inaccurate perception of dynamic obstacles with the Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm. We dubbed this new method as PL-TD3. Firstly, we improve the convergence speed of the algorithm by introducing PER strategy. Secondly, we use LSTM neural network to achieve the improvement of the algorithm for dynamic obstacle perception. In order to verify the method of this paper, we design static environment, dynamic environment and adaptability to dynamic experiments to compare and analyze the methods before and after improvement. The experimental results show that PL-TD3 outperforms TD3 in terms of execution time and execution path length in all environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI