强化学习
计算机科学
人工智能
运动规划
路径(计算)
趋同(经济学)
避障
理论(学习稳定性)
机器人学
机器学习
机器人
数学优化
移动机器人
数学
经济增长
经济
程序设计语言
作者
Huiyan Han,Jiaqi Wang,Liqun Kuang,Xie Han,Hongxin Xue
出处
期刊:Sensors
[MDPI AG]
日期:2023-06-15
卷期号:23 (12): 5622-5622
摘要
With the advancement of robotics, the field of path planning is currently experiencing a period of prosperity. Researchers strive to address this nonlinear problem and have achieved remarkable results through the implementation of the Deep Reinforcement Learning (DRL) algorithm DQN (Deep Q-Network). However, persistent challenges remain, including the curse of dimensionality, difficulties of model convergence and sparsity in rewards. To tackle these problems, this paper proposes an enhanced DDQN (Double DQN) path planning approach, in which the information after dimensionality reduction is fed into a two-branch network that incorporates expert knowledge and an optimized reward function to guide the training process. The data generated during the training phase are initially discretized into corresponding low-dimensional spaces. An “expert experience” module is introduced to facilitate the model’s early-stage training acceleration in the Epsilon–Greedy algorithm. To tackle navigation and obstacle avoidance separately, a dual-branch network structure is presented. We further optimize the reward function enabling intelligent agents to receive prompt feedback from the environment after performing each action. Experiments conducted in both virtual and real-world environments have demonstrated that the enhanced algorithm can accelerate model convergence, improve training stability and generate a smooth, shorter and collision-free path.
科研通智能强力驱动
Strongly Powered by AbleSci AI