一般化
计算机科学
路径(计算)
点(几何)
算法
功能(生物学)
数学优化
数学
数学分析
几何学
进化生物学
生物
程序设计语言
作者
Jinlong Chen,Yun Jiang,Hongren Pan,Minghao Yang
出处
期刊:Electronics
[Multidisciplinary Digital Publishing Institute]
日期:2024-09-20
卷期号:13 (18): 3746-3746
标识
DOI:10.3390/electronics13183746
摘要
The traditional Deep Deterministic Policy Gradient (DDPG) algorithm frequently exhibits a notable reduction in success rate when transferred to new environments after being trained in complex simulation settings. To address these issues, this paper adopts a Multi-Environment (Multi-Env) parallel training approach and integrates Multi-Head Attention (MHA) and Prioritized Experience Replay (PER) into the DDPG framework, optimizing the reward function to form the MAP-DDPG algorithm. This approach enhances the algorithm’s generalization capability and execution efficiency. Through comparative training and testing of the DDPG and MAP-DDPG algorithms in both simulation and real-world environments, the experimental results demonstrate that MAP-DDPG significantly improves generalization and execution efficiency over the DDPG algorithm. Specifically, in simulation environment tests, the MAP-DDPG algorithm achieved an average 30% increase in success rate and reduced the average time to reach the target point by 23.7 s compared to the DDPG algorithm. These results indicate that the MAP-DDPG algorithm significantly enhances path planning generalization and execution efficiency, providing a more effective solution for path planning in complex environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI