强化学习
计算机科学
运动规划
机器人
移动机器人
人工智能
路径(计算)
机器人学习
增强学习
移动机器人导航
任务(项目管理)
社交机器人
避障
人机交互
人工神经网络
深度学习
出处
期刊:International Conference on Mechatronics and Automation
日期:2020-10-13
标识
DOI:10.1109/icma49215.2020.9233738
摘要
Path planning based on deep reinforcement learning has been a hot topic in the field of mobile robots in recent years, but there are still many shortcomings in its application. Such as the lack of agents' generalization ability, the loss of targets in the process of exploration and the inefficiency of data exploration. In this paper, we propose an improved path planning algorithm to solve the above three problems. We take the Depth Deterministic Policy Gradient (DDPG) algorithm as the basic algorithm, and increase generalization ability of agents by adding adaptive gaussian noise, then, in order to solve the problem of losing target in exploration process, we redesign the reward functions based on the curiosity mechanism. We also add the memory units for agents to make the exploration process more efficient. Finally, we verify that the improved algorithm has good adaptability in different obstacle avoidance environments through simulation experiments.
科研通智能强力驱动
Strongly Powered by AbleSci AI