计算机科学
避障
路径(计算)
强化学习
适应性
趋同(经济学)
一般化
人工智能
障碍物
算法
机器人
移动机器人
数学
生态学
数学分析
经济
生物
程序设计语言
经济增长
政治学
法学
作者
Weilai Jiang,Chong Bao,Guoqiang Xu,Yaonan Wang
标识
DOI:10.1109/cac53003.2021.9728707
摘要
In order to overcome the shortcomings of slow convergence, low success rate and sub-optimal tracking path of deep reinforcement learning algorithms in solving the problems of unmanned aerial vehicle (UAV) autonomous obstacle avoidance and target tracking. This paper adopts Dueling Deep Q-network (Dueling DQN) algorithm with improved network structure based on the Deep Q-network (DQN) algorithm, combined with ε-inspire exploration strategy and experience replay buffer idea, a Multi Pool Dueling Deep Q-network (MP-Dueling DQN) algorithm is proposed to optimize the success rate and path of UAV target tracking. Furthermore, UAV is endowed with the ability of environment perception, and the environmental generalization ability of UAV is improved by designing a continuous reward function. The simulation results show that MP-Dueling DQN algorithm has shorter motion path, stronger environmental adaptability and better performance compared with DQN, Double Deep Q-network (DDQN) and Dueling DQN.
科研通智能强力驱动
Strongly Powered by AbleSci AI