计算机科学
符号
启发式
功能(生物学)
人工智能
数学
算法
算术
进化生物学
生物
作者
Muleilan Pei,Hao An,Bo Liu,Changhong Wang
出处
期刊:IEEE transactions on systems, man, and cybernetics
[Institute of Electrical and Electronics Engineers]
日期:2022-07-01
卷期号:52 (7): 4415-4425
被引量:37
标识
DOI:10.1109/tsmc.2021.3096935
摘要
This article deals with the problem of mobile robot path planning in an unknown environment that contains both static and dynamic obstacles, utilizing a reinforcement learning approach. We propose an improved Dyna- ${Q}$ algorithm, which incorporates heuristic search strategies, simulated annealing mechanism, and reactive navigation principle into ${Q}$ -learning based on the Dyna architecture. A novel action-selection strategy combining $\varepsilon $ -greedy policy with the cooling schedule control is presented, which, together with the heuristic reward function and heuristic actions, can tackle the exploration-exploitation dilemma and enhance the performance of global searching, convergence property, and learning efficiency for path planning. The proposed method is superior to the classical ${Q}$ -learning and Dyna- ${Q}$ algorithms in an unknown static environment, and it is successfully applied to an uncertain environment with multiple dynamic obstacles in simulations. Further, practical experiments are conducted by integrating MATLAB and robot operating system (ROS) on a physical robot platform, and the mobile robot manages to find a collision-free path, thus fulfilling autonomous navigation tasks in the real world.
科研通智能强力驱动
Strongly Powered by AbleSci AI