强化学习
稳健性(进化)
无人机
马尔可夫决策过程
运动学
计算机科学
控制理论(社会学)
趋同(经济学)
动态定位
控制工程
人工智能
工程类
马尔可夫过程
控制(管理)
数学
生物化学
化学
统计
物理
经典力学
海洋工程
经济
基因
经济增长
标识
DOI:10.1016/j.compeleceng.2023.108858
摘要
Dynamic positioning (DP) system is of great significance for the unmanned surface vehicle (USV) to achieve fully autonomous navigation. Traditional control schemes have problems such as model accuracy, parameter tuning, and complex design. In addition, although the deep reinforcement learning (DRL) is widely used in the field of vessel motion control, the learning efficiency is not high, and insufficient robustness in the face of changing environmental. In order to improve the anti-disturbance ability, robustness and convergence speed of the controller during training, a deep reinforcement learning control method based on priority experience replay (PER) is proposed for dynamic positioning of the USV. The mathematical models are established based on the kinematic and dynamic of the USV. Markov decision process (MDP) models are constructed according to the DP tasks. The simulation results show that compared with other DRL algorithms, the proposed method has higher reward value, faster convergence speed, higher control precision and smoother control output.
科研通智能强力驱动
Strongly Powered by AbleSci AI