汉密尔顿-雅各比-贝尔曼方程
控制理论(社会学)
强化学习
控制器(灌溉)
计算机科学
反推
弹道
非线性系统
最优控制
有界函数
人工神经网络
跟踪误差
自适应控制
数学优化
数学
控制(管理)
人工智能
数学分析
物理
生物
量子力学
农学
天文
作者
Zhifu Li,Ming Wang,Ge Ma
标识
DOI:10.1016/j.isatra.2022.12.003
摘要
In this paper, an adaptive model-free optimal reinforcement learning (RL) neural network (NN) control scheme based on filter error is proposed for the trajectory tracking control problem of an autonomous underwater vehicle (AUV) with input saturation. Generally, the optimal control is realized by solving the Hamilton-Jacobi-Bellman (HJB) equation. However, due to its inherent nonlinearity and complexity, the HJB equation of AUV dynamics is challenging to solve. To deal with this problem, an RL strategy based on an actor-critic framework is proposed to approximate the solution of the HJB equation, where actor and critic NNs are used to perform control behavior and evaluate control performance, respectively. In addition, for the AUV system with the second-order strict-feedback dynamic model, the optimal controller design method based on filtering errors is proposed for the first time to simplify the controller design and accelerate the response speed of the system. Then, to solve the model-dependent problem, an extended state observer (ESO) is designed to estimate the unknown nonlinear dynamics, and an adaptive law is designed to estimate the unknown model parameters. To deal with the input saturation, an auxiliary variable system is utilized in the control law. The strict Lyapunov analysis guarantees that all signals of the system are semi-global uniformly ultimately bounded (SGUUB). Finally, the superiority of the proposed method is verified by comparative experiments.
科研通智能强力驱动
Strongly Powered by AbleSci AI