控制理论(社会学)
代数Riccati方程
计算机科学
趋同(经济学)
最优控制
马尔可夫链
跟踪(教育)
李雅普诺夫函数
强化学习
马尔可夫过程
控制(管理)
数学优化
数学
Riccati方程
微分方程
非线性系统
人工智能
物理
心理学
经济
数学分析
机器学习
统计
教育学
量子力学
经济增长
作者
Kun Zhang,Huaguang Zhang,Yuliang Cai,Rong Su
标识
DOI:10.1109/tase.2019.2948431
摘要
This article is concerned with the optimal tracking control problem of the coupled Markov jump system (CMJS) by using the reinforcement learning (RL) technique. Based on the conventional optimal tracking architecture, an offline tracking iteration algorithm is first designed to solve the coupled algebraic Riccati equation that can hardly be solved by mathematical methods directly. To overcome the crucial requirements and existing shortcomings in the offline tracking method, a novel integral RL (IRL) tracking algorithm is first proposed for CMJS, which develops a transition-probability-free optimal tracking control scheme with a reconstructed augmented system and discounted cost function. Both the requirements of transition probability πij and system matrix Ai are avoided via the designed IRL algorithm. The stability and convergence of the novel schemes are proved by the Lyapunov theory, and the tracking objective is achieved as desired. Finally, we apply the designed algorithms in a fourth-order Markov jump control problem and the stochastic mass, spring, and damper system to track continuous sinusoidal waveforms, and the simulation results are provided to show the effectiveness and applicability.
科研通智能强力驱动
Strongly Powered by AbleSci AI