强化学习
最优化问题
最优控制
计算机科学
离散时间和连续时间
控制理论(社会学)
噪音(视频)
离散优化
数学优化
前馈
数学
国家(计算机科学)
动力学(音乐)
控制(管理)
贝尔曼方程
系统动力学
非线性系统
线性系统
控制器(灌溉)
控制工程
工程类
人工智能
图像(数学)
统计
作者
Yi Jiang,Bahare Kiumarsi,Jialu Fan,Tianyou Chai,Jinna Li,Frank L. Lewis
标识
DOI:10.1109/tcyb.2018.2890046
摘要
This paper presents a model-free optimal approach based on reinforcement learning for solving the output regulation problem for discrete-time systems under disturbances. This problem is first broken down into two optimization problems: 1) a constrained static optimization problem is established to find the solution to the output regulator equations (i.e., the feedforward control input) and 2) a dynamic optimization problem is established to find the optimal feedback control input. Solving these optimization problems requires the knowledge of the system dynamics. To obviate this requirement, a model-free off-policy algorithm is presented to find the solution to the dynamic optimization problem using only measured data. Then, based on the solution to the dynamic optimization problem, a model-free approach is provided for the static optimization problem. It is shown that the proposed algorithm is insensitive to the probing noise added to the control input for satisfying the persistence of excitation condition. Simulation results are provided to verify the effectiveness of the proposed approach.
科研通智能强力驱动
Strongly Powered by AbleSci AI