计算机科学
国家(计算机科学)
增强学习
不变(物理)
控制(管理)
跟踪(教育)
控制理论(社会学)
LTI系统理论
人工智能
强化学习
算法
数学
线性系统
心理学
数学分析
教育学
数学物理
作者
Huiyuan Shi,Mei Lv,Xueying Jiang,Chengli Su,Ping Li
标识
DOI:10.1016/j.engappai.2024.108006
摘要
Given that conventional model-based control methods have some limitations for dynamic systems with unknown model parameters and existing reinforcement learning methods do not take batch and time delay information into account, a novel data-based adaptive Q-learning approach with two-dimensional (2D) state and control policy is proposed to address the optimal tracking control issue for batch processes with time-invariant state delay. The extended delay state space equation, value function, Q function and optimal performance index are initially presented along the time and batch directions. By examining the correlation between the 2D value function and the 2D Q function, a delay-dependent 2D Bellman equation is designed independent of the process model, which is solved to obtain the expression of the control law. Without requiring prior knowledge of the system, the optimal gain matrices of the control law are further learned by using the current and historical state, output error values and time delay information of the timewise and batchwise. It is feasible to achieve accelerated convergence and reduced errors between the optimal control gain matrices and the learning gain matrices, hence enhancing the tracking capabilities of the systems. At the same time, the unbiasedness and convergence of the given adaptive Q-learning approach are strictly proved. The effectiveness of the proposed algorithm is ultimately validated by simulation comparisons of injection molding, specifically regarding the convergence of control gains and the tracking of output.
科研通智能强力驱动
Strongly Powered by AbleSci AI