模型预测控制
强化学习
参数化复杂度
地平线
计算机科学
控制理论(社会学)
控制(管理)
人工智能
数学优化
算法
数学
几何学
作者
Dean Brandner,Torben Talis,Erik Esche,Jens‐Uwe Repke,Sergio Lucia
出处
期刊:Computer-aided chemical engineering
日期:2023-01-01
卷期号:: 595-600
标识
DOI:10.1016/b978-0-443-15274-0.50094-9
摘要
Model predictive control (MPC) and reinforcement learning (RL) are two powerful optimal control methods. However, the performance of MPC depends mainly on the accuracy of the underlying model and the prediction horizon. Classic RL needs an excessive amount of data and cannot consider constraints explicitly. This work combines both approaches and uses Q-learning to improve the closed-loop performance of a parameterized MPC structure with a surrogate model and a short prediction horizon. The parameterized MPC structure provides a suitable starting point for RL training, which keeps the required data in a reasonable amount. Moreover, constraints are considered explicitly. The solution can be obtained in real-time due to the surrogate model and the short prediction horizon. The method is applied for control of a flash separation unit and compared to a MPC structure that uses a rigorous model and a large prediction horizon.
科研通智能强力驱动
Strongly Powered by AbleSci AI