ADAPTIVE TUNING OF MODEL PREDICTIVE CONTROL PARAMETERS VIA REINFORCEMENT LEARNING
Loading...
Date
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
This thesis presents a reinforcement learning (RL)-assisted model predictive control (MPC) framework for multivariable chemical processes that experience external time varying disturbances. MPC is widely used in industry for its ability to predict future behaviour and enforce operating constraints. However, its performance depends on tuning parameters, in this case, the prediction and control horizons. They are usually selected offline and kept fixed during operation. Even with offset-free formulations, the chosen horizons remain fixed, which can result in degraded plant performance. The thesis presents a control framework in which a Deep Q-Network agent dynamically updates the MPC prediction and control horizons in real time. The resulting improvement in set-point tracking is demonstrated by comparing it with the industrial MPC with fixed horizons. A variation of the formulation is also introduced with the reward function explicitly penalizing changes in the agents actions, resulting in more stable agent behavior under the closed-loop MPC.