TY - JOUR
T1 - A reinforcement learning-based economic model predictive control framework for autonomous operation of chemical reactors
AU - Alhazmi, Khalid
AU - Albalawi, Fahad
AU - Sarathy, Mani
N1 - KAUST Repository Item: Exported on 2021-07-15
Acknowledged KAUST grant number(s): OSR-2019-CRG7-4077
Acknowledgements: This work was funded by the KAUST, Saudi Arabia Office of Sponsored Research (Grant OSR-2019-CRG7-4077).
PY - 2021/7/3
Y1 - 2021/7/3
N2 - Economic model predictive control (EMPC) is a promising methodology for optimal operation of dynamical processes that has been shown to improve process economics considerably. However, EMPC performance relies heavily on the accuracy of the process model used. As an alternative to model-based control strategies, reinforcement learning (RL) has been investigated as a model-free control methodology, but issues regarding its safety and stability remain an open research challenge. This work presents a novel framework for integrating EMPC and RL for online model parameter estimation of a class of nonlinear systems. In this framework, EMPC optimally operates the closed loop system while maintaining closed loop stability and recursive feasibility. At the same time, to optimize the process, the RL agent continuously compares the measured state of the process with the model's predictions (nominal states), and modifies model parameters accordingly. The major advantage of this framework is its simplicity; state-of-the-art RL algorithms and EMPC schemes can be employed with minimal modifications. The performance of the proposed framework is illustrated on a network of reactions with challenging dynamics and practical significance. This framework allows control, optimization, and model correction to be performed online and continuously, making autonomous reactor operation more attainable.
AB - Economic model predictive control (EMPC) is a promising methodology for optimal operation of dynamical processes that has been shown to improve process economics considerably. However, EMPC performance relies heavily on the accuracy of the process model used. As an alternative to model-based control strategies, reinforcement learning (RL) has been investigated as a model-free control methodology, but issues regarding its safety and stability remain an open research challenge. This work presents a novel framework for integrating EMPC and RL for online model parameter estimation of a class of nonlinear systems. In this framework, EMPC optimally operates the closed loop system while maintaining closed loop stability and recursive feasibility. At the same time, to optimize the process, the RL agent continuously compares the measured state of the process with the model's predictions (nominal states), and modifies model parameters accordingly. The major advantage of this framework is its simplicity; state-of-the-art RL algorithms and EMPC schemes can be employed with minimal modifications. The performance of the proposed framework is illustrated on a network of reactions with challenging dynamics and practical significance. This framework allows control, optimization, and model correction to be performed online and continuously, making autonomous reactor operation more attainable.
UR - http://hdl.handle.net/10754/669169
UR - https://linkinghub.elsevier.com/retrieve/pii/S1385894721025766
UR - http://www.scopus.com/inward/record.url?scp=85109162947&partnerID=8YFLogxK
U2 - 10.1016/j.cej.2021.130993
DO - 10.1016/j.cej.2021.130993
M3 - Article
SN - 1385-8947
VL - 428
SP - 130993
JO - Chemical Engineering Journal
JF - Chemical Engineering Journal
ER -