(522f) Integration of Control with Process Operation through Reinforcement Learning | AIChE

(522f) Integration of Control with Process Operation through Reinforcement Learning

Authors 

Petsagkourakis, P. - Presenter, University College London
Charitopoulos, V., University College London
Galvanin, F., University College London
del Rio Chanona, A., Imperial College London
Increased market volatility, sustainability concerns and process flexibility are a few of the main challenges contemporary process industries are facing. At the same time major manufacturing areas, like the pharmaceutical industry, have started to investigate the continuous processing paradigm as more profitable and efficient. Moving in these directions, integration of control with operations has received increasing attention over the last decade [1,2]. Integrated planning, scheduling and control (iPSC) aims to exploit the inherent reciprocal information flow among the three problems so as to enhance feasibility, robustness and profitability of the operations in the process industries. The resulting problem tends to be computationally intensive and so far has been studied mostly under deterministic assumptions [3,4] with few exceptions who can be employed under restrictive assumptions [5]. As far as the incorporation of control considerations within these problems, the use of model predictive control [2,3], multi-parametric programming [6,7] as well as surrogate-based and data-driven approaches have been proposed [8,9].

Despite the research effort in this area the development of systematic, computationally efficient and data-driven methods remains an open challenge. On top of that, the underlying dynamic optimization of the control problem suffers from three conditions: (i) there is no precise known model for most industrial-scale processes (plant model mismatch), leading to inaccurate predictions and convergence to suboptimal solutions; (ii) the process is affected by endogenous uncertainty (i.e. the system is stochastic) & (iii) state constraints must be satisfied due to operational and safety concerns. Therefore constraint violation can be detrimental. To solve the above problems, Reinforcement Learning (RL) Policy Gradient method is proposed, which satisfies chance constraints with probabilistic guarantees [10]. The resulting optimal policy is a neural network designed to satisfy the optimality conditions, and the optimal control actions can be evaluated fast as the policy requires only function evaluations. In this work, a novel framework for closed-loop integration iPSC under dynamic disturbances and uncertainty is proposed. The key elements are the integration of novel RL-based optimal policy for the control task of an uncertain dynamic physical system with the optimization-based algorithm for the efficient rescheduling that mitigates the impact of exogenous disturbances on the real-time implementation. Finally, the proposed framework is tested on the iPSC of a nonlinear industrial process illustrating its merits and providing key insights about the interdependence of the iPSC decisions and the importance of RL for its real-time implementation.

  1. Grossmann, I. E. (2012). Advances in mathematical programming models for enterprise-wide optimization. Comput. Chem. Eng., 47, 2-18.
  2. Chu, Y., and F. You. (2015) Model based integration of control and operations: Overview, challenges, advances, and opportunities. Comput. Chem. Eng. , 83, 2-20.
  3. Dias, L. S., & Ierapetritou, M. G. (2016). Integration of scheduling and control under uncertainties: Review and challenges. Chem. Eng. Res. Des., 116, 98-113.
  4. Georgiadis, G.P., Elekidis, A.P. and Georgiadis, M.C. (2019). Optimization-Based Scheduling for the Process Industries: From Theory to Real-Life Industrial Applications. Processes, 7, 438.
  5. Charitopoulos, V. M., Aguirre, A. M., Papageorgiou, L. G., & Dua, V. (2018). Uncertainty aware integration of planning, scheduling and multi-parametric control. Comput. Aid. Chem. Eng. 44, 1171-1176.
  6. Charitopoulos, V. M., Papageorgiou, L. G., & Dua, V. (2019). Closed-loop integration of planning, scheduling and multi-parametric nonlinear control. Comput. Chem. Eng., 122, 172-192.
  7. Burnak, B., Katz, J., Diangelakis, N. A., & Pistikopoulos, E. N. (2018). Simultaneous process scheduling and control: a multiparametric programming-based approach. Ind. Eng. Chem. Res., 57(11), 3963-3976.

  1. Du, J., Park, J., Harjunkoski, I., & Baldea, M. (2015). A time scale-bridging approach for integrating production scheduling and process control. Comput. Chem. Eng., 79, 59-69.

  1. Dias, L. S., & Ierapetritou, M. G. (2019). Data-driven feasibility analysis for the integration of planning and scheduling problems. Optim. Eng., 20(4), 1029-1066.

  1. Petsagkourakis, P., Sandoval, I. O., Bradford E., Zhang, D. & del Rio-Chanona, E. A. (2020). Constrained Reinforcement Learning for Dynamic Optimization under Uncertainty. Accepted for publication in 21st IFAC World Congress

Topics