(149ad) A Constraint-Based Modeling Framework with Deep Reinforcement Learning and Multi-Objective Optimization for Control of Mammalian Cell Cultures
AIChE Annual Meeting
2023
2023 AIChE Annual Meeting
Computing and Systems Technology Division
Interactive Session: Systems and Process Control
Tuesday, November 7, 2023 - 3:30pm to 5:00pm
In this work, we used the Recurrent Neural Network Long-Short Term Memory (LSTM) method for modeling mammalian cell cultures. The LSTM method uses neural network algorithms to model and analyze sequential data and is capable of handling large and nonlinear experimental databases. This method has been improved in the past decade, and with the availability of large databases and better computing power, we are able to model and analyze mammalian cell cultures. The models developed here used data generated from in silico experiments involving a large number of process inputs and outputs. The model simulations predict the trends observed in batch and fed-batch mammalian cell cultures for key nutrients glucose and glutamine, viable cell density, target product (MAb) titer, and inhibitory metabolites lactate and ammonia with high accuracy. The optimization of mammalian cell cultures is a crucial step in biopharmaceutical production. However, the nonlinearity and complexity of cell cultures make it difficult to model and optimize these processes. To overcome these hindrances, we demonstrate the use of Deep Reinforcement Learning with Deep Deterministic Policy Gradient (DRL-DDPG) method for multi-objective optimization of fed-batch mammalian cell cultures. The DRL-DDPG method is a model-free, actor-critic algorithm that can find optimal control policies in dynamic systems such as bioprocesses. To represent the actor-critic in the DRL-DDPG algorithm, we employed the LSTM method to model and analyze experimental databases for fed-batch cultures. LSTM is capable of handling large sets of data with nonlinearities and can be tuned further for improved predictive ability. The model in this study is trained using in-silico data generated using 11 inputs and outputs, including MAb titer, viable cell density, and concentrations of glucose, glutamine, lactate, and ammonia. The proposed framework predicts the trajectories of these process variables satisfactorily and with sufficient accuracy and reliability. In addition to multi-objective optimization techniques, we utilized appropriate constraints on state and manipulated variables to ensure that the proposed control policy satisfies constraints on nutrients, target metabolites, and undesired metabolites (waste products). The results presented here reveal that the proposed framework can effectively optimize multiple objectives while satisfying appropriate constraints. The study represents a promising approach for multi-objective optimization and control of mammalian cell cultures using DRL-DDPG. In the near future, we will focus on further improving the accuracy and efficiency of the proposed framework by incorporating additional process variables and optimizing the DRL-DDPG algorithm.