(434a) End-to-End Design and Implementation of Robust MPC on Resource-Limited Hardware Using Multi-Objective Bayesian Optimization and Deep Learning | AIChE

(434a) End-to-End Design and Implementation of Robust MPC on Resource-Limited Hardware Using Multi-Objective Bayesian Optimization and Deep Learning

Authors 

Chan, K. - Presenter, University of California, Berkeley
Mesbah, A., University of California, Berkeley
Paulson, J., The Ohio State University
The increasing need for optimal control of complex and safety-critical systems has been a driving force for the development of robust and stochastic model predictive control (MPC) strategies [1]–[3], which can systematically account for uncertainties within the system and its environment. Despite their development, there remains a crucial hurdle in their ultimate implementation, in particular for fast-sampling systems. Inevitably, such robust strategies add a layer of complexity to the optimal control problem, increasing the computational requirements of the controller. On the other hand, the desire to implement such controllers on embedded hardware poses a unique challenge in controller design that involves a trade-off between performance and robustness and/or physical hardware restrictions. In particular, expansion of the processing power of modern hardware relies on the use of so-called hardware accelerators (e.g., graphics processing units (GPUs), field programmable gate arrays (FPGAs), etc.), which must be programmed specifically to speed up particular portions of computational algorithms [4]. Thus, to take full advantage of the speedup offered by these hardware accelerators, the hardware-software co-design process should be simultaneously optimized with respect to performance, robustness, and hardware utilization.

In this work, we propose a framework for the automated generation of approximated robust MPCs on FPGAs for advanced control of constrained, nonlinear systems. This framework provides an automatic means to efficiently select parameters that affect the performance, robustness, and embeddability of a controller on resource-restricted hardware. Our framework includes the use of approximate controllers to replace the online evaluation of MPC [5]–[8], the use of the high-level synthesis workflow to automatically generate hardware-ready code [9], and the use of multi-objective Bayesian optimization [10], [11], which balances the trade-off between performance, robustness, and embeddability, to optimally select controller design and implementation parameters. We use deep learning to approximate a robust MPC [6], [12], which offers two key advantages: (i) an explicit approximation of the complex robust MPC structure that can be quickly evaluated and (ii) a predefined structure for which the hardware design can be tailored for easier deployment. Once an approximate controller has be generated, embedded implementation on a FPGA requires the translation of the software-based algorithm to the hardware level. Here, we rely on the high-level synthesis workflow, which is a two-step process that takes in a high level description of the controller algorithm which should be sped up and converts it to a hardware description language (HDL) [9]. Then, the HDL code can be used to program the FPGA device. As opposed to a completely software-driven approach, the inclusion of a hardware accelerator allows for the use of more complex algorithms without significant loss of computational speed. Furthermore, the combination of prescribing a computational architecture with deep neural networks (DNNs) and the hardware synthesis provides an end-to-end controller design workflow, of which allows for fully automated optimization of the implemented controller.

Finally, within this design process, we encounter many controller design parameters related to the algorithm development, including robust MPC tuning parameters, data generation for the DNN approximation, DNN structure, DNN training, and hardware-specific design parameters, including numerical representation and parallelization of computations. The choice of such parameters is not trivial since each can affect each of the competing metrics (i.e., performance, robustness, and ultimate embeddability) in differing, implicit, and non-intuitive ways. Furthermore, the entire design process (from software-based controller synthesis to hardware implementation) can take an extended period of time to complete, making the systematic selection of design parameters highly valuable. For this purpose, we use Bayesian optimization (BO) [13], which is a data-driven optimization strategy to handle black-box and expensive-to-evaluate metrics. Recently, BO has been used for a variety of tuning functions in this framework, particularly in hyperparameter selection for deep learning [14] and auto-tuning of controllers [15]. Furthermore, to highlight the trade-off between each of the controller design metrics, we use a multi-objective approach to directly compute the Pareto frontier [10]. In this manner, we find the set optimal configurations and the ultimate decision between performance, robustness, and embeddability of a particular controller can be determined based on the system requirements.

We demonstrate the utility of this framework with closed-loop simulations and real-time experiments on an atmospheric pressure plasma jet (APPJ) for safety-critical plasma applications (e.g., plasma medicine). For closed-loop simulations, we target biomedical applications where a desired amount of thermal effect is delivered to a substrate. In simulation, we determined a Pareto frontier of optimal control implementations with respect to controller performance and constraint satisfaction. In computing the Pareto frontier for the controller design, we systematically realize trade-offs between the performance and robustness of the hardware controller for various configurations. The proposed framework for the end-to-end design and implementation of hardware controller is general. As such, it can be readily adapted for any control design method irrespective of its complexity.

[1] A. Bemporad and M. Morari, “Robust model predictive control: A survey,” in Robustness in Identification and Control. Springer, 1999, pp. 207–226.

[2] D. Q. Mayne, “Model predictive control: Recent developments and future promise,” Automatica, vol. 50, no. 12, pp. 2967–2986, 2014.

[3] A. Mesbah, “Stochastic model predictive control: An overview and perspectives for future research,” IEEE Control Systems Magazine, vol. 36, no. 6, pp. 30–44, 2016.

[4] S. Lucia, D. Navarro, O. Lucia, P. Zometa, and R. Findeisen, “Optimized FPGA implementation of model predictive control for embedded systems using high-level synthesis tool,” IEEE Transactions on Industrial Informatics, vol. 14, no. 1, pp. 137–145, 2017.

[5] S. Chen, K. Saulnier, N. Atanasov, D. D. Lee, V. Kumar, G. J. Pappas, and M. Morari, “Approximating explicit model predictive control using constrained neural networks,” in American Control Conference, Milwaukee, 2018, pp. 1520–1527.

[6] B. Karg and S. Lucia, “Efficient representation and approximation of model predictive control laws via deep learning,” arXiv preprint arXiv:1806.10644, 2018.

[7] A. D. Bonzanini, J. A. Paulson, D. B. Graves, and A. Mesbah, “Toward safe dose delivery in plasma medicine using projected neural network-based fast approximate NMPC,” IFAC-PapersOnLine, vol. 53, no. 2, pp. 5279–5285, 2020.

[8] J. A. Paulson and A. Mesbah, “Approximate closed-loop robust model predictive control with guaranteed stability and constraint satisfaction,” IEEE Control Systems Letters, vol. 4, pp. 719–724, 2020.

[9] M. H. Ghasemi, O.Lucia, and S.Lucia, “Computing in the blink of an eye: Current possibilities for edge computing and hardware-agnostic programming,” IEEE Access, 2020.

[10] N. Khan, D. E. Goldberg, and M. Pelikan, “Multi-objective Bayesian optimization algorithm,” in Proceedings of the 4th Annual Conference on Genetic and Evolutionary Computation. Citeseer, 2002, pp. 684–684.

[11] S. Daulton, M. Balandat, and E. Bakshy,“Differentiable expected hypervolume improvement for parallel multi-objective Bayesian optimization,” Advances in Neural Information Processing Systems, vol. 33, pp. 9851–9864, 2020.

[12] A. D. Bonzanini, J. A. Paulson, G. Makrygiorgos, and A. Mesbah, “Fast approximate learning-based multistage nonlinear model predictive control using Gaussian processes and deep neural networks,” Computers & Chemical Engineering, vol. 145, p. 107174, 2021.

[13] B. Shahriari, K. Swersky, Z. Wang, R. P. Adams, and N. DeFreitas, “Taking the human out of the loop: A review of Bayesian optimization,” Proceedings of the IEEE, vol. 104, pp. 148–175, 2015.

[14] J. Snoek, H. Larochelle, and R. P. Adams, “Practical Bayesian optimization of machine learning algorithms,” in Advances in Neural Information Processing Systems, 2012, pp. 2951–2959.

[15] F. Sorourifar, G. Makrygirgos, A. Mesbah, and J. A. Paulson, “A data-driven automatic tuning method for MPC under uncertainty using constrained Bayesian optimization,” IFAC-PapersOnLine, vol. 54, no. 3, pp. 243–250, 2021.