Nonintrusive parameter adaptation of chemical process models with reinforcement learning
Embargo End Date:
KAUST DepartmentChemical Engineering Program
Physical Science and Engineering (PSE) Division
Clean Combustion Research Center
KAUST Grant NumberOSR-2019-CRG7-4077
Embargo End Date2025-02-07
Permanent link to this recordhttp://hdl.handle.net/10754/690061
MetadataShow full item record
AbstractModel-based control is one of the most prevalent techniques for designing and controlling engineering systems. However, many of these systems are complex and characterized by changing dynamics. Hence, online system identification is required to achieve optimum adaptive control performance for such complex systems. This work proposes an algorithm for nonintrusive, online, nonlinear parameter estimation of physical models using deep reinforcement learning (RL). The problem of training a neural network for parameter estimation is formulated as a reinforcement learning problem. The RL-based parameter estimation policy is tested on a simulation of the selective hydrogenation of acetylene, which is a highly nonlinear system. The learned model estimation policy is able to correctly predict the states of the system with a prediction error of less than 1% in various conditions, such as in the presence of measurement noise and structural differences in models.
CitationAlhazmi, K., & Sarathy, S. M. (2023). Nonintrusive parameter adaptation of chemical process models with reinforcement learning. Journal of Process Control, 123, 87–95. https://doi.org/10.1016/j.jprocont.2023.02.001
SponsorsThis work was supported by King Abdullah University of Science and Technology (KAUST) Office of Sponsored Research under the award number OSR-2019-CRG7-4077.
JournalJournal of Process Control