Traditional reinforcement learning (RL) methods for optimal control of nonlinear processes often face challenges, such as high computational demands, long training times, and difficulties in ensuring the safety of closed-loop systems during training. To address these issues, this work proposes a safe transfer RL (TRL) framework. The TRL algorithm leverages knowledge from pretrained source tasks to accelerate learning in a new, related target task, significantly reducing both learning time and computational resources required for optimizing control policies. To ensure safety during knowledge transfer and training, data collection and optimization of the control policy are performed within a control invariant set (CIS) throughout the learning process. Furthermore, we theoretically analyze the errors between the approximate and optimal control policies by accounting for the differences between source and target tasks. Finally, the proposed TRL method is applied to the case studies of chemical processes to demonstrate its effectiveness in solving the optimal control problem with improved computational efficiency and guaranteed safety.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TCYB.2024.3485697 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!