π€ AI Summary
Deep reinforcement learning (DRL) in process industries suffers from low sample efficiency, poor generalization, and training instability. Method: This work establishes, for the first time, a taxonomy and adaptation principles for DRL transfer learning from both methodological and engineering deployment perspectives, proposing three core paradigms: domain adaptation, policy reuse, and simulation-to-reality transfer. It integrates domain-adversarial neural networks (DANN), policy distillation, meta-reinforcement learning, and high-fidelity process simulation to enable knowledge transfer across operating conditions, equipment configurations, and control tasks. Contribution/Results: Experiments demonstrate over 70% reduction in required interaction data for policy training under novel operating conditions. The frameworkβs effectiveness and robustness are validated on industrial-scale chemical distillation and power plant boiler control tasks.
π Abstract
This paper provides insights into deep reinforcement learning (DRL) for process control from the perspective of transfer learning. We analyze the challenges of applying DRL in the field of process industries and the necessity of introducing transfer learning. Furthermore, recommendations and prospects are provided for future research directions on how transfer learning can be integrated with DRL to empower process control.