🤖 AI Summary
Optimal power flow (OPF) control in smart grid energy management suffers from low sample efficiency and prohibitively high simulation costs. To address this, we propose a surrogate modeling framework that integrates reinforcement learning (RL) with physics-informed neural networks (PINNs): PINNs serve as high-fidelity, differentiable surrogates of grid dynamics, replacing computationally expensive high-resolution simulators and enabling efficient, low-cost environment interactions for RL policy training. This approach significantly reduces reliance on real-world systems or detailed simulations while preserving physical consistency, thereby accelerating training and improving sample efficiency. Experiments demonstrate several-fold reduction in policy convergence time, substantial decrease in computational overhead, and control performance comparable to baseline methods. The key contribution lies in the first systematic integration of PINNs into an RL architecture for OPF, enabling physics-constrained, data-efficient policy learning.
📝 Abstract
Optimizing the energy management within a smart grids scenario presents significant challenges, primarily due to the complexity of real-world systems and the intricate interactions among various components. Reinforcement Learning (RL) is gaining prominence as a solution for addressing the challenges of Optimal Power Flow in smart grids. However, RL needs to iterate compulsively throughout a given environment to obtain the optimal policy. This means obtaining samples from a, most likely, costly simulator, which can lead to a sample efficiency problem. In this work, we address this problem by substituting costly smart grid simulators with surrogate models built using Phisics-informed Neural Networks (PINNs), optimizing the RL policy training process by arriving to convergent results in a fraction of the time employed by the original environment.