Proximal Policy Optimization with Graph Neural Networks for Optimal Power Flow

📅 2022-12-23
🏛️ arXiv.org
📈 Citations: 4
Influential: 0
📄 PDF
🤖 AI Summary
Real-time solution of the non-convex, nonlinear, and computationally intensive AC optimal power flow (ACOPF) problem remains a critical challenge in power systems. Method: This paper proposes a deep reinforcement learning framework that tightly integrates graph neural networks (GNNs) with proximal policy optimization (PPO). The GNN explicitly encodes grid topology as structural prior knowledge, while PPO learns, end-to-end, physically feasible control policies satisfying AC power flow constraints—enabling zero-shot generalization across varying network topologies. Contribution/Results: Evaluated on the IEEE 30-bus system, the method solves ACOPF in milliseconds, achieving significantly lower operational costs than conventional DCOPF. It guarantees 100% convergence and feasibility across all test cases, demonstrating exceptional speed, robustness, and strong generalization capability—including seamless adaptation to unseen topologies without retraining.
📝 Abstract
Optimal Power Flow (OPF) is a very traditional research area within the power systems field that seeks for the optimal operation point of electric power plants, and which needs to be solved every few minutes in real-world scenarios. However, due to the nonconvexities that arise in power generation systems, there is not yet a fast, robust solution technique for the full Alternating Current Optimal Power Flow (ACOPF). In the last decades, power grids have evolved into a typical dynamic, non-linear and large-scale control system, known as the power system, so searching for better and faster ACOPF solutions is becoming crucial. Appearance of Graph Neural Networks (GNN) has allowed the natural use of Machine Learning (ML) algorithms on graph data, such as power networks. On the other hand, Deep Reinforcement Learning (DRL) is known for its powerful capability to solve complex decision-making problems. Although solutions that use these two methods separately are beginning to appear in the literature, none has yet combined the advantages of both. We propose a novel architecture based on the Proximal Policy Optimization algorithm with Graph Neural Networks to solve the Optimal Power Flow. The objective is to design an architecture that learns how to solve the optimization problem and that is at the same time able to generalize to unseen scenarios. We compare our solution with the DCOPF in terms of cost after having trained our DRL agent on IEEE 30 bus system and then computing the OPF on that base network with topology changes
Problem

Research questions and friction points this paper is trying to address.

Solves nonconvex ACOPF problems in power systems
Combines GNN and DRL for better OPF solutions
Generalizes to unseen power grid scenarios
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines Proximal Policy Optimization with GNNs
Solves Optimal Power Flow using DRL
Generalizes to unseen power grid scenarios
🔎 Similar Papers
No similar papers found.
Á
Ángela López-Cardona
Universitat Politècnica de Catalunya, Barcelona, Spain
G
Guillermo Bernárdez
UC Santa Barbara, California, US
P
P. Barlet-Ros
Universitat Politècnica de Catalunya, Barcelona, Spain; Barcelona Neural Networking Center, Barcelona, Spain
A
A. Cabellos-Aparicio
Universitat Politècnica de Catalunya, Barcelona, Spain; Barcelona Neural Networking Center, Barcelona, Spain