Energy and AI (Jan 2025)

Harnessing distributed GPU computing for generalizable graph convolutional networks in power grid reliability assessments

  • Somayajulu L.N. Dhulipala,
  • Nicholas Casaprima,
  • Audrey Olivier,
  • Bjorn C. Vaagensmith,
  • Timothy R. McJunkin,
  • Ryan C. Hruska

Journal volume & issue
Vol. 19
p. 100471

Abstract

Read online

Although machine learning (ML) has emerged as a powerful tool for rapidly assessing grid contingencies, prior studies have largely considered a static grid topology in their analyses. This limits their application, since they need to be re-trained for every new topology. This paper explores the development of generalizable graph convolutional network (GCN) models by pre-training them across a wide range of grid topologies and contingency types. We found that a GCN model with auto-regressive moving average (ARMA) layers with a line graph representation of the grid offered the best predictive performance in predicting voltage magnitudes (VM) and voltage angles (VA). We introduced the concept of phantom nodes to consider disparate grid topologies with a varying number of nodes and lines. For pre-training the GCN ARMA model across a variety of topologies, distributed graphics processing unit (GPU) computing afforded us significant training scalability. The predictive performance of this model on grid topologies that were part of the training data is substantially better than the direct current (DC) approximation. Although direct application of the pre-trained model to topologies that are not part of the grid is not particularly satisfactory, fine-tuning with small amounts of data from a specific topology of interest significantly improves predictive performance. In the context of foundational models in ML, this paper highlights the feasibility of training large-scale GNN models to assess the reliability of power grids by considering a wide variety of grid topologies and contingency types.

Keywords