Machine Learning: Science and Technology (Jan 2023)

Massively parallel fitting of Gaussian approximation potentials

  • Sascha Klawohn,
  • James R Kermode,
  • Albert P Bartók

DOI
https://doi.org/10.1088/2632-2153/aca743
Journal volume & issue
Vol. 4, no. 1
p. 015020

Abstract

Read online

We present a data-parallel software package for fitting Gaussian approximation potentials (GAPs) on multiple nodes using the ScaLAPACK library with MPI and OpenMP. Until now the maximum training set size for GAP models has been limited by the available memory on a single compute node. In our new implementation, descriptor evaluation is carried out in parallel with no communication requirement. The subsequent linear solve required to determine the model coefficients is parallelised with ScaLAPACK. Our approach scales to thousands of cores, lifting the memory limitation and also delivering substantial speedups. This development expands the applicability of the GAP approach to more complex systems as well as opening up opportunities for efficiently embedding GAP model fitting within higher-level workflows such as committee models or hyperparameter optimisation.

Keywords