Machine Learning: Science and Technology (Jan 2024)

Feature selection for high-dimensional neural network potentials with the adaptive group lasso

  • Johannes Sandberg,
  • Thomas Voigtmann,
  • Emilie Devijver,
  • Noel Jakse

DOI
https://doi.org/10.1088/2632-2153/ad450e
Journal volume & issue
Vol. 5, no. 2
p. 025043

Abstract

Read online

Neural network potentials are a powerful tool for atomistic simulations, allowing to accurately reproduce ab initio potential energy surfaces with computational performance approaching classical force fields. A central component of such potentials is the transformation of atomic positions into a set of atomic features in a most efficient and informative way. In this work, a feature selection method is introduced for high dimensional neural network potentials, based on the adaptive group lasso (AGL) approach. It is shown that the use of an embedded method, taking into account the interplay between features and their action in the estimator, is necessary to optimize the number of features. The method’s efficiency is tested on three different monoatomic systems, including Lennard–Jones as a simple test case, Aluminium as a system characterized by predominantly radial interactions, and Boron as representative of a system with strongly directional components in the interactions. The AGL is compared with unsupervised filter methods and found to perform consistently better in reducing the number of features needed to reproduce the reference simulation data at a similar level of accuracy as the starting feature set. In particular, our results show the importance of taking into account model predictions in feature selection for interatomic potentials.

Keywords