Nature Communications (Oct 2024)

Analytical ab initio hessian from a deep learning potential for transition state optimization

  • Eric C.-Y. Yuan,
  • Anup Kumar,
  • Xingyi Guan,
  • Eric D. Hermes,
  • Andrew S. Rosen,
  • Judit Zádor,
  • Teresa Head-Gordon,
  • Samuel M. Blau

DOI
https://doi.org/10.1038/s41467-024-52481-5
Journal volume & issue
Vol. 15, no. 1
pp. 1 – 9

Abstract

Read online

Abstract Identifying transition states—saddle points on the potential energy surface connecting reactant and product minima—is central to predicting kinetic barriers and understanding chemical reaction mechanisms. In this work, we train a fully differentiable equivariant neural network potential, NewtonNet, on thousands of organic reactions and derive the analytical Hessians. By reducing the computational cost by several orders of magnitude relative to the density functional theory (DFT) ab initio source, we can afford to use the learned Hessians at every step for the saddle point optimizations. We show that the full machine learned (ML) Hessian robustly finds the transition states of 240 unseen organic reactions, even when the quality of the initial guess structures are degraded, while reducing the number of optimization steps to convergence by 2–3× compared to the quasi-Newton DFT and ML methods. All data generation, NewtonNet model, and ML transition state finding methods are available in an automated workflow.