npj Computational Materials (May 2024)

JARVIS-Leaderboard: a large scale benchmark of materials design methods

  • Kamal Choudhary,
  • Daniel Wines,
  • Kangming Li,
  • Kevin F. Garrity,
  • Vishu Gupta,
  • Aldo H. Romero,
  • Jaron T. Krogel,
  • Kayahan Saritas,
  • Addis Fuhr,
  • Panchapakesan Ganesh,
  • Paul R. C. Kent,
  • Keqiang Yan,
  • Yuchao Lin,
  • Shuiwang Ji,
  • Ben Blaiszik,
  • Patrick Reiser,
  • Pascal Friederich,
  • Ankit Agrawal,
  • Pratyush Tiwary,
  • Eric Beyerle,
  • Peter Minch,
  • Trevor David Rhone,
  • Ichiro Takeuchi,
  • Robert B. Wexler,
  • Arun Mannodi-Kanakkithodi,
  • Elif Ertekin,
  • Avanish Mishra,
  • Nithin Mathew,
  • Mitchell Wood,
  • Andrew Dale Rohskopf,
  • Jason Hattrick-Simpers,
  • Shih-Han Wang,
  • Luke E. K. Achenie,
  • Hongliang Xin,
  • Maureen Williams,
  • Adam J. Biacchi,
  • Francesca Tavazza

DOI
https://doi.org/10.1038/s41524-024-01259-w
Journal volume & issue
Vol. 10, no. 1
pp. 1 – 17

Abstract

Read online

Abstract Lack of rigorous reproducibility and validation are significant hurdles for scientific development across many fields. Materials science, in particular, encompasses a variety of experimental and theoretical approaches that require careful benchmarking. Leaderboard efforts have been developed previously to mitigate these issues. However, a comprehensive comparison and benchmarking on an integrated platform with multiple data modalities with perfect and defect materials data is still lacking. This work introduces JARVIS-Leaderboard, an open-source and community-driven platform that facilitates benchmarking and enhances reproducibility. The platform allows users to set up benchmarks with custom tasks and enables contributions in the form of dataset, code, and meta-data submissions. We cover the following materials design categories: Artificial Intelligence (AI), Electronic Structure (ES), Force-fields (FF), Quantum Computation (QC), and Experiments (EXP). For AI, we cover several types of input data, including atomic structures, atomistic images, spectra, and text. For ES, we consider multiple ES approaches, software packages, pseudopotentials, materials, and properties, comparing results to experiment. For FF, we compare multiple approaches for material property predictions. For QC, we benchmark Hamiltonian simulations using various quantum algorithms and circuits. Finally, for experiments, we use the inter-laboratory approach to establish benchmarks. There are 1281 contributions to 274 benchmarks using 152 methods with more than 8 million data points, and the leaderboard is continuously expanding. The JARVIS-Leaderboard is available at the website: https://pages.nist.gov/jarvis_leaderboard/