Machine Learning: Science and Technology (Jan 2023)

FAIR AI models in high energy physics

  • Javier Duarte,
  • Haoyang Li,
  • Avik Roy,
  • Ruike Zhu,
  • E A Huerta,
  • Daniel Diaz,
  • Philip Harris,
  • Raghav Kansal,
  • Daniel S Katz,
  • Ishaan H Kavoori,
  • Volodymyr V Kindratenko,
  • Farouk Mokhtar,
  • Mark S Neubauer,
  • Sang Eon Park,
  • Melissa Quinnan,
  • Roger Rusack,
  • Zhizhen Zhao

DOI
https://doi.org/10.1088/2632-2153/ad12e3
Journal volume & issue
Vol. 4, no. 4
p. 045062

Abstract

Read online

The findable, accessible, interoperable, and reusable (FAIR) data principles provide a framework for examining, evaluating, and improving how data is shared to facilitate scientific discovery. Generalizing these principles to research software and other digital products is an active area of research. Machine learning models—algorithms that have been trained on data without being explicitly programmed—and more generally, artificial intelligence (AI) models, are an important target for this because of the ever-increasing pace with which AI is transforming scientific domains, such as experimental high energy physics (HEP). In this paper, we propose a practical definition of FAIR principles for AI models in HEP and describe a template for the application of these principles. We demonstrate the template’s use with an example AI model applied to HEP, in which a graph neural network is used to identify Higgs bosons decaying to two bottom quarks. We report on the robustness of this FAIR AI model, its portability across hardware architectures and software frameworks, and its interpretability.

Keywords