npj Computational Materials (Aug 2024)

A comparative study of predicting high entropy alloy phase fractions with traditional machine learning and deep neural networks

  • Shusen Liu,
  • Brandon Bocklund,
  • James Diffenderfer,
  • Shreya Chaganti,
  • Bhavya Kailkhura,
  • Scott K. McCall,
  • Brian Gallagher,
  • Aurélien Perron,
  • Joseph T. McKeown

DOI
https://doi.org/10.1038/s41524-024-01335-1
Journal volume & issue
Vol. 10, no. 1
pp. 1 – 14

Abstract

Read online

Abstract Predicting phase stability in high entropy alloys (HEAs), such as phase fractions as functions of composition and temperature, is essential for understanding alloy properties and screening desirable materials. Traditional methods like CALPHAD are computationally intensive for exploring high-dimensional compositional spaces. To address such a challenge, this study explored and compared the effectiveness of random forests (RF) and deep neural networks (DNN) for accelerating materials discovery by building surrogate models of phase stability prediction. For interpolation scenarios (testing on the same order of system as trained), RF models generally produce smaller errors than DNN models. However, for extrapolation scenarios (training on lower-order systems and testing on higher order systems), DNNs generalize more effectively than traditional ML models. DNN demonstrate the potential to predict topologically relevant phase composition when data were missing, making it a powerful predictive tool in materials discovery frameworks. The study uses a CALPHAD dataset of 480 million data points generated from a custom database, available for further model development and benchmarking. Experiments show that DNN models are data-efficient, achieving similar performance with a fraction of the dataset. This work highlights the potential of DNNs in materials discovery, providing a powerful tool for predicting phase stability in HEAs, particularly within the Cr-Hf-Mo-Nb-Ta-Ti-V-W-Zr composition space.