Results in Engineering (Sep 2024)

Enhancing interpretability of tree-based models for downstream salinity prediction: Decomposing feature importance using the Shapley additive explanation approach

  • Guang-yao Zhao,
  • Kenji Ohsu,
  • Henry Kasmanhadi Saputra,
  • Teruhisa Okada,
  • Jumpei Suzuki,
  • Yuji Kuwahara,
  • Masafumi Fujita

Journal volume & issue
Vol. 23
p. 102373

Abstract

Read online

To improve the interpretability of estimation processes in machine learning, we applied the Shapley additive explanation method to six tree-based models for predicting downstream salinity. Decision tree-based adaptive boosting (AdaBoost) demonstrated the highest performance, followed by extreme gradient boosting (XGBoost). Global post-decomposition in XGBoost identified water level, a major cause of salinity intrusion in actual environments, as the main contributor of salinity intrusion, while an interaction effect between water temperature and sea level pressure was identified as the main contributor in AdaBoost. Local post-decomposition in XGBoost enables visual interpretation of most of the salinity behaviors. However, the interaction effect between water level and sea level pressure identified as the main contributor by XGBoost was unlikely to drive the actual increase in salinity immediately after a heavy rainfall event. We conclude that highly accurate estimations are not necessarily interpretable and reliable and local post-decomposition can detect inconsistencies in estimation processes.

Keywords