Clinical and Translational Science (Sep 2024)

Explainable machine learning prediction of edema adverse events in patients treated with tepotinib

  • Federico Amato,
  • Rainer Strotmann,
  • Roberto Castello,
  • Rolf Bruns,
  • Vishal Ghori,
  • Andreas Johne,
  • Karin Berghoff,
  • Karthik Venkatakrishnan,
  • Nadia Terranova

DOI
https://doi.org/10.1111/cts.70010
Journal volume & issue
Vol. 17, no. 9
pp. n/a – n/a

Abstract

Read online

Abstract Tepotinib is approved for the treatment of patients with non‐small‐cell lung cancer harboring MET exon 14 skipping alterations. While edema is the most prevalent adverse event (AE) and a known class effect of MET inhibitors including tepotinib, there is still limited understanding about the factors contributing to its occurrence. Herein, we apply machine learning (ML)‐based approaches to predict the likelihood of occurrence of edema in patients undergoing tepotinib treatment, and to identify factors influencing its development over time. Data from 612 patients receiving tepotinib in five Phase I/II studies were modeled with two ML algorithms, Random Forest, and Gradient Boosting Trees, to predict edema AE incidence and severity. Probability calibration was applied to give a realistic estimation of the likelihood of edema AE. Best model was tested on follow‐up data and on data from clinical studies unused while training. Results showed high performances across all the tested settings, with F1 scores up to 0.961 when retraining the model with the most relevant covariates. The use of ML explainability methods identified serum albumin as the most informative longitudinal covariate, and higher age as associated with higher probabilities of more severe edema. The developed methodological framework enables the use of ML algorithms for analyzing clinical safety data and exploiting longitudinal information through various covariate engineering approaches. Probability calibration ensures the accurate estimation of the likelihood of the AE occurrence, while explainability tools can identify factors contributing to model predictions, hence supporting population and individual patient‐level interpretation.