npj Digital Medicine (Oct 2024)

A scoping review of reporting gaps in FDA-approved AI medical devices

  • Vijaytha Muralidharan,
  • Boluwatife Adeleye Adewale,
  • Caroline J. Huang,
  • Mfon Thelma Nta,
  • Peter Oluwaduyilemi Ademiju,
  • Pirunthan Pathmarajah,
  • Man Kien Hang,
  • Oluwafolajimi Adesanya,
  • Ridwanullah Olamide Abdullateef,
  • Abdulhammed Opeyemi Babatunde,
  • Abdulquddus Ajibade,
  • Sonia Onyeka,
  • Zhou Ran Cai,
  • Roxana Daneshjou,
  • Tobi Olatunji

DOI
https://doi.org/10.1038/s41746-024-01270-x
Journal volume & issue
Vol. 7, no. 1
pp. 1 – 9

Abstract

Read online

Abstract Machine learning and artificial intelligence (AI/ML) models in healthcare may exacerbate health biases. Regulatory oversight is critical in evaluating the safety and effectiveness of AI/ML devices in clinical settings. We conducted a scoping review on the 692 FDA-approved AI/ML-enabled medical devices approved from 1995-2023 to examine transparency, safety reporting, and sociodemographic representation. Only 3.6% of approvals reported race/ethnicity, 99.1% provided no socioeconomic data. 81.6% did not report the age of study subjects. Only 46.1% provided comprehensive detailed results of performance studies; only 1.9% included a link to a scientific publication with safety and efficacy data. Only 9.0% contained a prospective study for post-market surveillance. Despite the growing number of market-approved medical devices, our data shows that FDA reporting data remains inconsistent. Demographic and socioeconomic characteristics are underreported, exacerbating the risk of algorithmic bias and health disparity.