BMC Medical Imaging (Oct 2021)

The reporting quality of natural language processing studies: systematic review of studies of radiology reports

  • Emma M. Davidson,
  • Michael T. C. Poon,
  • Arlene Casey,
  • Andreas Grivas,
  • Daniel Duma,
  • Hang Dong,
  • Víctor Suárez-Paniagua,
  • Claire Grover,
  • Richard Tobin,
  • Heather Whalley,
  • Honghan Wu,
  • Beatrice Alex,
  • William Whiteley

DOI
https://doi.org/10.1186/s12880-021-00671-8
Journal volume & issue
Vol. 21, no. 1
pp. 1 – 13

Abstract

Read online

Abstract Background Automated language analysis of radiology reports using natural language processing (NLP) can provide valuable information on patients’ health and disease. With its rapid development, NLP studies should have transparent methodology to allow comparison of approaches and reproducibility. This systematic review aims to summarise the characteristics and reporting quality of studies applying NLP to radiology reports. Methods We searched Google Scholar for studies published in English that applied NLP to radiology reports of any imaging modality between January 2015 and October 2019. At least two reviewers independently performed screening and completed data extraction. We specified 15 criteria relating to data source, datasets, ground truth, outcomes, and reproducibility for quality assessment. The primary NLP performance measures were precision, recall and F1 score. Results Of the 4,836 records retrieved, we included 164 studies that used NLP on radiology reports. The commonest clinical applications of NLP were disease information or classification (28%) and diagnostic surveillance (27.4%). Most studies used English radiology reports (86%). Reports from mixed imaging modalities were used in 28% of the studies. Oncology (24%) was the most frequent disease area. Most studies had dataset size > 200 (85.4%) but the proportion of studies that described their annotated, training, validation, and test set were 67.1%, 63.4%, 45.7%, and 67.7% respectively. About half of the studies reported precision (48.8%) and recall (53.7%). Few studies reported external validation performed (10.8%), data availability (8.5%) and code availability (9.1%). There was no pattern of performance associated with the overall reporting quality. Conclusions There is a range of potential clinical applications for NLP of radiology reports in health services and research. However, we found suboptimal reporting quality that precludes comparison, reproducibility, and replication. Our results support the need for development of reporting standards specific to clinical NLP studies.

Keywords