PLoS ONE (Jan 2014)

Feature engineering and a proposed decision-support system for systematic reviewers of medical evidence.

  • Tanja Bekhuis,
  • Eugene Tseytlin,
  • Kevin J Mitchell,
  • Dina Demner-Fushman

DOI
https://doi.org/10.1371/journal.pone.0086277
Journal volume & issue
Vol. 9, no. 1
p. e86277

Abstract

Read online

ObjectivesEvidence-based medicine depends on the timely synthesis of research findings. An important source of synthesized evidence resides in systematic reviews. However, a bottleneck in review production involves dual screening of citations with titles and abstracts to find eligible studies. For this research, we tested the effect of various kinds of textual information (features) on performance of a machine learning classifier. Based on our findings, we propose an automated system to reduce screeing burden, as well as offer quality assurance.MethodsWe built a database of citations from 5 systematic reviews that varied with respect to domain, topic, and sponsor. Consensus judgments regarding eligibility were inferred from published reports. We extracted 5 feature sets from citations: alphabetic, alphanumeric(+), indexing, features mapped to concepts in systematic reviews, and topic models. To simulate a two-person team, we divided the data into random halves. We optimized the parameters of a Bayesian classifier, then trained and tested models on alternate data halves. Overall, we conducted 50 independent tests.ResultsAll tests of summary performance (mean F3) surpassed the corresponding baseline, PConclusionsA computer-assisted, decision support system based on our methods could substantially reduce the burden of screening citations for systematic review teams and solo reviewers. Additionally, such a system could deliver quality assurance both by confirming concordant decisions and by naming studies associated with discordant decisions for further consideration.