Scientific Reports (Mar 2022)

A framework for rigorous evaluation of human performance in human and machine learning comparison studies

  • Hannah P. Cowley,
  • Mandy Natter,
  • Karla Gray-Roncal,
  • Rebecca E. Rhodes,
  • Erik C. Johnson,
  • Nathan Drenkow,
  • Timothy M. Shead,
  • Frances S. Chance,
  • Brock Wester,
  • William Gray-Roncal

DOI
https://doi.org/10.1038/s41598-022-08078-3
Journal volume & issue
Vol. 12, no. 1
pp. 1 – 11

Abstract

Read online

Abstract Rigorous comparisons of human and machine learning algorithm performance on the same task help to support accurate claims about algorithm success rates and advances understanding of their performance relative to that of human performers. In turn, these comparisons are critical for supporting advances in artificial intelligence. However, the machine learning community has lacked a standardized, consensus framework for performing the evaluations of human performance necessary for comparison. We demonstrate common pitfalls in a designing the human performance evaluation and propose a framework for the evaluation of human performance, illustrating guiding principles for a successful comparison. These principles are first, to design the human evaluation with an understanding of the differences between human and algorithm cognition; second, to match trials between human participants and the algorithm evaluation, and third, to employ best practices for psychology research studies, such as the collection and analysis of supplementary and subjective data and adhering to ethical review protocols. We demonstrate our framework’s utility for designing a study to evaluate human performance on a one-shot learning task. Adoption of this common framework may provide a standard approach to evaluate algorithm performance and aid in the reproducibility of comparisons between human and machine learning algorithm performance.