BMC Women's Health (Sep 2024)

An assessment of ChatGPT’s responses to frequently asked questions about cervical and breast cancer

  • Zichen Ye,
  • Bo Zhang,
  • Kun Zhang,
  • María José González Méndez,
  • Huijiao Yan,
  • Tong Wu,
  • Yimin Qu,
  • Yu Jiang,
  • Peng Xue,
  • Youlin Qiao

DOI
https://doi.org/10.1186/s12905-024-03320-8
Journal volume & issue
Vol. 24, no. 1
pp. 1 – 10

Abstract

Read online

Abstract Background Cervical cancer (CC) and breast cancer (BC) threaten women’s well-being, influenced by health-related stigma and a lack of reliable information, which can cause late diagnosis and early death. ChatGPT is likely to become a key source of health information, although quality concerns could also influence health-seeking behaviours. Methods This cross-sectional online survey compared ChatGPT’s responses to five physicians specializing in mammography and five specializing in gynaecology. Twenty frequently asked questions about CC and BC were asked on 26th and 29th of April, 2023. A panel of seven experts assessed the accuracy, consistency, and relevance of ChatGPT’s responses using a 7-point Likert scale. Responses were analyzed for readability, reliability, and efficiency. ChatGPT’s responses were synthesized, and findings are presented as a radar chart. Results ChatGPT had an accuracy score of 7.0 (range: 6.6-7.0) for CC and BC questions, surpassing the highest-scoring physicians (P < 0.05). ChatGPT took an average of 13.6 s (range: 7.6-24.0) to answer each of the 20 questions presented. Readability was comparable to that of experts and physicians involved, but ChatGPT generated more extended responses compared to physicians. The consistency of repeated answers was 5.2 (range: 3.4-6.7). With different contexts combined, the overall ChatGPT relevance score was 6.5 (range: 4.8-7.0). Radar plot analysis indicated comparably good accuracy, efficiency, and to a certain extent, relevance. However, there were apparent inconsistencies, and the reliability and readability be considered inadequate. Conclusions ChatGPT shows promise as an initial source of information for CC and BC. ChatGPT is also highly functional and appears to be superior to physicians, and aligns with expert consensus, although there is room for improvement in readability, reliability, and consistency. Future efforts should focus on developing advanced ChatGPT models explicitly designed to improve medical practice and for those with concerns about symptoms.

Keywords