Current Research in Biotechnology (Jan 2024)

ChatGPT’s scorecard after the performance in a series of tests conducted at the multi-country level: A pattern of responses of generative artificial intelligence or large language models

  • Manojit Bhattacharya,
  • Soumen Pal,
  • Srijan Chatterjee,
  • Abdulrahman Alshammari,
  • Thamer H. Albekairi,
  • Supriya Jagga,
  • Elijah Ige Ohimain,
  • Hatem Zayed,
  • Siddappa N. Byrareddy,
  • Sang-Soo Lee,
  • Zhi-Hong Wen,
  • Govindasamy Agoramoorthy,
  • Prosun Bhattacharya,
  • Chiranjib Chakraborty

Journal volume & issue
Vol. 7
p. 100194

Abstract

Read online

Recently, researchers have shown concern about the ChatGPT-derived answers. Here, we conducted a series of tests using ChatGPT by individual researcher at multi-country level to understand the pattern of its answer accuracy, reproducibility, answer length, plagiarism, and in-depth using two questionnaires (the first set with 15 MCQs and the second 15 KBQ). Among 15 MCQ-generated answers, 13 ± 70 were correct (Median : 82.5; Coefficient variance : 4.85), 3 ± 0.77 were incorrect (Median: 3, Coefficient variance: 25.81), and 1 to 10 were reproducible, and 11 to 15 were not. Among 15 KBQ, the length of each question (in words) is about 294.5 ± 97.60 (mean range varies from 138.7 to 438.09), and the mean similarity index (in words) is about 29.53 ± 11.40 (Coefficient variance: 38.62) for each question. The statistical models were also developed using analyzed parameters of answers. The study shows a pattern of ChatGPT-derive answers with correctness and incorrectness and urges for an error-free, next-generation LLM to avoid users’ misguidance.

Keywords