ChatGPT-4 Beat Doctors at Diagnosing Illness, Study Finds
Dr. Adam Rodman, a Boston-based internal medicine expert, helped design a study testing 50 licensed physicians to see whether ChatGPT improved their diagnoses, reports the New York TImes. The results? “Doctors who were given ChatGPT-4 along with conventional resources did only slightly better than doctors who did not have access to the bot.
“And, to the researchers’ surprise, ChatGPT alone outperformed the doctors.”
[ChatGPT-4] scored an average of 90 percent when diagnosing a medical condition from a case report and explaining its reasoning. Doctors randomly assigned to use the chatbot got an average score of 76 percent. Those randomly assigned not to use it had an average score of 74 percent.
The study showed more than just the chatbot’s superior performance. It unveiled doctors’ sometimes unwavering belief in a diagnosis they made, even when a chatbot potentially suggests a better one.
And the study illustrated that while doctors are being exposed to the tools of artificial intelligence for their work, few know how to exploit the abilities of chatbots. As a result, they failed to take advantage of A.I. systems’ ability to solve complex diagnostic problems and offer explanations for their diagnoses. A.I. systems should be “doctor extenders,” Dr. Rodman said, offering valuable second opinions on diagnoses.
“The results were similar across subgroups of different training levels and experience with the chatbot,” the study concludes. “These results suggest that access alone to LLMs will not improve overall physician diagnostic reasoning in practice.
“These findings are particularly relevant now that many health systems offer Health Insurance Portability and Accountability Act-compliant chatbots that physicians can use in clinical settings, often with no to minimal training on how to use these tools.”
Read more of this story at Slashdot.
Dr. Adam Rodman, a Boston-based internal medicine expert, helped design a study testing 50 licensed physicians to see whether ChatGPT improved their diagnoses, reports the New York TImes. The results? “Doctors who were given ChatGPT-4 along with conventional resources did only slightly better than doctors who did not have access to the bot.
“And, to the researchers’ surprise, ChatGPT alone outperformed the doctors.”
[ChatGPT-4] scored an average of 90 percent when diagnosing a medical condition from a case report and explaining its reasoning. Doctors randomly assigned to use the chatbot got an average score of 76 percent. Those randomly assigned not to use it had an average score of 74 percent.
The study showed more than just the chatbot’s superior performance. It unveiled doctors’ sometimes unwavering belief in a diagnosis they made, even when a chatbot potentially suggests a better one.
And the study illustrated that while doctors are being exposed to the tools of artificial intelligence for their work, few know how to exploit the abilities of chatbots. As a result, they failed to take advantage of A.I. systems’ ability to solve complex diagnostic problems and offer explanations for their diagnoses. A.I. systems should be “doctor extenders,” Dr. Rodman said, offering valuable second opinions on diagnoses.
“The results were similar across subgroups of different training levels and experience with the chatbot,” the study concludes. “These results suggest that access alone to LLMs will not improve overall physician diagnostic reasoning in practice.
“These findings are particularly relevant now that many health systems offer Health Insurance Portability and Accountability Act-compliant chatbots that physicians can use in clinical settings, often with no to minimal training on how to use these tools.”
Read more of this story at Slashdot.