-
GPT-4 outperforms doctors in clinical reasoning, but also makes mistakes more often, study finds
In a new study, scientists at Beth Israel Medical Center (BIDMC) compared the clinical reasoning abilities of a large language model with those of human doctors. The researchers used the revised IDEA (r-IDEA) score, a commonly used tool for assessing clinical reasoning abilities. The study involved giving a GPT-4-powered chatbot, 21 attending physicians, and 18 residents 20 clinical cases to build diagnostic reasoning and solve problems. The r-IDEA scores of the three sets of answers were then evaluated. The researchers found that…- 1.7k