TY - JOUR
T1 - Evaluation and mitigation of the limitations of large language models in clinical decision-making
AU - Hager, Paul
AU - Jungmann, Friederike
AU - Holland, Robbie
AU - Bhagat, Kunal
AU - Hubrecht, Inga
AU - Knauer, Manuel
AU - Vielhauer, Jakob
AU - Makowski, Marcus
AU - Braren, Rickmer
AU - Kaissis, Georgios
AU - Rueckert, Daniel
N1 - Publisher Copyright:
© The Author(s) 2024.
PY - 2024/9
Y1 - 2024/9
N2 - Clinical decision-making is one of the most impactful parts of a physician’s responsibilities and stands to benefit greatly from artificial intelligence solutions and large language models (LLMs) in particular. However, while LLMs have achieved excellent performance on medical licensing exams, these tests fail to assess many skills necessary for deployment in a realistic clinical decision-making environment, including gathering information, adhering to guidelines, and integrating into clinical workflows. Here we have created a curated dataset based on the Medical Information Mart for Intensive Care database spanning 2,400 real patient cases and four common abdominal pathologies as well as a framework to simulate a realistic clinical setting. We show that current state-of-the-art LLMs do not accurately diagnose patients across all pathologies (performing significantly worse than physicians), follow neither diagnostic nor treatment guidelines, and cannot interpret laboratory results, thus posing a serious risk to the health of patients. Furthermore, we move beyond diagnostic accuracy and demonstrate that they cannot be easily integrated into existing workflows because they often fail to follow instructions and are sensitive to both the quantity and order of information. Overall, our analysis reveals that LLMs are currently not ready for autonomous clinical decision-making while providing a dataset and framework to guide future studies.
AB - Clinical decision-making is one of the most impactful parts of a physician’s responsibilities and stands to benefit greatly from artificial intelligence solutions and large language models (LLMs) in particular. However, while LLMs have achieved excellent performance on medical licensing exams, these tests fail to assess many skills necessary for deployment in a realistic clinical decision-making environment, including gathering information, adhering to guidelines, and integrating into clinical workflows. Here we have created a curated dataset based on the Medical Information Mart for Intensive Care database spanning 2,400 real patient cases and four common abdominal pathologies as well as a framework to simulate a realistic clinical setting. We show that current state-of-the-art LLMs do not accurately diagnose patients across all pathologies (performing significantly worse than physicians), follow neither diagnostic nor treatment guidelines, and cannot interpret laboratory results, thus posing a serious risk to the health of patients. Furthermore, we move beyond diagnostic accuracy and demonstrate that they cannot be easily integrated into existing workflows because they often fail to follow instructions and are sensitive to both the quantity and order of information. Overall, our analysis reveals that LLMs are currently not ready for autonomous clinical decision-making while providing a dataset and framework to guide future studies.
UR - http://www.scopus.com/inward/record.url?scp=85197736357&partnerID=8YFLogxK
U2 - 10.1038/s41591-024-03097-1
DO - 10.1038/s41591-024-03097-1
M3 - Article
AN - SCOPUS:85197736357
SN - 1078-8956
VL - 30
SP - 2613
EP - 2622
JO - Nature Medicine
JF - Nature Medicine
IS - 9
ER -