Literatur vom gleichen Autor/der gleichen Autor*in
plus bei Google Scholar

Bibliografische Daten exportieren
 

The Effect of Medical Explanations From Large Language Models on Diagnostic Accuracy in Radiology

Titelangaben

Spitzer, Philipp ; Hendriks, Daniel ; Rudolph, Jan ; Schlaeger, Sarah ; Ricke, Jens ; Kühl, Niklas ; Hoppe, Boj Friedrich ; Feuerriegel, Stefan:
The Effect of Medical Explanations From Large Language Models on Diagnostic Accuracy in Radiology.
In: npj Digital Medicine. Bd. 9 (2026) . - 333.
ISSN 2398-6352
DOI: https://doi.org/10.1038/s41746-026-02619-0

Volltext

Link zum Volltext (externe URL): Volltext

Abstract

Large language models (LLMs) are increasingly used by physicians for diagnostic support. A key advantage of LLMs is the ability to generate explanations that can help physicians understand the reasoning behind a diagnosis. However, the best-suited format for LLM-generated explanations remains unclear. In this large-scale study, we examined the effect of different formats for LLM explanations on clinical decision-making. For this, we conducted a randomized experiment with radiologists reviewing patient cases with radiological images (N = 2020 assessments). Participants received either no LLM support (control group) or were supported by one of three LLM-generated explanations: (1) a standard output providing the diagnosis without explanation; (2) a differential diagnosis comparing multiple possible diagnoses; or (3) a chain-of-thought explanation offering a detailed reasoning process for the diagnosis. We find that the format of explanations significantly influences diagnostic accuracy. The chain-of-thought explanations yielded the best performance, improving the diagnostic accuracy by 12.2% compared to the control condition without LLM support (P = 0.001). The chain-of-thought explanations are also superior to the standard output without explanation ( + 7.2%; P = 0.040) and the differential diagnosis format ( + 9.7%; P = 0.004). We further assessed the robustness of these findings across case difficulty and different physician backgrounds, such as general vs. specialized radiologists. Evidently, in the controlled setting of our vignette study, explaining the reasoning for a diagnosis helps physicians to identify and correct potential errors in LLM predictions and thus improve overall decisions. Altogether, the results highlight the importance of explanations in medical LLMs to support the reasoning processes of physicians, so that medical LLMs can improve diagnostic performance and, ultimately, patient outcomes.

Weitere Angaben

Publikationsform: Artikel in einer Zeitschrift
Begutachteter Beitrag: Ja
Keywords: LLM; AI; LLM-generated explanations; Medical LLM; Medical Explanations; Diagnostic Support;
Institutionen der Universität: Fakultäten > Rechts- und Wirtschaftswissenschaftliche Fakultät > Fachgruppe Betriebswirtschaftslehre
Fakultäten > Rechts- und Wirtschaftswissenschaftliche Fakultät > Fachgruppe Betriebswirtschaftslehre > Lehrstuhl Wirtschaftsinformatik und humanzentrische Künstliche Intelligenz
Fakultäten > Rechts- und Wirtschaftswissenschaftliche Fakultät > Fachgruppe Betriebswirtschaftslehre > Lehrstuhl Wirtschaftsinformatik und humanzentrische Künstliche Intelligenz > Lehrstuhl Wirtschaftsinformatik und humanzentrische Künstliche Intelligenz - Univ.-Prof. Dr.-Ing. Niklas Kühl
Forschungseinrichtungen
Forschungseinrichtungen > Institute in Verbindung mit der Universität
Forschungseinrichtungen > Institute in Verbindung mit der Universität > Institutsteil Wirtschaftsinformatik des Fraunhofer FIT
Forschungseinrichtungen > Institute in Verbindung mit der Universität > FIM Forschungsinstitut für Informationsmanagement
Titel an der UBT entstanden: Ja
Themengebiete aus DDC: 000 Informatik,Informationswissenschaft, allgemeine Werke > 004 Informatik
300 Sozialwissenschaften > 330 Wirtschaft
Eingestellt am: 28 Apr 2026 08:24
Letzte Änderung: 28 Apr 2026 08:24
URI: https://eref.uni-bayreuth.de/id/eprint/96942