Sanna E. Herwald MD, PhD , Preya Shah MD, PhD , Andrew Johnston MD, MBA , Cameron Olsen MD , Jean-Benoit Delbrouck PhD , Curtis P. Langlotz MD, PhD
{"title":"RadGPT: A System Based on a Large Language Model That Generates Sets of Patient-Centered Materials to Explain Radiology Report Information","authors":"Sanna E. Herwald MD, PhD , Preya Shah MD, PhD , Andrew Johnston MD, MBA , Cameron Olsen MD , Jean-Benoit Delbrouck PhD , Curtis P. Langlotz MD, PhD","doi":"10.1016/j.jacr.2025.06.013","DOIUrl":null,"url":null,"abstract":"<div><h3>Purpose</h3><div>The 21st Century Cures Act final rule requires that patients have real-time access to their radiology reports, which contain technical language. The objective of this study to was to use a novel system called RadGPT, which integrates concept extraction and a large language model (LLM), to help patients understand their radiology reports.</div></div><div><h3>Methods</h3><div>RadGPT generated 150 concept explanations and 390 question-and-answer pairs from 30 radiology report impressions from between 2012 and 2020. The extracted concepts were used to create concept-based explanations, as well as concept-based question-and-answer pairs for which questions were generated using either a fixed template or an LLM. Additionally, report-based question-and-answer pairs were generated directly from the impression using an LLM without concept extraction. One board-certified radiologist and four radiology residents rated the material quality using a standardized rubric.</div></div><div><h3>Results</h3><div>Concept-based LLM-generated questions were of significantly higher quality than concept-based template-generated questions (<em>P</em> < .001). Excluding those template-based question-and-answer pairs from further analysis, nearly all (>95%) of RadGPT-generated materials were rated highly, with at least 50% receiving the highest possible ranking from all five raters. No answers or explanations were rated as likely to affect the safety or effectiveness of patient care. Report-level LLM-based questions and answers were rated particularly highly, with 92% of report-level LLM-based questions and 61% of the corresponding report-level answers receiving the highest rating from all raters.</div></div><div><h3>Conclusions</h3><div>The educational tool RadGPT generated high-quality explanations and question-and-answer pairs that were personalized for each radiology report, unlikely to produce harmful explanations, and likely to enhance patient understanding of radiology information.</div></div>","PeriodicalId":49044,"journal":{"name":"Journal of the American College of Radiology","volume":"22 9","pages":"Pages 1050-1059"},"PeriodicalIF":5.1000,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the American College of Radiology","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1546144025003217","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose
The 21st Century Cures Act final rule requires that patients have real-time access to their radiology reports, which contain technical language. The objective of this study to was to use a novel system called RadGPT, which integrates concept extraction and a large language model (LLM), to help patients understand their radiology reports.
Methods
RadGPT generated 150 concept explanations and 390 question-and-answer pairs from 30 radiology report impressions from between 2012 and 2020. The extracted concepts were used to create concept-based explanations, as well as concept-based question-and-answer pairs for which questions were generated using either a fixed template or an LLM. Additionally, report-based question-and-answer pairs were generated directly from the impression using an LLM without concept extraction. One board-certified radiologist and four radiology residents rated the material quality using a standardized rubric.
Results
Concept-based LLM-generated questions were of significantly higher quality than concept-based template-generated questions (P < .001). Excluding those template-based question-and-answer pairs from further analysis, nearly all (>95%) of RadGPT-generated materials were rated highly, with at least 50% receiving the highest possible ranking from all five raters. No answers or explanations were rated as likely to affect the safety or effectiveness of patient care. Report-level LLM-based questions and answers were rated particularly highly, with 92% of report-level LLM-based questions and 61% of the corresponding report-level answers receiving the highest rating from all raters.
Conclusions
The educational tool RadGPT generated high-quality explanations and question-and-answer pairs that were personalized for each radiology report, unlikely to produce harmful explanations, and likely to enhance patient understanding of radiology information.
期刊介绍:
The official journal of the American College of Radiology, JACR informs its readers of timely, pertinent, and important topics affecting the practice of diagnostic radiologists, interventional radiologists, medical physicists, and radiation oncologists. In so doing, JACR improves their practices and helps optimize their role in the health care system. By providing a forum for informative, well-written articles on health policy, clinical practice, practice management, data science, and education, JACR engages readers in a dialogue that ultimately benefits patient care.