生成式人工智能辅助x线片报告的效率和质量。

IF 10.5 1区 医学 Q1 MEDICINE, GENERAL & INTERNAL
Jonathan Huang, Matthew T Wittbrodt, Caitlin N Teague, Eric Karl, Galal Galal, Michael Thompson, Ajay Chapa, Ming-Lun Chiu, Bradley Herynk, Richard Linchangco, Ali Serhal, J Alex Heller, Samir F Abboud, Mozziyar Etemadi
{"title":"生成式人工智能辅助x线片报告的效率和质量。","authors":"Jonathan Huang, Matthew T Wittbrodt, Caitlin N Teague, Eric Karl, Galal Galal, Michael Thompson, Ajay Chapa, Ming-Lun Chiu, Bradley Herynk, Richard Linchangco, Ali Serhal, J Alex Heller, Samir F Abboud, Mozziyar Etemadi","doi":"10.1001/jamanetworkopen.2025.13921","DOIUrl":null,"url":null,"abstract":"<p><strong>Importance: </strong>Diagnostic imaging interpretation involves distilling multimodal clinical information into text form, a task well-suited to augmentation by generative artificial intelligence (AI). However, to our knowledge, impacts of AI-based draft radiological reporting remain unstudied in clinical settings.</p><p><strong>Objective: </strong>To prospectively evaluate the association of radiologist use of a workflow-integrated generative model capable of providing draft radiological reports for plain radiographs across a tertiary health care system with documentation efficiency, the clinical accuracy and textual quality of final radiologist reports, and the model's potential for detecting unexpected, clinically significant pneumothorax.</p><p><strong>Design, setting, and participants: </strong>This prospective cohort study was conducted from November 15, 2023, to April 24, 2024, at a tertiary care academic health system. The association between use of the generative model and radiologist documentation efficiency was evaluated for radiographs documented with model assistance compared with a baseline set of radiographs without model use, matched by study type (chest or nonchest). Peer review was performed on model-assisted interpretations. Flagging of pneumothorax requiring intervention was performed on radiographs prospectively.</p><p><strong>Main outcomes and measures: </strong>The primary outcomes were association of use of the generative model with radiologist documentation efficiency, assessed by difference in documentation time with and without model use using a linear mixed-effects model; for peer review of model-assisted reports, the difference in Likert-scale ratings using a cumulative-link mixed model; and for flagging pneumothorax requiring intervention, sensitivity and specificity.</p><p><strong>Results: </strong>A total of 23 960 radiographs (11 980 each with and without model use) were used to analyze documentation efficiency. Interpretations with model assistance (mean [SE], 159.8 [27.0] seconds) were faster than the baseline set of those without (mean [SE], 189.2 [36.2] seconds) (P = .02), representing a 15.5% documentation efficiency increase. Peer review of 800 studies showed no difference in clinical accuracy (χ2 = 0.68; P = .41) or textual quality (χ2 = 3.62; P = .06) between model-assisted interpretations and nonmodel interpretations. Moreover, the model flagged studies containing a clinically significant, unexpected pneumothorax with a sensitivity of 72.7% and specificity of 99.9% among 97 651 studies screened.</p><p><strong>Conclusions and relevance: </strong>In this prospective cohort study of clinical use of a generative model for draft radiological reporting, model use was associated with improved radiologist documentation efficiency while maintaining clinical quality and demonstrated potential to detect studies containing a pneumothorax requiring immediate intervention. This study suggests the potential for radiologist and generative AI collaboration to improve clinical care delivery.</p>","PeriodicalId":14694,"journal":{"name":"JAMA Network Open","volume":"8 6","pages":"e2513921"},"PeriodicalIF":10.5000,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12142447/pdf/","citationCount":"0","resultStr":"{\"title\":\"Efficiency and Quality of Generative AI-Assisted Radiograph Reporting.\",\"authors\":\"Jonathan Huang, Matthew T Wittbrodt, Caitlin N Teague, Eric Karl, Galal Galal, Michael Thompson, Ajay Chapa, Ming-Lun Chiu, Bradley Herynk, Richard Linchangco, Ali Serhal, J Alex Heller, Samir F Abboud, Mozziyar Etemadi\",\"doi\":\"10.1001/jamanetworkopen.2025.13921\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Importance: </strong>Diagnostic imaging interpretation involves distilling multimodal clinical information into text form, a task well-suited to augmentation by generative artificial intelligence (AI). However, to our knowledge, impacts of AI-based draft radiological reporting remain unstudied in clinical settings.</p><p><strong>Objective: </strong>To prospectively evaluate the association of radiologist use of a workflow-integrated generative model capable of providing draft radiological reports for plain radiographs across a tertiary health care system with documentation efficiency, the clinical accuracy and textual quality of final radiologist reports, and the model's potential for detecting unexpected, clinically significant pneumothorax.</p><p><strong>Design, setting, and participants: </strong>This prospective cohort study was conducted from November 15, 2023, to April 24, 2024, at a tertiary care academic health system. The association between use of the generative model and radiologist documentation efficiency was evaluated for radiographs documented with model assistance compared with a baseline set of radiographs without model use, matched by study type (chest or nonchest). Peer review was performed on model-assisted interpretations. Flagging of pneumothorax requiring intervention was performed on radiographs prospectively.</p><p><strong>Main outcomes and measures: </strong>The primary outcomes were association of use of the generative model with radiologist documentation efficiency, assessed by difference in documentation time with and without model use using a linear mixed-effects model; for peer review of model-assisted reports, the difference in Likert-scale ratings using a cumulative-link mixed model; and for flagging pneumothorax requiring intervention, sensitivity and specificity.</p><p><strong>Results: </strong>A total of 23 960 radiographs (11 980 each with and without model use) were used to analyze documentation efficiency. Interpretations with model assistance (mean [SE], 159.8 [27.0] seconds) were faster than the baseline set of those without (mean [SE], 189.2 [36.2] seconds) (P = .02), representing a 15.5% documentation efficiency increase. Peer review of 800 studies showed no difference in clinical accuracy (χ2 = 0.68; P = .41) or textual quality (χ2 = 3.62; P = .06) between model-assisted interpretations and nonmodel interpretations. Moreover, the model flagged studies containing a clinically significant, unexpected pneumothorax with a sensitivity of 72.7% and specificity of 99.9% among 97 651 studies screened.</p><p><strong>Conclusions and relevance: </strong>In this prospective cohort study of clinical use of a generative model for draft radiological reporting, model use was associated with improved radiologist documentation efficiency while maintaining clinical quality and demonstrated potential to detect studies containing a pneumothorax requiring immediate intervention. This study suggests the potential for radiologist and generative AI collaboration to improve clinical care delivery.</p>\",\"PeriodicalId\":14694,\"journal\":{\"name\":\"JAMA Network Open\",\"volume\":\"8 6\",\"pages\":\"e2513921\"},\"PeriodicalIF\":10.5000,\"publicationDate\":\"2025-06-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12142447/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"JAMA Network Open\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1001/jamanetworkopen.2025.13921\",\"RegionNum\":1,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MEDICINE, GENERAL & INTERNAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"JAMA Network Open","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1001/jamanetworkopen.2025.13921","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MEDICINE, GENERAL & INTERNAL","Score":null,"Total":0}
引用次数: 0

摘要

重要性:诊断成像解释涉及将多模式临床信息提炼成文本形式,这是一项非常适合通过生成式人工智能(AI)增强的任务。然而,据我们所知,基于人工智能的放射学报告草案在临床环境中的影响尚未得到研究。目的:前瞻性评估放射科医生使用工作流集成生成模型的关联,该模型能够为三级卫生保健系统的x线平片提供放射学报告草稿,具有文档效率、临床准确性和最终放射科医生报告的文本质量,以及该模型检测意外的、临床意义重大的气胸的潜力。设计、环境和参与者:这项前瞻性队列研究于2023年11月15日至2024年4月24日在三级保健学术卫生系统进行。将使用模型辅助记录的x线片与未使用模型的基线x线片进行比较,并根据研究类型(胸部或非胸部)匹配,评估生成模型的使用与放射科医生记录效率之间的关系。对模型辅助解释进行同行评议。对需要干预的气胸进行前瞻性的x线片检查。主要结果和测量:主要结果是生成模型的使用与放射科医生记录效率的关联,使用线性混合效应模型通过使用和不使用模型的记录时间差异来评估;对于模型辅助报告的同行评审,使用累积链接混合模型的李克特量表评分的差异;以及标记需要干预的气胸,敏感性和特异性。结果:共使用23张 960张x线片(使用和未使用模型各11张 980张)分析文献效率。有模型辅助的口译(平均[SE], 159.8[27.0]秒)比没有模型辅助的基线组(平均[SE], 189.2[36.2]秒)快(P = 0.02),文件效率提高了15.5%。同行评议的800项研究显示,临床准确性无差异(χ2 = 0.68;P = .41)或文本质量(χ2 = 3.62;P = .06)在模型辅助解释和非模型解释之间。此外,在筛选的97 651项研究中,该模型标记了含有临床意义重大的意外气胸的研究,敏感性为72.7%,特异性为99.9%。结论和相关性:在本前瞻性队列研究中,生成模型用于放射学报告草稿的临床应用,模型的使用与提高放射科医生记录效率相关,同时保持临床质量,并显示出检测需要立即干预的气胸研究的潜力。这项研究表明,放射科医生和生成式人工智能协作在改善临床护理服务方面具有潜力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Efficiency and Quality of Generative AI-Assisted Radiograph Reporting.

Importance: Diagnostic imaging interpretation involves distilling multimodal clinical information into text form, a task well-suited to augmentation by generative artificial intelligence (AI). However, to our knowledge, impacts of AI-based draft radiological reporting remain unstudied in clinical settings.

Objective: To prospectively evaluate the association of radiologist use of a workflow-integrated generative model capable of providing draft radiological reports for plain radiographs across a tertiary health care system with documentation efficiency, the clinical accuracy and textual quality of final radiologist reports, and the model's potential for detecting unexpected, clinically significant pneumothorax.

Design, setting, and participants: This prospective cohort study was conducted from November 15, 2023, to April 24, 2024, at a tertiary care academic health system. The association between use of the generative model and radiologist documentation efficiency was evaluated for radiographs documented with model assistance compared with a baseline set of radiographs without model use, matched by study type (chest or nonchest). Peer review was performed on model-assisted interpretations. Flagging of pneumothorax requiring intervention was performed on radiographs prospectively.

Main outcomes and measures: The primary outcomes were association of use of the generative model with radiologist documentation efficiency, assessed by difference in documentation time with and without model use using a linear mixed-effects model; for peer review of model-assisted reports, the difference in Likert-scale ratings using a cumulative-link mixed model; and for flagging pneumothorax requiring intervention, sensitivity and specificity.

Results: A total of 23 960 radiographs (11 980 each with and without model use) were used to analyze documentation efficiency. Interpretations with model assistance (mean [SE], 159.8 [27.0] seconds) were faster than the baseline set of those without (mean [SE], 189.2 [36.2] seconds) (P = .02), representing a 15.5% documentation efficiency increase. Peer review of 800 studies showed no difference in clinical accuracy (χ2 = 0.68; P = .41) or textual quality (χ2 = 3.62; P = .06) between model-assisted interpretations and nonmodel interpretations. Moreover, the model flagged studies containing a clinically significant, unexpected pneumothorax with a sensitivity of 72.7% and specificity of 99.9% among 97 651 studies screened.

Conclusions and relevance: In this prospective cohort study of clinical use of a generative model for draft radiological reporting, model use was associated with improved radiologist documentation efficiency while maintaining clinical quality and demonstrated potential to detect studies containing a pneumothorax requiring immediate intervention. This study suggests the potential for radiologist and generative AI collaboration to improve clinical care delivery.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
JAMA Network Open
JAMA Network Open Medicine-General Medicine
CiteScore
16.00
自引率
2.90%
发文量
2126
审稿时长
16 weeks
期刊介绍: JAMA Network Open, a member of the esteemed JAMA Network, stands as an international, peer-reviewed, open-access general medical journal.The publication is dedicated to disseminating research across various health disciplines and countries, encompassing clinical care, innovation in health care, health policy, and global health. JAMA Network Open caters to clinicians, investigators, and policymakers, providing a platform for valuable insights and advancements in the medical field. As part of the JAMA Network, a consortium of peer-reviewed general medical and specialty publications, JAMA Network Open contributes to the collective knowledge and understanding within the medical community.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信