Medical validity and layperson interpretation of emergency visit recommendations by the GPT model: A cross-sectional study

IF 1.5 Q2 MEDICINE, GENERAL & INTERNAL
Chie Tanaka, Takahiro Kinoshita, Yohei Okada, Kasumi Satoh, Yosuke Homma, Kensuke Suzuki, Shoji Yokobori, Jun Oda, Yasuhiro Otomo, Takashi Tagami, Special Committee on the Utilization of Advanced Technology in Emergency Medicine, Japanese Association for Acute Medicine
{"title":"Medical validity and layperson interpretation of emergency visit recommendations by the GPT model: A cross-sectional study","authors":"Chie Tanaka,&nbsp;Takahiro Kinoshita,&nbsp;Yohei Okada,&nbsp;Kasumi Satoh,&nbsp;Yosuke Homma,&nbsp;Kensuke Suzuki,&nbsp;Shoji Yokobori,&nbsp;Jun Oda,&nbsp;Yasuhiro Otomo,&nbsp;Takashi Tagami,&nbsp;Special Committee on the Utilization of Advanced Technology in Emergency Medicine, Japanese Association for Acute Medicine","doi":"10.1002/ams2.70042","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Aim</h3>\n \n <p>In Japan, emergency ambulance dispatches involve minor cases requiring outpatient services, emphasizing the need for improved public guidance regarding emergency care. This study evaluated both the medical plausibility of the GPT model in aiding laypersons to determine the need for emergency medical care and the laypersons' interpretations of its outputs.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>This cross-sectional study was conducted from December 10, 2023, to March 7, 2024. We input clinical scenarios into the GPT model and evaluated the need for emergency visits based on the outputs. A total of 314 scenarios were labeled with red tags (emergency, immediate emergency department [ED] visit) and 152 with green tags (less urgent). Seven medical specialists assessed the outputs' validity, and 157 laypersons interpreted them via a web-based questionnaire.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>Experts reported that the GPT model accurately identified important information in 95.9% (301/314) of red-tagged scenarios and recommended immediate ED visits in 96.5% (303/314). However, only 43.0% (135/314) of laypersons interpreted those outputs as indicating urgent hospital visits. The model identified important information in 99.3% (151/152) of green-tagged scenarios and advised against immediate visits in 88.8% (135/152). However, only 32.2% (49/152) of laypersons considered them routine follow-ups.</p>\n </section>\n \n <section>\n \n <h3> Conclusions</h3>\n \n <p>Expert evaluations revealed that the GPT model could be highly accurate in advising on emergency visits. However, laypersons frequently misinterpreted its recommendations, highlighting a substantial gap in understanding AI-generated medical advice.</p>\n </section>\n </div>","PeriodicalId":7196,"journal":{"name":"Acute Medicine & Surgery","volume":"12 1","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ams2.70042","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Acute Medicine & Surgery","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/ams2.70042","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MEDICINE, GENERAL & INTERNAL","Score":null,"Total":0}
引用次数: 0

Abstract

Aim

In Japan, emergency ambulance dispatches involve minor cases requiring outpatient services, emphasizing the need for improved public guidance regarding emergency care. This study evaluated both the medical plausibility of the GPT model in aiding laypersons to determine the need for emergency medical care and the laypersons' interpretations of its outputs.

Methods

This cross-sectional study was conducted from December 10, 2023, to March 7, 2024. We input clinical scenarios into the GPT model and evaluated the need for emergency visits based on the outputs. A total of 314 scenarios were labeled with red tags (emergency, immediate emergency department [ED] visit) and 152 with green tags (less urgent). Seven medical specialists assessed the outputs' validity, and 157 laypersons interpreted them via a web-based questionnaire.

Results

Experts reported that the GPT model accurately identified important information in 95.9% (301/314) of red-tagged scenarios and recommended immediate ED visits in 96.5% (303/314). However, only 43.0% (135/314) of laypersons interpreted those outputs as indicating urgent hospital visits. The model identified important information in 99.3% (151/152) of green-tagged scenarios and advised against immediate visits in 88.8% (135/152). However, only 32.2% (49/152) of laypersons considered them routine follow-ups.

Conclusions

Expert evaluations revealed that the GPT model could be highly accurate in advising on emergency visits. However, laypersons frequently misinterpreted its recommendations, highlighting a substantial gap in understanding AI-generated medical advice.

Abstract Image

GPT模型对急诊就诊建议的医学效度和外行人解释:一项横断面研究
目的在日本,紧急救护车调度涉及需要门诊服务的小病例,强调需要改进关于紧急护理的公共指导。本研究评估了GPT模型在帮助外行人确定紧急医疗护理需求方面的医学合理性,以及外行人对其输出的解释。方法横断面研究时间为2023年12月10日至2024年3月7日。我们将临床情景输入GPT模型,并根据输出评估紧急访问的需求。共有314种情况被标记为红色标签(紧急情况,立即急诊部门[ED]就诊),152种被标记为绿色标签(不太紧急)。7名医学专家评估了产出的有效性,157名外行人通过基于网络的问卷对产出进行了解读。结果专家报告,GPT模型在95.9%(301/314)的红色标记场景中准确识别了重要信息,并在96.5%(303/314)的情况下建议立即就诊。然而,只有43.0%(135/314)的非专业人员将这些产出解释为表明紧急医院就诊。该模型在99.3%(151/152)的绿色标记场景中识别出重要信息,并在88.8%(135/152)的场景中建议不要立即访问。然而,只有32.2%(49/152)的外行人认为他们是常规随访。结论专家评价表明,GPT模型在急诊就诊建议中具有较高的准确性。然而,外行经常误解其建议,突出表明在理解人工智能生成的医疗建议方面存在巨大差距。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Acute Medicine & Surgery
Acute Medicine & Surgery MEDICINE, GENERAL & INTERNAL-
自引率
12.50%
发文量
87
审稿时长
53 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信