关于 "使用 ChatGPT 确定腰椎间盘突出症伴根性病变的临床和手术治疗 "的评论:北美脊柱学会指南比较"

IF 3.8 2区 医学 Q1 CLINICAL NEUROLOGY
Neurospine Pub Date : 2024-03-01 DOI:10.14245/ns.2448248.124
A. Seas, Muhammad M. Abd-El-Barr
{"title":"关于 \"使用 ChatGPT 确定腰椎间盘突出症伴根性病变的临床和手术治疗 \"的评论:北美脊柱学会指南比较\"","authors":"A. Seas, Muhammad M. Abd-El-Barr","doi":"10.14245/ns.2448248.124","DOIUrl":null,"url":null,"abstract":"Clinical medicine is a constantly changing field. However, no change is perhaps as drastic as the integration of machine learning (ML) and artificial intelligence (AI) into clinical practice. This rapid adaptation has recently been stretched with the introduction of the chat generative pre-trained transformer (ChatGPT) in 2022. Unlike many other complex tools for ML, ChatGPT is a large language model (LLM) developed with the intent for rapid use by the lay audience. The tremendously low barrier to entry—namely involving generation of an account—has led to expansive interest in the use of ChatGPT in nearly every subfield of surgery, including spine surgery and low back pain. The goal of the study by Mejia et al. 1 was to assess the ability of ChatGPT to provide accurate medical information regarding the care of patients with lumbar disk herniation with radiculopathy. The research team developed a series of questions related to lumbar disk herniation, us-ing the 2012 North American Spine Society (NASS) guidelines as a gold standard. 2 They then collected responses from both ChatGPT-3.5, and ChatGPT-4.0. They quantified several metrics for each response. A response was considered accurate if it did not contradict the NASS guidelines. It was considered overconclusive if it provided a recommendation when the NASS guidelines did not provide sufficient evidence. A response was supplementary if it included additional relevant information for the question. Finally, a response was considered incomplete if it was accurate but omitted relevant information included within the NASS guidelines. Both ChatGPT-3.5 and -4.0 provided accurate responses to just over 50% of questions. Nearly half of all responses were also overconclusive, providing recommendations without direct","PeriodicalId":19269,"journal":{"name":"Neurospine","volume":null,"pages":null},"PeriodicalIF":3.8000,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Commentary on “Use of ChatGPT for Determining Clinical and Surgical Treatment of Lumbar Disc Herniation With Radiculopathy: A North American Spine Society Guideline Comparison”\",\"authors\":\"A. Seas, Muhammad M. Abd-El-Barr\",\"doi\":\"10.14245/ns.2448248.124\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Clinical medicine is a constantly changing field. However, no change is perhaps as drastic as the integration of machine learning (ML) and artificial intelligence (AI) into clinical practice. This rapid adaptation has recently been stretched with the introduction of the chat generative pre-trained transformer (ChatGPT) in 2022. Unlike many other complex tools for ML, ChatGPT is a large language model (LLM) developed with the intent for rapid use by the lay audience. The tremendously low barrier to entry—namely involving generation of an account—has led to expansive interest in the use of ChatGPT in nearly every subfield of surgery, including spine surgery and low back pain. The goal of the study by Mejia et al. 1 was to assess the ability of ChatGPT to provide accurate medical information regarding the care of patients with lumbar disk herniation with radiculopathy. The research team developed a series of questions related to lumbar disk herniation, us-ing the 2012 North American Spine Society (NASS) guidelines as a gold standard. 2 They then collected responses from both ChatGPT-3.5, and ChatGPT-4.0. They quantified several metrics for each response. A response was considered accurate if it did not contradict the NASS guidelines. It was considered overconclusive if it provided a recommendation when the NASS guidelines did not provide sufficient evidence. A response was supplementary if it included additional relevant information for the question. Finally, a response was considered incomplete if it was accurate but omitted relevant information included within the NASS guidelines. Both ChatGPT-3.5 and -4.0 provided accurate responses to just over 50% of questions. Nearly half of all responses were also overconclusive, providing recommendations without direct\",\"PeriodicalId\":19269,\"journal\":{\"name\":\"Neurospine\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":3.8000,\"publicationDate\":\"2024-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurospine\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.14245/ns.2448248.124\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CLINICAL NEUROLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurospine","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.14245/ns.2448248.124","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CLINICAL NEUROLOGY","Score":null,"Total":0}
引用次数: 0

摘要

临床医学是一个不断变化的领域。然而,没有什么变化能像机器学习(ML)和人工智能(AI)融入临床实践这样剧烈。最近,随着 2022 年聊天生成预训练转换器(ChatGPT)的推出,这种快速适应得到了延伸。与许多其他复杂的 ML 工具不同,ChatGPT 是一种大型语言模型(LLM),其开发目的是为了让非专业人士也能快速使用。由于入门门槛极低--只需生成一个账户,因此几乎所有外科子领域都对 ChatGPT 的使用产生了浓厚的兴趣,包括脊柱外科和腰背痛领域。Mejia 等人 1 的研究目标是评估 ChatGPT 在为腰椎间盘突出症合并根神经病患者提供准确医疗信息方面的能力。研究小组以 2012 年北美脊柱协会 (NASS) 指南为黄金标准,提出了一系列与腰椎间盘突出症相关的问题。2 然后,他们收集了 ChatGPT-3.5 和 ChatGPT-4.0 的回答。他们对每个回复进行了量化。如果答复与 NASS 指南不矛盾,则认为答复准确。如果在 NASS 指南没有提供足够证据的情况下提供了建议,则被视为过度结论。如果答复中包含与问题相关的补充信息,则被视为补充性答复。最后,如果回答准确但遗漏了 NASS 指南中的相关信息,则被视为不完整。ChatGPT-3.5 和 -4.0 都对略高于 50% 的问题做出了准确的回答。近一半的回答也是过度结论性的,提供的建议并没有直接反映出问题的严重性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Commentary on “Use of ChatGPT for Determining Clinical and Surgical Treatment of Lumbar Disc Herniation With Radiculopathy: A North American Spine Society Guideline Comparison”
Clinical medicine is a constantly changing field. However, no change is perhaps as drastic as the integration of machine learning (ML) and artificial intelligence (AI) into clinical practice. This rapid adaptation has recently been stretched with the introduction of the chat generative pre-trained transformer (ChatGPT) in 2022. Unlike many other complex tools for ML, ChatGPT is a large language model (LLM) developed with the intent for rapid use by the lay audience. The tremendously low barrier to entry—namely involving generation of an account—has led to expansive interest in the use of ChatGPT in nearly every subfield of surgery, including spine surgery and low back pain. The goal of the study by Mejia et al. 1 was to assess the ability of ChatGPT to provide accurate medical information regarding the care of patients with lumbar disk herniation with radiculopathy. The research team developed a series of questions related to lumbar disk herniation, us-ing the 2012 North American Spine Society (NASS) guidelines as a gold standard. 2 They then collected responses from both ChatGPT-3.5, and ChatGPT-4.0. They quantified several metrics for each response. A response was considered accurate if it did not contradict the NASS guidelines. It was considered overconclusive if it provided a recommendation when the NASS guidelines did not provide sufficient evidence. A response was supplementary if it included additional relevant information for the question. Finally, a response was considered incomplete if it was accurate but omitted relevant information included within the NASS guidelines. Both ChatGPT-3.5 and -4.0 provided accurate responses to just over 50% of questions. Nearly half of all responses were also overconclusive, providing recommendations without direct
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurospine
Neurospine Multiple-
CiteScore
5.80
自引率
18.80%
发文量
93
审稿时长
10 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信