Use of large language models to identify pseudo-information: Implications for health information.

IF 2.2 4区 医学 Q2 INFORMATION SCIENCE & LIBRARY SCIENCE
Boris Schmitz
{"title":"Use of large language models to identify pseudo-information: Implications for health information.","authors":"Boris Schmitz","doi":"10.1111/hir.12569","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Open-access scientific research is an essential source of health-related information and self-education. Artificial intelligence-based large language models (LMMs) may be used to identify erroneous health information.</p><p><strong>Objective: </strong>To investigate to what extent LMMs can be used to identify pseudo-information.</p><p><strong>Methods: </strong>Four common LMM applications (ChatGPT-4o, Claude 3.5 Sonnet, Gemini and Copilot) were used to investigate their capability to indicate erroneous information provided in an open-access article.</p><p><strong>Results: </strong>Initially, ChatGPT-4o and Claude were able to mark the provided article as an unreliable information source, identifying most of the inaccuracy problems. The assessments provided by Gemini and Copilot were inaccurate, as several critical aspects were not identified or were misinterpreted. During the validation phase, the initially accurate assessment of ChatGPT-4o was not reproducible, and only Claude was able to detect several critical issues in this phase. The verdicts of Copilot and Gemini remained largely unaltered.</p><p><strong>Discussion: </strong>Large heterogeneity exists between LMMs in identifying inaccurate pseudo-information. Replication in LMM output may constitute a significant hurdle in their application.</p><p><strong>Conclusion: </strong>The accuracy of LMMs needs to be further improved until they can be reliably used by patients for health-related online information and as assistant tools for health information and library services workers without restriction.</p>","PeriodicalId":47580,"journal":{"name":"Health Information and Libraries Journal","volume":" ","pages":""},"PeriodicalIF":2.2000,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Health Information and Libraries Journal","FirstCategoryId":"91","ListUrlMain":"https://doi.org/10.1111/hir.12569","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"INFORMATION SCIENCE & LIBRARY SCIENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Open-access scientific research is an essential source of health-related information and self-education. Artificial intelligence-based large language models (LMMs) may be used to identify erroneous health information.

Objective: To investigate to what extent LMMs can be used to identify pseudo-information.

Methods: Four common LMM applications (ChatGPT-4o, Claude 3.5 Sonnet, Gemini and Copilot) were used to investigate their capability to indicate erroneous information provided in an open-access article.

Results: Initially, ChatGPT-4o and Claude were able to mark the provided article as an unreliable information source, identifying most of the inaccuracy problems. The assessments provided by Gemini and Copilot were inaccurate, as several critical aspects were not identified or were misinterpreted. During the validation phase, the initially accurate assessment of ChatGPT-4o was not reproducible, and only Claude was able to detect several critical issues in this phase. The verdicts of Copilot and Gemini remained largely unaltered.

Discussion: Large heterogeneity exists between LMMs in identifying inaccurate pseudo-information. Replication in LMM output may constitute a significant hurdle in their application.

Conclusion: The accuracy of LMMs needs to be further improved until they can be reliably used by patients for health-related online information and as assistant tools for health information and library services workers without restriction.

使用大型语言模型识别伪信息:对健康信息的影响。
背景:开放获取的科学研究是健康相关信息和自我教育的重要来源。基于人工智能的大语言模型(lmm)可用于识别错误的健康信息。目的:探讨lmm在多大程度上可用于鉴别假信息。方法:使用四种常见的LMM应用程序(chatgpt - 40、Claude 3.5 Sonnet、Gemini和Copilot)来研究它们对开放获取文章中提供的错误信息的指示能力。结果:最初,chatgpt - 40和Claude能够将所提供的文章标记为不可靠的信息源,识别出大多数不准确的问题。Gemini和Copilot提供的评估是不准确的,因为几个关键方面没有被识别或被误解。在验证阶段,chatgpt - 40最初的准确评估是不可重复的,只有Claude能够在这个阶段检测到几个关键问题。副驾驶和双子座的判决基本上没有改变。讨论:lmm在识别不准确伪信息方面存在较大的异质性。LMM输出中的复制可能构成其应用程序中的一个重大障碍。结论:lmm的准确性有待进一步提高,直至患者可以不受限制地可靠地将其用于健康相关的在线信息,并作为卫生信息和图书馆工作人员的辅助工具。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Health Information and Libraries Journal
Health Information and Libraries Journal INFORMATION SCIENCE & LIBRARY SCIENCE-
CiteScore
6.70
自引率
10.50%
发文量
52
期刊介绍: Health Information and Libraries Journal (HILJ) provides practitioners, researchers, and students in library and health professions an international and interdisciplinary forum. Its objectives are to encourage discussion and to disseminate developments at the frontiers of information management and libraries. A major focus is communicating practices that are evidence based both in managing information and in supporting health care. The Journal encompasses: - Identifying health information needs and uses - Managing programmes and services in the changing health environment - Information technology and applications in health - Educating and training health information professionals - Outreach to health user groups
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信