Evaluating GPT Models for Automated Literature Screening in Wastewater-Based Epidemiology

IF 6.7 Q1 ENGINEERING, ENVIRONMENTAL
Kaseba Chibwe, David Mantilla-Calderon and Fangqiong Ling*, 
{"title":"Evaluating GPT Models for Automated Literature Screening in Wastewater-Based Epidemiology","authors":"Kaseba Chibwe,&nbsp;David Mantilla-Calderon and Fangqiong Ling*,&nbsp;","doi":"10.1021/acsenvironau.4c0004210.1021/acsenvironau.4c00042","DOIUrl":null,"url":null,"abstract":"<p >Methods to quantitatively synthesize findings across multiple studies is an emerging need in wastewater-based epidemiology (WBE), where disease tracking through wastewater analysis is performed at broad geographical locations using various techniques to facilitate public health responses. Meta-analysis provides a rigorous statistical procedure for research synthesis, yet the manual process of screening large volumes of literature remains a hurdle for its application in timely evidence-based public health responses. Here, we evaluated the performance of GPT-3, GPT-3.5, and GPT-4 models in automated screening of publications for meta-analysis in the WBE literature. We show that the chat completion model in GPT-4 accurately differentiates papers that contain original data from those that did not with texts of the Abstract as the input at a Precision of 0.96 and Recall of 1.00, exceeding current quality standards for manual screening (Recall = 0.95) while costing less than $0.01 per paper. GPT models performed less accurately in detecting studies reporting relevant sampling location, highlighting the value of maintaining human intervention in AI-assisted literature screening. Importantly, we show that certain formulation and model choices generated nonsensical answers to the screening tasks, while others did not, urging the attention to robustness when employing AI-assisted literature screening. This study provided novel performance evaluation data on GPT models for document screening as a step in meta-analysis, suggesting AI-assisted literature screening a useful complementary technique to speed up research synthesis in WBE.</p>","PeriodicalId":29801,"journal":{"name":"ACS Environmental Au","volume":"5 1","pages":"61–68 61–68"},"PeriodicalIF":6.7000,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://pubs.acs.org/doi/epdf/10.1021/acsenvironau.4c00042","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Environmental Au","FirstCategoryId":"1085","ListUrlMain":"https://pubs.acs.org/doi/10.1021/acsenvironau.4c00042","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ENVIRONMENTAL","Score":null,"Total":0}
引用次数: 0

Abstract

Methods to quantitatively synthesize findings across multiple studies is an emerging need in wastewater-based epidemiology (WBE), where disease tracking through wastewater analysis is performed at broad geographical locations using various techniques to facilitate public health responses. Meta-analysis provides a rigorous statistical procedure for research synthesis, yet the manual process of screening large volumes of literature remains a hurdle for its application in timely evidence-based public health responses. Here, we evaluated the performance of GPT-3, GPT-3.5, and GPT-4 models in automated screening of publications for meta-analysis in the WBE literature. We show that the chat completion model in GPT-4 accurately differentiates papers that contain original data from those that did not with texts of the Abstract as the input at a Precision of 0.96 and Recall of 1.00, exceeding current quality standards for manual screening (Recall = 0.95) while costing less than $0.01 per paper. GPT models performed less accurately in detecting studies reporting relevant sampling location, highlighting the value of maintaining human intervention in AI-assisted literature screening. Importantly, we show that certain formulation and model choices generated nonsensical answers to the screening tasks, while others did not, urging the attention to robustness when employing AI-assisted literature screening. This study provided novel performance evaluation data on GPT models for document screening as a step in meta-analysis, suggesting AI-assisted literature screening a useful complementary technique to speed up research synthesis in WBE.

求助全文
约1分钟内获得全文 求助全文
来源期刊
ACS Environmental Au
ACS Environmental Au 环境科学-
CiteScore
7.10
自引率
0.00%
发文量
0
期刊介绍: ACS Environmental Au is an open access journal which publishes experimental research and theoretical results in all aspects of environmental science and technology both pure and applied. Short letters comprehensive articles reviews and perspectives are welcome in the following areas:Alternative EnergyAnthropogenic Impacts on Atmosphere Soil or WaterBiogeochemical CyclingBiomass or Wastes as ResourcesContaminants in Aquatic and Terrestrial EnvironmentsEnvironmental Data ScienceEcotoxicology and Public HealthEnergy and ClimateEnvironmental Modeling Processes and Measurement Methods and TechnologiesEnvironmental Nanotechnology and BiotechnologyGreen ChemistryGreen Manufacturing and EngineeringRisk assessment Regulatory Frameworks and Life-Cycle AssessmentsTreatment and Resource Recovery and Waste Management
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信