利用人工智能对神经外科研究文章进行同行评审。

IF 2.5 3区 医学 Q2 CLINICAL NEUROLOGY
Ali A Mohamed, Daniel Colome, Jack Yang, Emma C Sargent, Gabriel Flores-Milan, Zachary Sorrentino, Akshay Sharma, Owoicho Adogwa, Stephen Pirris, Brandon Lucke-Wold
{"title":"利用人工智能对神经外科研究文章进行同行评审。","authors":"Ali A Mohamed, Daniel Colome, Jack Yang, Emma C Sargent, Gabriel Flores-Milan, Zachary Sorrentino, Akshay Sharma, Owoicho Adogwa, Stephen Pirris, Brandon Lucke-Wold","doi":"10.1007/s10143-025-03783-9","DOIUrl":null,"url":null,"abstract":"<p><p>The traditional peer review process is time-consuming and can delay the dissemination of critical research. This study evaluates the effectiveness of artificial intelligence (AI) in predicting the acceptance or rejection of neurosurgical manuscripts, offering a possible solution to optimize the process. Neurosurgical preprints from Preprint.org and medRxiv.org were analyzed. Published preprints were compared to those presumed not accepted after remaining on preprint servers for over 12 months. Each article was uploaded to ChatGPT 4o, Gemini, and Copilot with the prompt: \"Based on the literature up to the date this article was posted, will it be accepted or rejected for publication following peer review? Please provide a yes or no answer.\" AI predictive accuracy and journal metrics were assessed between preprints that were accepted or presumed to be not accepted. A total of 51 preprints (31 skull base, 20 spine) were included, with 28 published and 23 presumed not accepted. The average impact factor and cite score for accepted preprints were 4.36 ± 2.07 and 6.38 ± 3.67 for skull base and 3.48 ± 1.08 and 4.83 ± 1.37 for spine topics. Across all AI models, there were no significant differences in journal metrics between preprints predicted to be accepted or not accepted (p > 0.05). Overall, AI models had significantly low performance, with accuracy ranging from 40 to 66.67% (p < 0.001). Current AI models exhibit moderate accuracy in predicting peer review outcomes. Future AI models, developed in collaboration with journals and with authors' consent, could access a more balanced dataset, enhancing accuracy and streamlining the peer review process.</p>","PeriodicalId":19184,"journal":{"name":"Neurosurgical Review","volume":"48 1","pages":"631"},"PeriodicalIF":2.5000,"publicationDate":"2025-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Leveraging artificial intelligence in the peer review of neurosurgical research articles.\",\"authors\":\"Ali A Mohamed, Daniel Colome, Jack Yang, Emma C Sargent, Gabriel Flores-Milan, Zachary Sorrentino, Akshay Sharma, Owoicho Adogwa, Stephen Pirris, Brandon Lucke-Wold\",\"doi\":\"10.1007/s10143-025-03783-9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>The traditional peer review process is time-consuming and can delay the dissemination of critical research. This study evaluates the effectiveness of artificial intelligence (AI) in predicting the acceptance or rejection of neurosurgical manuscripts, offering a possible solution to optimize the process. Neurosurgical preprints from Preprint.org and medRxiv.org were analyzed. Published preprints were compared to those presumed not accepted after remaining on preprint servers for over 12 months. Each article was uploaded to ChatGPT 4o, Gemini, and Copilot with the prompt: \\\"Based on the literature up to the date this article was posted, will it be accepted or rejected for publication following peer review? Please provide a yes or no answer.\\\" AI predictive accuracy and journal metrics were assessed between preprints that were accepted or presumed to be not accepted. A total of 51 preprints (31 skull base, 20 spine) were included, with 28 published and 23 presumed not accepted. The average impact factor and cite score for accepted preprints were 4.36 ± 2.07 and 6.38 ± 3.67 for skull base and 3.48 ± 1.08 and 4.83 ± 1.37 for spine topics. Across all AI models, there were no significant differences in journal metrics between preprints predicted to be accepted or not accepted (p > 0.05). Overall, AI models had significantly low performance, with accuracy ranging from 40 to 66.67% (p < 0.001). Current AI models exhibit moderate accuracy in predicting peer review outcomes. Future AI models, developed in collaboration with journals and with authors' consent, could access a more balanced dataset, enhancing accuracy and streamlining the peer review process.</p>\",\"PeriodicalId\":19184,\"journal\":{\"name\":\"Neurosurgical Review\",\"volume\":\"48 1\",\"pages\":\"631\"},\"PeriodicalIF\":2.5000,\"publicationDate\":\"2025-09-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurosurgical Review\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1007/s10143-025-03783-9\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"CLINICAL NEUROLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurosurgical Review","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s10143-025-03783-9","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"CLINICAL NEUROLOGY","Score":null,"Total":0}
引用次数: 0

摘要

传统的同行评议过程非常耗时,而且会延迟批判性研究的传播。本研究评估了人工智能(AI)在预测神经外科稿件的接受或拒绝方面的有效性,为优化这一过程提供了可能的解决方案。分析来自Preprint.org和medRxiv.org的神经外科预印本。将已发表的预印本与在预印本服务器上保存超过12个月后假定未被接受的预印本进行比较。每篇文章都被上传到ChatGPT 40、Gemini和Copilot,并提示:“根据本文发布日期的文献,在同行评议后,它是被接受还是被拒绝发表?”请回答是或不是。”人工智能预测准确性和期刊指标在被接受或假定不被接受的预印本之间进行评估。共纳入51篇预印本(31篇颅底,20篇脊柱),其中28篇已发表,23篇推定未被接受。接受预印本的平均影响因子和引用得分分别为颅底(4.36±2.07)和6.38±3.67,脊柱(3.48±1.08)和4.83±1.37。在所有AI模型中,预印本预测被接受或未被接受的期刊指标没有显著差异(p > 0.05)。总体而言,人工智能模型的性能明显较低,准确率在40%到66.67%之间(p
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Leveraging artificial intelligence in the peer review of neurosurgical research articles.

The traditional peer review process is time-consuming and can delay the dissemination of critical research. This study evaluates the effectiveness of artificial intelligence (AI) in predicting the acceptance or rejection of neurosurgical manuscripts, offering a possible solution to optimize the process. Neurosurgical preprints from Preprint.org and medRxiv.org were analyzed. Published preprints were compared to those presumed not accepted after remaining on preprint servers for over 12 months. Each article was uploaded to ChatGPT 4o, Gemini, and Copilot with the prompt: "Based on the literature up to the date this article was posted, will it be accepted or rejected for publication following peer review? Please provide a yes or no answer." AI predictive accuracy and journal metrics were assessed between preprints that were accepted or presumed to be not accepted. A total of 51 preprints (31 skull base, 20 spine) were included, with 28 published and 23 presumed not accepted. The average impact factor and cite score for accepted preprints were 4.36 ± 2.07 and 6.38 ± 3.67 for skull base and 3.48 ± 1.08 and 4.83 ± 1.37 for spine topics. Across all AI models, there were no significant differences in journal metrics between preprints predicted to be accepted or not accepted (p > 0.05). Overall, AI models had significantly low performance, with accuracy ranging from 40 to 66.67% (p < 0.001). Current AI models exhibit moderate accuracy in predicting peer review outcomes. Future AI models, developed in collaboration with journals and with authors' consent, could access a more balanced dataset, enhancing accuracy and streamlining the peer review process.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurosurgical Review
Neurosurgical Review 医学-临床神经学
CiteScore
5.60
自引率
7.10%
发文量
191
审稿时长
6-12 weeks
期刊介绍: The goal of Neurosurgical Review is to provide a forum for comprehensive reviews on current issues in neurosurgery. Each issue contains up to three reviews, reflecting all important aspects of one topic (a disease or a surgical approach). Comments by a panel of experts within the same issue complete the topic. By providing comprehensive coverage of one topic per issue, Neurosurgical Review combines the topicality of professional journals with the indepth treatment of a monograph. Original papers of high quality are also welcome.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信