同行评议——人工智能能帮上忙吗?

IF 3.4 2区 农林科学 Q2 FOOD SCIENCE & TECHNOLOGY
Richard Hartel
{"title":"同行评议——人工智能能帮上忙吗?","authors":"Richard Hartel","doi":"10.1111/1750-3841.70537","DOIUrl":null,"url":null,"abstract":"<p>One of my biggest concerns related to journal editing is the inconsistency in peer review. The fate of a manuscript often seems to depend on who is assigned as AE and who is called on to provide peer review comments. Let me qualify that. A really top-end paper will get accepted no matter what, and a really poor submission will be rejected accordingly. It's the manuscripts that fall in between where there is variability.</p><p>In my experience, each editor has a slightly different bar by which to judge a manuscript. First, the scientific editors evaluate each manuscript as assigned. They look for a variety of parameters to judge: new and novel work, robust experimental design with appropriate replications and statistical assessment, clear and concise presentation and discussion, appropriate abstract and conclusions, low similarity index, among others.</p><p>Each editor has their own bar regarding grounds for rejection, some much lower than others. This is evident in the rejection rates of individual editors. On average, SEs reject about half of the manuscripts that come to them, the other half being forwarded to AEs to process. But the variation among SEs runs from a low of 3.6% rejected immediately to as high as 84%. That is, one SE essentially sends everything to an AE, while another sends less than 2 out of 10 forward. Although some of this variation may be related to the topic area of interest for each SE, the variation among editors is still high.</p><p>Of those manuscripts sent to an AE, on average, about 50% more are rejected, but again, each AE has their own bar. A couple of AEs reject only about 15% of their manuscripts, while a couple of others reject over 80%. One AE rejects virtually everything sent to them.</p><p>Then we need to factor in the quality of the peer reviewers who evaluate a manuscript. Again, in my experience, this is widely variable; some reviewers provide 2–3 pages of insightful commentary, while others barely provide a sentence with little to no justification of their decision. We strive to get three good critical reviews for each manuscript, but often we don't reach that goal. AEs get to rate each review, so we know who does a consistently good job and who not to trust as much.</p><p>It's these numbers that raise concerns about the equity of our peer review system.</p><p>Some have suggested that we use AI to help make the peer review process easier and perhaps more consistent. But there are some huge roadblocks to this practice, and we still do not allow its use in peer review. The main issue is that it is not appropriate to feed a manuscript into a program like ChatGPT, since that breaks the confidentiality of the peer review process itself. And still, can you really trust everything that ChatGPT says?</p><p>Some LLMs do not require material to be added to the database, like Copilot. Still, we do not allow Copilot's use to conduct the review. While AI can provide interesting insights, depending on the questions asked, human judgment is still required to differentiate the fine points.</p><p>Over the past year or so, we have started asking reviewers to declare any AI usage when they submit their reviews. We recently reviewed their responses to this question. The vast majority (over 95%) said no AI sources were used. Those who declared its use stated it was simply for polishing the grammar of their own review comments; one of these even said they wrote their review in their native language and then used AI to translate to English.</p><p>Still, AI is here, and applications are growing rapidly. Why not harness the power of AI to assist in peer review and other aspects of scholarly publishing? This is, in fact, the topic of Peer Review Week, an annual event scheduled for September 15–19 this year. This event, a series of meetings, podcasts, webinars, and blog posts from a range of experts, is intended to “foster collaboration and awareness” on a new topic chosen each year. It's not surprising that the use of AI was chosen as the topic for this year.</p><p>The main points for discussion include ethics, literacy and training, fairness and transparency, human and machine judgment to provide a basis for how AI can best be used to enhance peer review. Hopefully, some new guidance about AI use in peer review will emerge from these discussions.</p><p>Whether we're ready or not, AI is here. Although reviewers generally do not use AI to assist them with the critical aspects of evaluating science, we need to find ways to optimize its use without breaching confidentiality. And maintain the most important aspect of peer review, human critical thinking.</p><p>Sincerely,</p><p></p><p>Rich Hartel, PhD</p><p>Editor in Chief, <i>Journal of Food Science</i></p><p>Professor, University of Wisconsin–Madison</p>","PeriodicalId":193,"journal":{"name":"Journal of Food Science","volume":"90 9","pages":""},"PeriodicalIF":3.4000,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ift.onlinelibrary.wiley.com/doi/epdf/10.1111/1750-3841.70537","citationCount":"0","resultStr":"{\"title\":\"Peer Review—Can AI Help?\",\"authors\":\"Richard Hartel\",\"doi\":\"10.1111/1750-3841.70537\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>One of my biggest concerns related to journal editing is the inconsistency in peer review. The fate of a manuscript often seems to depend on who is assigned as AE and who is called on to provide peer review comments. Let me qualify that. A really top-end paper will get accepted no matter what, and a really poor submission will be rejected accordingly. It's the manuscripts that fall in between where there is variability.</p><p>In my experience, each editor has a slightly different bar by which to judge a manuscript. First, the scientific editors evaluate each manuscript as assigned. They look for a variety of parameters to judge: new and novel work, robust experimental design with appropriate replications and statistical assessment, clear and concise presentation and discussion, appropriate abstract and conclusions, low similarity index, among others.</p><p>Each editor has their own bar regarding grounds for rejection, some much lower than others. This is evident in the rejection rates of individual editors. On average, SEs reject about half of the manuscripts that come to them, the other half being forwarded to AEs to process. But the variation among SEs runs from a low of 3.6% rejected immediately to as high as 84%. That is, one SE essentially sends everything to an AE, while another sends less than 2 out of 10 forward. Although some of this variation may be related to the topic area of interest for each SE, the variation among editors is still high.</p><p>Of those manuscripts sent to an AE, on average, about 50% more are rejected, but again, each AE has their own bar. A couple of AEs reject only about 15% of their manuscripts, while a couple of others reject over 80%. One AE rejects virtually everything sent to them.</p><p>Then we need to factor in the quality of the peer reviewers who evaluate a manuscript. Again, in my experience, this is widely variable; some reviewers provide 2–3 pages of insightful commentary, while others barely provide a sentence with little to no justification of their decision. We strive to get three good critical reviews for each manuscript, but often we don't reach that goal. AEs get to rate each review, so we know who does a consistently good job and who not to trust as much.</p><p>It's these numbers that raise concerns about the equity of our peer review system.</p><p>Some have suggested that we use AI to help make the peer review process easier and perhaps more consistent. But there are some huge roadblocks to this practice, and we still do not allow its use in peer review. The main issue is that it is not appropriate to feed a manuscript into a program like ChatGPT, since that breaks the confidentiality of the peer review process itself. And still, can you really trust everything that ChatGPT says?</p><p>Some LLMs do not require material to be added to the database, like Copilot. Still, we do not allow Copilot's use to conduct the review. While AI can provide interesting insights, depending on the questions asked, human judgment is still required to differentiate the fine points.</p><p>Over the past year or so, we have started asking reviewers to declare any AI usage when they submit their reviews. We recently reviewed their responses to this question. The vast majority (over 95%) said no AI sources were used. Those who declared its use stated it was simply for polishing the grammar of their own review comments; one of these even said they wrote their review in their native language and then used AI to translate to English.</p><p>Still, AI is here, and applications are growing rapidly. Why not harness the power of AI to assist in peer review and other aspects of scholarly publishing? This is, in fact, the topic of Peer Review Week, an annual event scheduled for September 15–19 this year. This event, a series of meetings, podcasts, webinars, and blog posts from a range of experts, is intended to “foster collaboration and awareness” on a new topic chosen each year. It's not surprising that the use of AI was chosen as the topic for this year.</p><p>The main points for discussion include ethics, literacy and training, fairness and transparency, human and machine judgment to provide a basis for how AI can best be used to enhance peer review. Hopefully, some new guidance about AI use in peer review will emerge from these discussions.</p><p>Whether we're ready or not, AI is here. Although reviewers generally do not use AI to assist them with the critical aspects of evaluating science, we need to find ways to optimize its use without breaching confidentiality. And maintain the most important aspect of peer review, human critical thinking.</p><p>Sincerely,</p><p></p><p>Rich Hartel, PhD</p><p>Editor in Chief, <i>Journal of Food Science</i></p><p>Professor, University of Wisconsin–Madison</p>\",\"PeriodicalId\":193,\"journal\":{\"name\":\"Journal of Food Science\",\"volume\":\"90 9\",\"pages\":\"\"},\"PeriodicalIF\":3.4000,\"publicationDate\":\"2025-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ift.onlinelibrary.wiley.com/doi/epdf/10.1111/1750-3841.70537\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Food Science\",\"FirstCategoryId\":\"97\",\"ListUrlMain\":\"https://ift.onlinelibrary.wiley.com/doi/10.1111/1750-3841.70537\",\"RegionNum\":2,\"RegionCategory\":\"农林科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"FOOD SCIENCE & TECHNOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Food Science","FirstCategoryId":"97","ListUrlMain":"https://ift.onlinelibrary.wiley.com/doi/10.1111/1750-3841.70537","RegionNum":2,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"FOOD SCIENCE & TECHNOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

我对期刊编辑最大的担忧之一是同行评议的不一致性。手稿的命运似乎往往取决于谁被指定为AE,谁被要求提供同行评审意见。我来解释一下。一篇真正顶级的论文无论如何都会被接受,而一篇非常糟糕的论文也会被拒绝。这是介于两者之间的手稿,有可变性。根据我的经验,每个编辑评判稿件的标准都略有不同。首先,科学编辑按指定的顺序评估每份稿件。他们寻找各种各样的参数来判断:新的和新颖的工作,稳健的实验设计,适当的重复和统计评估,清晰简洁的演示和讨论,适当的抽象和结论,低相似指数等。对于退稿理由,每个编辑都有自己的标准,有些要低得多。这在个别编辑的退稿率上是显而易见的。平均而言,se拒绝了大约一半的手稿,另一半被转发给ae进行处理。但se之间的差异从3.6%的低到84%的高不等。也就是说,一个SE基本上将所有内容发送给AE,而另一个SE转发的内容不到2 / 10。尽管其中一些差异可能与每个SE感兴趣的主题领域有关,但编辑之间的差异仍然很大。在那些送到AE的手稿中,平均有50%以上被拒绝,但是每个AE都有自己的酒吧。一些ae只拒绝了大约15%的手稿,而另一些则拒绝了80%以上的手稿。一个AE几乎拒绝所有发送给他们的东西。然后我们需要考虑评估稿件的同行审稿人的质量。再一次,根据我的经验,这是有很大差异的;一些评论者提供了2-3页富有洞察力的评论,而其他人几乎没有提供一句话来证明他们的决定。我们努力为每篇稿件争取三篇好的评论,但我们常常达不到这个目标。我们会给每条评论打分,这样我们就知道谁一直做得很好,谁不值得信任。正是这些数字引起了人们对我们同行评议制度公平性的担忧。一些人建议我们使用人工智能来帮助使同行评审过程更容易,也许更一致。但是这种做法有一些巨大的障碍,我们仍然不允许在同行评审中使用它。主要的问题是,将手稿输入ChatGPT这样的程序是不合适的,因为这会破坏同行评审过程本身的机密性。然而,你真的能相信ChatGPT所说的一切吗?一些法学硕士不需要向数据库中添加材料,比如Copilot。不过,我们不允许副驾驶进行审查。虽然人工智能可以根据提出的问题提供有趣的见解,但仍然需要人类的判断来区分细微之处。在过去一年左右的时间里,我们开始要求评论者在提交他们的评论时声明任何人工智能的使用。我们最近回顾了他们对这个问题的回答。绝大多数(超过95%)表示没有使用人工智能来源。那些宣称使用它的人说它只是为了完善他们自己的复习评论的语法;其中一人甚至说他们用母语写评论,然后用人工智能翻译成英语。然而,人工智能就在这里,应用正在迅速增长。为什么不利用人工智能的力量来协助同行评审和学术出版的其他方面呢?事实上,这正是“同行评议周”的主题,该活动定于今年9月15日至19日举行。这一活动包括一系列会议、播客、网络研讨会和一系列专家的博客文章,旨在就每年选定的一个新主题“促进合作和意识”。人工智能的使用被选为今年的主题并不奇怪。讨论的要点包括道德、素养和培训、公平和透明度、人类和机器的判断,为如何最好地利用人工智能来加强同行评审提供基础。希望这些讨论能够产生一些关于同行评审中使用人工智能的新指导。不管我们是否准备好了,人工智能就在这里。虽然审稿人通常不会使用人工智能来协助他们评估科学的关键方面,但我们需要找到在不违反保密的情况下优化其使用的方法。保持同行评议最重要的方面,人类的批判性思维。尊敬的Rich Hartel,威斯康星大学麦迪逊分校食品科学杂志教授,首席博士
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Peer Review—Can AI Help?

Peer Review—Can AI Help?

Peer Review—Can AI Help?

Peer Review—Can AI Help?

One of my biggest concerns related to journal editing is the inconsistency in peer review. The fate of a manuscript often seems to depend on who is assigned as AE and who is called on to provide peer review comments. Let me qualify that. A really top-end paper will get accepted no matter what, and a really poor submission will be rejected accordingly. It's the manuscripts that fall in between where there is variability.

In my experience, each editor has a slightly different bar by which to judge a manuscript. First, the scientific editors evaluate each manuscript as assigned. They look for a variety of parameters to judge: new and novel work, robust experimental design with appropriate replications and statistical assessment, clear and concise presentation and discussion, appropriate abstract and conclusions, low similarity index, among others.

Each editor has their own bar regarding grounds for rejection, some much lower than others. This is evident in the rejection rates of individual editors. On average, SEs reject about half of the manuscripts that come to them, the other half being forwarded to AEs to process. But the variation among SEs runs from a low of 3.6% rejected immediately to as high as 84%. That is, one SE essentially sends everything to an AE, while another sends less than 2 out of 10 forward. Although some of this variation may be related to the topic area of interest for each SE, the variation among editors is still high.

Of those manuscripts sent to an AE, on average, about 50% more are rejected, but again, each AE has their own bar. A couple of AEs reject only about 15% of their manuscripts, while a couple of others reject over 80%. One AE rejects virtually everything sent to them.

Then we need to factor in the quality of the peer reviewers who evaluate a manuscript. Again, in my experience, this is widely variable; some reviewers provide 2–3 pages of insightful commentary, while others barely provide a sentence with little to no justification of their decision. We strive to get three good critical reviews for each manuscript, but often we don't reach that goal. AEs get to rate each review, so we know who does a consistently good job and who not to trust as much.

It's these numbers that raise concerns about the equity of our peer review system.

Some have suggested that we use AI to help make the peer review process easier and perhaps more consistent. But there are some huge roadblocks to this practice, and we still do not allow its use in peer review. The main issue is that it is not appropriate to feed a manuscript into a program like ChatGPT, since that breaks the confidentiality of the peer review process itself. And still, can you really trust everything that ChatGPT says?

Some LLMs do not require material to be added to the database, like Copilot. Still, we do not allow Copilot's use to conduct the review. While AI can provide interesting insights, depending on the questions asked, human judgment is still required to differentiate the fine points.

Over the past year or so, we have started asking reviewers to declare any AI usage when they submit their reviews. We recently reviewed their responses to this question. The vast majority (over 95%) said no AI sources were used. Those who declared its use stated it was simply for polishing the grammar of their own review comments; one of these even said they wrote their review in their native language and then used AI to translate to English.

Still, AI is here, and applications are growing rapidly. Why not harness the power of AI to assist in peer review and other aspects of scholarly publishing? This is, in fact, the topic of Peer Review Week, an annual event scheduled for September 15–19 this year. This event, a series of meetings, podcasts, webinars, and blog posts from a range of experts, is intended to “foster collaboration and awareness” on a new topic chosen each year. It's not surprising that the use of AI was chosen as the topic for this year.

The main points for discussion include ethics, literacy and training, fairness and transparency, human and machine judgment to provide a basis for how AI can best be used to enhance peer review. Hopefully, some new guidance about AI use in peer review will emerge from these discussions.

Whether we're ready or not, AI is here. Although reviewers generally do not use AI to assist them with the critical aspects of evaluating science, we need to find ways to optimize its use without breaching confidentiality. And maintain the most important aspect of peer review, human critical thinking.

Sincerely,

Rich Hartel, PhD

Editor in Chief, Journal of Food Science

Professor, University of Wisconsin–Madison

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Journal of Food Science
Journal of Food Science 工程技术-食品科技
CiteScore
7.10
自引率
2.60%
发文量
412
审稿时长
3.1 months
期刊介绍: The goal of the Journal of Food Science is to offer scientists, researchers, and other food professionals the opportunity to share knowledge of scientific advancements in the myriad disciplines affecting their work, through a respected peer-reviewed publication. The Journal of Food Science serves as an international forum for vital research and developments in food science. The range of topics covered in the journal include: -Concise Reviews and Hypotheses in Food Science -New Horizons in Food Research -Integrated Food Science -Food Chemistry -Food Engineering, Materials Science, and Nanotechnology -Food Microbiology and Safety -Sensory and Consumer Sciences -Health, Nutrition, and Food -Toxicology and Chemical Food Safety The Journal of Food Science publishes peer-reviewed articles that cover all aspects of food science, including safety and nutrition. Reviews should be 15 to 50 typewritten pages (including tables, figures, and references), should provide in-depth coverage of a narrowly defined topic, and should embody careful evaluation (weaknesses, strengths, explanation of discrepancies in results among similar studies) of all pertinent studies, so that insightful interpretations and conclusions can be presented. Hypothesis papers are especially appropriate in pioneering areas of research or important areas that are afflicted by scientific controversy.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信