{"title":"Are there accurate and legitimate ways to machine-quantify predatoriness, or an urgent need for an automated online tool?","authors":"Bor Luen Tang","doi":"10.1080/08989621.2023.2253425","DOIUrl":null,"url":null,"abstract":"<p><p>Yamada and Teixeira da Silva voiced valid concerns with the inadequacies of an online machine learning-based tool to detect predatory journals, and stressed on the urgent need for an automated, open, online-based semi-quantitative system that measures \"predatoriness\". We agree that the said machine learning-based tool lacks accuracy in its demarcation and identification of journals outside those already found within existing black and white lists, and that its use could have undesirable impact on the community. We note further that the key characteristic of journals being predatory, namely a lack of stringent peer review, would normally not have the visibility necessary for training and informing machine learning-based online tools. This, together with the gray zone of inadequate scholarly practice and the plurality in authors' perception of predatoriness, makes it desirable for any machine-based, quantitative assessment to be complemented or moderated by a community-based, qualitative assessment that would do more justice to both journals and authors.</p>","PeriodicalId":50927,"journal":{"name":"Accountability in Research-Policies and Quality Assurance","volume":" ","pages":"182-187"},"PeriodicalIF":2.8000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Accountability in Research-Policies and Quality Assurance","FirstCategoryId":"98","ListUrlMain":"https://doi.org/10.1080/08989621.2023.2253425","RegionNum":1,"RegionCategory":"哲学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2023/8/31 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"MEDICAL ETHICS","Score":null,"Total":0}
引用次数: 0
Abstract
Yamada and Teixeira da Silva voiced valid concerns with the inadequacies of an online machine learning-based tool to detect predatory journals, and stressed on the urgent need for an automated, open, online-based semi-quantitative system that measures "predatoriness". We agree that the said machine learning-based tool lacks accuracy in its demarcation and identification of journals outside those already found within existing black and white lists, and that its use could have undesirable impact on the community. We note further that the key characteristic of journals being predatory, namely a lack of stringent peer review, would normally not have the visibility necessary for training and informing machine learning-based online tools. This, together with the gray zone of inadequate scholarly practice and the plurality in authors' perception of predatoriness, makes it desirable for any machine-based, quantitative assessment to be complemented or moderated by a community-based, qualitative assessment that would do more justice to both journals and authors.
Yamada和Teixeira da Silva对一种基于在线机器学习的工具在检测掠夺性期刊方面的不足表示了合理的担忧,并强调迫切需要一种自动化、开放、基于在线的半定量系统来衡量“掠夺性”。我们同意上述基于机器学习的工具在划分和识别现有黑白名单之外的期刊方面缺乏准确性,并且它的使用可能会对社区产生不良影响。我们进一步注意到,期刊掠夺性的关键特征,即缺乏严格的同行评议,通常不具备培训和通知基于机器学习的在线工具所必需的可见性。这一点,再加上学术实践不足的灰色地带和作者对掠夺性看法的多样性,使得任何基于机器的定量评估都需要由基于社区的定性评估来补充或缓和,这将对期刊和作者都更加公正。
期刊介绍:
Accountability in Research: Policies and Quality Assurance is devoted to the examination and critical analysis of systems for maximizing integrity in the conduct of research. It provides an interdisciplinary, international forum for the development of ethics, procedures, standards policies, and concepts to encourage the ethical conduct of research and to enhance the validity of research results.
The journal welcomes views on advancing the integrity of research in the fields of general and multidisciplinary sciences, medicine, law, economics, statistics, management studies, public policy, politics, sociology, history, psychology, philosophy, ethics, and information science.
All submitted manuscripts are subject to initial appraisal by the Editor, and if found suitable for further consideration, to peer review by independent, anonymous expert referees.