Elena Ierardi, J Chris Eilbeck, Frederike van Wijck, Myzoon Ali, Fiona Coupar
{"title":"Data mining versus manual screening to select papers for inclusion in systematic reviews: a novel method to increase efficiency.","authors":"Elena Ierardi, J Chris Eilbeck, Frederike van Wijck, Myzoon Ali, Fiona Coupar","doi":"10.1097/MRR.0000000000000595","DOIUrl":null,"url":null,"abstract":"<p><p>Systematic reviews rely on identification of studies, initially through electronic searches yielding potentially thousands of studies, and then reviewer-led screening studies for inclusion. This standard method is time- and resource-intensive. We designed and applied an algorithm written in Python involving computer-aided identification of keywords within each paper for an exemplar systematic review of arm impairment after stroke. The standard method involved reading each abstract searching for these keywords. We compared the methods in terms of accuracy in identification of keywords, abstracts' eligibility, and time taken to make a decision about eligibility. For external validation, we adapted the algorithm for a different systematic review, and compared eligible studies using the algorithm with those included in that review. For the exemplar systematic review, the algorithm failed on 72 out of 2,789 documents retrieved (2.6%). Both methods identified the same 610 studies for inclusion. Based on a sample of 21 randomly selected abstracts, the standard screening took 1.58 ± 0.26 min per abstract. Computer output screening took 0.43 ± 0.14 min per abstract. The mean difference between the two methods was 1.15 min ( P < 0.0001), saving 73% per abstract. For the other systematic review, use of the algorithm resulted in the same studies being identified. One study was excluded based on the interpretation of the comparison intervention. Our purpose-built software was an accurate and significantly time-saving method for identifying eligible abstracts for inclusion in systematic reviews. This novel method could be adapted for other systematic reviews in future for the benefit of authors, reviewers and editors.</p>","PeriodicalId":14301,"journal":{"name":"International Journal of Rehabilitation Research","volume":null,"pages":null},"PeriodicalIF":1.5000,"publicationDate":"2023-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Rehabilitation Research","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1097/MRR.0000000000000595","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2023/7/24 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"REHABILITATION","Score":null,"Total":0}
引用次数: 0
Abstract
Systematic reviews rely on identification of studies, initially through electronic searches yielding potentially thousands of studies, and then reviewer-led screening studies for inclusion. This standard method is time- and resource-intensive. We designed and applied an algorithm written in Python involving computer-aided identification of keywords within each paper for an exemplar systematic review of arm impairment after stroke. The standard method involved reading each abstract searching for these keywords. We compared the methods in terms of accuracy in identification of keywords, abstracts' eligibility, and time taken to make a decision about eligibility. For external validation, we adapted the algorithm for a different systematic review, and compared eligible studies using the algorithm with those included in that review. For the exemplar systematic review, the algorithm failed on 72 out of 2,789 documents retrieved (2.6%). Both methods identified the same 610 studies for inclusion. Based on a sample of 21 randomly selected abstracts, the standard screening took 1.58 ± 0.26 min per abstract. Computer output screening took 0.43 ± 0.14 min per abstract. The mean difference between the two methods was 1.15 min ( P < 0.0001), saving 73% per abstract. For the other systematic review, use of the algorithm resulted in the same studies being identified. One study was excluded based on the interpretation of the comparison intervention. Our purpose-built software was an accurate and significantly time-saving method for identifying eligible abstracts for inclusion in systematic reviews. This novel method could be adapted for other systematic reviews in future for the benefit of authors, reviewers and editors.
系统评价依赖于对研究的识别,最初通过电子搜索产生可能数以千计的研究,然后由审稿人主导筛选研究以纳入。这种标准方法耗时耗力。我们设计并应用了一个用Python编写的算法,涉及计算机辅助识别每篇论文中的关键字,用于中风后手臂损伤的范例系统综述。标准的方法包括阅读每个摘要,搜索这些关键词。我们从关键词识别的准确性、摘要的合格性和决定是否合格所花费的时间三个方面对这些方法进行了比较。为了进行外部验证,我们将该算法用于不同的系统综述,并将使用该算法的符合条件的研究与该综述中包含的研究进行了比较。对于范例系统评价,该算法在检索的2,789份文件中有72份(2.6%)失败。两种方法都确定了同样的610项研究。以随机抽取的21篇摘要为样本,标准筛选时间为1.58±0.26分钟。计算机输出筛选每篇摘要耗时0.43±0.14 min。两种方法的平均差异为1.15 min (P
期刊介绍:
International Journal of Rehabilitation Research is a quarterly, peer-reviewed, interdisciplinary forum for the publication of research into functioning, disability and contextual factors experienced by persons of all ages in both developed and developing societies. The wealth of information offered makes the journal a valuable resource for researchers, practitioners, and administrators in such fields as rehabilitation medicine, outcome measurement nursing, social and vocational rehabilitation/case management, return to work, special education, social policy, social work and social welfare, sociology, psychology, psychiatry assistive technology and environmental factors/disability. Areas of interest include functioning and disablement throughout the life cycle; rehabilitation programmes for persons with physical, sensory, mental and developmental disabilities; measurement of functioning and disability; special education and vocational rehabilitation; equipment access and transportation; information technology; independent living; consumer, legal, economic and sociopolitical aspects of functioning, disability and contextual factors.