Madelin R Siedler, Reem A Mustafa, Lifeng Lin, Rebecca L Morgan, Yngve Falck-Ytter, Philipp Dahm, Shahnaz Sultan, Mohammad Hassan Murad
{"title":"Meta-analysis of continuous outcomes: a user's guide for analysis and interpretation.","authors":"Madelin R Siedler, Reem A Mustafa, Lifeng Lin, Rebecca L Morgan, Yngve Falck-Ytter, Philipp Dahm, Shahnaz Sultan, Mohammad Hassan Murad","doi":"10.1136/bmjebm-2024-113031","DOIUrl":"10.1136/bmjebm-2024-113031","url":null,"abstract":"","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"340-346"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142104128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Su Jin Yim, Sevil Yasar, Nancy Schoenborn, Eddy Lang
{"title":"Expanded disease definitions in Alzheimer's disease and the new era of disease-modifying drugs.","authors":"Su Jin Yim, Sevil Yasar, Nancy Schoenborn, Eddy Lang","doi":"10.1136/bmjebm-2023-112588","DOIUrl":"10.1136/bmjebm-2023-112588","url":null,"abstract":"","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"288-290"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143405685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Matthew G Wilson, Folkert W Asselbergs, Nausheen Saleem, Lelia Jeilani, David Brealey, Matthew R Sydes, Steve Harris
{"title":"Digital integration of research conduct into clinical care: results of the PROSPECTOR randomised feasibility study.","authors":"Matthew G Wilson, Folkert W Asselbergs, Nausheen Saleem, Lelia Jeilani, David Brealey, Matthew R Sydes, Steve Harris","doi":"10.1136/bmjebm-2024-113081","DOIUrl":"10.1136/bmjebm-2024-113081","url":null,"abstract":"<p><strong>Objectives: </strong>To evaluate the feasibility of conducting a clinically integrated randomised comparative effectiveness trial using digital clinical trial infrastructure within an electronic patient record (EPR).</p><p><strong>Design: </strong>A mixed-methods, unblinded, feasibility study of digital clinical trial system incorporating testing of two designs of electronic point-of-care randomisation prompt.</p><p><strong>Setting: </strong>The study was conducted at University College London Hospitals NHS Trust between March and November 2022. The study used a real clinical research question for context, comparing liberal vs restrictive strategies for magnesium supplementation to prevent new-onset atrial fibrillation in critical care.</p><p><strong>Participants: </strong>Adult patients undergoing elective, non-cardiac surgical procedures expecting postoperative admission to critical care were recruited.</p><p><strong>Interventions: </strong>A digital trial system screened participants continuously against eligibility criteria. Participants were automatically randomised (1:1) to (1) magnesium supplementation strategy and (2) one of two electronic randomisation prompt designs (nudge or preference).Electronic point-of-care randomisation prompts displayed to clinicians at regular intervals, inviting them to follow a randomised magnesium supplementation suggestion.</p><p><strong>Main outcome measures: </strong>The primary outcome measure was a composite determination of study design feasibility (including recruitment, technical performance and concordance between the randomised suggestion and the observed clinician action).</p><p><strong>Results: </strong>23 patients were recruited and 11 successfully randomised. The implemented digital systems for automated eligibility screening, randomisation, data collection and follow-up demonstrated technical feasibility. 47 electronic point-of-care randomisation prompts successfully deployed across 11 patients. Clinician actions were concordant with randomised suggestions in 32 prompts (68%).Technical and implementational barriers to delivering the electronic point-of-care randomisation prompts were identified. Patients were followed up to 30 days following discharge from hospital, with no serious adverse events attributable to participation identified.There was insufficient data to make a quantitative determination on the superiority of either prompt design. Clinician feedback suggested the simplified design (nudge) had greater utility.</p><p><strong>Conclusions: </strong>This study demonstrates that digitally embedding clinical trial infrastructure into a site-level EPR and integrating conduct into clinical care is safe and feasible. Future work will focus on improving and expanding the integrated digital trial design across multiple centres.</p><p><strong>Trial registration number: </strong>NCT05149820.</p>","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"323-332"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12505030/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143964069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Anton Barchuk, Niko K Nordlund, Alex L E Halme, Kari A O Tikkinen
{"title":"Evidence categories in systematic assessment of cancer overdiagnosis.","authors":"Anton Barchuk, Niko K Nordlund, Alex L E Halme, Kari A O Tikkinen","doi":"10.1136/bmjebm-2024-113529","DOIUrl":"10.1136/bmjebm-2024-113529","url":null,"abstract":"<p><p>The phenomenon of cancer overdiagnosis, the diagnosis of a malignant tumour that, without detection, would never lead to adverse health effects, has been reported for several cancer types in different populations. There has been an increase in studies focused on overdiagnosis, creating an opportunity to synthesise evidence on specific cancer types. However, studies that systematically assess evidence across different research domains remain scarce, with most of them relying on data from studies that already mentioned overdiagnosis as a potential concern. In this review, we consider several evidence categories that are used to systematically assess the presence and magnitude of overdiagnosis, including (1) data from cancer surveillance, (2) studies exploring the 'true' prevalence of cancer in the population, (3) studies that explore the use of diagnostics and its effect on incidence and mortality and (4) studies that explore changes and progress in cancer management and its effect on cancer mortality. This article highlights the strengths and weaknesses of different evidence categories, provides examples of studies on different cancer types and discusses how these categories can help synthesise evidence on cancer overdiagnosis.</p>","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"333-339"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12505034/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144180727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Identifying actionable statements in Chinese health guidelines: a cross-sectional study.","authors":"Xiangying Ren, Tamara Lotfi, Jiyu Chen, Yuling Lei, Chenyibei Zhou, Wei Zhang, Qiao Huang, Yongbo Wang, Siyu Yan, Shichun Wang, Siyuan Ruan, Wanru Wang, Qiyi Zhang, Xiaomei Yao, Yinghui Jin, Holger J Schuenemann","doi":"10.1136/bmjebm-2024-113050","DOIUrl":"10.1136/bmjebm-2024-113050","url":null,"abstract":"<p><strong>Objective: </strong>The purpose of this study is to validate the taxonomy and framework using Chinese guidelines and identify actionable statements.</p><p><strong>Design and setting: </strong>We searched five databases, to identify the health guidelines from 1 January 2020 to 1 May 2023. Five researchers categorised statements into six types: formal recommendations (Type I) with clear direction and strength, with explicit and direct evidence; good practice statements (GPS) (Type II), actionable in isolation with a significant benefit; remarks (Type III), an inseparable unit belonging to a formal recommendation or GPS that provides additional clarification; research only recommendations (Type IV) for specific populations; implementation considerations, tools and tips (Type V), that describe the how, who, where, what and when, in relation to implementing a recommendation and lacking a direct evidence link; and informal recommendations (Type VI), unrelated to evidence and not meeting GPS criteria.</p><p><strong>Results: </strong>We included 116 guidelines, including 74 Western medicine guidelines, 12 traditional Chinese medicine guidelines and 30 integrated Chinese and Western medicine guidelines. 99 guidelines (85.3%) used the Grading of Recommendations Assessment, Development and Evaluation criteria. Medical specialty societies developed the highest number of guidelines (53.4%). Of all the statements, 4422 statements were extracted from the guidelines. Among them, 2154 (48.7%) were formal recommendations, 197 (4.4%) were GPS, 394 (8.9%) were remarks, 16 (0.4%) were research only recommendations, 1106 (25.0%) were implementation considerations, tools and tips, and 555 (12.6%) were informal recommendations.</p><p><strong>Conclusions: </strong>Up to date, the Chinese guideline developers tend to overestimate the number of formal recommendations and underestimate the number of GPS, remarks, research only recommendations, implementation considerations, tools and tips, and informal recommendations. Thus the current quality of actionable statements in Chinese health guidelines requires further enhancement.</p>","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"305-312"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12505086/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143630009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mohammad Hassan Murad, Zhen Wang, Yngve Falck-Ytter
{"title":"Facilitating GRADE judgements about the inconsistency of effects using a novel visualisation approach.","authors":"Mohammad Hassan Murad, Zhen Wang, Yngve Falck-Ytter","doi":"10.1136/bmjebm-2024-113038","DOIUrl":"10.1136/bmjebm-2024-113038","url":null,"abstract":"","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"347-350"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12505098/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142280158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jin Kyu Kim, Michael Erlano Chua, Tian Ge Li, Mandy Rickard, Armando J Lorenzo
{"title":"Novel AI applications in systematic review: GPT-4 assisted data extraction, analysis, review of bias.","authors":"Jin Kyu Kim, Michael Erlano Chua, Tian Ge Li, Mandy Rickard, Armando J Lorenzo","doi":"10.1136/bmjebm-2024-113066","DOIUrl":"10.1136/bmjebm-2024-113066","url":null,"abstract":"<p><strong>Objective: </strong>To assess custom GPT-4 performance in extracting and evaluating data from medical literature to assist in the systematic review (SR) process.</p><p><strong>Design: </strong>A proof-of-concept comparative study was conducted to assess the accuracy and precision of custom GPT-4 models against human-performed reviews of randomised controlled trials (RCTs).</p><p><strong>Setting: </strong>Four custom GPT-4 models were developed, each specialising in one of the following areas: (1) extraction of study characteristics, (2) extraction of outcomes, (3) extraction of bias assessment domains and (4) evaluation of risk of bias using results from the third GPT-4 model. Model outputs were compared against data from four SRs conducted by human authors. The evaluation focused on accuracy in data extraction, precision in replicating outcomes and agreement levels in risk of bias assessments.</p><p><strong>Participants: </strong>Among four SRs chosen, 43 studies were retrieved for data extraction evaluation. Additionally, 17 RCTs were selected for comparison of risk of bias assessments, where both human comparator SRs and an analogous SR provided assessments for comparison.</p><p><strong>Intervention: </strong>Custom GPT-4 models were deployed to extract data and evaluate risk of bias from selected studies, and their outputs were compared to those generated by human reviewers.</p><p><strong>Main outcome measures: </strong>Concordance rates between GPT-4 outputs and human-performed SRs in data extraction, effect size comparability and inter/intra-rater agreement in risk of bias assessments.</p><p><strong>Results: </strong>When comparing the automatically extracted data to the first table of study characteristics from the published review, GPT-4 showed 88.6% concordance with the original review, with <5% discrepancies due to inaccuracies or omissions. It exceeded human accuracy in 2.5% of instances. Study outcomes were extracted and pooling of results showed comparable effect sizes to comparator SRs. A review of bias assessment using GPT-4 showed fair-moderate but significant intra-rater agreement (ICC=0.518, p<0.001) and inter-rater agreements between human comparator SR (weighted kappa=0.237) and the analogous SR (weighted kappa=0.296). In contrast, there was a poor agreement between the two human-performed SRs (weighted kappa=0.094).</p><p><strong>Conclusion: </strong>Customized GPT-4 models perform well in extracting precise data from medical literature with potential for utilization in review of bias. While the evaluated tasks are simpler than the broader range of SR methodologies, they provide an important initial assessment of GPT-4's capabilities.</p>","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"313-322"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143810374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wendy Levinson, Karen Born, Juan Victor Ariel Franco, Karin Silvana Kopitowski
{"title":"Top 15 Choosing Wisely international campaign recommendations to reduce low-value care.","authors":"Wendy Levinson, Karen Born, Juan Victor Ariel Franco, Karin Silvana Kopitowski","doi":"10.1136/bmjebm-2025-113804","DOIUrl":"10.1136/bmjebm-2025-113804","url":null,"abstract":"","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"355-357"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144224274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Implementing AI models in clinical workflows: a roadmap.","authors":"Fei Wang, Ashley Beecy","doi":"10.1136/bmjebm-2023-112727","DOIUrl":"10.1136/bmjebm-2023-112727","url":null,"abstract":"","PeriodicalId":9059,"journal":{"name":"BMJ Evidence-Based Medicine","volume":" ","pages":"285-287"},"PeriodicalIF":7.6,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11666800/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141445397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}