Karel G M Moons, Johanna A A Damen, Tabea Kaul, Lotty Hooft, Constanza Andaur Navarro, Paula Dhiman, Andrew L Beam, Ben Van Calster, Leo Anthony Celi, Spiros Denaxas, Alastair K Denniston, Marzyeh Ghassemi, Georg Heinze, André Pascal Kengne, Lena Maier-Hein, Xiaoxuan Liu, Patricia Logullo, Melissa D McCradden, Nan Liu, Lauren Oakden-Rayner, Karandeep Singh, Daniel S Ting, Laure Wynants, Bada Yang, Johannes B Reitsma, Richard D Riley, Gary S Collins, Maarten van Smeden
{"title":"PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods","authors":"Karel G M Moons, Johanna A A Damen, Tabea Kaul, Lotty Hooft, Constanza Andaur Navarro, Paula Dhiman, Andrew L Beam, Ben Van Calster, Leo Anthony Celi, Spiros Denaxas, Alastair K Denniston, Marzyeh Ghassemi, Georg Heinze, André Pascal Kengne, Lena Maier-Hein, Xiaoxuan Liu, Patricia Logullo, Melissa D McCradden, Nan Liu, Lauren Oakden-Rayner, Karandeep Singh, Daniel S Ting, Laure Wynants, Bada Yang, Johannes B Reitsma, Richard D Riley, Gary S Collins, Maarten van Smeden","doi":"10.1136/bmj-2024-082505","DOIUrl":null,"url":null,"abstract":"The Prediction model Risk Of Bias ASsessment Tool (PROBAST) is used to assess the quality, risk of bias, and applicability of prediction models or algorithms and of prediction model/algorithm studies. Since PROBAST’s introduction in 2019, much progress has been made in the methodology for prediction modelling and in the use of artificial intelligence, including machine learning, techniques. An update to PROBAST-2019 is thus needed. This article describes the development of PROBAST+AI. PROBAST+AI consists of two distinctive parts: model development and model evaluation. For model development, PROBAST+AI users assess quality and applicability using 16 targeted signalling questions. For model evaluation, PROBAST+AI users assess the risk of bias and applicability using 18 targeted signalling questions. Both parts contain four domains: participants and data sources, predictors, outcome, and analysis. Applicability of the prediction model is rated for the participants and data sources, predictors, and outcome domains. PROBAST+AI may replace the original PROBAST tool and allows all key stakeholders (eg, model developers, AI companies, researchers, editors, reviewers, healthcare professionals, guideline developers, and policy organisations) to examine the quality, risk of bias, and applicability of any type of prediction model in the healthcare sector, irrespective of whether regression modelling or AI techniques are used. In healthcare, prediction models or algorithms (hereafter referred to as prediction models) estimate the probability of a health outcome for individuals. In the diagnostic setting—including screening and monitoring—the model typically aims to predict or classify the presence of a particular outcome, such as a disease or disorder. In the prognostic setting the model aims to predict a future outcome—typically health related—in patients with a diagnosis of a particular disease or disorder, or in the general population. The primary use of a prediction model in healthcare is to support individual healthcare counselling and shared decision making on, for example, subsequent medical testing, …","PeriodicalId":22388,"journal":{"name":"The BMJ","volume":"54 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The BMJ","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1136/bmj-2024-082505","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods
The Prediction model Risk Of Bias ASsessment Tool (PROBAST) is used to assess the quality, risk of bias, and applicability of prediction models or algorithms and of prediction model/algorithm studies. Since PROBAST’s introduction in 2019, much progress has been made in the methodology for prediction modelling and in the use of artificial intelligence, including machine learning, techniques. An update to PROBAST-2019 is thus needed. This article describes the development of PROBAST+AI. PROBAST+AI consists of two distinctive parts: model development and model evaluation. For model development, PROBAST+AI users assess quality and applicability using 16 targeted signalling questions. For model evaluation, PROBAST+AI users assess the risk of bias and applicability using 18 targeted signalling questions. Both parts contain four domains: participants and data sources, predictors, outcome, and analysis. Applicability of the prediction model is rated for the participants and data sources, predictors, and outcome domains. PROBAST+AI may replace the original PROBAST tool and allows all key stakeholders (eg, model developers, AI companies, researchers, editors, reviewers, healthcare professionals, guideline developers, and policy organisations) to examine the quality, risk of bias, and applicability of any type of prediction model in the healthcare sector, irrespective of whether regression modelling or AI techniques are used. In healthcare, prediction models or algorithms (hereafter referred to as prediction models) estimate the probability of a health outcome for individuals. In the diagnostic setting—including screening and monitoring—the model typically aims to predict or classify the presence of a particular outcome, such as a disease or disorder. In the prognostic setting the model aims to predict a future outcome—typically health related—in patients with a diagnosis of a particular disease or disorder, or in the general population. The primary use of a prediction model in healthcare is to support individual healthcare counselling and shared decision making on, for example, subsequent medical testing, …