Thomas J Steinbach, Debra A Tokarz, Caroll A Co, Shawn F Harris, Sandra J McBride, Keith R Shockley, Avinash Lokhande, Gargi Srivastava, Rajesh Ugalmugle, Arshad Kazi, Emily Singletary, Mark F Cesta, Heath C Thomas, Vivian S Chen, Kristen Hobbie, Torrie A Crabbs
{"title":"啮齿动物心肌病严重程度评分中评分者间和评分者内的一致性以及与基于人工智能的评分的关系。","authors":"Thomas J Steinbach, Debra A Tokarz, Caroll A Co, Shawn F Harris, Sandra J McBride, Keith R Shockley, Avinash Lokhande, Gargi Srivastava, Rajesh Ugalmugle, Arshad Kazi, Emily Singletary, Mark F Cesta, Heath C Thomas, Vivian S Chen, Kristen Hobbie, Torrie A Crabbs","doi":"10.1177/01926233241259998","DOIUrl":null,"url":null,"abstract":"<p><p>We previously developed a computer-assisted image analysis algorithm to detect and quantify the microscopic features of rodent progressive cardiomyopathy (PCM) in rat heart histologic sections and validated the results with a panel of five veterinary toxicologic pathologists using a multinomial logistic model. In this study, we assessed both the inter-rater and intra-rater agreement of the pathologists and compared pathologists' ratings to the artificial intelligence (AI)-predicted scores. Pathologists and the AI algorithm were presented with 500 slides of rodent heart. They quantified the amount of cardiomyopathy in each slide. A total of 200 of these slides were novel to this study, whereas 100 slides were intentionally selected for repetition from the previous study. After a washout period of more than six months, the repeated slides were examined to assess intra-rater agreement among pathologists. We found the intra-rater agreement to be substantial, with weighted Cohen's kappa values ranging from k = 0.64 to 0.80. Intra-rater variability is not a concern for the deterministic AI. The inter-rater agreement across pathologists was moderate (Cohen's kappa k = 0.56). These results demonstrate the utility of AI algorithms as a tool for pathologists to increase sensitivity and specificity for the histopathologic assessment of the heart in toxicology studies.</p>","PeriodicalId":23113,"journal":{"name":"Toxicologic Pathology","volume":" ","pages":"258-265"},"PeriodicalIF":1.4000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11412787/pdf/","citationCount":"0","resultStr":"{\"title\":\"Inter-Rater and Intra-Rater Agreement in Scoring Severity of Rodent Cardiomyopathy and Relation to Artificial Intelligence-Based Scoring.\",\"authors\":\"Thomas J Steinbach, Debra A Tokarz, Caroll A Co, Shawn F Harris, Sandra J McBride, Keith R Shockley, Avinash Lokhande, Gargi Srivastava, Rajesh Ugalmugle, Arshad Kazi, Emily Singletary, Mark F Cesta, Heath C Thomas, Vivian S Chen, Kristen Hobbie, Torrie A Crabbs\",\"doi\":\"10.1177/01926233241259998\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>We previously developed a computer-assisted image analysis algorithm to detect and quantify the microscopic features of rodent progressive cardiomyopathy (PCM) in rat heart histologic sections and validated the results with a panel of five veterinary toxicologic pathologists using a multinomial logistic model. In this study, we assessed both the inter-rater and intra-rater agreement of the pathologists and compared pathologists' ratings to the artificial intelligence (AI)-predicted scores. Pathologists and the AI algorithm were presented with 500 slides of rodent heart. They quantified the amount of cardiomyopathy in each slide. A total of 200 of these slides were novel to this study, whereas 100 slides were intentionally selected for repetition from the previous study. After a washout period of more than six months, the repeated slides were examined to assess intra-rater agreement among pathologists. We found the intra-rater agreement to be substantial, with weighted Cohen's kappa values ranging from k = 0.64 to 0.80. Intra-rater variability is not a concern for the deterministic AI. The inter-rater agreement across pathologists was moderate (Cohen's kappa k = 0.56). These results demonstrate the utility of AI algorithms as a tool for pathologists to increase sensitivity and specificity for the histopathologic assessment of the heart in toxicology studies.</p>\",\"PeriodicalId\":23113,\"journal\":{\"name\":\"Toxicologic Pathology\",\"volume\":\" \",\"pages\":\"258-265\"},\"PeriodicalIF\":1.4000,\"publicationDate\":\"2024-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11412787/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Toxicologic Pathology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1177/01926233241259998\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2024/6/22 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q3\",\"JCRName\":\"PATHOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Toxicologic Pathology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1177/01926233241259998","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/6/22 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"PATHOLOGY","Score":null,"Total":0}
引用次数: 0
摘要
我们之前开发了一种计算机辅助图像分析算法,用于检测和量化大鼠心脏组织切片中啮齿动物进行性心肌病 (PCM) 的显微特征,并使用多项式逻辑模型与五位兽医毒理学病理学家小组验证了结果。在这项研究中,我们评估了病理学家的评分者之间和评分者内部的一致性,并将病理学家的评分与人工智能(AI)预测的分数进行了比较。我们向病理学家和人工智能算法展示了 500 张啮齿动物心脏的切片。他们对每张切片中心肌病的数量进行量化。其中共有 200 张幻灯片是本研究的新内容,而 100 张幻灯片则是特意从之前的研究中挑选出来的重复内容。经过 6 个多月的缓冲期后,我们对重复的切片进行了检查,以评估病理学家之间的评分内一致性。我们发现评分者内部的一致性非常高,加权科恩卡帕值从 k = 0.64 到 0.80 不等。对于确定性人工智能来说,评分者内部的变异性并不令人担忧。病理学家之间的评分者间一致性为中等(科恩卡帕 k = 0.56)。这些结果证明了人工智能算法作为病理学家提高毒理学研究中心脏组织病理学评估灵敏度和特异性的工具的实用性。
Inter-Rater and Intra-Rater Agreement in Scoring Severity of Rodent Cardiomyopathy and Relation to Artificial Intelligence-Based Scoring.
We previously developed a computer-assisted image analysis algorithm to detect and quantify the microscopic features of rodent progressive cardiomyopathy (PCM) in rat heart histologic sections and validated the results with a panel of five veterinary toxicologic pathologists using a multinomial logistic model. In this study, we assessed both the inter-rater and intra-rater agreement of the pathologists and compared pathologists' ratings to the artificial intelligence (AI)-predicted scores. Pathologists and the AI algorithm were presented with 500 slides of rodent heart. They quantified the amount of cardiomyopathy in each slide. A total of 200 of these slides were novel to this study, whereas 100 slides were intentionally selected for repetition from the previous study. After a washout period of more than six months, the repeated slides were examined to assess intra-rater agreement among pathologists. We found the intra-rater agreement to be substantial, with weighted Cohen's kappa values ranging from k = 0.64 to 0.80. Intra-rater variability is not a concern for the deterministic AI. The inter-rater agreement across pathologists was moderate (Cohen's kappa k = 0.56). These results demonstrate the utility of AI algorithms as a tool for pathologists to increase sensitivity and specificity for the histopathologic assessment of the heart in toxicology studies.
期刊介绍:
Toxicologic Pathology is dedicated to the promotion of human, animal, and environmental health through the dissemination of knowledge, techniques, and guidelines to enhance the understanding and practice of toxicologic pathology. Toxicologic Pathology, the official journal of the Society of Toxicologic Pathology, will publish Original Research Articles, Symposium Articles, Review Articles, Meeting Reports, New Techniques, and Position Papers that are relevant to toxicologic pathology.