基于最大似然的Gwet机会协议模型互连器可靠性估计。

统计学期刊(英文) Pub Date : 2024-10-01 Epub Date: 2024-10-28 DOI:10.4236/ojs.2024.145021
Alek M Westover, Tara M Westover, M Brandon Westover
{"title":"基于最大似然的Gwet机会协议模型互连器可靠性估计。","authors":"Alek M Westover, Tara M Westover, M Brandon Westover","doi":"10.4236/ojs.2024.145021","DOIUrl":null,"url":null,"abstract":"<p><p>Interrater reliability (IRR) statistics, like Cohen's kappa, measure agreement between raters beyond what is expected by chance when classifying items into categories. While Cohen's kappa has been widely used, it has several limitations, prompting development of Gwet's agreement statistic, an alternative \"kappa\"statistic which models chance agreement via an \"occasional guessing\" model. However, we show that Gwet's formula for estimating the proportion of agreement due to chance is itself biased for intermediate levels of agreement, despite overcoming limitations of Cohen's kappa at high and low agreement levels. We derive a maximum likelihood estimator for the occasional guessing model that yields an unbiased estimator of the IRR, which we call the maximum likelihood kappa ( <math> <msub><mrow><mi>κ</mi></mrow> <mrow><mtext>ML</mtext></mrow> </msub> </math> ). The key result is that the chance agreement probability under the occasional guessing model is simply equal to the observed rate of disagreement between raters. The <math> <msub><mrow><mi>κ</mi></mrow> <mrow><mtext>ML</mtext></mrow> </msub> </math> statistic provides a theoretically principled approach to quantifying IRR that addresses limitations of previous <math><mi>κ</mi></math> coefficients. Given the widespread use of IRR measures, having an unbiased estimator is important for reliable inference across domains where rater judgments are analyzed.</p>","PeriodicalId":59624,"journal":{"name":"统计学期刊(英文)","volume":"14 5","pages":"481-491"},"PeriodicalIF":0.0000,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12061032/pdf/","citationCount":"0","resultStr":"{\"title\":\"Interrater Reliability Estimation via Maximum Likelihood for Gwet's Chance Agreement Model.\",\"authors\":\"Alek M Westover, Tara M Westover, M Brandon Westover\",\"doi\":\"10.4236/ojs.2024.145021\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Interrater reliability (IRR) statistics, like Cohen's kappa, measure agreement between raters beyond what is expected by chance when classifying items into categories. While Cohen's kappa has been widely used, it has several limitations, prompting development of Gwet's agreement statistic, an alternative \\\"kappa\\\"statistic which models chance agreement via an \\\"occasional guessing\\\" model. However, we show that Gwet's formula for estimating the proportion of agreement due to chance is itself biased for intermediate levels of agreement, despite overcoming limitations of Cohen's kappa at high and low agreement levels. We derive a maximum likelihood estimator for the occasional guessing model that yields an unbiased estimator of the IRR, which we call the maximum likelihood kappa ( <math> <msub><mrow><mi>κ</mi></mrow> <mrow><mtext>ML</mtext></mrow> </msub> </math> ). The key result is that the chance agreement probability under the occasional guessing model is simply equal to the observed rate of disagreement between raters. The <math> <msub><mrow><mi>κ</mi></mrow> <mrow><mtext>ML</mtext></mrow> </msub> </math> statistic provides a theoretically principled approach to quantifying IRR that addresses limitations of previous <math><mi>κ</mi></math> coefficients. Given the widespread use of IRR measures, having an unbiased estimator is important for reliable inference across domains where rater judgments are analyzed.</p>\",\"PeriodicalId\":59624,\"journal\":{\"name\":\"统计学期刊(英文)\",\"volume\":\"14 5\",\"pages\":\"481-491\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12061032/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"统计学期刊(英文)\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.4236/ojs.2024.145021\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2024/10/28 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"统计学期刊(英文)","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.4236/ojs.2024.145021","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/10/28 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

像科恩的kappa一样,评估者间信度(IRR)统计数据衡量的是在对项目进行分类时,评估者之间超出预期的一致性。虽然Cohen的kappa被广泛使用,但它有一些局限性,促使Gwet的协议统计的发展,这是一种替代的“kappa”统计,通过“偶尔猜测”模型来模拟偶然协议。然而,我们表明,尽管克服了Cohen的kappa在高和低一致性水平上的局限性,但Gwet用于估计偶然一致性比例的公式本身对于中间一致性水平是有偏差的。我们为偶尔猜测模型导出了一个极大似然估计量,该模型产生了IRR的无偏估计量,我们称之为极大似然kappa (κ ML)。关键的结果是,偶然猜测模型下的机会一致概率等于观察到的评分者之间的不一致率。κ ML统计量提供了一种理论上有原则的方法来量化IRR,解决了以前κ系数的局限性。考虑到IRR度量的广泛使用,拥有一个无偏估计量对于跨领域的可靠推断是很重要的,在跨领域的推断中分析比较的判断。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Interrater Reliability Estimation via Maximum Likelihood for Gwet's Chance Agreement Model.

Interrater reliability (IRR) statistics, like Cohen's kappa, measure agreement between raters beyond what is expected by chance when classifying items into categories. While Cohen's kappa has been widely used, it has several limitations, prompting development of Gwet's agreement statistic, an alternative "kappa"statistic which models chance agreement via an "occasional guessing" model. However, we show that Gwet's formula for estimating the proportion of agreement due to chance is itself biased for intermediate levels of agreement, despite overcoming limitations of Cohen's kappa at high and low agreement levels. We derive a maximum likelihood estimator for the occasional guessing model that yields an unbiased estimator of the IRR, which we call the maximum likelihood kappa ( κ ML ). The key result is that the chance agreement probability under the occasional guessing model is simply equal to the observed rate of disagreement between raters. The κ ML statistic provides a theoretically principled approach to quantifying IRR that addresses limitations of previous κ coefficients. Given the widespread use of IRR measures, having an unbiased estimator is important for reliable inference across domains where rater judgments are analyzed.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
571
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信