Should we trust our judgments about the proficiency of Motivational Interviewing counselors? A glimpse at the impact of low inter-rater reliability.

Chris Dunn, Doyanne Darnell, Sheng Kung Michael Yi, Mark Steyvers, Kristin Bumgardner, Sarah Peregrine Lord, Zac Imel, David C Atkins
{"title":"Should we trust our judgments about the proficiency of Motivational Interviewing counselors? A glimpse at the impact of low inter-rater reliability.","authors":"Chris Dunn,&nbsp;Doyanne Darnell,&nbsp;Sheng Kung Michael Yi,&nbsp;Mark Steyvers,&nbsp;Kristin Bumgardner,&nbsp;Sarah Peregrine Lord,&nbsp;Zac Imel,&nbsp;David C Atkins","doi":"10.5195/mitrip.2014.43","DOIUrl":null,"url":null,"abstract":"<p><p>Standardized rating systems are often used to evaluate the proficiency of Motivational Interviewing (MI) counselors. The published inter-rater reliability (degree of coder agreement) in many studies using these instruments has varied a great deal; some studies report MI proficiency scores that have only fair inter-rater reliability, and others report scores with excellent reliability. How much can we to trust the scores with fair versus excellent reliability? Using a Monte Carlo statistical simulation, we compared the impact of fair (0.50) versus excellent (0.90) reliability on the error rates of falsely judging a given counselor as MI proficient or not proficient. We found that improving the inter-rater reliability of any given score from 0.5 to 0.9 would cause a marked reduction in proficiency judgment errors, a reduction that in some MI evaluation situations would be critical. We discuss some practical tradeoffs inherent in various MI evaluation situations, and offer suggestions for applying findings from formal MI research to problems faced by real-world MI evaluators, to help them minimize the MI proficiency judgment errors bearing the greatest cost.</p>","PeriodicalId":89699,"journal":{"name":"Motivational interviewing : training, research, implementation, practice","volume":"1 3","pages":"38-41"},"PeriodicalIF":0.0000,"publicationDate":"2014-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5008854/pdf/nihms812516.pdf","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Motivational interviewing : training, research, implementation, practice","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5195/mitrip.2014.43","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Standardized rating systems are often used to evaluate the proficiency of Motivational Interviewing (MI) counselors. The published inter-rater reliability (degree of coder agreement) in many studies using these instruments has varied a great deal; some studies report MI proficiency scores that have only fair inter-rater reliability, and others report scores with excellent reliability. How much can we to trust the scores with fair versus excellent reliability? Using a Monte Carlo statistical simulation, we compared the impact of fair (0.50) versus excellent (0.90) reliability on the error rates of falsely judging a given counselor as MI proficient or not proficient. We found that improving the inter-rater reliability of any given score from 0.5 to 0.9 would cause a marked reduction in proficiency judgment errors, a reduction that in some MI evaluation situations would be critical. We discuss some practical tradeoffs inherent in various MI evaluation situations, and offer suggestions for applying findings from formal MI research to problems faced by real-world MI evaluators, to help them minimize the MI proficiency judgment errors bearing the greatest cost.

Abstract Image

我们应该相信我们对动机性访谈顾问熟练程度的判断吗?低评级机构间可靠性的影响。
标准化的评级系统经常被用来评估动机访谈(MI)咨询师的熟练程度。在使用这些工具的许多研究中,已发表的编码器间可靠性(编码器一致性程度)变化很大;一些研究报告的MI熟练程度分数只有一般的评分者间信度,而另一些研究报告的分数具有优异的信度。我们能在多大程度上相信信度一般的分数和信度极佳的分数?使用蒙特卡罗统计模拟,我们比较了公平(0.50)和优秀(0.90)可靠性对错误判断给定咨询师为MI精通或不精通的错误率的影响。我们发现,将任何给定分数的评分者间信度从0.5提高到0.9都会显著减少熟练度判断错误,这种减少在某些MI评估情况下是至关重要的。我们讨论了各种MI评估情况中固有的一些实际权衡,并提供了将正式MI研究结果应用于实际MI评估者所面临的问题的建议,以帮助他们最大限度地减少承担最大成本的MI熟练度判断错误。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信