Board scores in the spotlight: Public reporting and the unintended consequences

IF 1.7 Q2 EDUCATION, SCIENTIFIC DISCIPLINES
Matthew E Kelleher MD, MEd, Sally A Santen MD, PhD, Christiana Draper MD, PhD, Jaime Jordan MD, MA, Michael Gottlieb MD, Benjamin Kinnear MD, MEd
{"title":"Board scores in the spotlight: Public reporting and the unintended consequences","authors":"Matthew E Kelleher MD, MEd,&nbsp;Sally A Santen MD, PhD,&nbsp;Christiana Draper MD, PhD,&nbsp;Jaime Jordan MD, MA,&nbsp;Michael Gottlieb MD,&nbsp;Benjamin Kinnear MD, MEd","doi":"10.1002/aet2.70006","DOIUrl":null,"url":null,"abstract":"<p>The American Board of Emergency Medicine (ABEM) recently announced plans to publicly report program-level board certification examination pass rates.<span><sup>1</sup></span> This initiative will present program-level board pass rates for public viewing. Multiple arguments can be made for such a change. Medical education is increasingly seen as a service for which trainees have paid large sums of money and sacrifice a significant amount of time and effort. Transparency and accountability to learners that show services are high quality is important. Additionally, public reporting of program board pass rates provides a mechanism for accountability to society at large and aligns with ABEM's mission “To ensure the highest standards in the specialty of Emergency Medicine.”<span><sup>2</sup></span> Graduate medical education (GME) is largely funded through tax dollars, and an argument can be made that the public should be able to see the degree to which different training programs are helping their graduates pass certifying examinations.<span><sup>3</sup></span> Finally, this change would align ABEM with multiple other major American Board of Medical Specialties (ABMS) member boards who publicly report program pass rates, such as the American Board of Internal Medicine, American Board of Pediatrics, and American Board of Family Medicine.<span><sup>4</sup></span> These rationales have merit, but unintended consequences lurk around the corner. In this perspective, we describe the potential negative impact of publicly reporting program-level certifying examination pass rates. Specifically, we explore how public reporting could disincentivize holistic review of applicants during residency recruitment. We propose actionable strategies that various stakeholders may consider for balancing transparency with the broader mission of holistic review and inclusive recruitment practices in GME.</p><p>Humans are influenced by incentive structures. As rational actors, we tend to (either implicitly or explicitly) alter our thinking and behavior when different incentive structures are presented to us. When such structures lead to unwanted consequences, they are labeled “perverse incentives.” A colloquial term for the impact of a perverse incentive is sometimes called the <i>Cobra Effect</i>.<span><sup>5</sup></span> The Cobra Effect draws from an anecdote in which a governmental effort to reduce the number of cobras in Delhi, India, backfired. The initiative offered a bounty for dead cobras, so people began breeding cobras to turn in more dead snakes. The bounty was intended to incentivize the killing of cobras, hence decreasing the overall population. However, people quickly realized they could game the system for financial gain, leading to an overall increase in the cobra population. The incentive structure led to unwanted consequences that could potentially have been predicted by considering how people would respond.</p><p>We believe that publicly reporting program certifying examination pass rates may lead to its own Cobra Effect, potentially altering program incentives in ways that undermine holistic recruitment efforts. Holistic review in residency selection has gradually shifted programs away from an overreliance on academic metrics like test scores and class rank in favor of a broader (i.e., more holistic) evaluation of an applicant's strengths, interests, and value alignment with a residency.<span><sup>6</sup></span> While academic metrics still play a role, they are no longer the sole focus, thanks to growing awareness of their limitations. Many traditional academic metrics lack validity evidence for use in GME recruitment and do not predict future residency performance.<span><sup>7-9</sup></span> Additionally, concerns about inequity and bias in these metrics—such as the systematic disadvantages faced by minoritized groups—have highlighted the need for more inclusive strategies.<span><sup>10, 11</sup></span> Holistic review, while not a standardized approach, is recommended by most national organizations (AAMC, AMA, ACGME) because of its potential to improve equity in the residency selection process.<span><sup>12-16</sup></span> The competency-based medical education (CBME) movement has further underscored that knowledge and test-taking ability are only parts of a broader framework of competence.<span><sup>17</sup></span> We fear that publicly reporting program board pass rates for the sake of transparency and accountability may underestimate the looming Cobra Effect of incentivizing programs to return to <i>overemphasizing</i> test scores when evaluating applicants.</p><p>To be clear, we are not suggesting that passing the certifying examination is unimportant. When compared to individuals who are not board certified, board certification has been linked to meaningful patient care outcomes in future practice.<span><sup>18-21</sup></span> Thus, the certifying examination pass rate presents a metric that has some validity evidence to support its use in assessing <i>individual</i> graduates’ readiness for practice. The key issue lies in differentiating whether the residency program's quality or the individual's test-taking abilities has more influence on passing or failing the examination. The rationale for sharing program pass rates relies on a false premise: that all programs recruit learners who enter residency with a similar chance of passing the certifying examination. However, it is very unlikely that testing ability and medical knowledge is equally among matriculating cohorts across programs. The best predictor of passing a standardized certifying examination is performance on previous standardized examinations.<span><sup>22-31</sup></span> In contrast, the evidence is much less clear about whether the quality of residency training is associated with board passage rates.<span><sup>32-40</sup></span> Use of program-level board pass metrics as a signal of residency training quality fails to account for inherent differences in programs’ incoming residents' <i>prior</i> standardized examination performance, making intraprogram comparisons potentially misleading.</p><p>The public reporting of board certification pass rates will likely exert significant pressure on program directors (PDs) to mitigate the risk of examination failures. This is a good goal to aspire toward; PDs should help their graduates pass the certifying examination. However, they may feel pressured to avoid failures not only by improving education and training, but also by changing recruiting practices. Even a small number of individual examination failures could greatly alter a program's overall pass rates given the relatively low number of graduates per year from ABEM training programs. Programs might fear risk of impaired recruiting should applicants use program board pass rates as a metric to evaluate and rank programs. The most straightforward way for PDs to minimize risk is to recruit applicants with the least risk of examination failure. This aligns with Campbell's Law, which suggests that when any measure is emphasized as a marker of social standing (e.g., high pass rates equates to a better residency training program), it will be prone to distortion or corruption.<span><sup>41</sup></span> In this context, program board pass rates taking on such a visible and elevated status will no doubt make prior testing performance a more important metric for PDs to consider when making decisions on whom to invite for an interview and how to rank them on final lists. PDs, as rational actors, may make decisions to “protect” their programs as well as their personal reputation by prioritizing metrics that maximize pass rates and minimize examination failures. If programs focus significantly on recruiting applicants with proven test-taking abilities, the entire system shifts toward this strategy. In turn, this could lead to students and schools responding by placing even more emphasis on testing metrics. If a residency program chooses to resist this shift on principle, maintaining a more holistic review, they are opening themselves up to scrutiny from the public and applicants if their pass rates begin to lag behind other programs. Ultimately, a cycle of competitive defensiveness is perpetuated over time.</p><p>Public reporting of program board pass rates risks undermining national momentum made toward programs using holistic review for applicant selection. For many programs, the added pressure of transparency may act as a disincentive to invest in resource-intensive, nuanced holistic review processes. Instead, it may nudge programs—if not outright shove them—back toward relying heavily on easily sortable academic metrics like test scores. Programs that are already stretched thin and searching for simpler application review strategies may find this shift particularly appealing. Holistic review has led to an emphasis on increasing the number of diverse residents in programs nationwide. A return to prior selection strategies has potential to undermine the diverse representation of physicians, which we know has a positive effect on patient outcomes.<span><sup>42-44</sup></span> Public reporting could inadvertently set back holistic review efforts, prioritizing expediency over the broader, equity-driven goals that have shaped medical education reform in recent years.</p><p>Medical education has made great strides toward holistic recruitment strategies. The potential unintended consequences highlighted by Campbell's Law and the Cobra Effect reveal the risk of incentivizing programs to return to an overemphasis on test-taking abilities in resident selection. More research is needed to explore the complex alchemy of how certifying examination results can be transparently reported for stakeholders without disincentivizing holistic recruitment and selection practices. While we encourage residency programs to share program board pass rates with prospective applicants during the recruitment process when there is an opportunity to provide appropriate context and nuanced discussion, we recommend that ABEM not make board pass rates public.</p><p>The authors declare no conflicts of interest.</p>","PeriodicalId":37032,"journal":{"name":"AEM Education and Training","volume":"9 2","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aet2.70006","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"AEM Education and Training","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/aet2.70006","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"EDUCATION, SCIENTIFIC DISCIPLINES","Score":null,"Total":0}
引用次数: 0

Abstract

The American Board of Emergency Medicine (ABEM) recently announced plans to publicly report program-level board certification examination pass rates.1 This initiative will present program-level board pass rates for public viewing. Multiple arguments can be made for such a change. Medical education is increasingly seen as a service for which trainees have paid large sums of money and sacrifice a significant amount of time and effort. Transparency and accountability to learners that show services are high quality is important. Additionally, public reporting of program board pass rates provides a mechanism for accountability to society at large and aligns with ABEM's mission “To ensure the highest standards in the specialty of Emergency Medicine.”2 Graduate medical education (GME) is largely funded through tax dollars, and an argument can be made that the public should be able to see the degree to which different training programs are helping their graduates pass certifying examinations.3 Finally, this change would align ABEM with multiple other major American Board of Medical Specialties (ABMS) member boards who publicly report program pass rates, such as the American Board of Internal Medicine, American Board of Pediatrics, and American Board of Family Medicine.4 These rationales have merit, but unintended consequences lurk around the corner. In this perspective, we describe the potential negative impact of publicly reporting program-level certifying examination pass rates. Specifically, we explore how public reporting could disincentivize holistic review of applicants during residency recruitment. We propose actionable strategies that various stakeholders may consider for balancing transparency with the broader mission of holistic review and inclusive recruitment practices in GME.

Humans are influenced by incentive structures. As rational actors, we tend to (either implicitly or explicitly) alter our thinking and behavior when different incentive structures are presented to us. When such structures lead to unwanted consequences, they are labeled “perverse incentives.” A colloquial term for the impact of a perverse incentive is sometimes called the Cobra Effect.5 The Cobra Effect draws from an anecdote in which a governmental effort to reduce the number of cobras in Delhi, India, backfired. The initiative offered a bounty for dead cobras, so people began breeding cobras to turn in more dead snakes. The bounty was intended to incentivize the killing of cobras, hence decreasing the overall population. However, people quickly realized they could game the system for financial gain, leading to an overall increase in the cobra population. The incentive structure led to unwanted consequences that could potentially have been predicted by considering how people would respond.

We believe that publicly reporting program certifying examination pass rates may lead to its own Cobra Effect, potentially altering program incentives in ways that undermine holistic recruitment efforts. Holistic review in residency selection has gradually shifted programs away from an overreliance on academic metrics like test scores and class rank in favor of a broader (i.e., more holistic) evaluation of an applicant's strengths, interests, and value alignment with a residency.6 While academic metrics still play a role, they are no longer the sole focus, thanks to growing awareness of their limitations. Many traditional academic metrics lack validity evidence for use in GME recruitment and do not predict future residency performance.7-9 Additionally, concerns about inequity and bias in these metrics—such as the systematic disadvantages faced by minoritized groups—have highlighted the need for more inclusive strategies.10, 11 Holistic review, while not a standardized approach, is recommended by most national organizations (AAMC, AMA, ACGME) because of its potential to improve equity in the residency selection process.12-16 The competency-based medical education (CBME) movement has further underscored that knowledge and test-taking ability are only parts of a broader framework of competence.17 We fear that publicly reporting program board pass rates for the sake of transparency and accountability may underestimate the looming Cobra Effect of incentivizing programs to return to overemphasizing test scores when evaluating applicants.

To be clear, we are not suggesting that passing the certifying examination is unimportant. When compared to individuals who are not board certified, board certification has been linked to meaningful patient care outcomes in future practice.18-21 Thus, the certifying examination pass rate presents a metric that has some validity evidence to support its use in assessing individual graduates’ readiness for practice. The key issue lies in differentiating whether the residency program's quality or the individual's test-taking abilities has more influence on passing or failing the examination. The rationale for sharing program pass rates relies on a false premise: that all programs recruit learners who enter residency with a similar chance of passing the certifying examination. However, it is very unlikely that testing ability and medical knowledge is equally among matriculating cohorts across programs. The best predictor of passing a standardized certifying examination is performance on previous standardized examinations.22-31 In contrast, the evidence is much less clear about whether the quality of residency training is associated with board passage rates.32-40 Use of program-level board pass metrics as a signal of residency training quality fails to account for inherent differences in programs’ incoming residents' prior standardized examination performance, making intraprogram comparisons potentially misleading.

The public reporting of board certification pass rates will likely exert significant pressure on program directors (PDs) to mitigate the risk of examination failures. This is a good goal to aspire toward; PDs should help their graduates pass the certifying examination. However, they may feel pressured to avoid failures not only by improving education and training, but also by changing recruiting practices. Even a small number of individual examination failures could greatly alter a program's overall pass rates given the relatively low number of graduates per year from ABEM training programs. Programs might fear risk of impaired recruiting should applicants use program board pass rates as a metric to evaluate and rank programs. The most straightforward way for PDs to minimize risk is to recruit applicants with the least risk of examination failure. This aligns with Campbell's Law, which suggests that when any measure is emphasized as a marker of social standing (e.g., high pass rates equates to a better residency training program), it will be prone to distortion or corruption.41 In this context, program board pass rates taking on such a visible and elevated status will no doubt make prior testing performance a more important metric for PDs to consider when making decisions on whom to invite for an interview and how to rank them on final lists. PDs, as rational actors, may make decisions to “protect” their programs as well as their personal reputation by prioritizing metrics that maximize pass rates and minimize examination failures. If programs focus significantly on recruiting applicants with proven test-taking abilities, the entire system shifts toward this strategy. In turn, this could lead to students and schools responding by placing even more emphasis on testing metrics. If a residency program chooses to resist this shift on principle, maintaining a more holistic review, they are opening themselves up to scrutiny from the public and applicants if their pass rates begin to lag behind other programs. Ultimately, a cycle of competitive defensiveness is perpetuated over time.

Public reporting of program board pass rates risks undermining national momentum made toward programs using holistic review for applicant selection. For many programs, the added pressure of transparency may act as a disincentive to invest in resource-intensive, nuanced holistic review processes. Instead, it may nudge programs—if not outright shove them—back toward relying heavily on easily sortable academic metrics like test scores. Programs that are already stretched thin and searching for simpler application review strategies may find this shift particularly appealing. Holistic review has led to an emphasis on increasing the number of diverse residents in programs nationwide. A return to prior selection strategies has potential to undermine the diverse representation of physicians, which we know has a positive effect on patient outcomes.42-44 Public reporting could inadvertently set back holistic review efforts, prioritizing expediency over the broader, equity-driven goals that have shaped medical education reform in recent years.

Medical education has made great strides toward holistic recruitment strategies. The potential unintended consequences highlighted by Campbell's Law and the Cobra Effect reveal the risk of incentivizing programs to return to an overemphasis on test-taking abilities in resident selection. More research is needed to explore the complex alchemy of how certifying examination results can be transparently reported for stakeholders without disincentivizing holistic recruitment and selection practices. While we encourage residency programs to share program board pass rates with prospective applicants during the recruitment process when there is an opportunity to provide appropriate context and nuanced discussion, we recommend that ABEM not make board pass rates public.

The authors declare no conflicts of interest.

美国急诊医学委员会(American Board of Emergency Medicine,ABEM)最近宣布计划公开报告项目级别的委员会认证考试通过率。这一改变有多种理由。医学教育越来越被视为一种服务,受训者为此支付了大笔费用,牺牲了大量时间和精力。向学员展示高质量服务的透明度和问责制非常重要。此外,公开报告项目委员会的通过率为整个社会提供了一个问责机制,也符合 ABEM 的使命:"确保急诊医学专业的最高标准 "2。毕业后医学教育(GME)的资金主要来自税款,因此可以说公众应该能够看到不同的培训项目在多大程度上帮助其毕业生通过认证考试3。最后,这一变化将使 ABEM 与其他多个公开报告项目通过率的主要美国医学专业委员会(ABMS)成员委员会保持一致,如美国内科医学委员会、美国儿科医学委员会和美国家庭医学委员会。在本文中,我们将阐述公开报告项目级别认证考试通过率的潜在负面影响。具体而言,我们探讨了公开报告如何在住院医师招聘过程中抑制对申请人的全面审查。我们提出了一些可操作的策略,供各利益相关方考虑,以平衡透明度与整体审查的更广泛使命以及 GME 中的包容性招聘实践。作为理性的行动者,当不同的激励结构出现在我们面前时,我们往往会(或隐或显地)改变我们的思维和行为。当这种激励结构导致不想要的后果时,就会被称为 "反常激励"。5 眼镜蛇效应源于一则轶事:印度德里政府为减少眼镜蛇数量所做的努力适得其反。这项举措为死亡的眼镜蛇提供赏金,于是人们开始饲养眼镜蛇,以上交更多的死蛇。悬赏旨在鼓励人们杀死眼镜蛇,从而减少眼镜蛇的总数量。然而,人们很快意识到,他们可以利用这一制度来获取经济利益,从而导致眼镜蛇数量的整体增加。我们认为,公开报告项目认证考试通过率可能会导致眼镜蛇效应,从而可能改变项目的激励机制,破坏整体招聘工作。住院医师遴选中的全面审查已逐渐使项目不再过度依赖考试分数和班级排名等学术指标,而是更广泛地(即更全面地)评估申请人的优势、兴趣以及与住院医师价值的一致性。许多传统的学术指标在 GME 招生中缺乏有效性证据,也不能预测未来的住院医师表现。7-9 此外,人们对这些指标中的不公平和偏见的担忧--如少数群体所面临的系统性劣势--凸显了对更具包容性策略的需求。10, 11 整体审查虽然不是一种标准化的方法,但由于其在改善住院医师遴选过程中的公平性方面的潜力,被大多数国家组织(AAMC、AMA、ACGME)所推荐。基于能力的医学教育(CBME)运动进一步强调,知识和应试能力只是更广泛的能力框架的一部分。17 我们担心,为了透明和问责而公开报告项目的委员会通过率可能会低估迫在眉睫的眼镜蛇效应,即激励项目在评估申请人时重新过度强调考试分数。18-21 因此,认证考试通过率是一个具有一定有效性的指标,支持用它来评估毕业生是否做好了执业准备。问题的关键在于区分是住院医师培训项目的质量还是个人的应试能力对通过或未通过考试的影响更大。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
AEM Education and Training
AEM Education and Training Nursing-Emergency Nursing
CiteScore
2.60
自引率
22.20%
发文量
89
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信