{"title":"Assessment centers do not measure competencies: why this is now beyond reasonable doubt","authors":"Chris Dewberry","doi":"10.1017/iop.2024.5","DOIUrl":null,"url":null,"abstract":"\n Although assessment centers (ACs) are usually designed to measure stable competencies (i.e., dimensions), doubt about whether or not they reliably do so has endured for 70 years. Addressing this issue in a novel way, several published Generalizability (G) theory studies have sought to isolate the multiple sources of variance in AC ratings, including variance specifically concerned with competencies. Unlike previous research, these studies can provide a definitive answer to the AC construct validity issue. In this article, the historical context for the construct validity debate is set out, and the results of four large-scale G-theory studies of ACs are reviewed. It is concluded that these studies demonstrate, beyond reasonable doubt, that ACs do not reliably measure stable competencies, but instead measure general, and exercise-related, performance. The possibility that ACs measure unstable competencies is considered, and it is suggested that evidence that they do so may reflect an artefact of typical AC design rather than a “real” effect. For ethical, individual, and organizational reasons, it is argued that the use of ACs to measure competencies can no longer be justified and should be halted.","PeriodicalId":515605,"journal":{"name":"Industrial and Organizational Psychology","volume":" 12","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Industrial and Organizational Psychology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1017/iop.2024.5","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Although assessment centers (ACs) are usually designed to measure stable competencies (i.e., dimensions), doubt about whether or not they reliably do so has endured for 70 years. Addressing this issue in a novel way, several published Generalizability (G) theory studies have sought to isolate the multiple sources of variance in AC ratings, including variance specifically concerned with competencies. Unlike previous research, these studies can provide a definitive answer to the AC construct validity issue. In this article, the historical context for the construct validity debate is set out, and the results of four large-scale G-theory studies of ACs are reviewed. It is concluded that these studies demonstrate, beyond reasonable doubt, that ACs do not reliably measure stable competencies, but instead measure general, and exercise-related, performance. The possibility that ACs measure unstable competencies is considered, and it is suggested that evidence that they do so may reflect an artefact of typical AC design rather than a “real” effect. For ethical, individual, and organizational reasons, it is argued that the use of ACs to measure competencies can no longer be justified and should be halted.
尽管测评中心(AC)的设计通常是为了测量稳定的能力(即维度),但70年来,人们一直在怀疑它们是否能可靠地做到这一点。为了以一种新颖的方式解决这个问题,一些已发表的通用性(G)理论研究试图分离出 AC 评级中的多种差异来源,包括与能力有关的差异。与以往的研究不同,这些研究可以为 AC 构建有效性问题提供一个明确的答案。在本文中,我们阐述了构造效度争论的历史背景,并回顾了四项大规模 AC G 理论研究的结果。文章的结论是,这些研究毫无疑问地证明,量表并不能可靠地测量稳定的能力,而只能测量一般的、与运动相关的表现。研究还考虑了测评不稳定能力的可能性,认为测评不稳定能力的证据可能反映了典型测评设计的假象,而非 "真实 "效果。出于道德、个人和组织方面的原因,有观点认为,不再有理由使用能力测验来衡量能力,并应停止使用能力测验。