Adapting Simulation Responses From Judgment-Based to Analytic-Based Scores: A Process Model, Case Study, and Empirical Evaluation of Managers’ Responses Among a Sample of Managers

IF 0.6 Q3 Business, Management and Accounting
Diana R. Sanchez, Saar Van Lysebetten, A. Gibbons
{"title":"Adapting Simulation Responses From Judgment-Based to Analytic-Based Scores: A Process Model, Case Study, and Empirical Evaluation of Managers’ Responses Among a Sample of Managers","authors":"Diana R. Sanchez, Saar Van Lysebetten, A. Gibbons","doi":"10.1037/mgr0000049","DOIUrl":null,"url":null,"abstract":"Workplace simulations, often used to assess or train employees, historically rely on human raters who use judgment to evaluate and score the behavior they observe (judgment-based scoring). Such judgments are often complex and holistic, raising concerns about their reliability and susceptibility to bias. Human raters are also resource-intensive; thus, organizations are interested in strategies for reducing the role of human judgment in simulations. For example, using a checklist of discrete, clearly observable behaviors with predefined point values (analytic scoring) might be expected to simplify the rating process and produce more consistent scores. With the use of good text- or voice-recognition software, such a checklist might even be amenable to automation, eliminating the need for human raters altogether. Although the possibility of such potential benefits may appeal to organizations, it is unclear how changing the scoring method in this way may affect the meaning of scores. The authors developed a framework for converting judgment-based scores to analytic scores, using the automated scoring and qualitative content analysis literatures, and applied this framework to the original constructed responses of 84 managers in a workplace simulation. The responses were adapted into discrete behaviors and scored analytically. Results indicated that responses could be adequately summarized using a reasonable number of discrete behaviors, and that analytic scores converged significantly but not strongly with the original judgment-based scores from human raters. We discuss implications for future research and provide recommendations for practitioners considering automated scores in workplace simulations.","PeriodicalId":44734,"journal":{"name":"Psychologist-Manager Journal","volume":null,"pages":null},"PeriodicalIF":0.6000,"publicationDate":"2017-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Psychologist-Manager Journal","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1037/mgr0000049","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Business, Management and Accounting","Score":null,"Total":0}
引用次数: 1

Abstract

Workplace simulations, often used to assess or train employees, historically rely on human raters who use judgment to evaluate and score the behavior they observe (judgment-based scoring). Such judgments are often complex and holistic, raising concerns about their reliability and susceptibility to bias. Human raters are also resource-intensive; thus, organizations are interested in strategies for reducing the role of human judgment in simulations. For example, using a checklist of discrete, clearly observable behaviors with predefined point values (analytic scoring) might be expected to simplify the rating process and produce more consistent scores. With the use of good text- or voice-recognition software, such a checklist might even be amenable to automation, eliminating the need for human raters altogether. Although the possibility of such potential benefits may appeal to organizations, it is unclear how changing the scoring method in this way may affect the meaning of scores. The authors developed a framework for converting judgment-based scores to analytic scores, using the automated scoring and qualitative content analysis literatures, and applied this framework to the original constructed responses of 84 managers in a workplace simulation. The responses were adapted into discrete behaviors and scored analytically. Results indicated that responses could be adequately summarized using a reasonable number of discrete behaviors, and that analytic scores converged significantly but not strongly with the original judgment-based scores from human raters. We discuss implications for future research and provide recommendations for practitioners considering automated scores in workplace simulations.
从基于判断的模拟反应到基于分析的分数:一个过程模型,案例研究,以及管理者样本中管理者反应的实证评估
通常用于评估或培训员工的工作场所模拟,在历史上依赖于人类评分员,他们使用判断来评估和评分他们观察到的行为(基于判断的评分)。这样的判断往往是复杂而全面的,引起了人们对其可靠性和偏见敏感性的担忧。人类评级员也是资源密集型的;因此,组织对减少模拟中人类判断作用的策略感兴趣。例如,使用带有预定义分值(分析评分)的离散的、清晰可观察的行为清单,可能会简化评分过程并产生更一致的评分。通过使用好的文本或语音识别软件,这样的清单甚至可以自动化,完全消除了对人类评分员的需求。虽然这种潜在的好处可能会吸引组织,但目前还不清楚以这种方式改变评分方法会如何影响分数的意义。作者开发了一个框架,利用自动评分和定性内容分析文献,将基于判断的分数转换为分析分数,并将该框架应用于84名管理人员在工作场所模拟中的原始构建反应。这些反应被改编成离散的行为,并进行分析打分。结果表明,使用合理数量的离散行为可以充分总结反应,并且分析得分与人类评分者的原始基于判断的得分显著收敛,但不强烈收敛。我们讨论了对未来研究的影响,并为考虑在工作场所模拟中自动评分的从业者提供了建议。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Psychologist-Manager Journal
Psychologist-Manager Journal PSYCHOLOGY, APPLIED-
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信