Jerome Andres, Ivy A. Huang, Areti Tillou, Justin P. Wagner, Catherine E. Lewis, Farin F. Amersi, Timothy R. Donahue, Formosa C. Chen, James X. Wu
{"title":"Item analysis of general surgery multi-institutional mock oral exam: opportunities for quality improvement","authors":"Jerome Andres, Ivy A. Huang, Areti Tillou, Justin P. Wagner, Catherine E. Lewis, Farin F. Amersi, Timothy R. Donahue, Formosa C. Chen, James X. Wu","doi":"10.1007/s44186-023-00173-x","DOIUrl":null,"url":null,"abstract":"Abstract Purpose Mock oral examinations (MOE) prepare general surgery residents for the American Board of Surgery Certifying Exam by assessing their medical knowledge and clinical judgement. There is no standard accepted process for quality analysis among MOE content items. Effective questions should correlate with mastery of MOE content, as well as exam passage. Our aim was to identify opportunities for question improvement via item analysis of a standardized MOE. Methods Retrospective review of testing data from the 2022 Southern California Virtual MOE, which examined 64 general surgery residents from six training programs. Each resident was assessed with 73 exam questions distributed through 12 standardized cases. Study authors indexed questions by clinical topic (e.g. breast, trauma) and competency category (e.g. professionalism, operative approach). We defined MOE passage as mean percentage correct and mean room score within 1 standard deviation of the mean or higher. Questions were assessed for difficulty, discrimination between PGY level, and correlation with MOE passage. Results Passage rate was 77% overall (49/64 residents), with no differences between postgraduate year (PGY) levels. PGY3 residents answered fewer questions correctly vs PGY4 residents (72% vs 78%, p < 0.001) and PGY5 residents (72% vs 82%, p < 0.001). Out of 73 total questions, 17 questions (23.2%) significantly correlated with MOE passage or failure. By competency category, these were predominantly related to patient care (52.9%) and operative approach (23.5%), with fewer related to diagnosis (11.8%), professional behavior (5.9%), and decision to operate (5.9%). By clinical topic these were equally distributed between trauma (17.7%), large intestine (17.7%), endocrine (17.7%), and surgical critical care (17.7%), with fewer in breast (11.8%), stomach (11.8%), and pediatric surgery (5.9%). We identified two types of ineffective questions: 1) questions answered correctly by 100% of test-takers with no discriminatory ability (n = 3); and 2) questions that varied inversely with exam passage (n = 11). In total, 19% (14/73) of exam questions were deemed ineffective. Conclusions Item analysis of multi-institutional mock oral exam found that 23% of questions correlated with exam passage or failure, effectively discriminating which examinees had mastery of MOE content. We also recognized 19% of questions as ineffective that can be targeted for improvement.","PeriodicalId":472275,"journal":{"name":"Global Surgical Education - Journal of the Association for Surgical Education","volume":"58 10","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Global Surgical Education - Journal of the Association for Surgical Education","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s44186-023-00173-x","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Abstract Purpose Mock oral examinations (MOE) prepare general surgery residents for the American Board of Surgery Certifying Exam by assessing their medical knowledge and clinical judgement. There is no standard accepted process for quality analysis among MOE content items. Effective questions should correlate with mastery of MOE content, as well as exam passage. Our aim was to identify opportunities for question improvement via item analysis of a standardized MOE. Methods Retrospective review of testing data from the 2022 Southern California Virtual MOE, which examined 64 general surgery residents from six training programs. Each resident was assessed with 73 exam questions distributed through 12 standardized cases. Study authors indexed questions by clinical topic (e.g. breast, trauma) and competency category (e.g. professionalism, operative approach). We defined MOE passage as mean percentage correct and mean room score within 1 standard deviation of the mean or higher. Questions were assessed for difficulty, discrimination between PGY level, and correlation with MOE passage. Results Passage rate was 77% overall (49/64 residents), with no differences between postgraduate year (PGY) levels. PGY3 residents answered fewer questions correctly vs PGY4 residents (72% vs 78%, p < 0.001) and PGY5 residents (72% vs 82%, p < 0.001). Out of 73 total questions, 17 questions (23.2%) significantly correlated with MOE passage or failure. By competency category, these were predominantly related to patient care (52.9%) and operative approach (23.5%), with fewer related to diagnosis (11.8%), professional behavior (5.9%), and decision to operate (5.9%). By clinical topic these were equally distributed between trauma (17.7%), large intestine (17.7%), endocrine (17.7%), and surgical critical care (17.7%), with fewer in breast (11.8%), stomach (11.8%), and pediatric surgery (5.9%). We identified two types of ineffective questions: 1) questions answered correctly by 100% of test-takers with no discriminatory ability (n = 3); and 2) questions that varied inversely with exam passage (n = 11). In total, 19% (14/73) of exam questions were deemed ineffective. Conclusions Item analysis of multi-institutional mock oral exam found that 23% of questions correlated with exam passage or failure, effectively discriminating which examinees had mastery of MOE content. We also recognized 19% of questions as ineffective that can be targeted for improvement.
摘要目的通过模拟口腔考试(MOE)评估普通外科住院医师的医学知识和临床判断能力,为其参加美国外科委员会认证考试做准备。对于MOE内容项的质量分析,目前还没有一个公认的标准流程。有效的问题应该与对教材内容的掌握以及考试的通过相关联。我们的目的是通过标准化MOE的项目分析来确定问题改进的机会。方法回顾性分析2022年南加州虚拟MOE的测试数据,该数据对来自6个培训项目的64名普通外科住院医生进行了检查。每位住院医师通过12个标准化案例分发73道试题进行评估。研究作者根据临床主题(如乳房,创伤)和能力类别(如专业,手术方法)对问题进行索引。我们将MOE通道定义为平均正确率和平均房间评分在平均值的1个标准差或更高。评估问题的难度、PGY水平之间的区别以及与MOE通过的相关性。结果总体通过率为77%(49/64名居民),研究生年级(PGY)水平之间无差异。PGY3的居民比PGY4的居民回答正确的问题更少(72%对78%,p <0.001)和PGY5居民(72% vs 82%, p <0.001)。在总共73个问题中,17个问题(23.2%)与MOE通过或不通过显著相关。从能力类别来看,这些主要与患者护理(52.9%)和手术方式(23.5%)相关,与诊断(11.8%)、专业行为(5.9%)和手术决定(5.9%)相关的较少。按临床主题划分,这些病例平均分布在创伤(17.7%)、大肠(17.7%)、内分泌(17.7%)和外科重症监护(17.7%),较少分布在乳腺(11.8%)、胃(11.8%)和儿科外科(5.9%)。我们确定了两种类型的无效问题:1)100%没有区别能力的考生正确回答的问题(n = 3);2)与考试内容成反比的问题(n = 11)。总共有19%(14/73)的试题被认为是无效的。结论多机构模拟口语考试的项目分析发现,23%的试题与考试通过或不通过相关,可以有效判别考生是否掌握了MOE的内容。我们还认识到19%的问题是无效的,可以有针对性地进行改进。