Monalisa Hassan MD , Marco Ayad MD , Christine Nembhard MD , Andrea Hayes-Dixon MD , Anna Lin MD , Mahin Janjua MBBS , Jan Franko MD, PhD, MMM , May Tee MD, MPH, FACS
{"title":"人工智能与人工遴选未来外科住院医师的比较","authors":"Monalisa Hassan MD , Marco Ayad MD , Christine Nembhard MD , Andrea Hayes-Dixon MD , Anna Lin MD , Mahin Janjua MBBS , Jan Franko MD, PhD, MMM , May Tee MD, MPH, FACS","doi":"10.1016/j.jsurg.2024.103308","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>Artificial Intelligence (AI) in the selection of residency program applicants is a new tool that is gaining traction, with the aim of screening high numbers of applicants while introducing objectivity and mitigating bias in a traditionally subjective process. This study aims to compare applicants screened by an AI software to a single Program Director (PD) for interview selection.</div></div><div><h3>Methods</h3><div>A single PD at an ACGME-accredited, academic general surgery program screened applicants. A parallel screen by AI software, programmed by the same PD, was conducted on the same pool of applicants. Weighted preferences were assigned in the following order: personal statement, research, medical school rankings, letters of recommendation, personal qualities, board scores, graduate degree, geographic preference, past experiences, program signal, honor society membership, and multilingualism. Statistical analyses were conducted by chi-square, ANOVA, and independent two-sided t-tests.</div></div><div><h3>Results</h3><div>Out of 1235 applications, 144 applications were PD-selected and 150 AI-selected (294 top applications). Twenty applications (7.3%) were both PD and AI selected for a total analysis cohort of 274 prospective residents. We performed two analyses: 1) PD-selected vs. AI-selected vs. Both and 2) PD-selected vs. AI-selected with the overlapping applicants censored. For the first analysis, AI selected significantly: more White/Hispanic applicants (p < 0.001), less signals (p < 0.001), more AOA honors society (p = 0.016), and more publications (p < 0.001). When censoring overlapping PD and AI selection, AI selected significantly: more White/Hispanic applicants (p < 0.001), less signals (p < 0.001), more US medical graduates (p = 0.027), less applicants needing visa sponsorship (p = 0.01), younger applicants (p = 0.024), higher USMLE Step 2 CK scores (p < 0.001), and more publications (p < 0.001).</div></div><div><h3>Conclusions</h3><div>There was only a 7% overlap between PD-selected and AI-selected applicants for interview screening in the same applicant pool. Despite the same PD educating the AI software, the 2 application pools differed significantly. In its present state, AI may be utilized as a tool in resident application selection but should not completely replace human review. We recommend careful analysis of the performance of each AI model in the respective environment of each institution applying it, as it may alter the group of interviewees.</div></div>","PeriodicalId":50033,"journal":{"name":"Journal of Surgical Education","volume":"82 1","pages":"Article 103308"},"PeriodicalIF":2.6000,"publicationDate":"2024-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Artificial Intelligence Compared to Manual Selection of Prospective Surgical Residents\",\"authors\":\"Monalisa Hassan MD , Marco Ayad MD , Christine Nembhard MD , Andrea Hayes-Dixon MD , Anna Lin MD , Mahin Janjua MBBS , Jan Franko MD, PhD, MMM , May Tee MD, MPH, FACS\",\"doi\":\"10.1016/j.jsurg.2024.103308\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Background</h3><div>Artificial Intelligence (AI) in the selection of residency program applicants is a new tool that is gaining traction, with the aim of screening high numbers of applicants while introducing objectivity and mitigating bias in a traditionally subjective process. This study aims to compare applicants screened by an AI software to a single Program Director (PD) for interview selection.</div></div><div><h3>Methods</h3><div>A single PD at an ACGME-accredited, academic general surgery program screened applicants. A parallel screen by AI software, programmed by the same PD, was conducted on the same pool of applicants. Weighted preferences were assigned in the following order: personal statement, research, medical school rankings, letters of recommendation, personal qualities, board scores, graduate degree, geographic preference, past experiences, program signal, honor society membership, and multilingualism. Statistical analyses were conducted by chi-square, ANOVA, and independent two-sided t-tests.</div></div><div><h3>Results</h3><div>Out of 1235 applications, 144 applications were PD-selected and 150 AI-selected (294 top applications). Twenty applications (7.3%) were both PD and AI selected for a total analysis cohort of 274 prospective residents. We performed two analyses: 1) PD-selected vs. AI-selected vs. Both and 2) PD-selected vs. AI-selected with the overlapping applicants censored. For the first analysis, AI selected significantly: more White/Hispanic applicants (p < 0.001), less signals (p < 0.001), more AOA honors society (p = 0.016), and more publications (p < 0.001). When censoring overlapping PD and AI selection, AI selected significantly: more White/Hispanic applicants (p < 0.001), less signals (p < 0.001), more US medical graduates (p = 0.027), less applicants needing visa sponsorship (p = 0.01), younger applicants (p = 0.024), higher USMLE Step 2 CK scores (p < 0.001), and more publications (p < 0.001).</div></div><div><h3>Conclusions</h3><div>There was only a 7% overlap between PD-selected and AI-selected applicants for interview screening in the same applicant pool. Despite the same PD educating the AI software, the 2 application pools differed significantly. In its present state, AI may be utilized as a tool in resident application selection but should not completely replace human review. We recommend careful analysis of the performance of each AI model in the respective environment of each institution applying it, as it may alter the group of interviewees.</div></div>\",\"PeriodicalId\":50033,\"journal\":{\"name\":\"Journal of Surgical Education\",\"volume\":\"82 1\",\"pages\":\"Article 103308\"},\"PeriodicalIF\":2.6000,\"publicationDate\":\"2024-11-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Surgical Education\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1931720424004562\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"EDUCATION, SCIENTIFIC DISCIPLINES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Surgical Education","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1931720424004562","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION, SCIENTIFIC DISCIPLINES","Score":null,"Total":0}
引用次数: 0
摘要
背景人工智能(AI)在住院医师项目申请者的筛选中是一种新的工具,正在受到越来越多的关注,其目的是筛选大量申请者,同时在传统的主观过程中引入客观性并减少偏见。本研究旨在将人工智能软件筛选的申请人与单一项目主任(PD)进行面试筛选的申请人进行比较。由同一项目主任编程的人工智能软件对同一申请者库进行了平行筛选。加权偏好按以下顺序排列:个人陈述、研究、医学院排名、推荐信、个人素质、委员会评分、研究生学位、地域偏好、过往经历、项目信号、荣誉协会会员资格和多语言能力。统计分析采用卡方检验、方差分析和独立双侧 t 检验。有 20 份申请(占 7.3%)同时入选了 PD 和 AI,总共有 274 名准住院医师接受了分析。我们进行了两项分析:1)PD 入选 vs. AI 入选 vs. 同时入选;2)PD 入选 vs. AI 入选,并剔除重叠申请者。在第一项分析中,AI 选拔的结果如下:白人/西班牙裔申请人更多(p <0.001)、信号更少(p <0.001)、AOA 荣誉学会更多(p = 0.016)、发表论文更多(p <0.001)。当剔除重叠的 PD 和 AI 选择时,AI 选择了以下显著指标:更多的白人/西班牙裔申请人(p < 0.001)、更少的信号(p < 0.001)、更多的美国医学毕业生(p = 0.027)、更少的需要签证担保的申请人(p = 0.结论在同一申请人库中,PD 筛选和 AI 筛选的申请人在面试筛选方面只有 7% 的重叠。尽管采用了相同的 PD 教育人工智能软件,但两个申请者库还是存在显著差异。从目前的情况来看,人工智能可以作为驻地申请筛选的一种工具,但不应完全取代人工审核。我们建议仔细分析每种人工智能模型在各机构应用环境中的表现,因为它可能会改变面试者群体。
Artificial Intelligence Compared to Manual Selection of Prospective Surgical Residents
Background
Artificial Intelligence (AI) in the selection of residency program applicants is a new tool that is gaining traction, with the aim of screening high numbers of applicants while introducing objectivity and mitigating bias in a traditionally subjective process. This study aims to compare applicants screened by an AI software to a single Program Director (PD) for interview selection.
Methods
A single PD at an ACGME-accredited, academic general surgery program screened applicants. A parallel screen by AI software, programmed by the same PD, was conducted on the same pool of applicants. Weighted preferences were assigned in the following order: personal statement, research, medical school rankings, letters of recommendation, personal qualities, board scores, graduate degree, geographic preference, past experiences, program signal, honor society membership, and multilingualism. Statistical analyses were conducted by chi-square, ANOVA, and independent two-sided t-tests.
Results
Out of 1235 applications, 144 applications were PD-selected and 150 AI-selected (294 top applications). Twenty applications (7.3%) were both PD and AI selected for a total analysis cohort of 274 prospective residents. We performed two analyses: 1) PD-selected vs. AI-selected vs. Both and 2) PD-selected vs. AI-selected with the overlapping applicants censored. For the first analysis, AI selected significantly: more White/Hispanic applicants (p < 0.001), less signals (p < 0.001), more AOA honors society (p = 0.016), and more publications (p < 0.001). When censoring overlapping PD and AI selection, AI selected significantly: more White/Hispanic applicants (p < 0.001), less signals (p < 0.001), more US medical graduates (p = 0.027), less applicants needing visa sponsorship (p = 0.01), younger applicants (p = 0.024), higher USMLE Step 2 CK scores (p < 0.001), and more publications (p < 0.001).
Conclusions
There was only a 7% overlap between PD-selected and AI-selected applicants for interview screening in the same applicant pool. Despite the same PD educating the AI software, the 2 application pools differed significantly. In its present state, AI may be utilized as a tool in resident application selection but should not completely replace human review. We recommend careful analysis of the performance of each AI model in the respective environment of each institution applying it, as it may alter the group of interviewees.
期刊介绍:
The Journal of Surgical Education (JSE) is dedicated to advancing the field of surgical education through original research. The journal publishes research articles in all surgical disciplines on topics relative to the education of surgical students, residents, and fellows, as well as practicing surgeons. Our readers look to JSE for timely, innovative research findings from the international surgical education community. As the official journal of the Association of Program Directors in Surgery (APDS), JSE publishes the proceedings of the annual APDS meeting held during Surgery Education Week.