ChatGPT能在医学院解剖考试中生成可接受的基于案例的选择题吗?题目难度与辨别力的初步研究。

IF 2.3 4区 医学 Q1 ANATOMY & MORPHOLOGY
Clinical Anatomy Pub Date : 2025-03-24 DOI:10.1002/ca.24271
Yavuz Selim Kıyak, Ayşe Soylu, Özlem Coşkun, Işıl İrem Budakoğlu, Tuncay Veysel Peker
{"title":"ChatGPT能在医学院解剖考试中生成可接受的基于案例的选择题吗?题目难度与辨别力的初步研究。","authors":"Yavuz Selim Kıyak,&nbsp;Ayşe Soylu,&nbsp;Özlem Coşkun,&nbsp;Işıl İrem Budakoğlu,&nbsp;Tuncay Veysel Peker","doi":"10.1002/ca.24271","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Developing high-quality multiple-choice questions (MCQs) for medical school exams is effortful and time-consuming. In this study, we investigated the ability of ChatGPT to generate case-based anatomy MCQs with acceptable levels of item difficulty and discrimination for medical school exams. We used ChatGPT to generate case-based anatomy MCQs for an endocrine and urogenital system exam based on a framework for artificial intelligence (AI)-assisted item generation. The questions were evaluated by experts, approved by the department, and administered to 502 second-year medical students (372 Turkish-language, 130 English-language). The items were analyzed to determine the discrimination and difficulty indices. The item discrimination indices ranged from 0.29 to 0.54, indicating acceptable differentiation between high- and low-performing students. All items in Turkish (six out of six) and five out of six in English met the higher discrimination threshold (≥ 0.30) required for large-scale standardized tests. The item difficulty indices ranged from 0.41 to 0.89, most items falling within the moderate difficulty range (0.20–0.80). Therefore, it was concluded that ChatGPT can generate case-based anatomy MCQs with acceptable psychometric properties, offering a promising tool for medical educators. However, human expertise remains crucial for reviewing and refining AI-generated assessment items. Future research should explore AI-generated MCQs across various anatomy topics and investigate different AI models for question generation.</p>\n </div>","PeriodicalId":50687,"journal":{"name":"Clinical Anatomy","volume":"38 4","pages":"505-510"},"PeriodicalIF":2.3000,"publicationDate":"2025-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Can ChatGPT Generate Acceptable Case-Based Multiple-Choice Questions for Medical School Anatomy Exams? A Pilot Study on Item Difficulty and Discrimination\",\"authors\":\"Yavuz Selim Kıyak,&nbsp;Ayşe Soylu,&nbsp;Özlem Coşkun,&nbsp;Işıl İrem Budakoğlu,&nbsp;Tuncay Veysel Peker\",\"doi\":\"10.1002/ca.24271\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <p>Developing high-quality multiple-choice questions (MCQs) for medical school exams is effortful and time-consuming. In this study, we investigated the ability of ChatGPT to generate case-based anatomy MCQs with acceptable levels of item difficulty and discrimination for medical school exams. We used ChatGPT to generate case-based anatomy MCQs for an endocrine and urogenital system exam based on a framework for artificial intelligence (AI)-assisted item generation. The questions were evaluated by experts, approved by the department, and administered to 502 second-year medical students (372 Turkish-language, 130 English-language). The items were analyzed to determine the discrimination and difficulty indices. The item discrimination indices ranged from 0.29 to 0.54, indicating acceptable differentiation between high- and low-performing students. All items in Turkish (six out of six) and five out of six in English met the higher discrimination threshold (≥ 0.30) required for large-scale standardized tests. The item difficulty indices ranged from 0.41 to 0.89, most items falling within the moderate difficulty range (0.20–0.80). Therefore, it was concluded that ChatGPT can generate case-based anatomy MCQs with acceptable psychometric properties, offering a promising tool for medical educators. However, human expertise remains crucial for reviewing and refining AI-generated assessment items. Future research should explore AI-generated MCQs across various anatomy topics and investigate different AI models for question generation.</p>\\n </div>\",\"PeriodicalId\":50687,\"journal\":{\"name\":\"Clinical Anatomy\",\"volume\":\"38 4\",\"pages\":\"505-510\"},\"PeriodicalIF\":2.3000,\"publicationDate\":\"2025-03-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Clinical Anatomy\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/ca.24271\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ANATOMY & MORPHOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Clinical Anatomy","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/ca.24271","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ANATOMY & MORPHOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

为医学院考试开发高质量的多项选择题(mcq)既费力又耗时。在这项研究中,我们调查了ChatGPT生成基于案例的解剖学mcq的能力,这些mcq具有可接受的项目难度和区分水平,适用于医学院考试。基于人工智能(AI)辅助项目生成框架,我们使用ChatGPT为内分泌和泌尿生殖系统检查生成基于病例的解剖mcq。这些问题由专家评估,经部门批准,并向502名二年级医科学生(372名土耳其语学生,130名英语学生)发放。对项目进行分析,确定识别度和难度指数。项目区分指数在0.29 ~ 0.54之间,表明高差生与低差生之间存在可接受的差异。土耳其语的所有项目(6个项目中的6个)和英语的6个项目中的5个都符合大规模标准化测试所需的较高判别阈值(≥0.30)。项目难度指数范围为0.41 ~ 0.89,大部分项目处于中等难度范围(0.20 ~ 0.80)。因此,ChatGPT可以生成具有可接受的心理测量属性的基于病例的解剖mcq,为医学教育者提供了一个有前途的工具。然而,人类的专业知识对于审查和完善人工智能生成的评估项目仍然至关重要。未来的研究应该探索人工智能在各种解剖学主题中生成的mcq,并研究不同的人工智能模型来生成问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Can ChatGPT Generate Acceptable Case-Based Multiple-Choice Questions for Medical School Anatomy Exams? A Pilot Study on Item Difficulty and Discrimination

Developing high-quality multiple-choice questions (MCQs) for medical school exams is effortful and time-consuming. In this study, we investigated the ability of ChatGPT to generate case-based anatomy MCQs with acceptable levels of item difficulty and discrimination for medical school exams. We used ChatGPT to generate case-based anatomy MCQs for an endocrine and urogenital system exam based on a framework for artificial intelligence (AI)-assisted item generation. The questions were evaluated by experts, approved by the department, and administered to 502 second-year medical students (372 Turkish-language, 130 English-language). The items were analyzed to determine the discrimination and difficulty indices. The item discrimination indices ranged from 0.29 to 0.54, indicating acceptable differentiation between high- and low-performing students. All items in Turkish (six out of six) and five out of six in English met the higher discrimination threshold (≥ 0.30) required for large-scale standardized tests. The item difficulty indices ranged from 0.41 to 0.89, most items falling within the moderate difficulty range (0.20–0.80). Therefore, it was concluded that ChatGPT can generate case-based anatomy MCQs with acceptable psychometric properties, offering a promising tool for medical educators. However, human expertise remains crucial for reviewing and refining AI-generated assessment items. Future research should explore AI-generated MCQs across various anatomy topics and investigate different AI models for question generation.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Clinical Anatomy
Clinical Anatomy 医学-解剖学与形态学
CiteScore
5.50
自引率
12.50%
发文量
154
审稿时长
3 months
期刊介绍: Clinical Anatomy is the Official Journal of the American Association of Clinical Anatomists and the British Association of Clinical Anatomists. The goal of Clinical Anatomy is to provide a medium for the exchange of current information between anatomists and clinicians. This journal embraces anatomy in all its aspects as applied to medical practice. Furthermore, the journal assists physicians and other health care providers in keeping abreast of new methodologies for patient management and informs educators of new developments in clinical anatomy and teaching techniques. Clinical Anatomy publishes original and review articles of scientific, clinical, and educational interest. Papers covering the application of anatomic principles to the solution of clinical problems and/or the application of clinical observations to expand anatomic knowledge are welcomed.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信