{"title":"人工智能聊天机器人的认知领域评估:ChatGPT与双子座解剖教育理解的比较研究","authors":"Arthi Ganapathy, Parul Kaushal","doi":"10.1007/s40670-025-02303-0","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>The integration of AI chatbots into education has gained traction, particularly in medical fields such as anatomy. This study aims to evaluate and compare the responses of ChatGPT 4o mini and Gemini across different cognitive domains of anatomy education.</p><p><strong>Methods: </strong>A cross-sectional study was conducted to assess the responses of these two AI chatbots to a set of anatomy questions selected from the Manual on Competency-Based Undergraduate Curriculum. Questions were categorized into knowledge, comprehension and application levels of cognitive domain. Responses were scored against an answer key prepared by anatomy experts. Relevant comparative statistical analysis was performed.</p><p><strong>Results: </strong>The overall performance of ChatGPT 4o mini (76.15%) was significantly superior to Gemini (72.84%). In application-level questions, ChatGPT 4o mini outperformed Gemini. Conversely, Gemini scored higher in comprehension-level questions (76.88% vs. 73.66%). Both chatbots exhibited factual inaccuracies and limitations in contextually accurate responses, particularly in application-level questions.</p><p><strong>Conclusion: </strong>Both ChatGPT 4o mini and Gemini demonstrate potential as educational tools in anatomy, with strengths and limitations varying by cognitive domain. While AI chatbots can supplement traditional learning methods, they require ongoing refinement and validation. To ensure the responsible integration of AI into medical education, close attention must be devoted to faculty and student training, setting up relevant IT environment and ethical issues. Future research should focus on expanding question pools, incorporating user feedback and comparing with traditional educational approaches to enhance their effectiveness.</p>","PeriodicalId":37113,"journal":{"name":"Medical Science Educator","volume":"35 3","pages":"1295-1304"},"PeriodicalIF":1.8000,"publicationDate":"2025-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12228882/pdf/","citationCount":"0","resultStr":"{\"title\":\"Cognitive Domain Assessment of Artificial Intelligence Chatbots: A Comparative Study Between ChatGPT and Gemini's Understanding of Anatomy Education.\",\"authors\":\"Arthi Ganapathy, Parul Kaushal\",\"doi\":\"10.1007/s40670-025-02303-0\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Purpose: </strong>The integration of AI chatbots into education has gained traction, particularly in medical fields such as anatomy. This study aims to evaluate and compare the responses of ChatGPT 4o mini and Gemini across different cognitive domains of anatomy education.</p><p><strong>Methods: </strong>A cross-sectional study was conducted to assess the responses of these two AI chatbots to a set of anatomy questions selected from the Manual on Competency-Based Undergraduate Curriculum. Questions were categorized into knowledge, comprehension and application levels of cognitive domain. Responses were scored against an answer key prepared by anatomy experts. Relevant comparative statistical analysis was performed.</p><p><strong>Results: </strong>The overall performance of ChatGPT 4o mini (76.15%) was significantly superior to Gemini (72.84%). In application-level questions, ChatGPT 4o mini outperformed Gemini. Conversely, Gemini scored higher in comprehension-level questions (76.88% vs. 73.66%). Both chatbots exhibited factual inaccuracies and limitations in contextually accurate responses, particularly in application-level questions.</p><p><strong>Conclusion: </strong>Both ChatGPT 4o mini and Gemini demonstrate potential as educational tools in anatomy, with strengths and limitations varying by cognitive domain. While AI chatbots can supplement traditional learning methods, they require ongoing refinement and validation. To ensure the responsible integration of AI into medical education, close attention must be devoted to faculty and student training, setting up relevant IT environment and ethical issues. Future research should focus on expanding question pools, incorporating user feedback and comparing with traditional educational approaches to enhance their effectiveness.</p>\",\"PeriodicalId\":37113,\"journal\":{\"name\":\"Medical Science Educator\",\"volume\":\"35 3\",\"pages\":\"1295-1304\"},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2025-02-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12228882/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical Science Educator\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/s40670-025-02303-0\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/6/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q2\",\"JCRName\":\"EDUCATION, SCIENTIFIC DISCIPLINES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical Science Educator","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s40670-025-02303-0","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/6/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"EDUCATION, SCIENTIFIC DISCIPLINES","Score":null,"Total":0}
Cognitive Domain Assessment of Artificial Intelligence Chatbots: A Comparative Study Between ChatGPT and Gemini's Understanding of Anatomy Education.
Purpose: The integration of AI chatbots into education has gained traction, particularly in medical fields such as anatomy. This study aims to evaluate and compare the responses of ChatGPT 4o mini and Gemini across different cognitive domains of anatomy education.
Methods: A cross-sectional study was conducted to assess the responses of these two AI chatbots to a set of anatomy questions selected from the Manual on Competency-Based Undergraduate Curriculum. Questions were categorized into knowledge, comprehension and application levels of cognitive domain. Responses were scored against an answer key prepared by anatomy experts. Relevant comparative statistical analysis was performed.
Results: The overall performance of ChatGPT 4o mini (76.15%) was significantly superior to Gemini (72.84%). In application-level questions, ChatGPT 4o mini outperformed Gemini. Conversely, Gemini scored higher in comprehension-level questions (76.88% vs. 73.66%). Both chatbots exhibited factual inaccuracies and limitations in contextually accurate responses, particularly in application-level questions.
Conclusion: Both ChatGPT 4o mini and Gemini demonstrate potential as educational tools in anatomy, with strengths and limitations varying by cognitive domain. While AI chatbots can supplement traditional learning methods, they require ongoing refinement and validation. To ensure the responsible integration of AI into medical education, close attention must be devoted to faculty and student training, setting up relevant IT environment and ethical issues. Future research should focus on expanding question pools, incorporating user feedback and comparing with traditional educational approaches to enhance their effectiveness.
期刊介绍:
Medical Science Educator is the successor of the journal JIAMSE. It is the peer-reviewed publication of the International Association of Medical Science Educators (IAMSE). The Journal offers all who teach in healthcare the most current information to succeed in their task by publishing scholarly activities, opinions, and resources in medical science education. Published articles focus on teaching the sciences fundamental to modern medicine and health, and include basic science education, clinical teaching, and the use of modern education technologies. The Journal provides the readership a better understanding of teaching and learning techniques in order to advance medical science education.