Can artificial intelligence read between the lines: Utilizing ChatGPT to evaluate medical students' implicit attitudes towards doctor-patient relationship.
{"title":"Can artificial intelligence read between the lines: Utilizing ChatGPT to evaluate medical students' implicit attitudes towards doctor-patient relationship.","authors":"Wenqi Geng, Yinan Jiang, Wei Zhai, Xiaohui Zhao, Qing Zhao, Jianqiang Li, Jinya Cao, Lili Shi","doi":"10.1080/0142159X.2025.2515971","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>To explore ChatGPT's utility in evaluating medical students' implicit attitudes toward the doctor-patient relationship (DPR).</p><p><strong>Materials and methods: </strong>This study analyzed interview transcripts from 10 medical students, categorizing implicit DPR attitudes into Care and Share dimensions, each with 5 levels. We first assessed ChatGPT's ability to identify DPR-related textual content, then compared grading results from experts, ChatGPT, and participants' self-evaluations. Finally, experts evaluated ChatGPT's performance acceptability.</p><p><strong>Results: </strong>ChatGPT annotated fewer DPR-related segments than human experts. In grading, pre-course scores from experts and ChatGPT were comparable but lower than self-assessments. Post-course, expert scores were lower than ChatGPT's and further below self-assessments. ChatGPT achieved an accuracy of 0.84-0.85, precision of 0.81-0.85, recall of 0.84-0.85, and F1 score of 0.82-0.84 for attitude classification, with an average acceptability score of 3.9/5.</p><p><strong>Conclusions: </strong>Large language models (LLMs) demonstrated high consistency with human experts in judging implicit attitudes. Future research should optimize LLMs and replicate this framework across diverse contexts with larger samples.</p>","PeriodicalId":18643,"journal":{"name":"Medical Teacher","volume":" ","pages":"1-8"},"PeriodicalIF":3.3000,"publicationDate":"2025-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical Teacher","FirstCategoryId":"95","ListUrlMain":"https://doi.org/10.1080/0142159X.2025.2515971","RegionNum":2,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION, SCIENTIFIC DISCIPLINES","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: To explore ChatGPT's utility in evaluating medical students' implicit attitudes toward the doctor-patient relationship (DPR).
Materials and methods: This study analyzed interview transcripts from 10 medical students, categorizing implicit DPR attitudes into Care and Share dimensions, each with 5 levels. We first assessed ChatGPT's ability to identify DPR-related textual content, then compared grading results from experts, ChatGPT, and participants' self-evaluations. Finally, experts evaluated ChatGPT's performance acceptability.
Results: ChatGPT annotated fewer DPR-related segments than human experts. In grading, pre-course scores from experts and ChatGPT were comparable but lower than self-assessments. Post-course, expert scores were lower than ChatGPT's and further below self-assessments. ChatGPT achieved an accuracy of 0.84-0.85, precision of 0.81-0.85, recall of 0.84-0.85, and F1 score of 0.82-0.84 for attitude classification, with an average acceptability score of 3.9/5.
Conclusions: Large language models (LLMs) demonstrated high consistency with human experts in judging implicit attitudes. Future research should optimize LLMs and replicate this framework across diverse contexts with larger samples.
期刊介绍:
Medical Teacher provides accounts of new teaching methods, guidance on structuring courses and assessing achievement, and serves as a forum for communication between medical teachers and those involved in general education. In particular, the journal recognizes the problems teachers have in keeping up-to-date with the developments in educational methods that lead to more effective teaching and learning at a time when the content of the curriculum—from medical procedures to policy changes in health care provision—is also changing. The journal features reports of innovation and research in medical education, case studies, survey articles, practical guidelines, reviews of current literature and book reviews. All articles are peer reviewed.