Arjun Watane, Brittany M Perzia, Madison E Weiss, Andrea A Tooley, Emily Li, Larissa A Habib, Phillip A Tenzel, Michelle M Maeng
{"title":"对患者进行上眼睑整形手术的常见问题进行了探讨。","authors":"Arjun Watane, Brittany M Perzia, Madison E Weiss, Andrea A Tooley, Emily Li, Larissa A Habib, Phillip A Tenzel, Michelle M Maeng","doi":"10.1080/01676830.2024.2435930","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Online health information seekers may access information produced by artificial intelligence language models such as ChatGPT (OpenAI). The medical field may pose a significant challenge for incorporating these applications given the training and experience needed to master clinical reasoning. The objective was to evaluate the performance of ChatGPT responses compared to human oculofacial plastic surgeon (OPS) responses to FAQs about an upper eyelid blepharoplasty procedure.</p><p><strong>Methods: </strong>A cross-sectional survey was conducted. Three OPS trained by the American Society of Ophthalmic Plastic and Reconstructive Surgery (ASOPRS) and three ChatGPT instances each answered 6 frequently asked questions (FAQs) about an upper eyelid blepharoplasty procedure. Two blinded ASOPRS-trained OPS evaluated each response for their accuracy, comprehensiveness, and personal answer similarity based on a Likert scale (1=strongly disagree; 5=strongly agree).</p><p><strong>Results: </strong>ChatGPT achieved a mean Likert scale score of 3.8 (SD 0.9) in accuracy, 3.6 (SD 1.1) in comprehensiveness, and 3.2 (SD 1.1) in personal answer similarity. In comparison, OPS achieved a mean score of 3.6 (SD 1.2) in accuracy (<i>p</i> = .72), 3.0 (SD 1.1) in comprehensiveness (<i>p</i> = .03), and 2.9 (SD 1.1) in personal answer similarity (<i>p</i> = .66).</p><p><strong>Conclusions: </strong>ChatGPT was non-inferior to OPS in answering upper eyelid blepharoplasty FAQs. Compared to OPS, ChatGPT achieved better comprehensiveness ratings and non-inferior accuracy and personal answer similarity ratings. This study poses the potential for ChatGPT to serve as an adjunct to OPS for patient education but not a replacement. However, safeguards to protect patients from possible harm must be implemented.</p>","PeriodicalId":47421,"journal":{"name":"Orbit-The International Journal on Orbital Disorders-Oculoplastic and Lacrimal Surgery","volume":" ","pages":"1-4"},"PeriodicalIF":0.9000,"publicationDate":"2024-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ChatGPT and frequently asked patient questions for upper eyelid blepharoplasty surgery.\",\"authors\":\"Arjun Watane, Brittany M Perzia, Madison E Weiss, Andrea A Tooley, Emily Li, Larissa A Habib, Phillip A Tenzel, Michelle M Maeng\",\"doi\":\"10.1080/01676830.2024.2435930\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Purpose: </strong>Online health information seekers may access information produced by artificial intelligence language models such as ChatGPT (OpenAI). The medical field may pose a significant challenge for incorporating these applications given the training and experience needed to master clinical reasoning. The objective was to evaluate the performance of ChatGPT responses compared to human oculofacial plastic surgeon (OPS) responses to FAQs about an upper eyelid blepharoplasty procedure.</p><p><strong>Methods: </strong>A cross-sectional survey was conducted. Three OPS trained by the American Society of Ophthalmic Plastic and Reconstructive Surgery (ASOPRS) and three ChatGPT instances each answered 6 frequently asked questions (FAQs) about an upper eyelid blepharoplasty procedure. Two blinded ASOPRS-trained OPS evaluated each response for their accuracy, comprehensiveness, and personal answer similarity based on a Likert scale (1=strongly disagree; 5=strongly agree).</p><p><strong>Results: </strong>ChatGPT achieved a mean Likert scale score of 3.8 (SD 0.9) in accuracy, 3.6 (SD 1.1) in comprehensiveness, and 3.2 (SD 1.1) in personal answer similarity. In comparison, OPS achieved a mean score of 3.6 (SD 1.2) in accuracy (<i>p</i> = .72), 3.0 (SD 1.1) in comprehensiveness (<i>p</i> = .03), and 2.9 (SD 1.1) in personal answer similarity (<i>p</i> = .66).</p><p><strong>Conclusions: </strong>ChatGPT was non-inferior to OPS in answering upper eyelid blepharoplasty FAQs. Compared to OPS, ChatGPT achieved better comprehensiveness ratings and non-inferior accuracy and personal answer similarity ratings. This study poses the potential for ChatGPT to serve as an adjunct to OPS for patient education but not a replacement. However, safeguards to protect patients from possible harm must be implemented.</p>\",\"PeriodicalId\":47421,\"journal\":{\"name\":\"Orbit-The International Journal on Orbital Disorders-Oculoplastic and Lacrimal Surgery\",\"volume\":\" \",\"pages\":\"1-4\"},\"PeriodicalIF\":0.9000,\"publicationDate\":\"2024-12-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Orbit-The International Journal on Orbital Disorders-Oculoplastic and Lacrimal Surgery\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1080/01676830.2024.2435930\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"OPHTHALMOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Orbit-The International Journal on Orbital Disorders-Oculoplastic and Lacrimal Surgery","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/01676830.2024.2435930","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
ChatGPT and frequently asked patient questions for upper eyelid blepharoplasty surgery.
Purpose: Online health information seekers may access information produced by artificial intelligence language models such as ChatGPT (OpenAI). The medical field may pose a significant challenge for incorporating these applications given the training and experience needed to master clinical reasoning. The objective was to evaluate the performance of ChatGPT responses compared to human oculofacial plastic surgeon (OPS) responses to FAQs about an upper eyelid blepharoplasty procedure.
Methods: A cross-sectional survey was conducted. Three OPS trained by the American Society of Ophthalmic Plastic and Reconstructive Surgery (ASOPRS) and three ChatGPT instances each answered 6 frequently asked questions (FAQs) about an upper eyelid blepharoplasty procedure. Two blinded ASOPRS-trained OPS evaluated each response for their accuracy, comprehensiveness, and personal answer similarity based on a Likert scale (1=strongly disagree; 5=strongly agree).
Results: ChatGPT achieved a mean Likert scale score of 3.8 (SD 0.9) in accuracy, 3.6 (SD 1.1) in comprehensiveness, and 3.2 (SD 1.1) in personal answer similarity. In comparison, OPS achieved a mean score of 3.6 (SD 1.2) in accuracy (p = .72), 3.0 (SD 1.1) in comprehensiveness (p = .03), and 2.9 (SD 1.1) in personal answer similarity (p = .66).
Conclusions: ChatGPT was non-inferior to OPS in answering upper eyelid blepharoplasty FAQs. Compared to OPS, ChatGPT achieved better comprehensiveness ratings and non-inferior accuracy and personal answer similarity ratings. This study poses the potential for ChatGPT to serve as an adjunct to OPS for patient education but not a replacement. However, safeguards to protect patients from possible harm must be implemented.
期刊介绍:
Orbit is the international medium covering developments and results from the variety of medical disciplines that overlap and converge in the field of orbital disorders: ophthalmology, otolaryngology, reconstructive and maxillofacial surgery, medicine and endocrinology, radiology, radiotherapy and oncology, neurology, neuroophthalmology and neurosurgery, pathology and immunology, haematology.