Rachel N Rohrich, Karen R Li, Christian X Lava, Isabel Snee, Sami Alahmadi, Richard C Youn, John S Steinberg, Jayson M Atves, Christopher E Attinger, Karen K Evans
{"title":"Consulting the Digital Doctor: Efficacy of ChatGPT-3.5 in Answering Questions Related to Diabetic Foot Ulcer Care.","authors":"Rachel N Rohrich, Karen R Li, Christian X Lava, Isabel Snee, Sami Alahmadi, Richard C Youn, John S Steinberg, Jayson M Atves, Christopher E Attinger, Karen K Evans","doi":"10.1097/ASW.0000000000000317","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Diabetic foot ulcer (DFU) care is a challenge in reconstructive surgery. Artificial intelligence (AI) tools represent a new resource for patients with DFUs to seek information.</p><p><strong>Objective: </strong>To evaluate the efficacy of ChatGPT-3.5 in responding to frequently asked questions related to DFU care.</p><p><strong>Methods: </strong>Researchers posed 11 DFU care questions to ChatGPT-3.5 in December 2023. Questions were divided into topic categories of wound care, concerning symptoms, and surgical management. Four plastic surgeons in the authors' wound care department evaluated responses on a 10-point Likert-type scale for accuracy, comprehensiveness, and danger, in addition to providing qualitative feedback. Readability was assessed using 10 readability indexes.</p><p><strong>Results: </strong>ChatGPT-3.5 answered questions with a mean accuracy of 8.7±0.3, comprehensiveness of 8.0±0.7, and danger of 2.2±0.6. ChatGPT-3.5 answered at the mean grade level of 11.9±1.8. Physician reviewers complimented the simplicity of the responses (n=11/11) and the AI's ability to provide general information (n=4/11). Three responses presented incorrect information, and the majority of responses (n=10/11) left out key information, such as deep vein thrombosis symptoms and comorbid conditions impacting limb salvage.</p><p><strong>Conclusions: </strong>The researchers observed that ChatGPT-3.5 provided misinformation, omitted crucial details, and responded at nearly 4 grade levels higher than the American average. However, ChatGPT-3.5 was sufficient in its ability to provide general information, which may enable patients with DFUs to make more informed decisions and better engage in their care. Physicians must proactively address the potential benefits and limitations of AI.</p>","PeriodicalId":7489,"journal":{"name":"Advances in Skin & Wound Care","volume":" ","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2025-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in Skin & Wound Care","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1097/ASW.0000000000000317","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"DERMATOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Background: Diabetic foot ulcer (DFU) care is a challenge in reconstructive surgery. Artificial intelligence (AI) tools represent a new resource for patients with DFUs to seek information.
Objective: To evaluate the efficacy of ChatGPT-3.5 in responding to frequently asked questions related to DFU care.
Methods: Researchers posed 11 DFU care questions to ChatGPT-3.5 in December 2023. Questions were divided into topic categories of wound care, concerning symptoms, and surgical management. Four plastic surgeons in the authors' wound care department evaluated responses on a 10-point Likert-type scale for accuracy, comprehensiveness, and danger, in addition to providing qualitative feedback. Readability was assessed using 10 readability indexes.
Results: ChatGPT-3.5 answered questions with a mean accuracy of 8.7±0.3, comprehensiveness of 8.0±0.7, and danger of 2.2±0.6. ChatGPT-3.5 answered at the mean grade level of 11.9±1.8. Physician reviewers complimented the simplicity of the responses (n=11/11) and the AI's ability to provide general information (n=4/11). Three responses presented incorrect information, and the majority of responses (n=10/11) left out key information, such as deep vein thrombosis symptoms and comorbid conditions impacting limb salvage.
Conclusions: The researchers observed that ChatGPT-3.5 provided misinformation, omitted crucial details, and responded at nearly 4 grade levels higher than the American average. However, ChatGPT-3.5 was sufficient in its ability to provide general information, which may enable patients with DFUs to make more informed decisions and better engage in their care. Physicians must proactively address the potential benefits and limitations of AI.
期刊介绍:
A peer-reviewed, multidisciplinary journal, Advances in Skin & Wound Care is highly regarded for its unique balance of cutting-edge original research and practical clinical management articles on wounds and other problems of skin integrity. Each issue features CME/CE for physicians and nurses, the first journal in the field to regularly offer continuing education for both disciplines.