Information about labor epidural analgesia: an updated evaluation on the readability, accuracy, and quality of ChatGPT responses incorporating patient preferences and complex clinical scenarios
C.W. Tan , J.C.Y. Chan , J.J.I. Chan , S. Nagarajan , B.L. Sng
{"title":"Information about labor epidural analgesia: an updated evaluation on the readability, accuracy, and quality of ChatGPT responses incorporating patient preferences and complex clinical scenarios","authors":"C.W. Tan , J.C.Y. Chan , J.J.I. Chan , S. Nagarajan , B.L. Sng","doi":"10.1016/j.ijoa.2025.104688","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>Recent studies evaluating frequently asked questions (FAQs) on labor epidural analgesia (LEA) only used generic questions without incorporating detailed clinical information that reflects patient-specific inputs. We investigated the performance of ChatGPT in addressing these questions related to LEA with an emphasis on individual preferences and clinical conditions.</div></div><div><h3>Methods</h3><div>Twenty-nine questions for the AI chatbot were generated from the commonly asked questions relating to LEA based on clinical conditions. The generation of responses was performed in January 2025 with each question under individual sub-topics initiated as a “New chat” in ChatGPT-4o. Upon having the first questions answered, subsequent question(s) in the same sub-topic were continued in the same chat following the sequences as predefined. The readability of each response was graded using six readability indices, while the accuracy, Patient Education Materials Assessment Tool for Print (PEMAT) understandability and actionability was assessed by four obstetric anesthesiologists.</div></div><div><h3>Results</h3><div>The mean readability indices of the ChatGPT-4o responses to the questions were generally rated as fairly difficult to very difficult, which corresponded to a US grade level between 11th grade to college level entry. The mean (± standard deviation) accuracy of the responses was 97.7% ± 8.1%. The PEMAT understandability and actionability scores were 97.9% ± 0.9%) and 98.0% ± 1.4%), respectively.</div></div><div><h3>Conclusions</h3><div>ChatGPT can provide accurate and readable information about LEA even under different clinical contexts. However, improvement is needed to refine the responses with suitable prompts to simplify the outputs and improve readability. These approaches will thereby meet the need for the effective delivery of reliable patient education information.</div></div>","PeriodicalId":14250,"journal":{"name":"International journal of obstetric anesthesia","volume":"63 ","pages":"Article 104688"},"PeriodicalIF":2.6000,"publicationDate":"2025-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International journal of obstetric anesthesia","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0959289X25002808","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ANESTHESIOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Background
Recent studies evaluating frequently asked questions (FAQs) on labor epidural analgesia (LEA) only used generic questions without incorporating detailed clinical information that reflects patient-specific inputs. We investigated the performance of ChatGPT in addressing these questions related to LEA with an emphasis on individual preferences and clinical conditions.
Methods
Twenty-nine questions for the AI chatbot were generated from the commonly asked questions relating to LEA based on clinical conditions. The generation of responses was performed in January 2025 with each question under individual sub-topics initiated as a “New chat” in ChatGPT-4o. Upon having the first questions answered, subsequent question(s) in the same sub-topic were continued in the same chat following the sequences as predefined. The readability of each response was graded using six readability indices, while the accuracy, Patient Education Materials Assessment Tool for Print (PEMAT) understandability and actionability was assessed by four obstetric anesthesiologists.
Results
The mean readability indices of the ChatGPT-4o responses to the questions were generally rated as fairly difficult to very difficult, which corresponded to a US grade level between 11th grade to college level entry. The mean (± standard deviation) accuracy of the responses was 97.7% ± 8.1%. The PEMAT understandability and actionability scores were 97.9% ± 0.9%) and 98.0% ± 1.4%), respectively.
Conclusions
ChatGPT can provide accurate and readable information about LEA even under different clinical contexts. However, improvement is needed to refine the responses with suitable prompts to simplify the outputs and improve readability. These approaches will thereby meet the need for the effective delivery of reliable patient education information.
期刊介绍:
The International Journal of Obstetric Anesthesia is the only journal publishing original articles devoted exclusively to obstetric anesthesia and bringing together all three of its principal components; anesthesia care for operative delivery and the perioperative period, pain relief in labour and care of the critically ill obstetric patient.
• Original research (both clinical and laboratory), short reports and case reports will be considered.
• The journal also publishes invited review articles and debates on topical and controversial subjects in the area of obstetric anesthesia.
• Articles on related topics such as perinatal physiology and pharmacology and all subjects of importance to obstetric anaesthetists/anesthesiologists are also welcome.
The journal is peer-reviewed by international experts. Scholarship is stressed to include the focus on discovery, application of knowledge across fields, and informing the medical community. Through the peer-review process, we hope to attest to the quality of scholarships and guide the Journal to extend and transform knowledge in this important and expanding area.