{"title":"Using Artificial Intelligence ChatGPT to Access Medical Information about Chemical Eye Injuries: A Comparative Study.","authors":"Layan Yousef Alharbi, Rema Rashed Alrashoud, Bader Shabib Alotaibi, Abdulaziz Meshal Al Dera, Raghad Saleh Alajlan, Reem Rashed AlHuthail, Dalal Ibrahim Alessa","doi":"10.2196/73642","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Background: Chemical ocular injuries are a major public health issue. They cause eye damage from harmful chemicals and can lead to severe vision loss or blindness if not treated promptly and effectively. Although medical knowledge has advanced, accessing reliable and understandable information on these injuries remains a challenge. This is due to unverified online content and complex terminology. Artificial Intelligence (AI) tools like ChatGPT provide a promising solution by simplifying medical information and making it more accessible to the general public.</p><p><strong>Objective: </strong>Objective: This study aims to assess the use of ChatGPT in providing reliable, accurate, and accessible medical information on chemical ocular injuries. It evaluates the correctness, thematic accuracy, and coherence of ChatGPT's responses compared to established medical guidelines and explores its potential for patient education.</p><p><strong>Methods: </strong>Methods: Nine questions were entered to ChatGPT regarding various aspects of chemical ocular injuries. These included the definition, prevalence, etiology, prevention, symptoms, diagnosis, treatment, follow-up, and complications. The responses provided by ChatGPT were compared to the ICD-9 and ICD-10 guidelines for chemical (alkali and acid) injuries of the conjunctiva and cornea. The evaluation focused on criteria such as correctness, thematic accuracy, coherence to assess the accuracy of ChatGPT's responses. The inputs were categorized into three distinct groups, and statistical analyses, including Flesch-Kincaid readability tests, ANOVA, and trend analysis, were conducted to assess their readability, complexity and trends.</p><p><strong>Results: </strong>Results: The results showed that ChatGPT provided accurate and coherent responses for most questions about chemical ocular injuries, demonstrating thematic relevance. However, the responses sometimes overlooked critical clinical details or guideline-specific elements, such as emphasizing the urgency of care, using precise classification systems, and addressing detailed diagnostic or management protocols. While the answers were generally valid, they occasionally included less relevant or overly generalized information. This reduced their consistency with established medical guidelines. The average FRES was 33.84 ± 2.97, indicating a fairly challenging reading level, while the FKGL averaged 14.21 ± 0.97, suitable for readers with college-level proficiency. Passive voice was used in 7.22% ± 5.60% of sentences, indicating moderate reliance. Statistical analysis showed no significant differences in FRES (p = .385), FKGL (p = .555), or passive sentence usage (p = .601) across categories, as determined by one-way ANOVA. Readability remained relatively constant across the three categories, as determined by trend analysis.</p><p><strong>Conclusions: </strong>Conclusions: ChatGPT shows strong potential in providing accurate and relevant information about chemical ocular injuries. However, its language complexity may prevent accessibility for individuals with lower health literacy and sometimes miss critical aspects. Future improvements should focus on enhancing readability, increasing context-specific accuracy, and tailoring responses to person needs and literacy levels.</p><p><strong>Clinicaltrial: </strong>This is not RCT.</p>","PeriodicalId":14841,"journal":{"name":"JMIR Formative Research","volume":" ","pages":""},"PeriodicalIF":2.0000,"publicationDate":"2025-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"JMIR Formative Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2196/73642","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
引用次数: 0
Abstract
Background: Background: Chemical ocular injuries are a major public health issue. They cause eye damage from harmful chemicals and can lead to severe vision loss or blindness if not treated promptly and effectively. Although medical knowledge has advanced, accessing reliable and understandable information on these injuries remains a challenge. This is due to unverified online content and complex terminology. Artificial Intelligence (AI) tools like ChatGPT provide a promising solution by simplifying medical information and making it more accessible to the general public.
Objective: Objective: This study aims to assess the use of ChatGPT in providing reliable, accurate, and accessible medical information on chemical ocular injuries. It evaluates the correctness, thematic accuracy, and coherence of ChatGPT's responses compared to established medical guidelines and explores its potential for patient education.
Methods: Methods: Nine questions were entered to ChatGPT regarding various aspects of chemical ocular injuries. These included the definition, prevalence, etiology, prevention, symptoms, diagnosis, treatment, follow-up, and complications. The responses provided by ChatGPT were compared to the ICD-9 and ICD-10 guidelines for chemical (alkali and acid) injuries of the conjunctiva and cornea. The evaluation focused on criteria such as correctness, thematic accuracy, coherence to assess the accuracy of ChatGPT's responses. The inputs were categorized into three distinct groups, and statistical analyses, including Flesch-Kincaid readability tests, ANOVA, and trend analysis, were conducted to assess their readability, complexity and trends.
Results: Results: The results showed that ChatGPT provided accurate and coherent responses for most questions about chemical ocular injuries, demonstrating thematic relevance. However, the responses sometimes overlooked critical clinical details or guideline-specific elements, such as emphasizing the urgency of care, using precise classification systems, and addressing detailed diagnostic or management protocols. While the answers were generally valid, they occasionally included less relevant or overly generalized information. This reduced their consistency with established medical guidelines. The average FRES was 33.84 ± 2.97, indicating a fairly challenging reading level, while the FKGL averaged 14.21 ± 0.97, suitable for readers with college-level proficiency. Passive voice was used in 7.22% ± 5.60% of sentences, indicating moderate reliance. Statistical analysis showed no significant differences in FRES (p = .385), FKGL (p = .555), or passive sentence usage (p = .601) across categories, as determined by one-way ANOVA. Readability remained relatively constant across the three categories, as determined by trend analysis.
Conclusions: Conclusions: ChatGPT shows strong potential in providing accurate and relevant information about chemical ocular injuries. However, its language complexity may prevent accessibility for individuals with lower health literacy and sometimes miss critical aspects. Future improvements should focus on enhancing readability, increasing context-specific accuracy, and tailoring responses to person needs and literacy levels.