Artificial intelligence-generated informed patient consent in various ophthalmological procedures: A comparative study of correctness, completeness, readability, and real-word application between Deepseek and Chatgpt 4o.
{"title":"Artificial intelligence-generated informed patient consent in various ophthalmological procedures: A comparative study of correctness, completeness, readability, and real-word application between Deepseek and Chatgpt 4o.","authors":"Deepsekhar Das, Bhavna Chawla, Neiwete Lomi, Preeti Tomar, Ananya Herle, Shivani Joshi","doi":"10.4103/IJO.IJO_1126_25","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>To study the correctness, completeness, language and readability, and real-world applicability of artificial intelligence chatbots-generated informed consent forms for various ophthalmological procedures and interventions.</p><p><strong>Methods: </strong>A cross-sectional observational study was performed by ophthalmology faculties of a tertiary care eye hospital. A list of popularly performed ophthalmological interventions in ophthalmological operation theaters was compiled. Questions were created asking for informed consents. Each question was standardized; the age and diagnosis were mentioned, which were eventually fed into two publicly available chatbots, namely, ChatGPT 4o and Deepseek. The answers obtained from these chatbots were then evaluated on the basis of correctness, completeness, language and readability, additional relevant information, irrelevant information, and real-world applicability of the consent (word to word) in Indian Scenario. Chi-square tests were used for performing analysis of categorical data, namely, correctness and completeness, whereas Mann-Whitney U test was performed for numerical data.</p><p><strong>Results: </strong>ChatGPT had less words and sentences compared to Deepseek; however, Deepseek offered a higher average readability score on both Flesch Kincaid calculator and Gunning Fog Index. Deepseek required more attempted to obtain the responses. However, 40% of the consents generated by both chatbots were not fit to be used in Indian scenarios.</p><p><strong>Conclusion: </strong>Deepseek offered significantly more elaborate readable informed consents than ChatGPT; however, both the chatbots at present failed 40% of the times to create informed consents which can be used in Indian scenarios.</p>","PeriodicalId":13329,"journal":{"name":"Indian Journal of Ophthalmology","volume":"73 10","pages":"1466-1470"},"PeriodicalIF":1.8000,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Indian Journal of Ophthalmology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.4103/IJO.IJO_1126_25","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/9/25 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"OPHTHALMOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: To study the correctness, completeness, language and readability, and real-world applicability of artificial intelligence chatbots-generated informed consent forms for various ophthalmological procedures and interventions.
Methods: A cross-sectional observational study was performed by ophthalmology faculties of a tertiary care eye hospital. A list of popularly performed ophthalmological interventions in ophthalmological operation theaters was compiled. Questions were created asking for informed consents. Each question was standardized; the age and diagnosis were mentioned, which were eventually fed into two publicly available chatbots, namely, ChatGPT 4o and Deepseek. The answers obtained from these chatbots were then evaluated on the basis of correctness, completeness, language and readability, additional relevant information, irrelevant information, and real-world applicability of the consent (word to word) in Indian Scenario. Chi-square tests were used for performing analysis of categorical data, namely, correctness and completeness, whereas Mann-Whitney U test was performed for numerical data.
Results: ChatGPT had less words and sentences compared to Deepseek; however, Deepseek offered a higher average readability score on both Flesch Kincaid calculator and Gunning Fog Index. Deepseek required more attempted to obtain the responses. However, 40% of the consents generated by both chatbots were not fit to be used in Indian scenarios.
Conclusion: Deepseek offered significantly more elaborate readable informed consents than ChatGPT; however, both the chatbots at present failed 40% of the times to create informed consents which can be used in Indian scenarios.
期刊介绍:
Indian Journal of Ophthalmology covers clinical, experimental, basic science research and translational research studies related to medical, ethical and social issues in field of ophthalmology and vision science. Articles with clinical interest and implications will be given preference.