Lily Nguyen , Viet Tran , Joy Li , Denise Baughn , Joseph Shotwell , Kimberly Gushanas , Sayyeda Hasan , Lisa Falls , Rocksheng Zhong
{"title":"Artificial intelligence chatbots as sources for patient education material on child abuse","authors":"Lily Nguyen , Viet Tran , Joy Li , Denise Baughn , Joseph Shotwell , Kimberly Gushanas , Sayyeda Hasan , Lisa Falls , Rocksheng Zhong","doi":"10.1016/j.chipro.2025.100167","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>The World Health Organization defines <strong>c</strong>hildhood maltreatment as any form of abuse or neglect affecting children under 18 years of age that can cause actual or potential harm. Child abuse is a form of interpersonal trauma that can critically impact neurodevelopment and increase the risk of developing psychiatric disorders. With the increasing power and accessibility of artificial intelligence (AI) large language models, patients may turn to these platforms as sources of medical information. To date, no studies have evaluated the use of AI in creating patient education materials in childhood maltreatment and the field of psychiatry.</div></div><div><h3>Methods</h3><div>Eight questions on child abuse from the National Child Traumatic Stress Network (NCTSN) were input into ChatGPT, Google Gemini, and Microsoft Copilot. A team of child psychiatrists and a pediatric psychologist reviewed and scored the responses by NCTSN and each AI, assessing quality, understandability, and actionability. Secondary outcomes included misinformation, readability, word count, and top references.</div></div><div><h3>Results</h3><div>The analysis of 32 responses showed good quality (mean DISCERN score 51.7) and moderate understandability (mean PEMAT 76.5 %). However, actionability was poor (mean PEMAT 64 %). Responses averaged a tenth-grade reading level, with ChatGPT being more difficult to read than NCTSN. AI-generated responses were significantly longer (p < 0.001).</div></div><div><h3>Conclusions</h3><div>Findings of this study suggest that AI chatbots may currently be able to provide accurate, quality information on child abuse comparable to authoritative sources, albeit of significantly greater length. However, all sources lack actionability and exceed recommended reading levels, which limits effectiveness. These constraints suggest that AI chatbots should supplement rather than replace primary medical information sources. Urgent efforts are needed to improve the accessibility, readability, and actionability of patient education materials generated by AI and standardized sources on topics like child abuse and neglect.</div></div>","PeriodicalId":100237,"journal":{"name":"Child Protection and Practice","volume":"5 ","pages":"Article 100167"},"PeriodicalIF":0.0000,"publicationDate":"2025-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Child Protection and Practice","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2950193825000749","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Background
The World Health Organization defines childhood maltreatment as any form of abuse or neglect affecting children under 18 years of age that can cause actual or potential harm. Child abuse is a form of interpersonal trauma that can critically impact neurodevelopment and increase the risk of developing psychiatric disorders. With the increasing power and accessibility of artificial intelligence (AI) large language models, patients may turn to these platforms as sources of medical information. To date, no studies have evaluated the use of AI in creating patient education materials in childhood maltreatment and the field of psychiatry.
Methods
Eight questions on child abuse from the National Child Traumatic Stress Network (NCTSN) were input into ChatGPT, Google Gemini, and Microsoft Copilot. A team of child psychiatrists and a pediatric psychologist reviewed and scored the responses by NCTSN and each AI, assessing quality, understandability, and actionability. Secondary outcomes included misinformation, readability, word count, and top references.
Results
The analysis of 32 responses showed good quality (mean DISCERN score 51.7) and moderate understandability (mean PEMAT 76.5 %). However, actionability was poor (mean PEMAT 64 %). Responses averaged a tenth-grade reading level, with ChatGPT being more difficult to read than NCTSN. AI-generated responses were significantly longer (p < 0.001).
Conclusions
Findings of this study suggest that AI chatbots may currently be able to provide accurate, quality information on child abuse comparable to authoritative sources, albeit of significantly greater length. However, all sources lack actionability and exceed recommended reading levels, which limits effectiveness. These constraints suggest that AI chatbots should supplement rather than replace primary medical information sources. Urgent efforts are needed to improve the accessibility, readability, and actionability of patient education materials generated by AI and standardized sources on topics like child abuse and neglect.