{"title":"Can ChatGPT be trusted? Evaluating AI responses to oral health questions among pregnant Arabic-speaking women.","authors":"Khalid Talal Aboalshamat, Jomanh Humied Alnafei, Lojain Ahmed Alkhattabi, Ghadi Yaqoub Alhawsawi, Shrooq Majed Alahmadi, Shatha Omar Almalki, Afnan Anas Nassar","doi":"10.1186/s12903-025-06909-z","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>ChatGPT, an artificial intelligence (AI) chatbot developed by OpenAI, is increasingly being used in healthcare, including dentistry, for patient education; this study aimed to assess the usability and quality of ChatGPT's responses to pregnancy-related oral health queries in Saudi Arabia.</p><p><strong>Method: </strong>This two-part cross-sectional study assessed pregnant Arabic women's perceptions of ChatGPT for oral health queries and evaluated its responses using an online questionnaire. Responses from ChatGPT-4o mini were rated by 5 dental experts with regard to accuracy, clarity, relevance, and acceptance using a 5-point Likert scale.</p><p><strong>Results: </strong>Among the 300 participants, 42.0% (126) knew about ChatGPT, 33.7% (101) had previously used it, 14.3% (43) had used it to obtain medical information, 8.7% (26) had used it for dental information, and 8.3% (25) had used it for dental information during pregnancy. Attitudes regarding ChatGPT were rated from 1 to 4. Except for 1 item, the means were all above the midpoint. Attitude ratings ranged from a mean of 2.71 (SD 0.76) for ChatGPT competency to a mean of 2.34 (SD 0.92) for its ability to replace human interactions. However, ChatGPT competency (P = .028), security (P = .015), willingness to use ChatGPT for inquiries (P = .021), ability to assist in informed decision-making (P = .01), willingness to make decisions based on recommendations (P = .024), and persuasiveness (P = .049) were significantly different based on educational level. Pregnant women with higher levels of education rated these aspects significantly lower than those with a high school diploma or bachelor's degree.</p><p><strong>Conclusion: </strong>ChatGPT provided useful oral health information for pregnant individuals, but its responses required revision and supervision by health professionals. Its usage among pregnant women in Saudi Arabia remained low.</p>","PeriodicalId":9072,"journal":{"name":"BMC Oral Health","volume":"25 1","pages":"1597"},"PeriodicalIF":3.1000,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12513007/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC Oral Health","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1186/s12903-025-06909-z","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"DENTISTRY, ORAL SURGERY & MEDICINE","Score":null,"Total":0}
引用次数: 0
Abstract
Background: ChatGPT, an artificial intelligence (AI) chatbot developed by OpenAI, is increasingly being used in healthcare, including dentistry, for patient education; this study aimed to assess the usability and quality of ChatGPT's responses to pregnancy-related oral health queries in Saudi Arabia.
Method: This two-part cross-sectional study assessed pregnant Arabic women's perceptions of ChatGPT for oral health queries and evaluated its responses using an online questionnaire. Responses from ChatGPT-4o mini were rated by 5 dental experts with regard to accuracy, clarity, relevance, and acceptance using a 5-point Likert scale.
Results: Among the 300 participants, 42.0% (126) knew about ChatGPT, 33.7% (101) had previously used it, 14.3% (43) had used it to obtain medical information, 8.7% (26) had used it for dental information, and 8.3% (25) had used it for dental information during pregnancy. Attitudes regarding ChatGPT were rated from 1 to 4. Except for 1 item, the means were all above the midpoint. Attitude ratings ranged from a mean of 2.71 (SD 0.76) for ChatGPT competency to a mean of 2.34 (SD 0.92) for its ability to replace human interactions. However, ChatGPT competency (P = .028), security (P = .015), willingness to use ChatGPT for inquiries (P = .021), ability to assist in informed decision-making (P = .01), willingness to make decisions based on recommendations (P = .024), and persuasiveness (P = .049) were significantly different based on educational level. Pregnant women with higher levels of education rated these aspects significantly lower than those with a high school diploma or bachelor's degree.
Conclusion: ChatGPT provided useful oral health information for pregnant individuals, but its responses required revision and supervision by health professionals. Its usage among pregnant women in Saudi Arabia remained low.
期刊介绍:
BMC Oral Health is an open access, peer-reviewed journal that considers articles on all aspects of the prevention, diagnosis and management of disorders of the mouth, teeth and gums, as well as related molecular genetics, pathophysiology, and epidemiology.