{"title":"Evaluation of Google and ChatGPT responses to common patient questions about scoliosis.","authors":"Sezgin Bahadır Tekin, Kamil Ince, Bedriye Gizem Tekin, Erkan Servet, Bahri Bozgeyik","doi":"10.1007/s43390-025-01169-x","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>Scoliosis is primarily seen during adolescence and often causes significant concern among patients and their families when the deformity becomes noticeable. With technological advancements, patients frequently search the Internet for information regarding their disease's diagnosis, treatment, prognosis, and potential complications. This study aims to assess the quality of Google and ChatGPT responses to questions about scoliosis.</p><p><strong>Methods: </strong>A search was conducted using Google with the keyword \"scoliosis.\" The first ten questions listed under the \"People Also Ask\" (FAQs) section were recorded. Responses to these questions from ChatGPT and Google were evaluated using a four-level rating system: \"Excellent response not requiring clarification,\" \"satisfactory requiring minimal clarification,\" \"satisfactory requiring moderate clarification,\" and \"unsatisfactory requiring substantial clarification.\" Additionally, the sources of the responses were categorized as academic, commercial, medical practice, governmental, or social media.</p><p><strong>Results: </strong>ChatGPT provided \"excellent responses requiring no explanation\" for 9 out of 10 questions (90%). In contrast, none of Google's responses were categorized as excellent; 50% were unsatisfactory, requiring substantial clarification; 40% were satisfactory, requiring moderate clarification, and 10% were satisfactory, requiring minimal clarification. ChatGPT sourced 60% of its responses from academic resources and 40% from medical practice websites. Conversely, Google did not use scholarly sources, with 50% of reactions derived from commercial websites, 30% from medical practice sources, and 20% from social media. When the agreement between the 4 raters, regardless of AI, was examined using Fleiss Multirater Kappa in the reliability analysis, a statistically significant (p < 0.001) moderate agreement (κ = 0.48) was found.</p><p><strong>Conclusion: </strong>ChatGPT outperformed Google by providing more accurate, well-referenced responses and utilizing more credible academic sources. This suggests its potential as a more reliable tool for obtaining health-related information.</p>","PeriodicalId":21796,"journal":{"name":"Spine deformity","volume":" ","pages":""},"PeriodicalIF":1.8000,"publicationDate":"2025-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Spine deformity","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s43390-025-01169-x","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"CLINICAL NEUROLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Objective: Scoliosis is primarily seen during adolescence and often causes significant concern among patients and their families when the deformity becomes noticeable. With technological advancements, patients frequently search the Internet for information regarding their disease's diagnosis, treatment, prognosis, and potential complications. This study aims to assess the quality of Google and ChatGPT responses to questions about scoliosis.
Methods: A search was conducted using Google with the keyword "scoliosis." The first ten questions listed under the "People Also Ask" (FAQs) section were recorded. Responses to these questions from ChatGPT and Google were evaluated using a four-level rating system: "Excellent response not requiring clarification," "satisfactory requiring minimal clarification," "satisfactory requiring moderate clarification," and "unsatisfactory requiring substantial clarification." Additionally, the sources of the responses were categorized as academic, commercial, medical practice, governmental, or social media.
Results: ChatGPT provided "excellent responses requiring no explanation" for 9 out of 10 questions (90%). In contrast, none of Google's responses were categorized as excellent; 50% were unsatisfactory, requiring substantial clarification; 40% were satisfactory, requiring moderate clarification, and 10% were satisfactory, requiring minimal clarification. ChatGPT sourced 60% of its responses from academic resources and 40% from medical practice websites. Conversely, Google did not use scholarly sources, with 50% of reactions derived from commercial websites, 30% from medical practice sources, and 20% from social media. When the agreement between the 4 raters, regardless of AI, was examined using Fleiss Multirater Kappa in the reliability analysis, a statistically significant (p < 0.001) moderate agreement (κ = 0.48) was found.
Conclusion: ChatGPT outperformed Google by providing more accurate, well-referenced responses and utilizing more credible academic sources. This suggests its potential as a more reliable tool for obtaining health-related information.
期刊介绍:
Spine Deformity the official journal of the?Scoliosis Research Society is a peer-refereed publication to disseminate knowledge on basic science and clinical research into the?etiology?biomechanics?treatment?methods and outcomes of all types of?spinal deformities. The international members of the Editorial Board provide a worldwide perspective for the journal's area of interest.The?journal?will enhance the mission of the Society which is to foster the optimal care of all patients with?spine?deformities worldwide. Articles published in?Spine Deformity?are Medline indexed in PubMed.? The journal publishes original articles in the form of clinical and basic research. Spine Deformity will only publish studies that have institutional review board (IRB) or similar ethics committee approval for human and animal studies and have strictly observed these guidelines. The minimum follow-up period for follow-up clinical studies is 24 months.