{"title":"ChatGPT-4会提高医学摘要的质量吗?","authors":"Jocelyn Gravel, Chloé Dion, Mandana Fadaei Kermani, Sarah Mousseau, Esli Osmanlliu","doi":"10.1093/pch/pxae062","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>ChatGPT received attention for medical writing. Our objective was to evaluate whether ChatGPT 4.0 could improve the quality of abstracts submitted to a medical conference by clinical researchers.</p><p><strong>Methods: </strong>This was an experimental study involving 24 international researchers (the participants) who provided one original abstract intended for submission at the 2024 Pediatric Academic Society (PAS) conference. We asked ChatGPT-4 to improve the quality of the abstract while adhering to PAS submission guidelines. Participants received the revised version and were tasked with creating a final abstract. The quality of each version (original, ChatGPT and final) was evaluated by the participants themselves using a numeric scale (0-100). Additionally, three co-investigators assessed abstracts blinded to the version. The primary analysis focused on the mean difference in scores between the final and original abstracts.</p><p><strong>Results: </strong>Abstract quality varied between the three versions with mean scores of 82, 65 and 90 for the original, ChatGPT and final versions, respectively. Overall, the final version displayed significantly improved quality compared to the original (mean difference 8.0 points; 95% CI: 5.6-10.3). Independent ratings by the co-investigators confirmed statistically significant improvements (mean difference 1.10 points; 95% CI: 0.54-1.66). Participants identified minor (n = 10) and major (n = 3) factual errors in ChatGPT's abstracts.</p><p><strong>Conclusion: </strong>ChatGPT 4.0 does not produce abstracts of better quality than the one crafted by researchers but it offers suggestions to help them improve their abstracts. It may be more useful for researchers encountering challenges in abstract generation due to limited experience or language barriers.</p>","PeriodicalId":19730,"journal":{"name":"Paediatrics & child health","volume":"30 3","pages":"116-121"},"PeriodicalIF":1.8000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12208364/pdf/","citationCount":"0","resultStr":"{\"title\":\"Will ChatGPT-4 improve the quality of medical abstracts?\",\"authors\":\"Jocelyn Gravel, Chloé Dion, Mandana Fadaei Kermani, Sarah Mousseau, Esli Osmanlliu\",\"doi\":\"10.1093/pch/pxae062\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>ChatGPT received attention for medical writing. Our objective was to evaluate whether ChatGPT 4.0 could improve the quality of abstracts submitted to a medical conference by clinical researchers.</p><p><strong>Methods: </strong>This was an experimental study involving 24 international researchers (the participants) who provided one original abstract intended for submission at the 2024 Pediatric Academic Society (PAS) conference. We asked ChatGPT-4 to improve the quality of the abstract while adhering to PAS submission guidelines. Participants received the revised version and were tasked with creating a final abstract. The quality of each version (original, ChatGPT and final) was evaluated by the participants themselves using a numeric scale (0-100). Additionally, three co-investigators assessed abstracts blinded to the version. The primary analysis focused on the mean difference in scores between the final and original abstracts.</p><p><strong>Results: </strong>Abstract quality varied between the three versions with mean scores of 82, 65 and 90 for the original, ChatGPT and final versions, respectively. Overall, the final version displayed significantly improved quality compared to the original (mean difference 8.0 points; 95% CI: 5.6-10.3). Independent ratings by the co-investigators confirmed statistically significant improvements (mean difference 1.10 points; 95% CI: 0.54-1.66). Participants identified minor (n = 10) and major (n = 3) factual errors in ChatGPT's abstracts.</p><p><strong>Conclusion: </strong>ChatGPT 4.0 does not produce abstracts of better quality than the one crafted by researchers but it offers suggestions to help them improve their abstracts. It may be more useful for researchers encountering challenges in abstract generation due to limited experience or language barriers.</p>\",\"PeriodicalId\":19730,\"journal\":{\"name\":\"Paediatrics & child health\",\"volume\":\"30 3\",\"pages\":\"116-121\"},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2024-09-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12208364/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Paediatrics & child health\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1093/pch/pxae062\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/6/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q2\",\"JCRName\":\"PEDIATRICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Paediatrics & child health","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1093/pch/pxae062","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/6/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"PEDIATRICS","Score":null,"Total":0}
Will ChatGPT-4 improve the quality of medical abstracts?
Background: ChatGPT received attention for medical writing. Our objective was to evaluate whether ChatGPT 4.0 could improve the quality of abstracts submitted to a medical conference by clinical researchers.
Methods: This was an experimental study involving 24 international researchers (the participants) who provided one original abstract intended for submission at the 2024 Pediatric Academic Society (PAS) conference. We asked ChatGPT-4 to improve the quality of the abstract while adhering to PAS submission guidelines. Participants received the revised version and were tasked with creating a final abstract. The quality of each version (original, ChatGPT and final) was evaluated by the participants themselves using a numeric scale (0-100). Additionally, three co-investigators assessed abstracts blinded to the version. The primary analysis focused on the mean difference in scores between the final and original abstracts.
Results: Abstract quality varied between the three versions with mean scores of 82, 65 and 90 for the original, ChatGPT and final versions, respectively. Overall, the final version displayed significantly improved quality compared to the original (mean difference 8.0 points; 95% CI: 5.6-10.3). Independent ratings by the co-investigators confirmed statistically significant improvements (mean difference 1.10 points; 95% CI: 0.54-1.66). Participants identified minor (n = 10) and major (n = 3) factual errors in ChatGPT's abstracts.
Conclusion: ChatGPT 4.0 does not produce abstracts of better quality than the one crafted by researchers but it offers suggestions to help them improve their abstracts. It may be more useful for researchers encountering challenges in abstract generation due to limited experience or language barriers.
期刊介绍:
Paediatrics & Child Health (PCH) is the official journal of the Canadian Paediatric Society, and the only peer-reviewed paediatric journal in Canada. Its mission is to advocate for the health and well-being of all Canadian children and youth and to educate child and youth health professionals across the country.
PCH reaches 8,000 paediatricians, family physicians and other child and youth health professionals, as well as ministers and officials in various levels of government who are involved with child and youth health policy in Canada.