自闭症谱系障碍机器学习诊断模型中错误的临床相关性:样本队列的影响。

IF 5.6 2区 心理学 Q1 PSYCHOLOGY, DEVELOPMENTAL
Autism Pub Date : 2025-08-05 DOI:10.1177/13623613251360271
Yen-Chin Wang, Chung-Yuan Cheng, Chi-Shin Wu, Chi-Chun Lee, Susan Shur-Fen Gau
{"title":"自闭症谱系障碍机器学习诊断模型中错误的临床相关性:样本队列的影响。","authors":"Yen-Chin Wang, Chung-Yuan Cheng, Chi-Shin Wu, Chi-Chun Lee, Susan Shur-Fen Gau","doi":"10.1177/13623613251360271","DOIUrl":null,"url":null,"abstract":"<p><p>Machine-learning models can assist in diagnosing autism but have biases. We examines the correlates of misclassifications and how training data affect model generalizability. The Social Responsive Scale data were collected from two cohorts in Taiwan: the clinical cohort comprised 1203 autistic participants and 1182 non-autistic comparisons, and the community cohort consisted of 35 autistic participants and 3297 non-autistic comparisons. Classification models were trained, and the misclassification cases were investigated regarding their associations with sex, age, intelligence quotient (IQ), symptoms from the child behavioral checklist (CBCL), and co-occurring psychiatric diagnosis. Models showed high within-cohort accuracy (clinical: sensitivity 0.91-0.95, specificity 0.93-0.94; community: sensitivity 0.91-1.00, specificity 0.89-0.96), but generalizability across cohorts was limited. When the community-trained model was applied to the clinical cohort, performance declined (sensitivity 0.65, specificity 0.95). In both models, non-autistic individuals misclassified as autistic showed elevated behavioral symptoms and attention-deficit hyperactivity disorder (ADHD) prevalence. Conversely, autistic individuals who were misclassified tended to show fewer behavioral symptoms and, in the community model, higher IQ and aggressive behavior but less social and attention problems. Error patterns of machine-learning model and the impact of training data warrant careful consideration in future research.Lay AbstractMachine-learning is a type of computer model that can help identify patterns in data and make predictions. In autism research, these models may support earlier or more accurate identification of autistic individuals. But to be useful, they need to make reliable predictions across different groups of people. In this study, we explored when and why these models might make mistakes-and how the kind of data used to train them affects their accuracy. Training models means using information to teach the computer model how to tell the difference between autistic and non-autistic individuals. We used the information from the Social Responsiveness Scale (SRS), which is a questionnaire that measures autistic features. We tested these models on two different groups: one from clinical settings and one from the general community. The models worked well when tested within the same type of group they were trained. However, a model trained on the community group did not perform as accurately when tested on the clinical group. Sometimes, the model got it wrong. For example, in the clinical group, some autistic individuals were mistakenly identified as non-autistic. These individuals tended to have fewer emotional or behavioral difficulties. In the community group, autistic individuals who were mistakenly identified as non-autistic had higher IQs and showed more aggressive behaviors but fewer attention or social problems. On the contrary, some non-autistic people were incorrectly identified as autistic. These people had more emotional or behavioral challenges and were more likely to have attention-deficit hyperactivity disorder (ADHD). These findings highlight that machine-learning models are sensitive to the type of data they are trained on. To build fair and accurate models for predicting autism, it is essential to consider where the training data come from and whether it represents the full diversity of individuals. Understanding these patterns of error can help improve future tools used in both research and clinical care.</p>","PeriodicalId":8724,"journal":{"name":"Autism","volume":" ","pages":"13623613251360271"},"PeriodicalIF":5.6000,"publicationDate":"2025-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Clinical correlates of errors in machine-learning diagnostic model of autism spectrum disorder: Impact of sample cohorts.\",\"authors\":\"Yen-Chin Wang, Chung-Yuan Cheng, Chi-Shin Wu, Chi-Chun Lee, Susan Shur-Fen Gau\",\"doi\":\"10.1177/13623613251360271\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Machine-learning models can assist in diagnosing autism but have biases. We examines the correlates of misclassifications and how training data affect model generalizability. The Social Responsive Scale data were collected from two cohorts in Taiwan: the clinical cohort comprised 1203 autistic participants and 1182 non-autistic comparisons, and the community cohort consisted of 35 autistic participants and 3297 non-autistic comparisons. Classification models were trained, and the misclassification cases were investigated regarding their associations with sex, age, intelligence quotient (IQ), symptoms from the child behavioral checklist (CBCL), and co-occurring psychiatric diagnosis. Models showed high within-cohort accuracy (clinical: sensitivity 0.91-0.95, specificity 0.93-0.94; community: sensitivity 0.91-1.00, specificity 0.89-0.96), but generalizability across cohorts was limited. When the community-trained model was applied to the clinical cohort, performance declined (sensitivity 0.65, specificity 0.95). In both models, non-autistic individuals misclassified as autistic showed elevated behavioral symptoms and attention-deficit hyperactivity disorder (ADHD) prevalence. Conversely, autistic individuals who were misclassified tended to show fewer behavioral symptoms and, in the community model, higher IQ and aggressive behavior but less social and attention problems. Error patterns of machine-learning model and the impact of training data warrant careful consideration in future research.Lay AbstractMachine-learning is a type of computer model that can help identify patterns in data and make predictions. In autism research, these models may support earlier or more accurate identification of autistic individuals. But to be useful, they need to make reliable predictions across different groups of people. In this study, we explored when and why these models might make mistakes-and how the kind of data used to train them affects their accuracy. Training models means using information to teach the computer model how to tell the difference between autistic and non-autistic individuals. We used the information from the Social Responsiveness Scale (SRS), which is a questionnaire that measures autistic features. We tested these models on two different groups: one from clinical settings and one from the general community. The models worked well when tested within the same type of group they were trained. However, a model trained on the community group did not perform as accurately when tested on the clinical group. Sometimes, the model got it wrong. For example, in the clinical group, some autistic individuals were mistakenly identified as non-autistic. These individuals tended to have fewer emotional or behavioral difficulties. In the community group, autistic individuals who were mistakenly identified as non-autistic had higher IQs and showed more aggressive behaviors but fewer attention or social problems. On the contrary, some non-autistic people were incorrectly identified as autistic. These people had more emotional or behavioral challenges and were more likely to have attention-deficit hyperactivity disorder (ADHD). These findings highlight that machine-learning models are sensitive to the type of data they are trained on. To build fair and accurate models for predicting autism, it is essential to consider where the training data come from and whether it represents the full diversity of individuals. Understanding these patterns of error can help improve future tools used in both research and clinical care.</p>\",\"PeriodicalId\":8724,\"journal\":{\"name\":\"Autism\",\"volume\":\" \",\"pages\":\"13623613251360271\"},\"PeriodicalIF\":5.6000,\"publicationDate\":\"2025-08-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Autism\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://doi.org/10.1177/13623613251360271\",\"RegionNum\":2,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHOLOGY, DEVELOPMENTAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Autism","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1177/13623613251360271","RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, DEVELOPMENTAL","Score":null,"Total":0}
引用次数: 0

摘要

机器学习模型可以帮助诊断自闭症,但存在偏见。我们研究了错误分类的相关性以及训练数据如何影响模型的泛化性。社会反应量表数据采自台湾两个队列:临床队列包括1203名自闭症参与者和1182名非自闭症参与者,社区队列包括35名自闭症参与者和3297名非自闭症参与者。训练分类模型,并调查错误分类病例与性别、年龄、智商(IQ)、儿童行为检查表(CBCL)症状和合并精神病学诊断的关系。模型显示出较高的队列内准确性(临床:敏感性0.91-0.95,特异性0.93-0.94;社区:敏感性0.91-1.00,特异性0.89-0.96),但跨队列的普遍性有限。当社区训练模型应用于临床队列时,表现下降(敏感性0.65,特异性0.95)。在这两个模型中,被错误归类为自闭症的非自闭症个体表现出更高的行为症状和注意力缺陷多动障碍(ADHD)患病率。相反,被错误分类的自闭症患者往往表现出较少的行为症状,在社区模型中,他们的智商和攻击行为更高,但社交和注意力问题较少。机器学习模型的错误模式和训练数据的影响值得在未来的研究中仔细考虑。机器学习是一种计算机模型,可以帮助识别数据中的模式并进行预测。在自闭症研究中,这些模型可能支持更早或更准确地识别自闭症个体。但要发挥作用,它们需要对不同人群做出可靠的预测。在这项研究中,我们探讨了这些模型何时以及为什么会出错,以及用于训练它们的数据类型如何影响它们的准确性。训练模型意味着使用信息来教计算机模型如何区分自闭症和非自闭症个体。我们使用了来自社会反应量表(SRS)的信息,这是一份测量自闭症特征的问卷。我们在两个不同的群体中测试了这些模型:一个来自临床环境,一个来自普通社区。这些模型在接受训练的同一类型的群体中进行测试时效果良好。然而,在社区组中训练的模型在临床组中测试时表现不准确。有时,模型会出错。例如,在临床组中,一些自闭症患者被错误地认定为非自闭症患者。这些人倾向于较少出现情绪或行为上的困难。在社区小组中,被错误地认定为非自闭症的自闭症个体智商更高,表现出更多的攻击性行为,但注意力较少或社会问题较少。相反,一些没有自闭症的人被错误地认定为自闭症。这些人有更多的情感或行为挑战,更有可能患有注意力缺陷多动障碍(ADHD)。这些发现突出表明,机器学习模型对它们所训练的数据类型非常敏感。为了建立公平和准确的预测自闭症的模型,必须考虑训练数据的来源以及它是否代表了个体的全部多样性。了解这些错误模式可以帮助改进未来在研究和临床护理中使用的工具。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Clinical correlates of errors in machine-learning diagnostic model of autism spectrum disorder: Impact of sample cohorts.

Machine-learning models can assist in diagnosing autism but have biases. We examines the correlates of misclassifications and how training data affect model generalizability. The Social Responsive Scale data were collected from two cohorts in Taiwan: the clinical cohort comprised 1203 autistic participants and 1182 non-autistic comparisons, and the community cohort consisted of 35 autistic participants and 3297 non-autistic comparisons. Classification models were trained, and the misclassification cases were investigated regarding their associations with sex, age, intelligence quotient (IQ), symptoms from the child behavioral checklist (CBCL), and co-occurring psychiatric diagnosis. Models showed high within-cohort accuracy (clinical: sensitivity 0.91-0.95, specificity 0.93-0.94; community: sensitivity 0.91-1.00, specificity 0.89-0.96), but generalizability across cohorts was limited. When the community-trained model was applied to the clinical cohort, performance declined (sensitivity 0.65, specificity 0.95). In both models, non-autistic individuals misclassified as autistic showed elevated behavioral symptoms and attention-deficit hyperactivity disorder (ADHD) prevalence. Conversely, autistic individuals who were misclassified tended to show fewer behavioral symptoms and, in the community model, higher IQ and aggressive behavior but less social and attention problems. Error patterns of machine-learning model and the impact of training data warrant careful consideration in future research.Lay AbstractMachine-learning is a type of computer model that can help identify patterns in data and make predictions. In autism research, these models may support earlier or more accurate identification of autistic individuals. But to be useful, they need to make reliable predictions across different groups of people. In this study, we explored when and why these models might make mistakes-and how the kind of data used to train them affects their accuracy. Training models means using information to teach the computer model how to tell the difference between autistic and non-autistic individuals. We used the information from the Social Responsiveness Scale (SRS), which is a questionnaire that measures autistic features. We tested these models on two different groups: one from clinical settings and one from the general community. The models worked well when tested within the same type of group they were trained. However, a model trained on the community group did not perform as accurately when tested on the clinical group. Sometimes, the model got it wrong. For example, in the clinical group, some autistic individuals were mistakenly identified as non-autistic. These individuals tended to have fewer emotional or behavioral difficulties. In the community group, autistic individuals who were mistakenly identified as non-autistic had higher IQs and showed more aggressive behaviors but fewer attention or social problems. On the contrary, some non-autistic people were incorrectly identified as autistic. These people had more emotional or behavioral challenges and were more likely to have attention-deficit hyperactivity disorder (ADHD). These findings highlight that machine-learning models are sensitive to the type of data they are trained on. To build fair and accurate models for predicting autism, it is essential to consider where the training data come from and whether it represents the full diversity of individuals. Understanding these patterns of error can help improve future tools used in both research and clinical care.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Autism
Autism PSYCHOLOGY, DEVELOPMENTAL-
CiteScore
9.80
自引率
11.50%
发文量
160
期刊介绍: Autism is a major, peer-reviewed, international journal, published 8 times a year, publishing research of direct and practical relevance to help improve the quality of life for individuals with autism or autism-related disorders. It is interdisciplinary in nature, focusing on research in many areas, including: intervention; diagnosis; training; education; translational issues related to neuroscience, medical and genetic issues of practical import; psychological processes; evaluation of particular therapies; quality of life; family needs; and epidemiological research. Autism provides a major international forum for peer-reviewed research of direct and practical relevance to improving the quality of life for individuals with autism or autism-related disorders. The journal''s success and popularity reflect the recent worldwide growth in the research and understanding of autistic spectrum disorders, and the consequent impact on the provision of treatment and care. Autism is interdisciplinary in nature, focusing on evaluative research in all areas, including: intervention, diagnosis, training, education, neuroscience, psychological processes, evaluation of particular therapies, quality of life issues, family issues and family services, medical and genetic issues, epidemiological research.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信