{"title":"用于自动检测和分类的自闭症语音声学特征表征","authors":"Abhijit Mohanta, Prerana Mukherjee, Vinay Kumar Mirtal","doi":"10.1109/NCC48643.2020.9056025","DOIUrl":null,"url":null,"abstract":"The verbal children affected with autism spectrum disorder (ASD) often shows some notable acoustic patterns. This paper represents the classification of autism speech, i.e., the speech signal of children affected with ASD. In addition, this work specifically aims to classify the speech signals of non-native Indo English speakers (children) affected with ASD. Previous studies, however, have focused only on native English speakers. Hence, for this study purpose a speech signal dataset of ASD children and a speech signal dataset of normal children were recorded in English, and all the children selected for the data collection were non-native Indo English speakers. Here, for the ASD and the normal children, the acoustic features explored for classification are namely, fundamental frequency (FO), strength of excitation (SoE), formants frequencies (F1 to F5), dominant frequencies (FD1, FD2), signal energy (E), zero-crossing rate (ZCR), mel-frequency cepstral coefficients (MFCC), and linear prediction cepstrum coefficients (LPCC). Further, these feature sets are classified by utilizing different classifiers. The KNN classifier model achieves the highest 96.5% accuracy with respect to other baseline models explored here.","PeriodicalId":183772,"journal":{"name":"2020 National Conference on Communications (NCC)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":"{\"title\":\"Acoustic Features Characterization of Autism Speech for Automated Detection and Classification\",\"authors\":\"Abhijit Mohanta, Prerana Mukherjee, Vinay Kumar Mirtal\",\"doi\":\"10.1109/NCC48643.2020.9056025\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The verbal children affected with autism spectrum disorder (ASD) often shows some notable acoustic patterns. This paper represents the classification of autism speech, i.e., the speech signal of children affected with ASD. In addition, this work specifically aims to classify the speech signals of non-native Indo English speakers (children) affected with ASD. Previous studies, however, have focused only on native English speakers. Hence, for this study purpose a speech signal dataset of ASD children and a speech signal dataset of normal children were recorded in English, and all the children selected for the data collection were non-native Indo English speakers. Here, for the ASD and the normal children, the acoustic features explored for classification are namely, fundamental frequency (FO), strength of excitation (SoE), formants frequencies (F1 to F5), dominant frequencies (FD1, FD2), signal energy (E), zero-crossing rate (ZCR), mel-frequency cepstral coefficients (MFCC), and linear prediction cepstrum coefficients (LPCC). Further, these feature sets are classified by utilizing different classifiers. The KNN classifier model achieves the highest 96.5% accuracy with respect to other baseline models explored here.\",\"PeriodicalId\":183772,\"journal\":{\"name\":\"2020 National Conference on Communications (NCC)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"9\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 National Conference on Communications (NCC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NCC48643.2020.9056025\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 National Conference on Communications (NCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NCC48643.2020.9056025","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Acoustic Features Characterization of Autism Speech for Automated Detection and Classification
The verbal children affected with autism spectrum disorder (ASD) often shows some notable acoustic patterns. This paper represents the classification of autism speech, i.e., the speech signal of children affected with ASD. In addition, this work specifically aims to classify the speech signals of non-native Indo English speakers (children) affected with ASD. Previous studies, however, have focused only on native English speakers. Hence, for this study purpose a speech signal dataset of ASD children and a speech signal dataset of normal children were recorded in English, and all the children selected for the data collection were non-native Indo English speakers. Here, for the ASD and the normal children, the acoustic features explored for classification are namely, fundamental frequency (FO), strength of excitation (SoE), formants frequencies (F1 to F5), dominant frequencies (FD1, FD2), signal energy (E), zero-crossing rate (ZCR), mel-frequency cepstral coefficients (MFCC), and linear prediction cepstrum coefficients (LPCC). Further, these feature sets are classified by utilizing different classifiers. The KNN classifier model achieves the highest 96.5% accuracy with respect to other baseline models explored here.