{"title":"Integration of Different Optimization Approaches for the Classification of Different Arabic and English Voice Commands","authors":"Karim Dabbabi, Abdelkarim Mars","doi":"10.1109/IC_ASET58101.2023.10151158","DOIUrl":null,"url":null,"abstract":"Several hyperparameters represent major sensitive factors for deep learning models. For this, different hyperparameter optimization approaches are proposed to accelerate the convergence towards the optimal configurations and to support the calculations during the long learning times, and thus to give improved performance. These approaches include Bayesian Optimization (BO), Hyperband and Tree Parzen Estimator (TPE), and they are proposed for optimization task in this paper. Also, Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN) are suggested as classifiers and Mel frequency cepstrum coefficients (MFCC) and Mel as features. Experiments showed that the best results in terms of evaluated performances (Precision = 94.96%, Recall = 94.85%, F1 = 94.85%) were obtained with the combination of LSTM and MFCC (LSTM-MFCC) with BO on English voice command database compared to those obtained with other combinations of features and classifiers with different optimization approaches. Moreover, there is evidence that BO converged faster than TPE and HB, and converged to better configurations.","PeriodicalId":272261,"journal":{"name":"2023 IEEE International Conference on Advanced Systems and Emergent Technologies (IC_ASET)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Conference on Advanced Systems and Emergent Technologies (IC_ASET)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IC_ASET58101.2023.10151158","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Several hyperparameters represent major sensitive factors for deep learning models. For this, different hyperparameter optimization approaches are proposed to accelerate the convergence towards the optimal configurations and to support the calculations during the long learning times, and thus to give improved performance. These approaches include Bayesian Optimization (BO), Hyperband and Tree Parzen Estimator (TPE), and they are proposed for optimization task in this paper. Also, Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN) are suggested as classifiers and Mel frequency cepstrum coefficients (MFCC) and Mel as features. Experiments showed that the best results in terms of evaluated performances (Precision = 94.96%, Recall = 94.85%, F1 = 94.85%) were obtained with the combination of LSTM and MFCC (LSTM-MFCC) with BO on English voice command database compared to those obtained with other combinations of features and classifiers with different optimization approaches. Moreover, there is evidence that BO converged faster than TPE and HB, and converged to better configurations.