S. Masood, J. S. Nayal, R. Jain, M. N. Doja, Musheer Ahmad
{"title":"MFCC, Spectral and Temporal Feature based Emotion Identification in Songs","authors":"S. Masood, J. S. Nayal, R. Jain, M. N. Doja, Musheer Ahmad","doi":"10.14257/ijhit.2017.10.5.03","DOIUrl":null,"url":null,"abstract":"This work aims for the solution of one of the challenging yet evolving problem of music information retrieval field i.e. identification of emotions in songs. A collection of four datasets of four separate song sample sizes were selected for the purpose of experiments. For each experiment features such as MFCC, spectral and temporal were extracted for each sample of the dataset. A multilayered sigmoidal feed-forward neural network was trained for construction of a model by using error back propagation algorithm. This helped in recognition of four emotion categories (sad, happy, peaceful and angry) from the song samples. The results obtained at the end of these experiments strongly suggest that this trained model was successfully able to identify the emotions in the selected song samples with an average class accuracy of 88.65%.","PeriodicalId":170772,"journal":{"name":"International Journal of Hybrid Information Technology","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Hybrid Information Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.14257/ijhit.2017.10.5.03","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
This work aims for the solution of one of the challenging yet evolving problem of music information retrieval field i.e. identification of emotions in songs. A collection of four datasets of four separate song sample sizes were selected for the purpose of experiments. For each experiment features such as MFCC, spectral and temporal were extracted for each sample of the dataset. A multilayered sigmoidal feed-forward neural network was trained for construction of a model by using error back propagation algorithm. This helped in recognition of four emotion categories (sad, happy, peaceful and angry) from the song samples. The results obtained at the end of these experiments strongly suggest that this trained model was successfully able to identify the emotions in the selected song samples with an average class accuracy of 88.65%.