Emotion classification for musical data using deep learning techniques

Gaurav Agarwal, Sachi Gupta, Shivani Agarwal, Aul Kumar Rai
{"title":"Emotion classification for musical data using deep learning techniques","authors":"Gaurav Agarwal, Sachi Gupta, Shivani Agarwal, Aul Kumar Rai","doi":"10.11591/ijres.v12.i2.pp240-247","DOIUrl":null,"url":null,"abstract":"This research is done based on the identification and thorough analyzing musical data that is extracted by the various method. This extracted information can be utilized in the deep learning algorithm to identify the emotion, based on the hidden features of the dataset. Deep learning-based convolutional neural network (CNN) and long short-term memory-gated recurrent unit (LSTM-GRU) models were developed to predict the information from the musical information. The musical dataset is extracted using the fast Fourier transform (FFT) models. The three deep learning models were developed in this work the first model was based on the information of extracted information such as zero-crossing rate, and spectral roll-off. Another model was developed on the information of Mel frequencybased cepstral coefficient (MFCC) features, the deep and wide CNN algorithm with LSTM-GRU bidirectional model was developed. The third model was developed on the extracted information from Mel-spectrographs and untied these graphs based on two-dimensional (2D) data information to the 2D CNN model alongside LSTM models. Proposed model performance on the information from Mel-spectrographs is compared on the F1 score, precision, and classification report of the models. Which shows better accuracy with improved F1 and recall values as compared with existing approaches.","PeriodicalId":158991,"journal":{"name":"International Journal of Reconfigurable and Embedded Systems (IJRES)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Reconfigurable and Embedded Systems (IJRES)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.11591/ijres.v12.i2.pp240-247","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

This research is done based on the identification and thorough analyzing musical data that is extracted by the various method. This extracted information can be utilized in the deep learning algorithm to identify the emotion, based on the hidden features of the dataset. Deep learning-based convolutional neural network (CNN) and long short-term memory-gated recurrent unit (LSTM-GRU) models were developed to predict the information from the musical information. The musical dataset is extracted using the fast Fourier transform (FFT) models. The three deep learning models were developed in this work the first model was based on the information of extracted information such as zero-crossing rate, and spectral roll-off. Another model was developed on the information of Mel frequencybased cepstral coefficient (MFCC) features, the deep and wide CNN algorithm with LSTM-GRU bidirectional model was developed. The third model was developed on the extracted information from Mel-spectrographs and untied these graphs based on two-dimensional (2D) data information to the 2D CNN model alongside LSTM models. Proposed model performance on the information from Mel-spectrographs is compared on the F1 score, precision, and classification report of the models. Which shows better accuracy with improved F1 and recall values as compared with existing approaches.
使用深度学习技术对音乐数据进行情感分类
本研究是在对各种方法提取的音乐数据进行识别和深入分析的基础上完成的。这些提取的信息可以用于深度学习算法,基于数据集的隐藏特征来识别情绪。建立了基于深度学习的卷积神经网络(CNN)和长短期记忆门控循环单元(LSTM-GRU)模型对音乐信息进行预测。使用快速傅里叶变换(FFT)模型提取音乐数据集。本文建立了三种深度学习模型,第一种模型基于提取的信息,如过零率和谱滚降。另一种基于Mel频基倒谱系数(MFCC)特征信息的模型,提出了基于LSTM-GRU双向模型的深宽CNN算法。第三个模型是基于mel -光谱仪提取的信息开发的,并将这些基于二维(2D)数据信息的图与LSTM模型一起解绑定到二维CNN模型中。从模型的F1分数、精度和分类报告三个方面比较了模型在mel -光谱仪信息上的性能。与现有方法相比,改进的F1和召回值显示出更好的准确率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
1.50
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信