Classifying musical instruments using speech signal processing methods

Seema Ghisingh, V. K. Mittal
{"title":"Classifying musical instruments using speech signal processing methods","authors":"Seema Ghisingh, V. K. Mittal","doi":"10.1109/INDICON.2016.7839034","DOIUrl":null,"url":null,"abstract":"Identification of musical instruments from the acoustic signal using speech signal processing methods is a challenging problem. Further, whether this identification can be carried out by a single musical note, like humans are able to do, is an interesting research issue that has several potential applications in the music industry. Attempts have been made earlier using the spectral and temporal features of the music acoustic signals. The process of identifying the musical instrument from monophonic audio recording basically involves three steps — pre-processing of music signal, extracting features from it and then classifying those. In this paper, we present an experiment-based comparative study of different features for classifying few musical instruments. The acoustic features, namely, the Mel-Frequency Cepstral Coefficients (MFCCs), Spectral Centroids (SC), Zero-Crossing Rate (ZCR) and signal energy are derived from the music acoustic signal using different speech signal processing methods. A Support Vector Machine (SVM) classifier is used with each feature for the relative comparisons. The classification results using different combinations of training by features from different music instrument and testing with another/same type of music instruments are compared. Our results indicate that the most significant feature for classifying Guitar, Violin and Drum is MFCC as it gives the better accurate results. Also, the feature which gives better accuracy results for the drum instrument is ZCR. Among the features used, after MFCC, ZCR proved to be the optimal feature for the classification of drum instrument.","PeriodicalId":283953,"journal":{"name":"2016 IEEE Annual India Conference (INDICON)","volume":"180 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE Annual India Conference (INDICON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INDICON.2016.7839034","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11

Abstract

Identification of musical instruments from the acoustic signal using speech signal processing methods is a challenging problem. Further, whether this identification can be carried out by a single musical note, like humans are able to do, is an interesting research issue that has several potential applications in the music industry. Attempts have been made earlier using the spectral and temporal features of the music acoustic signals. The process of identifying the musical instrument from monophonic audio recording basically involves three steps — pre-processing of music signal, extracting features from it and then classifying those. In this paper, we present an experiment-based comparative study of different features for classifying few musical instruments. The acoustic features, namely, the Mel-Frequency Cepstral Coefficients (MFCCs), Spectral Centroids (SC), Zero-Crossing Rate (ZCR) and signal energy are derived from the music acoustic signal using different speech signal processing methods. A Support Vector Machine (SVM) classifier is used with each feature for the relative comparisons. The classification results using different combinations of training by features from different music instrument and testing with another/same type of music instruments are compared. Our results indicate that the most significant feature for classifying Guitar, Violin and Drum is MFCC as it gives the better accurate results. Also, the feature which gives better accuracy results for the drum instrument is ZCR. Among the features used, after MFCC, ZCR proved to be the optimal feature for the classification of drum instrument.
用语音信号处理方法对乐器进行分类
利用语音信号处理方法从声信号中识别乐器是一个具有挑战性的问题。此外,这种识别是否能像人类一样,通过一个音符来完成,是一个有趣的研究问题,在音乐产业中有几个潜在的应用。早先已经尝试利用音乐声学信号的频谱和时间特征。从单声道录音中识别乐器的过程主要包括三个步骤:对音乐信号进行预处理,从中提取特征,然后对特征进行分类。在本文中,我们提出了一个基于实验的比较研究,不同的特征,以分类少数乐器。采用不同的语音信号处理方法,得到音乐声学信号的Mel-Frequency倒谱系数(MFCCs)、谱质心(SC)、过零率(ZCR)和信号能量等声学特征。使用支持向量机(SVM)分类器对每个特征进行相对比较。比较了不同乐器特征训练和其他/同类乐器测试的不同组合的分类结果。我们的研究结果表明,在吉他、小提琴和鼓的分类中,最显著的特征是MFCC,因为它给出了更好的准确结果。同时,对鼓形仪精度效果较好的特点是ZCR。在使用的特征中,MFCC之后,ZCR被证明是最适合鼓乐器分类的特征。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信