{"title":"Harmonic Blind Sound Source Isolation Enhanced by Spectrum Clustering","authors":"Cynthia Xin Zhang, Wenxin Jiang, Z. Ras","doi":"10.1109/ICDMW.2008.67","DOIUrl":null,"url":null,"abstract":"Automatic indexing of music by instruments and their types is a challenging problem, especially when multiple instruments are playing at the same time. We have built a database containing more than one million of music instrument sounds, each described by a large number o features including standard MPEG7 audio descriptors, features for speech recognition, and many new audio features developed by our team. Our previous research results show that all these features only lead to classifiers which successfully identify music instruments in monophonic music (only one instrument playing at a time). Their confidence for polyphonic music is much lower. This brought the need for blind sound source separation algorithms. In this paper, we present a new spectrum clustering enhanced method which improves the estimation of fundamental frequency as well as the balance of the categorization tree of training datasets, and therefore enhances the precision of automatic indexing. The system is recursively detecting the pitch of the predominant sound source, then calculates the features based on the estimated pitch, and then predicts the most similar spectrum by the corresponding classification tree, and finally subtracts the estimated predominant spectrum until silence is detected.","PeriodicalId":175955,"journal":{"name":"2008 IEEE International Conference on Data Mining Workshops","volume":"56 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 IEEE International Conference on Data Mining Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDMW.2008.67","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Automatic indexing of music by instruments and their types is a challenging problem, especially when multiple instruments are playing at the same time. We have built a database containing more than one million of music instrument sounds, each described by a large number o features including standard MPEG7 audio descriptors, features for speech recognition, and many new audio features developed by our team. Our previous research results show that all these features only lead to classifiers which successfully identify music instruments in monophonic music (only one instrument playing at a time). Their confidence for polyphonic music is much lower. This brought the need for blind sound source separation algorithms. In this paper, we present a new spectrum clustering enhanced method which improves the estimation of fundamental frequency as well as the balance of the categorization tree of training datasets, and therefore enhances the precision of automatic indexing. The system is recursively detecting the pitch of the predominant sound source, then calculates the features based on the estimated pitch, and then predicts the most similar spectrum by the corresponding classification tree, and finally subtracts the estimated predominant spectrum until silence is detected.