{"title":"Audio Signal Processing and Musical Instrument Detection using Deep Learning Techniques","authors":"S. Elghamrawy, Shehab Edin Ibrahim","doi":"10.1109/JAC-ECC54461.2021.9691427","DOIUrl":null,"url":null,"abstract":"The advance of deep learning and audio signal processing techniques has led to serious development on Musical Information retrieval (MIR). Effective audio processing can improve speed, reduce errors, and sometimes increase the accuracy of detecting musical instrument. Spectrographic data is also necessary for many mathematical tools common across Musical Information retrieval. A major aspect of MIR is the categorization of pieces of music. One of the main tools used for categorization tasks in recent years is deep learning, which has led to many advancements in MIR. One such categorization task that deep learning is useful for is the recognition of instruments in a piece of music. In this paper, a new architecture is proposed for audio processing and musical instrument detection using Multilayer Perceptron (MLPs), Convolution Neural Networks (CNN), and Recurrent Neural Networks - Long Short Term Memory (RNN-LSTM). In addition, a number of experiments are implemented using real dataset that contains 20,000 recording. The three deep learning techniques are implemented and compared to present potential new solutions. The usage of processing techniques unique to the field of deep learning is also discussed.","PeriodicalId":354908,"journal":{"name":"2021 9th International Japan-Africa Conference on Electronics, Communications, and Computations (JAC-ECC)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 9th International Japan-Africa Conference on Electronics, Communications, and Computations (JAC-ECC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/JAC-ECC54461.2021.9691427","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
The advance of deep learning and audio signal processing techniques has led to serious development on Musical Information retrieval (MIR). Effective audio processing can improve speed, reduce errors, and sometimes increase the accuracy of detecting musical instrument. Spectrographic data is also necessary for many mathematical tools common across Musical Information retrieval. A major aspect of MIR is the categorization of pieces of music. One of the main tools used for categorization tasks in recent years is deep learning, which has led to many advancements in MIR. One such categorization task that deep learning is useful for is the recognition of instruments in a piece of music. In this paper, a new architecture is proposed for audio processing and musical instrument detection using Multilayer Perceptron (MLPs), Convolution Neural Networks (CNN), and Recurrent Neural Networks - Long Short Term Memory (RNN-LSTM). In addition, a number of experiments are implemented using real dataset that contains 20,000 recording. The three deep learning techniques are implemented and compared to present potential new solutions. The usage of processing techniques unique to the field of deep learning is also discussed.