{"title":"Music Generation using Deep Generative Modelling","authors":"Advait Maduskar, Aniket Ladukar, Shubhankar Gore, Neha Patwari","doi":"10.1109/ICCDW45521.2020.9318683","DOIUrl":null,"url":null,"abstract":"Efficient synthesis of musical sequences is a challenging task from a machine learning perspective, as human perception is aware of the global context to shorter sequences as well of audio waveforms on a smaller scale. Autoregressive models such as WaveNet use iterative subsampling to generate short sequences that are a result of a localized modeling process but lacking in overall global structures. In juxtaposition, Generative Adversarial Networks (GANs) are effective for modeling globally coherent sequence structures, but struggle to generate localized sequences. Through this project, we aim to propose a system that combines the random subsampling approach of GANs with a recurrent autoregressive model. Such a model will help to model coherent musical structures effectively on both, global and local levels.","PeriodicalId":282429,"journal":{"name":"2020 International Conference on Convergence to Digital World - Quo Vadis (ICCDW)","volume":"2673 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Convergence to Digital World - Quo Vadis (ICCDW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCDW45521.2020.9318683","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
Efficient synthesis of musical sequences is a challenging task from a machine learning perspective, as human perception is aware of the global context to shorter sequences as well of audio waveforms on a smaller scale. Autoregressive models such as WaveNet use iterative subsampling to generate short sequences that are a result of a localized modeling process but lacking in overall global structures. In juxtaposition, Generative Adversarial Networks (GANs) are effective for modeling globally coherent sequence structures, but struggle to generate localized sequences. Through this project, we aim to propose a system that combines the random subsampling approach of GANs with a recurrent autoregressive model. Such a model will help to model coherent musical structures effectively on both, global and local levels.