Weipeng Wang, Xiaobing Li, Cong Jin, Di Lu, Qingwen Zhou, Tie Yun
{"title":"CPS: Full-Song and Style-Conditioned Music Generation with Linear Transformer","authors":"Weipeng Wang, Xiaobing Li, Cong Jin, Di Lu, Qingwen Zhou, Tie Yun","doi":"10.1109/ICMEW56448.2022.9859286","DOIUrl":null,"url":null,"abstract":"Many deep music generation algorithms have recently been able to produce good-sounding music, but there have been few studies on controlled generation. In this process, the human sense of participation is usually very weak, and it is difficult to integrate one’s own musical motivation into the creation. In this study, we will introduce CPS (Compound word with style), a model that can specify a target style and generate a complete musical composition from scratch. We first added the genre meta-information to the music representation and distinguished it from other low-level music representations, thus strengthening the influence of the control signal. We modeled with the linear transformer, while used an adaptive strategy with different settings for different types of music tokens to reduce the probability of disharmonic music. The experiments show that, when compared to the baseline model, our model performs better in terms of basic music metrics as well as metrics for evaluating controlled ability.","PeriodicalId":106759,"journal":{"name":"2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMEW56448.2022.9859286","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Many deep music generation algorithms have recently been able to produce good-sounding music, but there have been few studies on controlled generation. In this process, the human sense of participation is usually very weak, and it is difficult to integrate one’s own musical motivation into the creation. In this study, we will introduce CPS (Compound word with style), a model that can specify a target style and generate a complete musical composition from scratch. We first added the genre meta-information to the music representation and distinguished it from other low-level music representations, thus strengthening the influence of the control signal. We modeled with the linear transformer, while used an adaptive strategy with different settings for different types of music tokens to reduce the probability of disharmonic music. The experiments show that, when compared to the baseline model, our model performs better in terms of basic music metrics as well as metrics for evaluating controlled ability.
近年来,许多深度音乐生成算法都能够产生好听的音乐,但对控制生成的研究却很少。在这个过程中,人的参与感通常很弱,很难将自己的音乐动机融入到创作中。在本研究中,我们将介绍CPS (Compound word with style),这是一个可以指定目标风格并从头生成完整音乐作品的模型。我们首先在音乐表征中加入体裁元信息,并将其与其他低级音乐表征区分开来,从而加强控制信号的影响。我们使用线性变压器建模,同时对不同类型的音乐符号使用不同设置的自适应策略来减少不和谐音乐的概率。实验表明,与基线模型相比,我们的模型在基本音乐指标以及评估控制能力的指标方面表现得更好。