{"title":"Comparative Study of Recurrent Neural Networks for Virtual Analog Audio Effects Modeling","authors":"Riccardo Simionato, Stefano Fasciani","doi":"arxiv-2405.04124","DOIUrl":null,"url":null,"abstract":"Analog electronic circuits are at the core of an important category of\nmusical devices. The nonlinear features of their electronic components give\nanalog musical devices a distinctive timbre and sound quality, making them\nhighly desirable. Artificial neural networks have rapidly gained popularity for\nthe emulation of analog audio effects circuits, particularly recurrent\nnetworks. While neural approaches have been successful in accurately modeling\ndistortion circuits, they require architectural improvements that account for\nparameter conditioning and low latency response. In this article, we explore\nthe application of recent machine learning advancements for virtual analog\nmodeling. We compare State Space models and Linear Recurrent Units against the\nmore common Long Short Term Memory networks. These have shown promising ability\nin sequence to sequence modeling tasks, showing a notable improvement in signal\nhistory encoding. Our comparative study uses these black box neural modeling\ntechniques with a variety of audio effects. We evaluate the performance and\nlimitations using multiple metrics aiming to assess the models' ability to\naccurately replicate energy envelopes, frequency contents, and transients in\nthe audio signal. To incorporate control parameters we employ the Feature wise\nLinear Modulation method. Long Short Term Memory networks exhibit better\naccuracy in emulating distortions and equalizers, while the State Space model,\nfollowed by Long Short Term Memory networks when integrated in an encoder\ndecoder structure, outperforms others in emulating saturation and compression.\nWhen considering long time variant characteristics, the State Space model\ndemonstrates the greatest accuracy. The Long Short Term Memory and, in\nparticular, Linear Recurrent Unit networks present more tendency to introduce\naudio artifacts.","PeriodicalId":501178,"journal":{"name":"arXiv - CS - Sound","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Sound","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.04124","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Analog electronic circuits are at the core of an important category of
musical devices. The nonlinear features of their electronic components give
analog musical devices a distinctive timbre and sound quality, making them
highly desirable. Artificial neural networks have rapidly gained popularity for
the emulation of analog audio effects circuits, particularly recurrent
networks. While neural approaches have been successful in accurately modeling
distortion circuits, they require architectural improvements that account for
parameter conditioning and low latency response. In this article, we explore
the application of recent machine learning advancements for virtual analog
modeling. We compare State Space models and Linear Recurrent Units against the
more common Long Short Term Memory networks. These have shown promising ability
in sequence to sequence modeling tasks, showing a notable improvement in signal
history encoding. Our comparative study uses these black box neural modeling
techniques with a variety of audio effects. We evaluate the performance and
limitations using multiple metrics aiming to assess the models' ability to
accurately replicate energy envelopes, frequency contents, and transients in
the audio signal. To incorporate control parameters we employ the Feature wise
Linear Modulation method. Long Short Term Memory networks exhibit better
accuracy in emulating distortions and equalizers, while the State Space model,
followed by Long Short Term Memory networks when integrated in an encoder
decoder structure, outperforms others in emulating saturation and compression.
When considering long time variant characteristics, the State Space model
demonstrates the greatest accuracy. The Long Short Term Memory and, in
particular, Linear Recurrent Unit networks present more tendency to introduce
audio artifacts.