Sound Model Factory: An Integrated System Architecture for Generative Audio Modelling

06/27/2022
by   Lonce Wyse, et al.
12

We introduce a new system for data-driven audio sound model design built around two different neural network architectures, a Generative Adversarial Network(GAN) and a Recurrent Neural Network (RNN), that takes advantage of the unique characteristics of each to achieve the system objectives that neither is capable of addressing alone. The objective of the system is to generate interactively controllable sound models given (a) a range of sounds the model should be able to synthesize, and (b) a specification of the parametric controls for navigating that space of sounds. The range of sounds is defined by a dataset provided by the designer, while the means of navigation is defined by a combination of data labels and the selection of a sub-manifold from the latent space learned by the GAN. Our proposed system takes advantage of the rich latent space of a GAN that consists of sounds that fill out the spaces ”between" real data-like sounds. This augmented data from the GAN is then used to train an RNN for its ability to respond immediately and continuously to parameter changes and to generate audio over unlimited periods of time. Furthermore, we develop a self-organizing map technique for “smoothing" the latent space of GAN that results in perceptually smooth interpolation between audio timbres. We validate this process through user studies. The system contributes advances to the state of the art for generative sound model design that include system configuration and components for improving interpolation and the expansion of audio modeling capabilities beyond musical pitch and percussive instrument sounds into the more complex space of audio textures.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/23/2023

Towards Controllable Audio Texture Morphing

In this paper, we propose a data-driven approach to train a Generative A...
research
05/28/2018

Real-valued parametric conditioning of an RNN for interactive sound synthesis

A Recurrent Neural Network (RNN) for audio synthesis is trained by augme...
research
06/29/2022

DrumGAN VST: A Plugin for Drum Sound Analysis/Synthesis With Autoencoding Generative Adversarial Networks

In contemporary popular music production, drum sound design is commonly ...
research
03/26/2019

Conditioning a Recurrent Neural Network to synthesize musical instrument transients

A recurrent Neural Network (RNN) is trained to predict sound samples bas...
research
05/16/2019

Non-Parametric Priors For Generative Adversarial Networks

The advent of generative adversarial networks (GAN) has enabled new capa...
research
11/09/2021

CAESynth: Real-Time Timbre Interpolation and Pitch Control with Conditional Autoencoders

In this paper, we present a novel audio synthesizer, CAESynth, based on ...
research
08/23/2023

Example-Based Framework for Perceptually Guided Audio Texture Generation

Generative models for synthesizing audio textures explicitly encode cont...

Please sign up or login with your details

Forgot password? Click here to reset