Autoencoder Based Architecture For Fast & Real Time Audio Style Transfer

12/18/2018
by   Dhruv Ramani, et al.
0

Recently, there has been great interest in the field of audio style transfer, where a stylized audio is generated by imposing the style of a reference audio on the content of a target audio. We improve on the current approaches which use neural networks to extract the content and the style of the audio signal and propose a new autoencoder based architecture for the task. This network generates a stylized audio for a content audio in a single forward pass. The proposed network architecture proves to be advantageous over the quality of audio produced and the time taken to train the network. The network is experimented on speech signals to confirm the validity of our proposal.

READ FULL TEXT
research
10/31/2017

Audio style transfer

"Style transfer" among images has recently emerged as a very active rese...
research
11/29/2017

Time Domain Neural Audio Style Transfer

A recently published method for audio style transfer has shown how to ex...
research
09/06/2022

Read it to me: An emotionally aware Speech Narration Application

In this work we try to perform emotional style transfer on audios. In pa...
research
10/19/2020

MicAugment: One-shot Microphone Style Transfer

A crucial aspect for the successful deployment of audio-based models "in...
research
06/29/2017

Audio Spectrogram Representations for Processing with Convolutional Neural Networks

One of the decisions that arise when designing a neural network for any ...
research
01/29/2019

Applying Visual Domain Style Transfer and Texture Synthesis Techniques to Audio - Insights and Challenges

Style transfer is a technique for combining two images based on the acti...
research
01/04/2018

Neural Style Transfer for Audio Spectograms

There has been fascinating work on creating artistic transformations of ...

Please sign up or login with your details

Forgot password? Click here to reset