MoEVC: A Mixture-of-experts Voice Conversion System with Sparse Gating Mechanism for Accelerating Online Computation

12/27/2019
by   Yu-Tao Chang, et al.
0

With the recent advancements of deep learning technologies, the performance of voice conversion (VC) in terms of quality and similarity has been significantly improved. However, heavy computations are generally required for deep-learning-based VC systems, which can cause notable latency and thus confine their deployments in real-world applications. Therefore, increasing online computation efficiency has become an important task. In this study, we propose a novel mixture-of-experts (MoE) based VC system. The MoE model uses a gating mechanism to specify optimal weights to feature maps to increase VC performance. In addition, assigning sparse constraints on the gating mechanism can accelerate online computation by skipping the convolution process by zeroing out redundant feature maps. Experimental results show that by specifying suitable sparse constraints, we can effectively increase the online computation efficiency with a notable 70 second) reduction while improving the VC performance in both objective evaluations and human listening tests.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/26/2022

Invertible Voice Conversion

In this paper, we propose an invertible deep learning framework called I...
research
12/27/2022

Voice conversion with limited data and limitless data augmentations

Applying changes to an input speech signal to change the perceived speak...
research
06/15/2022

Streaming non-autoregressive model for any-to-many voice conversion

Voice conversion models have developed for decades, and current mainstre...
research
11/12/2021

AC-VC: Non-parallel Low Latency Phonetic Posteriorgrams Based Voice Conversion

This paper presents AC-VC (Almost Causal Voice Conversion), a phonetic p...
research
11/29/2022

MegaBlocks: Efficient Sparse Training with Mixture-of-Experts

We present MegaBlocks, a system for efficient Mixture-of-Experts (MoE) t...
research
06/22/2020

Split to Be Slim: An Overlooked Redundancy in Vanilla Convolution

Many effective solutions have been proposed to reduce the redundancy of ...

Please sign up or login with your details

Forgot password? Click here to reset