A Comparison of Discrete Latent Variable Models for Speech Representation Learning

10/24/2020
by   Henry Zhou, et al.
0

Neural latent variable models enable the discovery of interesting structure in speech audio data. This paper presents a comparison of two different approaches which are broadly based on predicting future time-steps or auto-encoding the input signal. Our study compares the representations learned by vq-vae and vq-wav2vec in terms of sub-word unit discovery and phoneme recognition performance. Results show that future time-step prediction with vq-wav2vec achieves better performance. The best system achieves an error rate of 13.22 on the ZeroSpeech 2019 ABX phoneme discrimination challenge

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2022

Improving VAE-based Representation Learning

Latent variable models like the Variational Auto-Encoder (VAE) are commo...
research
04/16/2019

Unsupervised acoustic unit discovery for speech synthesis using discrete latent-variable neural networks

For our submission to the ZeroSpeech 2019 challenge, we apply discrete l...
research
05/18/2020

Robust Training of Vector Quantized Bottleneck Models

In this paper we demonstrate methods for reliable and efficient training...
research
08/16/2020

Unsupervised Acoustic Unit Representation Learning for Voice Conversion using WaveNet Auto-encoders

Unsupervised representation learning of speech has been of keen interest...
research
02/22/2022

Benchmarking Generative Latent Variable Models for Speech

Stochastic latent variable models (LVMs) achieve state-of-the-art perfor...
research
05/19/2020

Bayesian Subspace HMM for the Zerospeech 2020 Challenge

In this paper we describe our submission to the Zerospeech 2020 challeng...
research
05/24/2020

Transformer VQ-VAE for Unsupervised Unit Discovery and Speech Synthesis: ZeroSpeech 2020 Challenge

In this paper, we report our submitted system for the ZeroSpeech 2020 ch...

Please sign up or login with your details

Forgot password? Click here to reset