Learning neural trans-dimensional random field language models with noise-contrastive estimation

10/30/2017
by   Bin Wang, et al.
0

Trans-dimensional random field language models (TRF LMs) where sentences are modeled as a collection of random fields, have shown close performance with LSTM LMs in speech recognition and are computationally more efficient in inference. However, the training efficiency of neural TRF LMs is not satisfactory, which limits the scalability of TRF LMs on large training corpus. In this paper, several techniques on both model formulation and parameter estimation are proposed to improve the training efficiency and the performance of neural TRF LMs. First, TRFs are reformulated in the form of exponential tilting of a reference distribution. Second, noise-contrastive estimation (NCE) is introduced to jointly estimate the model parameters and normalization constants. Third, we extend the neural TRF LMs by marrying the deep convolutional neural network (CNN) and the bidirectional LSTM into the potential function to extract the deep hierarchical features and bidirectionally sequential features. Utilizing all the above techniques enables the successful and efficient training of neural TRF LMs on a 40x larger training set with only 1/3 training time and further reduces the WER with relative reduction of 4.7

READ FULL TEXT

page 1

page 2

research
07/03/2018

Improved training of neural trans-dimensional random field language models with dynamic noise-contrastive estimation

A new whole-sentence language model - neural trans-dimensional random fi...
research
07/23/2017

Language modeling with Neural trans-dimensional random fields

Trans-dimensional random field language models (TRF LMs) have recently b...
research
03/30/2016

Model Interpolation with Trans-dimensional Random Field Language Models for Speech Recognition

The dominant language models (LMs) such as n-gram and neural network (NN...
research
02/14/2020

Integrating Discrete and Neural Features via Mixed-feature Trans-dimensional Random Field Language Models

There has been a long recognition that discrete features (n-gram feature...
research
08/20/2017

A Batch Noise Contrastive Estimation Approach for Training Large Vocabulary Language Models

Training large vocabulary Neural Network Language Models (NNLMs) is a di...
research
09/22/2017

Improving Language Modelling with Noise-contrastive estimation

Neural language models do not scale well when the vocabulary is large. N...
research
05/27/2022

Unsupervised learning of features and object boundaries from local prediction

A visual system has to learn both which features to extract from images ...

Please sign up or login with your details

Forgot password? Click here to reset