Compressing Neural Language Models by Sparse Word Representations

10/13/2016
by   Yunchuan Chen, et al.
0

Neural networks are among the state-of-the-art techniques for language modeling. Existing neural language models typically map discrete words to distributed, dense vector representations. After information processing of the preceding context words by hidden layers, an output layer estimates the probability of the next word. Such approaches are time- and memory-intensive because of the large numbers of parameters for word embeddings and the output layer. In this paper, we propose to compress neural language models by sparse word representations. In the experiments, the number of parameters in our model increases very slowly with the growth of the vocabulary size, which is almost imperceptible. Moreover, our approach not only reduces the parameter space to a large extent, but also improves the performance in terms of the perplexity measure.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/10/2017

Character-Word LSTM Language Models

We present a Character-Word Long Short-Term Memory Language Model which ...
research
11/21/2015

BlackOut: Speeding up Recurrent Neural Network Language Models With Very Large Vocabularies

We propose BlackOut, an approximation algorithm to efficiently train mas...
research
10/31/2016

LightRNN: Memory and Computation-Efficient Recurrent Neural Networks

Recurrent neural networks (RNNs) have achieved state-of-the-art performa...
research
08/27/2018

Predefined Sparseness in Recurrent Sequence Models

Inducing sparseness while training neural networks has been shown to yie...
research
07/02/2018

Neural Random Projections for Language Modelling

Neural network-based language models deal with data sparsity problems by...
research
06/12/2019

Putting words in context: LSTM language models and lexical ambiguity

In neural network models of language, words are commonly represented usi...
research
09/28/2018

Adaptive Input Representations for Neural Language Modeling

We introduce adaptive input representations for neural language modeling...

Please sign up or login with your details

Forgot password? Click here to reset