Auto-Sizing Neural Networks: With Applications to n-gram Language Models

08/20/2015
by   Kenton Murray, et al.
0

Neural networks have been shown to improve performance across a range of natural-language tasks. However, designing and training them can be complicated. Frequently, researchers resort to repeated experimentation to pick optimal settings. In this paper, we address the issue of choosing the correct number of units in hidden layers. We introduce a method for automatically adjusting network size by pruning out hidden units through ℓ_∞,1 and ℓ_2,1 regularization. We apply this method to language modeling and demonstrate its ability to correctly choose the number of hidden units while maintaining perplexity. We also include these models in a machine translation decoder and show that these smaller neural models maintain the significant improvements of their unpruned versions.

READ FULL TEXT
research
11/09/2020

Scaling Hidden Markov Language Models

The hidden Markov model (HMM) is a fundamental tool for sequence modelin...
research
12/22/2014

Pragmatic Neural Language Modelling in Machine Translation

This paper presents an in-depth investigation on integrating neural lang...
research
03/07/2017

Data Noising as Smoothing in Neural Network Language Models

Data noising is an effective technique for regularizing neural network m...
research
12/01/2015

LSTM Neural Reordering Feature for Statistical Machine Translation

Artificial neural networks are powerful models, which have been widely a...
research
06/10/2019

Improving Neural Language Modeling via Adversarial Training

Recently, substantial progress has been made in language modeling by usi...
research
08/28/2018

Rational Recurrences

Despite the tremendous empirical success of neural models in natural lan...
research
12/14/2017

Nonparametric Neural Networks

Automatically determining the optimal size of a neural network for a giv...

Please sign up or login with your details

Forgot password? Click here to reset