RNN Architecture Learning with Sparse Regularization

09/06/2019
by   Jesse Dodge, et al.
13

Neural models for NLP typically use large numbers of parameters to reach state-of-the-art performance, which can lead to excessive memory usage and increased runtime. We present a structure learning method for learning sparse, parameter-efficient NLP models. Our method applies group lasso to rational RNNs (Peng et al., 2018), a family of models that is closely connected to weighted finite-state automata (WFSAs). We take advantage of rational RNNs' natural grouping of the weights, so the group lasso penalty directly removes WFSA states, substantially reducing the number of parameters in the model. Our experiments on a number of sentiment analysis datasets, using both GloVe and BERT embeddings, show that our approach learns neural structures which have fewer parameters without sacrificing performance relative to parameter-rich baselines. Our method also highlights the interpretable properties of rational RNNs. We show that sparsifying such models makes them easier to visualize, and we present models that rely exclusively on as few as three WFSAs after pruning more than 90

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2017

Block-Sparse Recurrent Neural Networks

Recurrent Neural Networks (RNNs) are used in state-of-the-art models in ...
research
05/03/2018

Noisin: Unbiased Regularization for Recurrent Neural Networks

Recurrent neural networks (RNNs) are powerful models of sequential data....
research
04/18/2020

A Formal Hierarchy of RNN Architectures

We develop a formal hierarchy of the expressive capacity of RNN architec...
research
01/25/2019

State-Regularized Recurrent Neural Networks

Recurrent neural networks are a widely used class of neural architecture...
research
08/28/2018

Rational Recurrences

Despite the tremendous empirical success of neural models in natural lan...
research
12/16/2021

Explainable Natural Language Processing with Matrix Product States

Despite empirical successes of recurrent neural networks (RNNs) in natur...

Please sign up or login with your details

Forgot password? Click here to reset