Rational Recurrences

08/28/2018
by   Hao Peng, et al.
0

Despite the tremendous empirical success of neural models in natural language processing, many of them lack the strong intuitions that accompany classical machine learning approaches. Recently, connections have been shown between convolutional neural networks (CNNs) and weighted finite state automata (WFSAs), leading to new interpretations and insights. In this work, we show that some recurrent neural networks also share this connection to WFSAs. We characterize this connection formally, defining rational recurrences to be recurrent hidden state update functions that can be written as the Forward calculation of a finite set of WFSAs. We show that several recent neural models use rational recurrences. Our analysis provides a fresh view of these models and facilitates devising new neural architectures that draw inspiration from WFSAs. We present one such model, which performs better than two recent baselines on language modeling and text classification. Our results demonstrate that transferring intuitions from classical models like WFSAs can be an effective approach to designing and understanding neural models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2016

Combining Recurrent and Convolutional Neural Networks for Relation Classification

This paper investigates two different neural architectures for the task ...
research
06/27/2022

Extracting Weighted Finite Automata from Recurrent Neural Networks for Natural Languages

Recurrent Neural Networks (RNNs) have achieved tremendous success in seq...
research
05/15/2018

SoPa: Bridging CNNs, RNNs, and Weighted Finite-State Machines

Recurrent and convolutional neural networks comprise two distinct famili...
research
03/26/2017

Learning Simpler Language Models with the Differential State Framework

Learning useful information across long time lags is a critical and diff...
research
12/13/2021

Language Models are not Models of Language

Natural Language Processing (NLP) has become one of the leading applicat...
research
09/06/2019

RNN Architecture Learning with Sparse Regularization

Neural models for NLP typically use large numbers of parameters to reach...
research
08/20/2015

Auto-Sizing Neural Networks: With Applications to n-gram Language Models

Neural networks have been shown to improve performance across a range of...

Please sign up or login with your details

Forgot password? Click here to reset