Think Again Networks, the Delta Loss, and an Application in Language Modeling

04/26/2019
by   Alexandre Salle, et al.
0

This short paper introduces an abstraction called Think Again Networks (ThinkNet) which can be applied to any state-dependent function (such as a recurrent neural network). Here we show a simple application in Language Modeling which achieves state of the art perplexity on the Penn Treebank.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/26/2019

Think Again Networks and the Delta Loss

This short paper introduces an abstraction called Think Again Networks (...
research
02/25/2016

Recurrent Neural Network Grammars

We introduce recurrent neural network grammars, probabilistic models of ...
research
06/30/2019

Multiplicative Models for Recurrent Language Modeling

Recently, there has been interest in multiplicative recurrent neural net...
research
12/23/2016

Language Modeling with Gated Convolutional Networks

The pre-dominant approach to language modeling to date is based on recur...
research
12/14/2021

Towards Interactive Language Modeling

Interaction between caregivers and children plays a critical role in hum...
research
01/14/2020

Block-wise Dynamic Sparseness

Neural networks have achieved state of the art performance across a wide...
research
11/04/2016

Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling

Recurrent neural networks have been very successful at predicting sequen...

Please sign up or login with your details

Forgot password? Click here to reset