Why and when should you pool? Analyzing Pooling in Recurrent Architectures

05/01/2020
by   Pratyush Maini, et al.
0

Pooling-based recurrent neural architectures consistently outperform their counterparts without pooling. However, the reasons for their enhanced performance are largely unexamined. In this work, we examine three commonly used pooling techniques (mean-pooling, max-pooling, and attention), and propose max-attention, a novel variant that effectively captures interactions among predictive tokens in a sentence. We find that pooling-based architectures substantially differ from their non-pooling equivalents in their learning ability and positional biases–which elucidate their performance benefits. By analyzing the gradient propagation, we discover that pooling facilitates better gradient flow compared to BiLSTMs. Further, we expose how BiLSTMs are positionally biased towards tokens in the beginning and the end of a sequence. Pooling alleviates such biases. Consequently, we identify settings where pooling offers large benefits: (i) in low resource scenarios, and (ii) when important words lie towards the middle of the sentence. Among the pooling techniques studied, max-attention is the most effective, resulting in significant performance gains on several text classification tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/27/2018

Natural Language Inference with Hierarchical BiLSTM Max Pooling Architecture

Recurrent neural networks have proven to be very effective for natural l...
research
11/21/2016

Text Classification Improved by Integrating Bidirectional LSTM with Two-dimensional Max Pooling

Recurrent Neural Network (RNN) is one of the most popular architectures ...
research
06/02/2020

Studying The Effect of MIL Pooling Filters on MIL Tasks

There are different multiple instance learning (MIL) pooling filters use...
research
04/02/2020

ProxyNCA++: Revisiting and Revitalizing Proxy Neighborhood Component Analysis

We consider the problem of distance metric learning (DML), where the tas...
research
06/26/2018

Enhancing Sentence Embedding with Generalized Pooling

Pooling is an essential component of a wide variety of sentence represen...
research
01/30/2017

Emergence of Selective Invariance in Hierarchical Feed Forward Networks

Many theories have emerged which investigate how in- variance is generat...
research
11/29/2017

Colour Constancy: Biologically-inspired Contrast Variant Pooling Mechanism

Pooling is a ubiquitous operation in image processing algorithms that al...

Please sign up or login with your details

Forgot password? Click here to reset