Attentive Pooling Networks

02/11/2016
by   Cicero dos Santos, et al.
0

In this work, we propose Attentive Pooling (AP), a two-way attention mechanism for discriminative model training. In the context of pair-wise ranking or classification with neural networks, AP enables the pooling layer to be aware of the current input pair, in a way that information from the two input items can directly influence the computation of each other's representations. Along with such representations of the paired inputs, AP jointly learns a similarity measure over projected segments (e.g. trigrams) of the pair, and subsequently, derives the corresponding attention vector for each input to guide the pooling. Our two-way attention mechanism is a general framework independent of the underlying representation learning, and it has been applied to both convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in our studies. The empirical results, from three very different benchmark tasks of question answering/answer selection, demonstrate that our proposed models outperform a variety of strong baselines and achieve state-of-the-art performance in all the benchmarks.

READ FULL TEXT

page 8

page 9

research
10/02/2017

Attentive Convolution

In NLP, convolution neural networks (CNNs) have benefited less than recu...
research
04/24/2021

A Multi-Size Neural Network with Attention Mechanism for Answer Selection

Semantic matching is of central significance to the answer selection tas...
research
12/11/2022

Vision Transformer with Attentive Pooling for Robust Facial Expression Recognition

Facial Expression Recognition (FER) in the wild is an extremely challeng...
research
08/08/2018

Learning to Focus when Ranking Answers

One of the main challenges in ranking is embedding the query and documen...
research
06/03/2018

Multi-Cast Attention Networks for Retrieval-based Question Answering and Response Prediction

Attention is typically used to select informative sub-phrases that are u...
research
04/10/2018

Question Answering over Freebase via Attentive RNN with Similarity Matrix based CNN

With the rapid growth of knowledge bases (KBs), question answering over ...
research
12/08/2022

Group Generalized Mean Pooling for Vision Transformer

Vision Transformer (ViT) extracts the final representation from either c...

Please sign up or login with your details

Forgot password? Click here to reset