Adaptive Convolutional Filter Generation for Natural Language Understanding

09/25/2017
by   Dinghan Shen, et al.
0

Convolutional neural networks (CNNs) have recently emerged as a popular building block for natural language processing (NLP). Despite their success, most existing CNN models employed in NLP are not expressive enough, in the sense that all input sentences share the same learned (and static) set of filters. Motivated by this problem, we propose an adaptive convolutional filter generation framework for natural language understanding, by leveraging a meta network to generate input-aware filters. We further generalize our framework to model question-answer sentence pairs and propose an adaptive question answering (AdaQA) model; a novel two-way feature abstraction mechanism is introduced to encapsulate co-dependent sentence representations. We investigate the effectiveness of our framework on document categorization and answer sentence-selection tasks, achieving state-of-the-art performance on several benchmark datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/15/2019

Dual Supervised Learning for Natural Language Understanding and Generation

Natural language understanding (NLU) and natural language generation (NL...
research
11/10/2019

INSET: Sentence Infilling with Inter-sentential Generative Pre-training

Missing sentence generation (or sentence infilling) fosters a wide range...
research
08/28/2018

Convolutional Neural Networks with Recurrent Neural Filters

We introduce a class of convolutional neural networks (CNNs) that utiliz...
research
10/09/2016

Open-Ended Visual Question-Answering

This thesis report studies methods to solve Visual Question-Answering (V...
research
09/17/2019

Learning to Generate Questions with Adaptive Copying Neural Networks

Automatic question generation is an important problem in natural languag...
research
11/27/2018

GaterNet: Dynamic Filter Selection in Convolutional Neural Network via a Dedicated Global Gating Network

The concept of conditional computation for deep nets has been proposed p...
research
05/25/2022

Is a Question Decomposition Unit All We Need?

Large Language Models (LMs) have achieved state-of-the-art performance o...

Please sign up or login with your details

Forgot password? Click here to reset