Lightweight Convolutional Approaches to Reading Comprehension on SQuAD

10/19/2018
by   Tobin Bell, et al.
0

Current state-of-the-art reading comprehension models rely heavily on recurrent neural networks. We explored an entirely different approach to question answering: a convolutional model. By their nature, these convolutional models are fast to train and capture local dependencies well, though they can struggle with longer-range dependencies and thus require augmentation to achieve comparable performance to RNN-based models. We conducted over two dozen controlled experiments with convolutional models and various kernel/attention/regularization schemes to determine the precise performance gains of each strategy, while maintaining a focus on speed. We ultimately ensembled three models: crossconv (0.5398 dev F1), attnconv (0.5665), and maybeconv (0.5285). The ensembled model was able to achieve a 0.6238 F1 score using the official SQuAD evaluation script. Our individual convolutional model crossconv was able to exceed the performance of the RNN-plus-attention baseline by 25

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2017

Fast Reading Comprehension with ConvNets

State-of-the-art deep reading comprehension models are dominated by recu...
research
04/23/2018

QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension

Current end-to-end machine reading and question answering (Q&A) models a...
research
04/16/2018

Neural Models for Reasoning over Multiple Mentions using Coreference

Many problems in NLP require aggregating information from multiple menti...
research
10/22/2019

MRQA 2019 Shared Task: Evaluating Generalization in Reading Comprehension

We present the results of the Machine Reading for Question Answering (MR...
research
11/06/2016

Hierarchical Question Answering for Long Documents

We present a framework for question answering that can efficiently scale...
research
11/10/2018

Densely Connected Attention Propagation for Reading Comprehension

We propose DecaProp (Densely Connected Attention Propagation), a new den...
research
10/02/2019

AntMan: Sparse Low-Rank Compression to Accelerate RNN inference

Wide adoption of complex RNN based models is hindered by their inference...

Please sign up or login with your details

Forgot password? Click here to reset