A Decomposable Attention Model for Natural Language Inference

06/06/2016
by   Ankur P. Parikh, et al.
0

We propose a simple neural architecture for natural language inference. Our approach uses attention to decompose the problem into subproblems that can be solved separately, thus making it trivially parallelizable. On the Stanford Natural Language Inference (SNLI) dataset, we obtain state-of-the-art results with almost an order of magnitude fewer parameters than previous work and without relying on any word-order information. Adding intra-sentence attention that takes a minimum amount of order into account yields further improvements.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/21/2018

Stochastic Answer Networks for Natural Language Inference

We propose a stochastic answer network (SAN) to explore multi-step infer...
research
07/24/2017

Character-level Intra Attention Network for Natural Language Inference

Natural language inference (NLI) is a central problem in language unders...
research
11/15/2016

A Neural Architecture Mimicking Humans End-to-End for Natural Language Inference

In this work we use the recent advances in representation learning to pr...
research
08/04/2017

Recurrent Neural Network-Based Sentence Encoder with Gated Attention for Natural Language Inference

The RepEval 2017 Shared Task aims to evaluate natural language understan...
research
08/27/2018

Natural Language Inference with Hierarchical BiLSTM Max Pooling Architecture

Recurrent neural networks have proven to be very effective for natural l...
research
10/12/2022

Interactive Language: Talking to Robots in Real Time

We present a framework for building interactive, real-time, natural lang...
research
12/30/2017

A Compare-Propagate Architecture with Alignment Factorization for Natural Language Inference

This paper presents a new deep learning architecture for Natural Languag...

Please sign up or login with your details

Forgot password? Click here to reset