Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

06/24/2015
by   Ankit Kumar, et al.
0

Most tasks in natural language processing can be cast into question answering (QA) problems over language input. We introduce the dynamic memory network (DMN), a neural network architecture which processes input sequences and questions, forms episodic memories, and generates relevant answers. Questions trigger an iterative attention process which allows the model to condition its attention on the inputs and the result of previous iterations. These results are then reasoned over in a hierarchical recurrent sequence model to generate answers. The DMN can be trained end-to-end and obtains state-of-the-art results on several types of tasks and datasets: question answering (Facebook's bAbI dataset), text classification for sentiment analysis (Stanford Sentiment Treebank) and sequence modeling for part-of-speech tagging (WSJ-PTB). The training for these different tasks relies exclusively on trained word vector representations and input-question-answer triplets.

READ FULL TEXT
research
07/06/2017

Long-Term Memory Networks for Question Answering

Question answering is an important and difficult task in the natural lan...
research
03/11/2017

Ask Me Even More: Dynamic Memory Tensor Networks (Extended Model)

We examine Memory Networks for the task of question answering (QA), unde...
research
10/11/2020

End to End Binarized Neural Networks for Text Classification

Deep neural networks have demonstrated their superior performance in alm...
research
02/01/2018

Adaptive Memory Networks

We present Adaptive Memory Networks (AMN) that processes input-question ...
research
05/31/2023

Attention-Based Methods For Audio Question Answering

Audio question answering (AQA) is the task of producing natural language...
research
07/14/2016

Using Recurrent Neural Network for Learning Expressive Ontologies

Recently, Neural Networks have been proven extremely effective in many n...
research
03/31/2015

End-To-End Memory Networks

We introduce a neural network with a recurrent attention model over a po...

Please sign up or login with your details

Forgot password? Click here to reset