Learning to Organize Knowledge with N-Gram Machines

11/17/2017
by   Fan Yang, et al.
0

Deep neural networks (DNNs) had great success on NLP tasks such as language modeling, machine translation and certain question answering (QA) tasks. However, the success is limited at more knowledge intensive tasks such as QA from a big corpus. Existing end-to-end deep QA models (Miller et al., 2016; Weston et al., 2014) need to read the entire text after observing the question, and therefore their complexity in responding a question is linear in the text size. This is prohibitive for practical tasks such as QA from Wikipedia, a novel, or the Web. We propose to solve this scalability issue by using symbolic meaning representations, which can be indexed and retrieved efficiently with complexity that is independent of the text size. More specifically, we use sequence-to-sequence models to encode knowledge symbolically and generate programs to answer questions from the encoded knowledge. We apply our approach, called the N-Gram Machine (NGM), to the bAbI tasks (Weston et al., 2015) and a special version of them ("life-long bAbI") which has stories of up to 10 million sentences. Our experiments show that NGM can successfully solve both of these tasks accurately and efficiently. Unlike fully differentiable memory models, NGM's time complexity and answering quality are not affected by the story length. The whole system of NGM is trained end-to-end with REINFORCE (Williams, 1992). To avoid high variance in gradient estimation, which is typical in discrete latent variable models, we use beam search instead of sampling. To tackle the exponentially large search space, we use a stabilized auto-encoding objective and a structure tweak procedure to iteratively reduce and refine the search space.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/02/2020

BERT-kNN: Adding a kNN Search Component to Pretrained Language Models for Better QA

Khandelwal et al. (2020) show that a k-nearest-neighbor (kNN) component ...
research
11/19/2015

Skip-Thought Memory Networks

Question Answering (QA) is fundamental to natural language processing in...
research
11/12/2016

Leveraging Video Descriptions to Learn Video Question Answering

We propose a scalable approach to learn video-based question answering (...
research
10/26/2015

Empirical Study on Deep Learning Models for Question Answering

In this paper we explore deep learning models with memory component or a...
research
06/29/2020

Answering Questions on COVID-19 in Real-Time

The recent outbreak of the novel coronavirus is wreaking havoc on the wo...
research
03/31/2015

End-To-End Memory Networks

We introduce a neural network with a recurrent attention model over a po...
research
06/11/2022

Can the Language of the Collation be Translated into the Language of the Stemma? Using Machine Translation for Witness Localization

Stemmatology is a subfield of philology where one approach to understand...

Please sign up or login with your details

Forgot password? Click here to reset