Guessing What's Plausible But Remembering What's True: Accurate Neural Reasoning for Question-Answering

04/07/2020
by   Haitian Sun, et al.
7

Neural approaches to natural language processing (NLP) often fail at the logical reasoning needed for deeper language understanding. In particular, neural approaches to reasoning that rely on embedded generalizations of a knowledge base (KB) implicitly model which facts that are plausible, but may not model which facts are true, according to the KB. While generalizing the facts in a KB is useful for KB completion, the inability to distinguish between plausible inferences and logically entailed conclusions can be problematic in settings like as KB question answering (KBQA). We propose here a novel KB embedding scheme that supports generalization, but also allows accurate logical reasoning with a KB. Our approach introduces two new mechanisms for KB reasoning: neural retrieval over a set of embedded triples, and "memorization" of highly specific information with a compact sketch structure. Experimentally, this leads to substantial improvements over the state-of-the-art on two KBQA benchmarks.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset