Discrete Word Embedding for Logical Natural Language Understanding

08/26/2020
by   Masataro Asai, et al.
0

In this paper, we propose an unsupervised neural model for learning a discrete embedding of words. While being discrete, our embedding supports vector arithmetic operations similar to continuous embeddings by interpreting each word as a set of propositional statements describing a rule. The formulation of our vector arithmetic closely reflects the logical structure originating from the symbolic sequential decision making formalism (classical/STRIPS planning). Contrary to the conventional wisdom that discrete representation cannot perform well due to the lack of ability to capture the uncertainty, our representation is competitive against the continuous representations in several downstream tasks. We demonstrate that our embedding is directly compatible with the symbolic, classical planning solvers by performing a "paraphrasing" task. Due to the discrete/logical decision making in classical algorithms with deterministic (non-probabilistic) completeness, and also because it does not require additional training on the paraphrasing dataset, our system can negatively answer a paraphrasing query (inexistence of solutions), and can answer that only some approximate solutions exist – A feature that is missing in the recent, huge, purely neural language models such as GPT-3.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/02/2023

Tsetlin Machine Embedding: Representing Words Using Logical Expressions

Embedding words in vector space is a fundamental first step in state-of-...
research
02/21/2019

Unsupervised Grounding of Plannable First-Order Logic Representation from Images

Recently, there is an increasing interest in obtaining the relational st...
research
07/19/2022

Uncertainty in Contrastive Learning: On the Predictability of Downstream Performance

The superior performance of some of today's state-of-the-art deep learni...
research
08/03/2023

Reasoning in Large Language Models Through Symbolic Math Word Problems

Large language models (LLMs) have revolutionized NLP by solving downstre...
research
05/26/2016

The Symbolic Interior Point Method

A recent trend in probabilistic inference emphasizes the codification of...
research
11/30/2020

Extreme Model Compression for On-device Natural Language Understanding

In this paper, we propose and experiment with techniques for extreme com...
research
07/05/2023

Deductive Additivity for Planning of Natural Language Proofs

Current natural language systems designed for multi-step claim validatio...

Please sign up or login with your details

Forgot password? Click here to reset