Planning with Logical Graph-based Language Model for Instruction Generation

08/26/2023
by   Fan Zhang, et al.
0

Despite the superior performance of large language models to generate natural language texts, it is hard to generate texts with correct logic according to a given task, due to the difficulties for neural models to capture implied rules from free-form texts. In this paper, we propose a novel graph-based language model, Logical-GLM, to infuse logic into language models for more valid text generation and interpretability. Specifically, we first capture information from natural language instructions and construct logical bayes graphs that generally describe domains. Next, we generate logical skeletons to guide language model training, infusing domain knowledge into language models. Finally, we alternately optimize the searching policy of graphs and language models until convergence. The experimental results show that Logical-GLM is both effective and efficient compared with traditional language models, despite using smaller-scale training data and fewer parameters. Our approach can generate instructional texts with more correct logic owing to the internalized domain knowledge. Moreover, the usage of logical graphs reflects the inner mechanism of the language models, which improves the interpretability of black-box models.

READ FULL TEXT
research
04/27/2023

Controlled Text Generation with Natural Language Instructions

Large language models generate fluent texts and can follow natural langu...
research
12/10/2022

A Unified Knowledge Graph Service for Developing Domain Language Models in AI Software

Natural Language Processing (NLP) is one of the core techniques in AI so...
research
04/12/2023

Using large language models for (de-)formalization and natural argumentation exercises for beginner's students

We describe two systems that use text-davinci-003, a large language mode...
research
05/25/2023

RewriteLM: An Instruction-Tuned Large Language Model for Text Rewriting

Large Language Models (LLMs) have demonstrated impressive zero-shot capa...
research
04/02/2021

Humor@IITK at SemEval-2021 Task 7: Large Language Models for Quantifying Humor and Offensiveness

Humor and Offense are highly subjective due to multiple word senses, cul...
research
01/24/2022

Relational Memory Augmented Language Models

We present a memory-augmented approach to condition an autoregressive la...
research
10/28/2022

Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models

Fully-parametric language models generally require a huge number of mode...

Please sign up or login with your details

Forgot password? Click here to reset