Auto-Learning: An Adversarial Process of Two Pre-trained Models for Natural Language Generation

02/08/2023
by   Zhengqing Yuan, et al.
0

Pre-trained models have been used in many fields in recent years, ranging from natural language understanding to computer vision and natural language generation. Nowadays, the performance of these natural language generation models is overly dependent on the model's scale and the dataset's size. While the larger language model is excellent in some respects, it cannot learn up-to-date knowledge and is relatively difficult to relearn. In this paper, a new adversarial process learning method is called Auto-Learning, which can improve the performance of any natural language generation model without the help of additional datasets. Auto-Learning includes two models: G is a text generation model, and D can test whether the data generated by G is legitimate. Firstly, the fine-tuned D model is used as the brain's knowledge base before the process. Then the text generated by the G model is used as the input of D to determine whether the text is legitimate. Finally, G is fine-tuned according to the output of D. This adversarial process is like a self-escalation of the brain through some a priori knowledge. When this adversarial system wants to learn something new, simply fine-tune the D model. Our approach applies to Autoregressive Language Modeling for all Transformer classes. Auto-Learning enables 8 models to achieve stable improvement in 10 natural language processing tasks without any change in structure.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/13/2021

CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation

In this paper, we take the advantage of previous pre-trained models (PTM...
research
02/24/2022

From Natural Language to Simulations: Applying GPT-3 Codex to Automate Simulation Modeling of Logistics Systems

Our work is the first attempt to apply Natural Language Processing to au...
research
12/08/2020

Facts2Story: Controlling Text Generation by Key Facts

Recent advancements in self-attention neural network architectures have ...
research
05/23/2022

What Makes Data-to-Text Generation Hard for Pretrained Language Models?

Expressing natural language descriptions of structured facts or relation...
research
08/01/2022

Learning from flowsheets: A generative transformer model for autocompletion of flowsheets

We propose a novel method enabling autocompletion of chemical flowsheets...
research
12/26/2022

Biologically Inspired Design Concept Generation Using Generative Pre-Trained Transformers

Biological systems in nature have evolved for millions of years to adapt...
research
05/30/2023

LLM-BRAIn: AI-driven Fast Generation of Robot Behaviour Tree based on Large Language Model

This paper presents a novel approach in autonomous robot control, named ...

Please sign up or login with your details

Forgot password? Click here to reset