Evolution through Large Models

06/17/2022
by   Joel Lehman, et al.
6

This paper pursues the insight that large language models (LLMs) trained to generate code can vastly improve the effectiveness of mutation operators applied to programs in genetic programming (GP). Because such LLMs benefit from training data that includes sequential changes and modifications, they can approximate likely changes that humans would make. To highlight the breadth of implications of such evolution through large models (ELM), in the main experiment ELM combined with MAP-Elites generates hundreds of thousands of functional examples of Python programs that output working ambulating robots in the Sodarace domain, which the original LLM had never seen in pre-training. These examples then help to bootstrap training a new conditional language model that can output the right walker for a particular terrain. The ability to bootstrap new models that can output appropriate artifacts for a given context in a domain where zero training data was previously available carries implications for open-endedness, deep learning, and reinforcement learning. These implications are explored here in depth in the hope of inspiring new directions of research now opened up by ELM.

READ FULL TEXT

page 19

page 20

page 22

page 24

page 28

page 32

page 33

research
12/14/2020

Extracting Training Data from Large Language Models

It has become common to publish large (billion parameter) language model...
research
03/13/2023

Meet in the Middle: A New Pre-training Paradigm

Most language models (LMs) are trained and applied in an autoregressive ...
research
02/15/2023

Learning Performance-Improving Code Edits

The waning of Moore's Law has shifted the focus of the tech industry tow...
research
06/13/2023

Large Language Models Sometimes Generate Purely Negatively-Reinforced Text

When using adversarial training, it is common practice to train against ...
research
11/28/2022

CoNAL: Anticipating Outliers with Large Language Models

In many task settings, text classification models are likely to encounte...
research
04/19/2023

Towards Objective-Tailored Genetic Improvement Through Large Language Models

While Genetic Improvement (GI) is a useful paradigm to improve functiona...
research
05/29/2023

Controllable Path of Destruction

Path of Destruction (PoD) is a self-supervised method for learning itera...

Please sign up or login with your details

Forgot password? Click here to reset