LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Models

12/08/2022
by   Chan Hee Song, et al.
0

This study focuses on embodied agents that can follow natural language instructions to complete complex tasks in a visually-perceived environment. Existing methods rely on a large amount of (instruction, gold trajectory) pairs to learn a good policy. The high data cost and poor sample efficiency prevents the development of versatile agents that are capable of many tasks and can learn new tasks quickly. In this work, we propose a novel method, LLM-Planner, that harnesses the power of large language models (LLMs) such as GPT-3 to do few-shot planning for embodied agents. We further propose a simple but effective way to enhance LLMs with physical grounding to generate plans that are grounded in the current environment. Experiments on the ALFRED dataset show that our method can achieve very competitive few-shot performance, even outperforming several recent baselines that are trained using the full training data despite using less than 0.5 barely complete any task successfully under the same few-shot setting. Our work opens the door for developing versatile and sample-efficient embodied agents that can quickly learn many tasks.

READ FULL TEXT

page 1

page 3

page 7

page 8

research
09/29/2020

Visually-Grounded Planning without Vision: Language Models Infer Detailed Plans from High-level Instructions

The recently proposed ALFRED challenge task aims for a virtual robotic a...
research
07/04/2023

Embodied Task Planning with Large Language Models

Equipping embodied agents with commonsense is important for robots to su...
research
07/12/2023

SayPlan: Grounding Large Language Models using 3D Scene Graphs for Scalable Task Planning

Large language models (LLMs) have demonstrated impressive results in dev...
research
05/09/2021

gComm: An environment for investigating generalization in Grounded Language Acquisition

gComm is a step towards developing a robust platform to foster research ...
research
08/11/2023

Dynamic Planning with a LLM

While Large Language Models (LLMs) can solve many NLP tasks in zero-shot...
research
09/16/2021

Hierarchical Control of Situated Agents through Natural Language

When humans conceive how to perform a particular task, they do so hierar...
research
01/29/2023

Distilling Internet-Scale Vision-Language Models into Embodied Agents

Instruction-following agents must ground language into their observation...

Please sign up or login with your details

Forgot password? Click here to reset