Synapse: Leveraging Few-Shot Exemplars for Human-Level Computer Control

06/13/2023
by   Longtao Zheng, et al.
0

This paper investigates the design of few-shot exemplars for computer automation through prompting large language models (LLMs). While previous prompting approaches focus on self-correction, we find that well-structured exemplars alone are sufficient for human-level performance. We present Synapse, an in-context computer control agent demonstrating human-level performance on the MiniWob++ benchmark. Synapse consists of three main components: 1) state-conditional decomposition, which divides demonstrations into exemplar sets based on the agent's need for new environment states, enabling temporal abstraction; 2) structured prompting, which filters states and reformulates task descriptions for each set to improve planning correctness; and 3) exemplar retrieval, which associates incoming tasks with corresponding exemplars in an exemplar database for multi-task adaptation and generalization. Synapse overcomes context length limits, reduces errors in multi-step control, and allows for more exemplars within the context. Importantly, Synapse complements existing prompting approaches that enhance LLMs' reasoning and planning abilities. Synapse outperforms previous methods, including behavioral cloning, reinforcement learning, finetuning, and prompting, with an average success rate of 98.5% across 63 tasks in MiniWob++. Notably, Synapse relies on exemplars from only 47 tasks, demonstrating effective generalization to novel tasks. Our results highlight the potential of in-context learning to advance the integration of LLMs into practical tool automation.

READ FULL TEXT

page 2

page 6

page 7

page 16

page 17

page 18

page 19

research
03/30/2023

Language Models can Solve Computer Tasks

Agents capable of carrying out general tasks on a computer can improve e...
research
04/22/2023

Boosting Theory-of-Mind Performance in Large Language Models via Prompting

Large language models (LLMs) excel in many tasks in 2023, but they still...
research
05/22/2023

Meta-in-context learning in large language models

Large language models have shown tremendous performance in a variety of ...
research
05/03/2023

Visual Transformation Telling

In this paper, we propose a new visual reasoning task, called Visual Tra...
research
04/16/2019

Object-Oriented Dynamics Learning through Multi-Level Abstraction

Object-based approaches for learning action-conditioned dynamics has dem...
research
02/16/2022

A data-driven approach for learning to control computers

It would be useful for machines to use computers as humans do so that th...

Please sign up or login with your details

Forgot password? Click here to reset