proScript: Partially Ordered Scripts Generation via Pre-trained Language Models

04/16/2021
by   Keisuke Sakaguchi, et al.
0

Scripts - standardized event sequences describing typical everyday activities - have been shown to help understand narratives by providing expectations, resolving ambiguity, and filling in unstated information. However, to date they have proved hard to author or extract from text. In this work, we demonstrate for the first time that pre-trained neural language models (LMs) can be be finetuned to generate high-quality scripts, at varying levels of granularity, for a wide range of everyday scenarios (e.g., bake a cake). To do this, we collected a large (6.4k), crowdsourced partially ordered scripts (named proScript), which is substantially larger than prior datasets, and developed models that generate scripts with combining language generation and structure prediction. We define two complementary tasks: (i) edge prediction: given a scenario and unordered events, organize the events into a valid (possibly partial-order) script, and (ii) script generation: given only a scenario, generate events and organize them into a (possibly partial-order) script. Our experiments show that our models perform well (e.g., F1=75.7 in task (i)), illustrating a new approach to overcoming previous barriers to script collection. We also show that there is still significant room for improvement toward human level performance. Together, our tasks, dataset, and models offer a new research direction for learning script knowledge.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/27/2021

What do Large Language Models Learn about Scripts?

Script Knowledge (Schank and Abelson, 1975) has long been recognized as ...
research
09/21/2023

Choice-75: A Dataset on Decision Branching in Script Learning

Script learning studies how daily events unfold. Previous works tend to ...
research
04/16/2022

Probing Script Knowledge from Pre-Trained Models

Script knowledge is critical for humans to understand the broad daily ta...
research
12/31/2020

Conditional Generation of Temporally-ordered Event Sequences

Models encapsulating narrative schema knowledge have proven to be useful...
research
04/15/2021

Time-Stamped Language Model: Teaching Language Models to Understand the Flow of Events

Tracking entities throughout a procedure described in a text is challeng...
research
10/06/2020

Modeling Preconditions in Text with a Crowd-sourced Dataset

Preconditions provide a form of logical connection between events that e...
research
01/31/2021

Classification Models for Partially Ordered Sequences

Many models such as Long Short Term Memory (LSTMs), Gated Recurrent Unit...

Please sign up or login with your details

Forgot password? Click here to reset