Less is More: Summary of Long Instructions is Better for Program Synthesis

03/16/2022
by   Kirby Kuznia, et al.
2

Despite the success of large pre-trained language models (LMs) such as Codex, they show below-par performance on the larger and more complicated programming related questions. We show that LMs benefit from the summarized version of complicated questions. Our findings show that superfluous information often present in problem description such as human characters, background stories, names (which are included to help humans in understanding a task) does not help models in understanding a task. To this extent, we create a meta-dataset from the frequently used APPS dataset for the program synthesis task. Our meta-dataset consists of human and synthesized summary of the long and complicated programming questions. Experimental results on Codex show that our proposed approach outperforms baseline by 8.13 strict accuracy. Our analysis shows that summary significantly improve performance for introductory (9.86 questions. However, it shows improvement by a small margin ( 2 competitive programming questions, implying the scope for future research direction.

READ FULL TEXT

page 12

page 18

page 19

research
07/06/2018

NAPS: Natural Program Synthesis Dataset

We present a program synthesis-oriented dataset consisting of human writ...
research
06/11/2021

TellMeWhy: A Dataset for Answering Why-Questions in Narratives

Answering questions about why characters perform certain actions is cent...
research
08/04/2023

ChatGPT for GTFS: From Words to Information

The General Transit Feed Specification (GTFS) standard for publishing tr...
research
11/02/2021

Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks

Natural language understanding (NLU) has made massive progress driven by...
research
01/17/2023

Are Language Models Worse than Humans at Following Prompts? It's Complicated

Prompts have been the center of progress in advancing language models' z...
research
07/24/2023

A Real-World WebAgent with Planning, Long Context Understanding, and Program Synthesis

Pre-trained large language models (LLMs) have recently achieved better g...

Please sign up or login with your details

Forgot password? Click here to reset