Heads-up! Unsupervised Constituency Parsing via Self-Attention Heads

10/19/2020
by   Bowen Li, et al.
7

Transformer-based pre-trained language models (PLMs) have dramatically improved the state of the art in NLP across many tasks. This has led to substantial interest in analyzing the syntactic knowledge PLMs learn. Previous approaches to this question have been limited, mostly using test suites or probes. Here, we propose a novel fully unsupervised parsing approach that extracts constituency trees from PLM attention heads. We rank transformer attention heads based on their inherent properties, and create an ensemble of high-ranking heads to produce the final tree. Our method is adaptable to low-resource languages, as it does not rely on development sets, which can be expensive to annotate. Our experiments show that the proposed method often outperform existing approaches if there is no development set present. Our unsupervised parser can also be used as a tool to analyze the grammars PLMs learn implicitly. For this, we use the parse trees induced by our method to train a neural PCFG and compare it to a grammar derived from a human-annotated treebank.

READ FULL TEXT

page 14

page 15

research
06/10/2022

Unsupervised and Few-shot Parsing from Pretrained Language Models

Pretrained language models are generally acknowledged to be able to enco...
research
04/18/2015

Unsupervised Dependency Parsing: Let's Use Supervised Parsers

We present a self-training approach to unsupervised dependency parsing t...
research
10/04/2020

A Survey of Unsupervised Dependency Parsing

Syntactic dependency parsing is an important task in natural language pr...
research
09/21/2021

Something Old, Something New: Grammar-based CCG Parsing with Transformer Models

This report describes the parsing problem for Combinatory Categorial Gra...
research
06/01/2023

Contextual Distortion Reveals Constituency: Masked Language Models are Implicit Parsers

Recent advancements in pre-trained language models (PLMs) have demonstra...
research
10/21/2022

Syntax-guided Localized Self-attention by Constituency Syntactic Distance

Recent works have revealed that Transformers are implicitly learning the...
research
05/04/2023

G-MATT: Single-step Retrosynthesis Prediction using Molecular Grammar Tree Transformer

Various template-based and template-free approaches have been proposed f...

Please sign up or login with your details

Forgot password? Click here to reset