When Are Tree Structures Necessary for Deep Learning of Representations?

02/28/2015
by   Jiwei Li, et al.
0

Recursive neural models, which use syntactic parse trees to recursively generate representations bottom-up, are a popular architecture. But there have not been rigorous evaluations showing for exactly which tasks this syntax-based method is appropriate. In this paper we benchmark recursive neural models against sequential recurrent neural models (simple recurrent and LSTM models), enforcing apples-to-apples comparison as much as possible. We investigate 4 tasks: (1) sentiment classification at the sentence level and phrase level; (2) matching questions to answer-phrases; (3) discourse parsing; (4) semantic relation extraction (e.g., component-whole between nouns). Our goal is to understand better when, and why, recursive models can outperform simpler models. We find that recursive models help mainly on tasks (like semantic relation extraction) that require associating headwords across a long distance, particularly on very long sequences. We then introduce a method for allowing recurrent models to achieve similar performance: breaking long sentences into clause-like units at punctuation and processing them separately before combining. Our results thus help understand the limitations of both classes of models, and suggest directions for improving recurrent models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/18/2015

Combining Neural Networks and Log-linear Models to Improve Relation Extraction

The last decade has witnessed the success of the traditional feature-bas...
research
07/15/2016

Neural Tree Indexers for Text Understanding

Recurrent neural networks (RNNs) process input text sequentially and mod...
research
09/07/2018

Dynamic Compositionality in Recursive Neural Networks with Structure-aware Tag Representations

Most existing recursive neural network (RvNN) architectures utilize only...
research
04/10/2019

Simple BERT Models for Relation Extraction and Semantic Role Labeling

We present simple BERT-based models for relation extraction and semantic...
research
12/04/2019

Enhancing Relation Extraction Using Syntactic Indicators and Sentential Contexts

State-of-the-art methods for relation extraction consider the sentential...
research
01/01/2021

MrGCN: Mirror Graph Convolution Network for Relation Extraction with Long-Term Dependencies

The ability to capture complex linguistic structures and long-term depen...
research
05/14/2019

Correlating neural and symbolic representations of language

Analysis methods which enable us to better understand the representation...

Please sign up or login with your details

Forgot password? Click here to reset