Learning to Compose over Tree Structures via POS Tags

08/18/2018
by   Gehui Shen, et al.
0

Recursive Neural Network (RecNN), a type of models which compose words or phrases recursively over syntactic tree structures, has been proven to have superior ability to obtain sentence representation for a variety of NLP tasks. However, RecNN is born with a thorny problem that a shared compositional function for each node of trees can't capture the complex semantic compositionality so that the expressive power of model is limited. In this paper, in order to address this problem, we propose Tag-Guided HyperRecNN/TreeLSTM (TG-HRecNN/TreeLSTM), which introduces hypernetwork into RecNNs to take as inputs Part-of-Speech (POS) tags of word/phrase and generate the semantic composition parameters dynamically. Experimental results on five datasets for two typical NLP tasks show proposed models both obtain significant improvement compared with RecNN and TreeLSTM consistently. Our TG-HTreeLSTM outperforms all existing RecNN-based models and achieves or is competitive with state-of-the-art on four sentence classification benchmarks. The effectiveness of our models is also demonstrated by qualitative analysis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/07/2018

Dynamic Compositionality in Recursive Neural Networks with Structure-aware Tag Representations

Most existing recursive neural network (RvNN) architectures utilize only...
research
05/11/2017

Dynamic Compositional Neural Networks over Tree Structure

Tree-structured neural networks have proven to be effective in learning ...
research
11/06/2018

Learning to Embed Sentences Using Attentive Recursive Trees

Sentence embedding is an effective feature representation for most deep ...
research
04/15/2020

Coreferential Reasoning Learning for Language Representation

Language representation models such as BERT could effectively capture co...
research
02/18/2020

An enhanced Tree-LSTM architecture for sentence semantic modeling using typed dependencies

Tree-based Long short term memory (LSTM) network has become state-of-the...
research
09/18/2018

Analysis of Bag-of-n-grams Representation's Properties Based on Textual Reconstruction

Despite its simplicity, bag-of-n-grams sen- tence representation has bee...
research
11/02/2020

Learning from Non-Binary Constituency Trees via Tensor Decomposition

Processing sentence constituency trees in binarised form is a common and...

Please sign up or login with your details

Forgot password? Click here to reset