On Semi-Supervised Multiple Representation Behavior Learning

10/21/2019
by   Ruqian Lu, et al.
0

We propose a novel paradigm of semi-supervised learning (SSL)–the semi-supervised multiple representation behavior learning (SSMRBL). SSMRBL aims to tackle the difficulty of learning a grammar for natural language parsing where the data are natural language texts and the 'labels' for marking data are parsing trees and/or grammar rule pieces. We call such 'labels' as compound structured labels which require a hard work for training. SSMRBL is an incremental learning process that can learn more than one representation, which is an appropriate solution for dealing with the scarce of labeled training data in the age of big data and with the heavy workload of learning compound structured labels. We also present a typical example of SSMRBL, regarding behavior learning in form of a grammatical approach towards domain-based multiple text summarization (DBMTS). DBMTS works under the framework of rhetorical structure theory (RST). SSMRBL includes two representations: text embedding (for representing information contained in the texts) and grammar model (for representing parsing as a behavior). The first representation was learned as embedded digital vectors called impacts in a low dimensional space. The grammar model was learned in an iterative way. Then an automatic domain-oriented multi-text summarization approach was proposed based on the two representations discussed above. Experimental results on large-scale Chinese dataset SogouCA indicate that the proposed method brings a good performance even if only few labeled texts are used for training with respect to our defined automated metrics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/03/2019

Attributed Rhetorical Structure Grammar for Domain Text Summarization

This paper presents a new approach of automatic text summarization which...
research
01/05/2016

Joint learning of ontology and semantic parser from text

Semantic parsing methods are used for capturing and representing semanti...
research
10/04/2018

Semi-Supervised Methods for Out-of-Domain Dependency Parsing

Dependency parsing is one of the important natural language processing t...
research
10/14/2019

Knowledge-guided Unsupervised Rhetorical Parsing for Text Summarization

Automatic text summarization (ATS) has recently achieved impressive perf...
research
08/02/2015

PTE: Predictive Text Embedding through Large-scale Heterogeneous Text Networks

Unsupervised text embedding methods, such as Skip-gram and Paragraph Vec...
research
06/03/2019

Jointly Learning Semantic Parser and Natural Language Generator via Dual Information Maximization

Semantic parsing aims to transform natural language (NL) utterances into...
research
07/19/2011

Towards Open-Text Semantic Parsing via Multi-Task Learning of Structured Embeddings

Open-text (or open-domain) semantic parsers are designed to interpret an...

Please sign up or login with your details

Forgot password? Click here to reset