Semi-Supervised Sequence Modeling with Cross-View Training

09/22/2018
by   Kevin Clark, et al.
0

Unsupervised representation learning algorithms such as word2vec and ELMo improve the accuracy of many supervised NLP models, mainly because they can take advantage of large amounts of unlabeled text. However, the supervised models only learn from task-specific labeled data during the main training phase. We therefore propose Cross-View Training (CVT), a semi-supervised learning algorithm that improves the representations of a Bi-LSTM sentence encoder using a mix of labeled and unlabeled data. On labeled examples, standard supervised learning is used. On unlabeled examples, CVT teaches auxiliary prediction modules that see restricted views of the input (e.g., only part of a sentence) to match the predictions of the full model seeing the whole input. Since the auxiliary modules and the full model share intermediate representations, this in turn improves the full model. Moreover, we show that CVT is particularly effective when combined with multi-task learning. We evaluate CVT on five sequence tagging tasks, machine translation, and dependency parsing, achieving state-of-the-art results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/21/2018

Semi-Supervised Learning for Neural Keyphrase Generation

We study the problem of generating keyphrases that summarize the key poi...
research
08/04/2019

Semi-supervised Thai Sentence Segmentation Using Local and Distant Word Representations

A sentence is typically treated as the minimal syntactic unit used for e...
research
10/27/2020

To BERT or Not to BERT: Comparing Task-specific and Task-agnostic Semi-Supervised Approaches for Sequence Tagging

Leveraging large amounts of unlabeled data using Transformer-like archit...
research
11/08/2021

TAGLETS: A System for Automatic Semi-Supervised Learning with Auxiliary Data

Machine learning practitioners often have access to a spectrum of data: ...
research
10/16/2020

Auxiliary Task Reweighting for Minimum-data Learning

Supervised learning requires a large amount of training data, limiting i...
research
08/25/2022

Fix-A-Step: Effective Semi-supervised Learning from Uncurated Unlabeled Sets

Semi-supervised learning (SSL) promises gains in accuracy compared to tr...

Please sign up or login with your details

Forgot password? Click here to reset