Self-Training with Weak Supervision

04/12/2021
by   Giannis Karamanolakis, et al.
81

State-of-the-art deep neural networks require large-scale labeled training data that is often expensive to obtain or not available for many tasks. Weak supervision in the form of domain-specific rules has been shown to be useful in such settings to automatically generate weakly labeled training data. However, learning with weak rules is challenging due to their inherent heuristic and noisy nature. An additional challenge is rule coverage and overlap, where prior work on weak supervision only considers instances that are covered by weak rules, thus leaving valuable unlabeled data behind. In this work, we develop a weak supervision framework (ASTRA) that leverages all the available data for a given task. To this end, we leverage task-specific unlabeled data through self-training with a model (student) that considers contextualized representations and predicts pseudo-labels for instances that may not be covered by weak rules. We further develop a rule attention network (teacher) that learns how to aggregate student pseudo-labels with weak rule labels, conditioned on their fidelity and the underlying context of an instance. Finally, we construct a semi-supervised learning objective for end-to-end training with unlabeled data, domain-specific rules, and a small amount of labeled data. Extensive experiments on six benchmark datasets for text classification demonstrate the effectiveness of our approach with significant improvements over state-of-the-art baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/18/2023

Enhancing Self-Training Methods

Semi-supervised learning approaches train on small sets of labeled data ...
research
05/18/2022

Automatic Rule Induction for Efficient Semi-Supervised Learning

Semi-supervised learning has shown promise in allowing NLP models to gen...
research
04/13/2020

Learning from Rules Generalizing Labeled Exemplars

In many applications labeled data is not readily available, and needs to...
research
10/09/2020

Denoising Multi-Source Weak Supervision for Neural Text Classification

We study the problem of learning neural text classifiers without using a...
research
11/01/2017

Avoiding Your Teacher's Mistakes: Training Neural Networks with Controlled Weak Supervision

Training deep neural networks requires massive amounts of training data,...
research
08/01/2021

SPEAR : Semi-supervised Data Programming in Python

We present SPEAR, an open-source python library for data programming wit...
research
12/09/2021

How Universal is Genre in Universal Dependencies?

This work provides the first in-depth analysis of genre in Universal Dep...

Please sign up or login with your details

Forgot password? Click here to reset