Uncertainty-aware Self-training for Text Classification with Few Labels

06/27/2020
by   Subhabrata Mukherjee, et al.
0

Recent success of large-scale pre-trained language models crucially hinge on fine-tuning them on large amounts of labeled data for the downstream task, that are typically expensive to acquire. In this work, we study self-training as one of the earliest semi-supervised learning approaches to reduce the annotation bottleneck by making use of large-scale unlabeled data for the target task. Standard self-training mechanism randomly samples instances from the unlabeled pool to pseudo-label and augment labeled data. In this work, we propose an approach to improve self-training by incorporating uncertainty estimates of the underlying neural network leveraging recent advances in Bayesian deep learning. Specifically, we propose (i) acquisition functions to select instances from the unlabeled pool leveraging Monte Carlo (MC) Dropout, and (ii) learning mechanism leveraging model confidence for self-training. As an application, we focus on text classification on five benchmark datasets. We show our methods leveraging only 20-30 labeled samples per class for each task for training and for validation can perform within 3 models fine-tuned on thousands of labeled instances with an aggregate accuracy of 91

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2021

Self-Tuning for Data-Efficient Deep Learning

Deep learning has made revolutionary advances to diverse applications in...
research
12/16/2021

ATM: An Uncertainty-aware Active Self-training Framework for Label-efficient Text Classification

Despite the great success of pre-trained language models (LMs) in many n...
research
04/17/2018

Reinforced Co-Training

Co-training is a popular semi-supervised learning framework to utilize a...
research
01/30/2019

Learning Fast Matching Models from Weak Annotations

This paper proposes a novel training scheme for fast matching models in ...
research
07/03/2022

NP-Match: When Neural Processes meet Semi-Supervised Learning

Semi-supervised learning (SSL) has been widely explored in recent years,...
research
02/17/2023

Uncertainty-aware Self-training for Low-resource Neural Sequence Labeling

Neural sequence labeling (NSL) aims at assigning labels for input langua...
research
06/13/2023

Rank-Aware Negative Training for Semi-Supervised Text Classification

Semi-supervised text classification-based paradigms (SSTC) typically emp...

Please sign up or login with your details

Forgot password? Click here to reset