Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self-Training Approach

10/15/2020
by   Yue Yu, et al.
0

Fine-tuned pre-trained language models (LMs) achieve enormous success in many natural language processing (NLP) tasks, but they still require excessive labeled data in the fine-tuning stage. We study the problem of fine-tuning pre-trained LMs using only weak supervision, without any labeled data. This problem is challenging because the high capacity of LMs makes them prone to overfitting the noisy labels generated by weak supervision. To address this problem, we develop a contrastive self-training framework, COSINE, to enable fine-tuning LMs with weak supervision. Underpinned by contrastive regularization and confidence-based reweighting, this contrastive self-training framework can gradually improve model fitting while effectively suppressing error propagation. Experiments on sequence, token, and sentence pair classification tasks show that our model outperforms the strongest baseline by large margins on 7 benchmarks in 6 tasks, and achieves competitive performance with fully-supervised fine-tuning methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2021

Improved Regularization and Robustness for Fine-tuning in Neural Networks

A widely used algorithm for transfer learning is fine-tuning, where a pr...
research
10/29/2022

Differentiable Data Augmentation for Contrastive Sentence Representation Learning

Fine-tuning a pre-trained language model via the contrastive learning fr...
research
06/04/2021

Bi-Granularity Contrastive Learning for Post-Training in Few-Shot Scene

The major paradigm of applying a pre-trained language model to downstrea...
research
09/21/2023

Audio Contrastive based Fine-tuning

Audio classification plays a crucial role in speech and sound processing...
research
05/12/2022

SimCPSR: Simple Contrastive Learning for Paper Submission Recommendation System

The recommendation system plays a vital role in many areas, especially a...
research
10/11/2022

Contrastive Trajectory Similarity Learning with Dual-Feature Attention

Trajectory similarity measures act as query predicates in trajectory dat...
research
06/26/2020

Train and You'll Miss It: Interactive Model Iteration with Weak Supervision and Pre-Trained Embeddings

Our goal is to enable machine learning systems to be trained interactive...

Please sign up or login with your details

Forgot password? Click here to reset