Supervised Graph Contrastive Pretraining for Text Classification

12/21/2021
by   Samujjwal Ghosh, et al.
0

Contrastive pretraining techniques for text classification has been largely studied in an unsupervised setting. However, oftentimes labeled data from related tasks which share label semantics with current task is available. We hypothesize that using this labeled data effectively can lead to better generalization on current task. In this paper, we propose a novel way to effectively utilize labeled data from related tasks with a graph based supervised contrastive learning approach. We formulate a token-graph by extrapolating the supervised information from examples to tokens. Our formulation results in an embedding space where tokens with high/low probability of belonging to same class are near/further-away from one another. We also develop detailed theoretical insights which serve as a motivation for our method. In our experiments with 13 datasets, we show our method outperforms pretraining schemes by 2.5% and also example-level contrastive learning based formulation by 1.8% on average. In addition, we show cross-domain effectiveness of our method in a zero-shot setting by 3.91% on average. Lastly, we also demonstrate our method can be used as a noisy teacher in a knowledge distillation setting to significantly improve performance of transformer based models in low labeled data regime by 4.57% on average.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2023

PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification

We present PESCO, a novel contrastive learning framework that substantia...
research
05/16/2023

ContrastNet: A Contrastive Learning Framework for Few-Shot Text Classification

Few-shot text classification has recently been promoted by the meta-lear...
research
09/29/2022

Few-shot Text Classification with Dual Contrastive Consistency

In this paper, we explore how to utilize pre-trained language model to p...
research
09/02/2023

Contrastive Feature Masking Open-Vocabulary Vision Transformer

We present Contrastive Feature Masking Vision Transformer (CFM-ViT) - an...
research
10/02/2020

Long-Tail Zero and Few-Shot Learning via Contrastive Pretraining on and for Small Data

For natural language processing (NLP) tasks such as sentiment or topic c...
research
12/21/2022

Continual Contrastive Finetuning Improves Low-Resource Relation Extraction

Relation extraction (RE), which has relied on structurally annotated cor...
research
12/28/2020

Learning by Ignoring

Learning by ignoring, which identifies less important things and exclude...

Please sign up or login with your details

Forgot password? Click here to reset