WC-SBERT: Zero-Shot Text Classification via SBERT with Self-Training for Wikipedia Categories

07/28/2023
by   Te-Yu Chi, et al.
0

Our research focuses on solving the zero-shot text classification problem in NLP, with a particular emphasis on innovative self-training strategies. To achieve this objective, we propose a novel self-training strategy that uses labels rather than text for training, significantly reducing the model's training time. Specifically, we use categories from Wikipedia as our training set and leverage the SBERT pre-trained model to establish positive correlations between pairs of categories within the same text, facilitating associative training. For new test datasets, we have improved the original self-training approach, eliminating the need for prior training and testing data from each target dataset. Instead, we adopt Wikipedia as a unified training dataset to better approximate the zero-shot scenario. This modification allows for rapid fine-tuning and inference across different datasets, greatly reducing the time required for self-training. Our experimental results demonstrate that this method can adapt the model to the target dataset within minutes. Compared to other BERT-based transformer models, our approach significantly reduces the amount of training data by training only on labels, not the actual text, and greatly improves training efficiency by utilizing a unified training set. Additionally, our method achieves state-of-the-art results on both the Yahoo Topic and AG News datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2023

PESCO: Prompt-enhanced Self Contrastive Learning for Zero-shot Text Classification

We present PESCO, a novel contrastive learning framework that substantia...
research
05/25/2023

Label Agnostic Pre-training for Zero-shot Text Classification

Conventional approaches to text classification typically assume the exis...
research
12/16/2021

Extreme Zero-Shot Learning for Extreme Text Classification

The eXtreme Multi-label text Classification (XMC) problem concerns findi...
research
04/08/2022

Infusing Knowledge from Wikipedia to Enhance Stance Detection

Stance detection infers a text author's attitude towards a target. This ...
research
04/24/2023

Generation-driven Contrastive Self-training for Zero-shot Text Classification with Instruction-tuned GPT

Moreover, GPT-based zero-shot classification models tend to make indepen...
research
05/31/2022

Few-Shot Unlearning by Model Inversion

We consider the problem of machine unlearning to erase a target dataset,...
research
05/12/2021

Priberam Labs at the NTCIR-15 SHINRA2020-ML: Classification Task

Wikipedia is an online encyclopedia available in 285 languages. It compo...

Please sign up or login with your details

Forgot password? Click here to reset