Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot Text Classification Tasks

05/22/2023
by   Haoqi Zheng, et al.
0

Text classification tasks often encounter few shot scenarios with limited labeled data, and addressing data scarcity is crucial. Data augmentation with mixup has shown to be effective on various text classification tasks. However, most of the mixup methods do not consider the varying degree of learning difficulty in different stages of training and generate new samples with one hot labels, resulting in the model over confidence. In this paper, we propose a self evolution learning (SE) based mixup approach for data augmentation in text classification, which can generate more adaptive and model friendly pesudo samples for the model training. SE focuses on the variation of the model's learning ability. To alleviate the model confidence, we introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up. Through experimental analysis, in addition to improving classification accuracy, we demonstrate that SE also enhances the model's generalize ability.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2021

ALP: Data Augmentation using Lexicalized PCFGs for Few-Shot Text Classification

Data augmentation has been an important ingredient for boosting performa...
research
02/28/2022

Text Smoothing: Enhance Various Data Augmentation Methods on Text Classification Tasks

Before entering the neural network, a token is generally converted to th...
research
11/08/2019

Not Enough Data? Deep Learning to the Rescue!

Based on recent advances in natural language modeling and those in text ...
research
04/29/2021

Recognition and Processing of NATOM

In this paper we show how to process the NOTAM (Notice to Airmen) data o...
research
09/09/2022

Improving Model Training via Self-learned Label Representations

Modern neural network architectures have shown remarkable success in sev...
research
02/13/2023

Bag of Tricks for In-Distribution Calibration of Pretrained Transformers

While pre-trained language models (PLMs) have become a de-facto standard...
research
02/26/2022

Automated Data Augmentations for Graph Classification

Data augmentations are effective in improving the invariance of learning...

Please sign up or login with your details

Forgot password? Click here to reset