Applying SoftTriple Loss for Supervised Language Model Fine Tuning

12/15/2021
by   Witold Sosnowski, et al.
0

We introduce a new loss function TripleEntropy, to improve classification performance for fine-tuning general knowledge pre-trained language models based on cross-entropy and SoftTriple loss. This loss function can improve the robust RoBERTa baseline model fine-tuned with cross-entropy loss by about (0.02 2.29 samples in the training dataset, the higher gain – thus, for small-sized dataset it is 0.78 extra-large 0.04

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/28/2022

Distance Metric Learning Loss Functions in Few-Shot Scenarios of Supervised Language Models Fine-Tuning

This paper presents an analysis regarding an influence of the Distance M...
research
11/12/2018

Fine-tuning of Language Models with Discriminator

Cross-entropy loss is a common choice when it comes to multiclass classi...
research
12/21/2020

LQF: Linear Quadratic Fine-Tuning

Classifiers that are linear in their parameters, and trained by optimizi...
research
06/15/2022

Differentiable Top-k Classification Learning

The top-k classification accuracy is one of the core metrics in machine ...
research
11/28/2022

Revisiting Distance Metric Learning for Few-Shot Natural Language Classification

Distance Metric Learning (DML) has attracted much attention in image pro...
research
02/02/2021

Scaling Laws for Transfer

We study empirical scaling laws for transfer learning between distributi...
research
10/15/2020

LiteDepthwiseNet: An Extreme Lightweight Network for Hyperspectral Image Classification

Deep learning methods have shown considerable potential for hyperspectra...

Please sign up or login with your details

Forgot password? Click here to reset