Differentiable Top-k Classification Learning

06/15/2022
by   Felix Petersen, et al.
0

The top-k classification accuracy is one of the core metrics in machine learning. Here, k is conventionally a positive integer, such as 1 or 5, leading to top-1 or top-5 training objectives. In this work, we relax this assumption and optimize the model for multiple k simultaneously instead of using a single k. Leveraging recent advances in differentiable sorting and ranking, we propose a differentiable top-k cross-entropy classification loss. This allows training the network while not only considering the top-1 prediction, but also, e.g., the top-2 and top-5 predictions. We evaluate the proposed loss function for fine-tuning on state-of-the-art architectures, as well as for training from scratch. We find that relaxing k does not only produce better top-5 accuracies, but also leads to top-1 accuracy improvements. When fine-tuning publicly available ImageNet models, we achieve a new state-of-the-art for these models.

READ FULL TEXT

page 1

page 5

page 6

page 7

page 8

page 9

page 10

page 12

research
12/15/2021

Applying SoftTriple Loss for Supervised Language Model Fine Tuning

We introduce a new loss function TripleEntropy, to improve classificatio...
research
08/28/2023

When hard negative sampling meets supervised contrastive learning

State-of-the-art image models predominantly follow a two-stage strategy:...
research
05/09/2020

Domain-specific loss design for unsupervised physical training: A new approach to modeling medical ML solutions

Today, cataract surgery is the most frequently performed ophthalmic surg...
research
05/27/2016

SNN: Stacked Neural Networks

It has been proven that transfer learning provides an easy way to achiev...
research
11/28/2022

Distance Metric Learning Loss Functions in Few-Shot Scenarios of Supervised Language Models Fine-Tuning

This paper presents an analysis regarding an influence of the Distance M...
research
10/17/2022

ZooD: Exploiting Model Zoo for Out-of-Distribution Generalization

Recent advances on large-scale pre-training have shown great potentials ...
research
12/10/2018

Data Fine-tuning

In real-world applications, commercial off-the-shelf systems are utilize...

Please sign up or login with your details

Forgot password? Click here to reset