EvidentialMix: Learning with Combined Open-set and Closed-set Noisy Labels

11/11/2020
by   Ragav Sachdeva, et al.
6

The efficacy of deep learning depends on large-scale data sets that have been carefully curated with reliable data acquisition and annotation processes. However, acquiring such large-scale data sets with precise annotations is very expensive and time-consuming, and the cheap alternatives often yield data sets that have noisy labels. The field has addressed this problem by focusing on training models under two types of label noise: 1) closed-set noise, where some training samples are incorrectly annotated to a training label other than their known true class; and 2) open-set noise, where the training set includes samples that possess a true class that is (strictly) not contained in the set of known training labels. In this work, we study a new variant of the noisy label problem that combines the open-set and closed-set noisy labels, and introduce a benchmark evaluation to assess the performance of training algorithms under this setup. We argue that such problem is more general and better reflects the noisy label scenarios in practice. Furthermore, we propose a novel algorithm, called EvidentialMix, that addresses this problem and compare its performance with the state-of-the-art methods for both closed-set and open-set noise on the proposed benchmark. Our results show that our method produces superior classification results and better feature representations than previous state-of-the-art methods. The code is available at https://github.com/ragavsachdeva/EvidentialMix.

READ FULL TEXT
research
03/31/2018

Iterative Learning with Open-set Noisy Labels

Large-scale datasets possessing clean label annotations are crucial for ...
research
12/02/2020

Extended T: Learning with Mixed Closed-set and Open-set Noisy Labels

The label noise transition matrix T, reflecting the probabilities that t...
research
12/02/2022

Model and Data Agreement for Learning with Noisy Labels

Learning with noisy labels is a vital topic for practical deep learning ...
research
12/05/2020

A Survey on Deep Learning with Noisy Labels: How to train your model when you cannot trust on the annotations?

Noisy Labels are commonly present in data sets automatically collected f...
research
06/30/2021

Learning Bounds for Open-Set Learning

Traditional supervised learning aims to train a classifier in the closed...
research
08/25/2021

NGC: A Unified Framework for Learning with Open-World Noisy Data

The existence of noisy data is prevalent in both the training and testin...
research
03/21/2021

ScanMix: Learning from Severe Label Noise via Semantic Clustering and Semi-Supervised Learning

In this paper, we address the problem of training deep neural networks i...

Please sign up or login with your details

Forgot password? Click here to reset