Importance Reweighting for Biquality Learning

10/19/2020
by   Pierre Nodet, et al.
0

The field of Weakly Supervised Learning (WSL) has recently seen a surge of popularity, with numerous papers addressing different types of “supervision deficiencies”, namely: poor quality, non adaptability, and insufficient quantity of labels. Regarding quality, label noise can be of different kinds, including completely-at-random, at-random or even not-at-random. All these kinds of label noise are addressed separately in the literature, leading to highly specialized approaches. This paper proposes an original view of Weakly Supervised Learning, to design generic approaches capable of dealing with any kind of label noise. For this purpose, an alternative setting called “Biquality data” is used. This setting assumes that a small trusted dataset of correctly labeled examples is available, in addition to the untrusted dataset of noisy examples. In this paper, we propose a new reweigthing scheme capable of identifying noncorrupted examples in the untrusted dataset. This allows one to learn classifiers using both datasets. Extensive experiments demonstrate that the proposed approach outperforms baselines and state-of-the-art approaches, by simulating several kinds of label noise and varying the quality and quantity of untrusted examples.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2020

From Weakly Supervised Learning to Biquality Learning, a brief introduction

The field of Weakly Supervised Learning (WSL) has recently seen a surge ...
research
04/22/2019

Reliable Weakly Supervised Learning: Maximize Gain and Maintain Safeness

Weakly supervised data are widespread and have attracted much attention....
research
03/18/2023

CroSel: Cross Selection of Confident Pseudo Labels for Partial-Label Learning

Partial-label learning (PLL) is an important weakly supervised learning ...
research
07/13/2020

TrustNet: Learning from Trusted Data Against (A)symmetric Label Noise

Robustness to label noise is a critical property for weakly-supervised c...
research
01/19/2020

Weakly Supervised Learning Meets Ride-Sharing User Experience Enhancement

Weakly supervised learning aims at coping with scarce labeled data. Prev...
research
02/08/2016

Loss factorization, weakly supervised learning and label noise robustness

We prove that the empirical risk of most well-known loss functions facto...
research
10/18/2022

CNT (Conditioning on Noisy Targets): A new Algorithm for Leveraging Top-Down Feedback

We propose a novel regularizer for supervised learning called Conditioni...

Please sign up or login with your details

Forgot password? Click here to reset