Large Loss Matters in Weakly Supervised Multi-Label Classification

06/08/2022
by   Youngwook Kim, et al.
0

Weakly supervised multi-label classification (WSML) task, which is to learn a multi-label classification using partially observed labels per image, is becoming increasingly important due to its huge annotation cost. In this work, we first regard unobserved labels as negative labels, casting the WSML task into noisy multi-label classification. From this point of view, we empirically observe that memorization effect, which was first discovered in a noisy multi-class setting, also occurs in a multi-label setting. That is, the model first learns the representation of clean labels, and then starts memorizing noisy labels. Based on this finding, we propose novel methods for WSML which reject or correct the large loss samples to prevent model from memorizing the noisy label. Without heavy and complex components, our proposed methods outperform previous state-of-the-art WSML methods on several partial label settings including Pascal VOC 2012, MS COCO, NUSWIDE, CUB, and OpenImages V3 datasets. Various analysis also show that our methodology actually works well, validating that treating large loss properly matters in a weakly supervised multi-label classification. Our code is available at https://github.com/snucml/LargeLossMatters.

READ FULL TEXT
research
02/26/2019

Learning a Deep ConvNet for Multi-label Classification with Partial Labels

Deep ConvNets have shown great performance for single-label image classi...
research
04/04/2023

Bridging the Gap between Model Explanations in Partially Annotated Multi-label Classification

Due to the expensive costs of collecting labels in multi-label classific...
research
08/04/2021

Multi-Label Gold Asymmetric Loss Correction with Single-Label Regulators

Multi-label learning is an emerging extension of the multi-class classif...
research
10/21/2021

Multi-label Classification with Partial Annotations using Class-aware Selective Loss

Large-scale multi-label classification datasets are commonly, and perhap...
research
03/07/2023

GaussianMLR: Learning Implicit Class Significance via Calibrated Multi-Label Ranking

Existing multi-label frameworks only exploit the information deduced fro...
research
10/08/2022

A Survey on Extreme Multi-label Learning

Multi-label learning has attracted significant attention from both acade...
research
03/11/2022

Spatial Consistency Loss for Training Multi-Label Classifiers from Single-Label Annotations

As natural images usually contain multiple objects, multi-label image cl...

Please sign up or login with your details

Forgot password? Click here to reset