Webly Supervised Image Classification with Self-Contained Confidence

08/27/2020
by   Jingkang Yang, et al.
0

This paper focuses on webly supervised learning (WSL), where datasets are built by crawling samples from the Internet and directly using search queries as web labels. Although WSL benefits from fast and low-cost data collection, noises in web labels hinder better performance of the image classification model. To alleviate this problem, in recent works, self-label supervised loss ℒ_s is utilized together with webly supervised loss ℒ_w. ℒ_s relies on pseudo labels predicted by the model itself. Since the correctness of the web label or pseudo label is usually on a case-by-case basis for each web sample, it is desirable to adjust the balance between ℒ_s and ℒ_w on sample level. Inspired by the ability of Deep Neural Networks (DNNs) in confidence prediction, we introduce Self-Contained Confidence (SCC) by adapting model uncertainty for WSL setting, and use it to sample-wisely balance ℒ_s and ℒ_w. Therefore, a simple yet effective WSL framework is proposed. A series of SCC-friendly regularization approaches are investigated, among which the proposed graph-enhanced mixup is the most effective method to provide high-quality confidence to enhance our framework. The proposed WSL framework has achieved the state-of-the-art results on two large-scale WSL datasets, WebVision-1000 and Food101-N. Code is available at https://github.com/bigvideoresearch/SCC.

READ FULL TEXT
research
08/17/2023

MarginMatch: Improving Semi-Supervised Learning with Pseudo-Margins

We introduce MarginMatch, a new SSL approach combining consistency regul...
research
01/10/2023

Neighborhood-Regularized Self-Training for Learning with Few Labels

Training deep neural networks (DNNs) with limited supervision has been a...
research
08/26/2019

Confidence Regularized Self-Training

Recent advances in domain adaptation show that deep self-training presen...
research
10/19/2022

Pseudo-Label Noise Suppression Techniques for Semi-Supervised Semantic Segmentation

Semi-supervised learning (SSL) can reduce the need for large labelled da...
research
01/29/2023

Confidence-Aware Calibration and Scoring Functions for Curriculum Learning

Despite the great success of state-of-the-art deep neural networks, seve...
research
12/03/2021

Neural Pseudo-Label Optimism for the Bank Loan Problem

We study a class of classification problems best exemplified by the bank...
research
06/30/2022

ProSelfLC: Progressive Self Label Correction Towards A Low-Temperature Entropy State

To train robust deep neural networks (DNNs), we systematically study sev...

Please sign up or login with your details

Forgot password? Click here to reset