Cost-Sensitive Self-Training for Optimizing Non-Decomposable Metrics

04/28/2023
by   Harsh Rangwani, et al.
0

Self-training based semi-supervised learning algorithms have enabled the learning of highly accurate deep neural networks, using only a fraction of labeled data. However, the majority of work on self-training has focused on the objective of improving accuracy, whereas practical machine learning systems can have complex goals (e.g. maximizing the minimum of recall across classes, etc.) that are non-decomposable in nature. In this work, we introduce the Cost-Sensitive Self-Training (CSST) framework which generalizes the self-training-based methods for optimizing non-decomposable metrics. We prove that our framework can better optimize the desired non-decomposable metric utilizing unlabeled data, under similar data distribution assumptions made for the analysis of self-training. Using the proposed CSST framework, we obtain practical self-training methods (for both vision and NLP tasks) for optimizing different non-decomposable metrics using deep neural networks. Our results demonstrate that CSST achieves an improvement over the state-of-the-art in majority of the cases across datasets and objectives.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/21/2022

How does unlabeled data improve generalization in self-training? A one-hidden-layer theoretical analysis

Self-training, a semi-supervised learning algorithm, leverages a large a...
research
03/29/2022

Self-Contrastive Learning based Semi-Supervised Radio Modulation Classification

This paper presents a semi-supervised learning framework that is new in ...
research
03/21/2023

Adaptive Negative Evidential Deep Learning for Open-set Semi-supervised Learning

Semi-supervised learning (SSL) methods assume that labeled data, unlabel...
research
10/14/2018

Distributed learning of deep neural network over multiple agents

In domains such as health care and finance, shortage of labeled data and...
research
06/23/2022

Few-Shot Non-Parametric Learning with Deep Latent Variable Model

Most real-world problems that machine learning algorithms are expected t...
research
10/07/2020

Theoretical Analysis of Self-Training with Deep Networks on Unlabeled Data

Self-training algorithms, which train a model to fit pseudolabels predic...
research
07/09/2021

Training Over-parameterized Models with Non-decomposable Objectives

Many modern machine learning applications come with complex and nuanced ...

Please sign up or login with your details

Forgot password? Click here to reset