Towards a New Understanding of the Training of Neural Networks with Mislabeled Training Data

09/18/2019
by   Herbert Gish, et al.
0

We investigate the problem of machine learning with mislabeled training data. We try to make the effects of mislabeled training better understood through analysis of the basic model and equations that characterize the problem. This includes results about the ability of the noisy model to make the same decisions as the clean model and the effects of noise on model performance. In addition to providing better insights we also are able to show that the Maximum Likelihood (ML) estimate of the parameters of the noisy model determine those of the clean model. This property is obtained through the use of the ML invariance property and leads to an approach to developing a classifier when training has been mislabeled: namely train the classifier on noisy data and adjust the decision threshold based on the noise levels and/or class priors. We show how our approach to mislabeled training works with multi-layered perceptrons (MLPs).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2022

Inferring Class Label Distribution of Training Data from Classifiers: An Accuracy-Augmented Meta-Classifier Attack

Property inference attacks against machine learning (ML) models aim to i...
research
05/22/2021

Denoising Noisy Neural Networks: A Bayesian Approach with Compensation

Noisy neural networks (NoisyNNs) refer to the inference and training of ...
research
08/30/2019

Classifying single-qubit noise using machine learning

Quantum characterization, validation, and verification (QCVV) techniques...
research
10/02/2022

The Dynamic of Consensus in Deep Networks and the Identification of Noisy Labels

Deep neural networks have incredible capacity and expressibility, and ca...
research
02/26/2023

Training neural networks with structured noise improves classification and generalization

The beneficial role of noise in learning is nowadays a consolidated conc...
research
08/25/2021

NGC: A Unified Framework for Learning with Open-World Noisy Data

The existence of noisy data is prevalent in both the training and testin...
research
07/30/2021

Toward Robust Autotuning of Noisy Quantum Dot Devices

The current autotuning approaches for quantum dot (QD) devices, while sh...

Please sign up or login with your details

Forgot password? Click here to reset