The information of attribute uncertainties: what convolutional neural networks can learn about errors in input data

08/10/2021
by   Natália V. N. Rodrigues, et al.
0

Errors in measurements are key to weighting the value of data, but are often neglected in Machine Learning (ML). We show how Convolutional Neural Networks (CNNs) are able to learn about the context and patterns of signal and noise, leading to improvements in the performance of classification methods. We construct a model whereby two classes of objects follow an underlying Gaussian distribution, and where the features (the input data) have varying, but known, levels of noise. This model mimics the nature of scientific data sets, where the noises arise as realizations of some random processes whose underlying distributions are known. The classification of these objects can then be performed using standard statistical techniques (e.g., least-squares minimization or Markov-Chain Monte Carlo), as well as ML techniques. This allows us to take advantage of a maximum likelihood approach to object classification, and to measure the amount by which the ML methods are incorporating the information in the input data uncertainties. We show that, when each data point is subject to different levels of noise (i.e., noises with different distribution functions), that information can be learned by the CNNs, raising the ML performance to at least the same level of the least-squares method – and sometimes even surpassing it. Furthermore, we show that, with varying noise levels, the confidence of the ML classifiers serves as a proxy for the underlying cumulative distribution function, but only if the information about specific input data uncertainties is provided to the CNNs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2021

Examining convolutional feature extraction using Maximum Entropy (ME) and Signal-to-Noise Ratio (SNR) for image classification

Convolutional Neural Networks (CNNs) specialize in feature extraction ra...
research
11/10/2021

Automatically detecting data drift in machine learning classifiers

Classifiers and other statistics-based machine learning (ML) techniques ...
research
09/17/2022

Interrelation of equivariant Gaussian processes and convolutional neural networks

Currently there exists rather promising new trend in machine leaning (ML...
research
06/08/2020

Learning to Utilize Correlated Auxiliary Classical or Quantum Noise

This paper has two messages. First, we demonstrate that neural networks ...
research
05/01/2021

Non-asymptotic Excess Risk Bounds for Classification with Deep Convolutional Neural Networks

In this paper, we consider the problem of binary classification with a c...
research
03/16/2023

VFP: Converting Tabular Data for IIoT into Images Considering Correlations of Attributes for Convolutional Neural Networks

For tabular data generated from IIoT devices, traditional machine learni...
research
05/26/2015

An Empirical Evaluation of Current Convolutional Architectures' Ability to Manage Nuisance Location and Scale Variability

We conduct an empirical study to test the ability of Convolutional Neura...

Please sign up or login with your details

Forgot password? Click here to reset