Statistical Hypothesis Testing Based on Machine Learning: Large Deviations Analysis

07/22/2022
by   Paolo Braca, et al.
6

We study the performance – and specifically the rate at which the error probability converges to zero – of Machine Learning (ML) classification techniques. Leveraging the theory of large deviations, we provide the mathematical conditions for a ML classifier to exhibit error probabilities that vanish exponentially, say ∼exp(-n I + o(n) ), where n is the number of informative observations available for testing (or another relevant parameter, such as the size of the target in an image) and I is the error rate. Such conditions depend on the Fenchel-Legendre transform of the cumulant-generating function of the Data-Driven Decision Function (D3F, i.e., what is thresholded before the final binary decision is made) learned in the training phase. As such, the D3F and, consequently, the related error rate I, depend on the given training set, which is assumed of finite size. Interestingly, these conditions can be verified and tested numerically exploiting the available dataset, or a synthetic dataset, generated according to the available information on the underlying statistical model. In other words, the classification error probability convergence to zero and its rate can be computed on a portion of the dataset available for training. Coherently with the large deviations theory, we can also establish the convergence, for n large enough, of the normalized D3F statistic to a Gaussian distribution. This property is exploited to set a desired asymptotic false alarm probability, which empirically turns out to be accurate even for quite realistic values of n. Furthermore, approximate error probability curves ∼ζ_n exp(-n I ) are provided, thanks to the refined asymptotic derivation (often referred to as exact asymptotics), where ζ_n represents the most representative sub-exponential terms of the error probabilities.

READ FULL TEXT

page 1

page 4

page 11

page 17

page 18

page 19

page 20

page 23

research
01/16/2023

Large Deviations for Classification Performance Analysis of Machine Learning Systems

We study the performance of machine learning binary classification techn...
research
06/13/2021

Finite-Length Bounds on Hypothesis Testing Subject to Vanishing Type I Error Restrictions

A central problem in Binary Hypothesis Testing (BHT) is to determine the...
research
06/03/2018

Second-Order Asymptotically Optimal Statistical Classification

Motivated by real-world machine learning applications, we analyze approx...
research
01/23/2022

Asymptotics for Outlier Hypothesis Testing

We revisit the outlier hypothesis testing framework of Li et al. (TIT 20...
research
12/13/2017

Exponential convergence of testing error for stochastic gradient methods

We consider binary classification problems with positive definite kernel...
research
02/22/2018

Multidimensional multiscale scanning in Exponential Families: Limit theory and statistical consequences

In this paper we consider the problem of finding anomalies in a d-dimens...
research
07/18/2022

The Vocal Signature of Social Anxiety: Exploration using Hypothesis-Testing and Machine-Learning Approaches

Background - Social anxiety (SA) is a common and debilitating condition,...

Please sign up or login with your details

Forgot password? Click here to reset