Statistical optimality conditions for compressive ensembles

06/02/2021
by   Henry W J Reeve, et al.
0

We present a framework for the theoretical analysis of ensembles of low-complexity empirical risk minimisers trained on independent random compressions of high-dimensional data. First we introduce a general distribution-dependent upper-bound on the excess risk, framed in terms of a natural notion of compressibility. This bound is independent of the dimension of the original data representation, and explains the in-built regularisation effect of the compressive approach. We then instantiate this general bound to classification and regression tasks, considering Johnson-Lindenstrauss mappings as the compression scheme. For each of these tasks, our strategy is to develop a tight upper bound on the compressibility function, and by doing so we discover distributional conditions of geometric nature under which the compressive algorithm attains minimax-optimal rates up to at most poly-logarithmic factors. In the case of compressive classification, this is achieved with a mild geometric margin condition along with a flexible moment condition that is significantly more general than the assumption of bounded domain. In the case of regression with strongly convex smooth loss functions we find that compressive regression is capable of exploiting spectral decay with near-optimal guarantees. In addition, a key ingredient for our central upper bound is a high probability uniform upper bound on the integrated deviation of dependent empirical processes, which may be of independent interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/14/2018

Differentially Private Empirical Risk Minimization Revisited: Faster and More General

In this paper we study the differentially private Empirical Risk Minimiz...
research
02/21/2015

Learning with Square Loss: Localization through Offset Rademacher Complexity

We consider regression with square loss and general classes of functions...
research
07/31/2021

Faster Rates of Differentially Private Stochastic Convex Optimization

In this paper, we revisit the problem of Differentially Private Stochast...
research
05/13/2022

Improved Upper Bound on Independent Domination Number for Hypercubes

We revisit the problem of determining the independent domination number ...
research
05/19/2021

Localization, Convexity, and Star Aggregation

Offset Rademacher complexities have been shown to imply sharp, data-depe...
research
11/02/2022

Large deviations rates for stochastic gradient descent with strongly convex functions

Recent works have shown that high probability metrics with stochastic gr...
research
06/15/2023

Non-Asymptotic Performance of Social Machine Learning Under Limited Data

This paper studies the probability of error associated with the social m...

Please sign up or login with your details

Forgot password? Click here to reset