Global empirical risk minimizers with "shape constraints" are rate optimal in general dimensions

05/30/2019
by   Qiyang Han, et al.
0

Entropy integrals are widely used as a powerful tool to obtain upper bounds for the rates of convergence of global empirical risk minimizers (ERMs), in standard settings such as density estimation and regression. The upper bound for the convergence rates thus obtained typically matches the minimax lower bound when the entropy integral converges, but admits a strict gap compared with the lower bound when it diverges. [BM93] provided a striking example showing that such a gap is real with the entropy structure alone: for a variant of the natural Hölder class with low regularity, the global ERM actually converges at the rate predicted by the entropy integral that substantially deviates from the lower bound. The counter-example has spawned a long-standing negative position on the use of global ERMs in the regime where the entropy integral diverges, as they are heuristically believed to converge at a sub-optimal rate in a variety of models. The present paper demonstrates that this gap can be closed if the models admit certain degree of `shape constraints' in addition to the entropy structure. In other words, the global ERMs in such `shape-constrained' models will indeed be rate-optimal, matching the lower bound even when the entropy integral diverges. The models with `shape constraints' we investigate include (i) edge estimation with additive and multiplicative errors, (ii) binary classification, (iii) multiple isotonic regression, (iv) s-concave density estimation, all in general dimensions when the entropy integral diverges. Here `shape constraints' are interpreted broadly in the sense that the complexity of the underlying models can be essentially captured by the size of the empirical process over certain class of measurable sets, for which matching upper and lower bounds are obtained to facilitate the derivation of sharp convergence rates for the associated global ERMs.

READ FULL TEXT
research
11/02/2017

Measuring Quantum Entropy

The entropy of a quantum system is a measure of its randomness, and has ...
research
02/03/2012

Minimax Rates of Estimation for Sparse PCA in High Dimensions

We study sparse principal components analysis in the high-dimensional se...
research
09/21/2021

Minimax Rates for Conditional Density Estimation via Empirical Entropy

We consider the task of estimating a conditional density using i.i.d. sa...
research
07/25/2022

Minimax Rates for High-dimensional Double Sparse Structure over ℓ_q-balls

In this paper, we focus on the high-dimensional double sparse structure,...
research
08/05/2022

Malliavin calculus for the optimal estimation of the invariant density of discretely observed diffusions in intermediate regime

Let (X_t)_t ≥ 0 be solution of a one-dimensional stochastic differential...
research
06/10/2021

GBHT: Gradient Boosting Histogram Transform for Density Estimation

In this paper, we propose a density estimation algorithm called Gradient...
research
10/31/2020

Mixing it up: A general framework for Markovian statistics beyond reversibility and the minimax paradigm

Up to now, the nonparametric analysis of multidimensional continuous-tim...

Please sign up or login with your details

Forgot password? Click here to reset