High-Dimensional Learning under Approximate Sparsity: A Unifying Framework for Nonsmooth Learning and Regularized Neural Networks

03/02/2019
by   Hongcheng Liu, et al.
0

High-dimensional statistical learning (HDSL) has been widely applied in data analysis, operations research, and stochastic optimization. Despite the availability of multiple theoretical frameworks, most HDSL theories stipulate the following two conditions, which are sometimes overly critical: (a) the sparsity, and (b) the restricted strong convexity (RSC). This paper generalizes both conditions via the use of the folded concave penalty (FCP); we show that, for an M-estimation problem where (i) the (conventional) sparsity is relaxed into the approximate sparsity and (ii) the RSC is completely absent, the FCP-based regularization leads to poly-logarithmic sample complexity: the size of the training data is only required to be poly-logarithmic in the problem dimensionality. This finding allows us to further understand two important paradigms much less discussed formerly: the high-dimensional nonsmooth learning and the (deep) neural networks (NN). For both problems, we show that the poly-logarithmic sample complexity can be maintained. Furthermore, via integrating the NN with the FCP, the excess risk of a stationary point to the training formulation for the NN is strictly monotonic with respect to the solution's suboptimality gap, providing the first theoretical evidence for the empirically observed consistency between the generalization performance and the optimization quality in training an NN.

READ FULL TEXT
research
10/08/2020

Learning Partially Observed Linear Dynamical Systems from Logarithmic Number of Samples

In this work, we study the problem of learning partially observed linear...
research
06/11/2018

High Dimensional Data Enrichment: Interpretable, Fast, and Data-Efficient

High dimensional structured data enriched model describes groups of obse...
research
10/18/2021

Path Regularization: A Convexity and Sparsity Inducing Regularization for Parallel ReLU Networks

Despite several attempts, the fundamental mechanisms behind the success ...
research
01/20/2020

Generalization Bounds for High-dimensional M-estimation under Sparsity Constraint

The ℓ_0-constrained empirical risk minimization (ℓ_0-ERM) is a promising...
research
04/25/2020

Learning to Guide Random Search

We are interested in derivative-free optimization of high-dimensional fu...
research
10/19/2011

Is the k-NN classifier in high dimensions affected by the curse of dimensionality?

There is an increasing body of evidence suggesting that exact nearest ne...
research
06/08/2021

What training reveals about neural network complexity

This work explores the hypothesis that the complexity of the function a ...

Please sign up or login with your details

Forgot password? Click here to reset