What needles do sparse neural networks find in nonlinear haystacks

06/07/2020
by   Sylvain Sardy, et al.
0

Using a sparsity inducing penalty in artificial neural networks (ANNs) avoids over-fitting, especially in situations where noise is high and the training set is small in comparison to the number of features. For linear models, such an approach provably also recovers the important features with high probability in regimes for a well-chosen penalty parameter. The typical way of setting the penalty parameter is by splitting the data set and performing the cross-validation, which is (1) computationally expensive and (2) not desirable when the data set is already small to be further split (for example, whole-genome sequence data). In this study, we establish the theoretical foundation to select the penalty parameter without cross-validation based on bounding with a high probability the infinite norm of the gradient of the loss function at zero under the zero-feature assumption. Our approach is a generalization of the universal threshold of Donoho and Johnstone (1994) to nonlinear ANN learning. We perform a set of comprehensive Monte Carlo simulations on a simple model, and the numerical results show the effectiveness of the proposed approach.

READ FULL TEXT

page 6

page 7

research
03/10/2019

Lasso tuning through the flexible-weighted bootstrap

Regularized regression approaches such as the Lasso have been widely ado...
research
04/10/2021

Analytic and Bootstrap-after-Cross-Validation Methods for Selecting Penalty Parameters of High-Dimensional M-Estimators

We develop two new methods for selecting the penalty parameter for the ℓ...
research
06/06/2017

Shape Parameter Estimation

Performance of machine learning approaches depends strongly on the choic...
research
10/03/2019

Modeling of Electrical Resistivity of Soil Based on Geotechnical Properties

Determining the relationship between the electrical resistivity of soil ...
research
03/28/2019

An analysis of the cost of hyper-parameter selection via split-sample validation, with applications to penalized regression

In the regression setting, given a set of hyper-parameters, a model-esti...
research
03/28/2017

Gradient-based Regularization Parameter Selection for Problems with Non-smooth Penalty Functions

In high-dimensional and/or non-parametric regression problems, regulariz...
research
12/23/2021

Cooperative learning for multi-view analysis

We propose a new method for supervised learning with multiple sets of fe...

Please sign up or login with your details

Forgot password? Click here to reset