Group Sparse Regularization for Deep Neural Networks

07/02/2016
by   Simone Scardapane, et al.
0

In this paper, we consider the joint task of simultaneously optimizing (i) the weights of a deep neural network, (ii) the number of neurons for each hidden layer, and (iii) the subset of active input features (i.e., feature selection). While these problems are generally dealt with separately, we present a simple regularized formulation allowing to solve all three of them in parallel, using standard optimization routines. Specifically, we extend the group Lasso penalty (originated in the linear regression literature) in order to impose group-level sparsity on the network's connections, where each group is defined as the set of outgoing weights from a unit. Depending on the specific case, the weights can be related to an input variable, to a hidden neuron, or to a bias unit, thus performing simultaneously all the aforementioned tasks in order to obtain a compact network. We perform an extensive experimental evaluation, by comparing with classical weight decay and Lasso penalties. We show that a sparse version of the group Lasso penalty is able to achieve competitive performances, while at the same time resulting in extremely compact networks with a smaller number of input features. We evaluate both on a toy dataset for handwritten digit recognition, and on multiple realistic large-scale classification problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/24/2021

Adaptive Group Lasso Neural Network Models for Functions of Few Variables and Time-Dependent Data

In this paper, we propose an adaptive group Lasso deep neural network fo...
research
07/29/2019

A neural network with feature sparsity

We propose a neural network model, with a separate linear (residual) ter...
research
07/01/2023

Sparse-Input Neural Network using Group Concave Regularization

Simultaneous feature selection and non-linear function estimation are ch...
research
11/21/2017

Sparse-Input Neural Networks for High-dimensional Nonparametric Regression and Classification

Neural networks are usually not the tool of choice for nonparametric hig...
research
04/09/2020

Hierarchical Group Sparse Regularization for Deep Convolutional Neural Networks

In a deep neural network (DNN), the number of the parameters is usually ...
research
05/30/2018

How Important Is a Neuron?

The problem of attributing a deep network's prediction to its input/base...
research
09/03/2012

Proximal methods for the latent group lasso penalty

We consider a regularized least squares problem, with regularization by ...

Please sign up or login with your details

Forgot password? Click here to reset