Effective Neural Network L_0 Regularization With BinMask

04/21/2023
by   Kai Jia, et al.
0

L_0 regularization of neural networks is a fundamental problem. In addition to regularizing models for better generalizability, L_0 regularization also applies to selecting input features and training sparse neural networks. There is a large body of research on related topics, some with quite complicated methods. In this paper, we show that a straightforward formulation, BinMask, which multiplies weights with deterministic binary masks and uses the identity straight-through estimator for backpropagation, is an effective L_0 regularizer. We evaluate BinMask on three tasks: feature selection, network sparsification, and model regularization. Despite its simplicity, BinMask achieves competitive performance on all the benchmarks without task-specific tuning compared to methods designed for each task. Our results suggest that decoupling weights from mask optimization, which has been widely adopted by previous work, is a key component for effective L_0 regularization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/08/2020

Binary Stochastic Filtering: feature selection and beyond

Feature selection is one of the most decisive tools in understanding dat...
research
07/01/2023

Sparse-Input Neural Network using Group Concave Regularization

Simultaneous feature selection and non-linear function estimation are ch...
research
01/13/2022

Automatic Sparse Connectivity Learning for Neural Networks

Since sparse neural networks usually contain many zero weights, these un...
research
06/14/2021

The Flip Side of the Reweighted Coin: Duality of Adaptive Dropout and Regularization

Among the most successful methods for sparsifying deep (neural) networks...
research
08/14/2023

HyperSparse Neural Networks: Shifting Exploration to Exploitation through Adaptive Regularization

Sparse neural networks are a key factor in developing resource-efficient...
research
03/09/2023

TANGOS: Regularizing Tabular Neural Networks through Gradient Orthogonalization and Specialization

Despite their success with unstructured data, deep neural networks are n...
research
05/20/2019

Interpretable Neural Predictions with Differentiable Binary Variables

The success of neural networks comes hand in hand with a desire for more...

Please sign up or login with your details

Forgot password? Click here to reset