SHRIMP: Sparser Random Feature Models via Iterative Magnitude Pruning

12/07/2021
by   Yuege Xie, et al.
3

Sparse shrunk additive models and sparse random feature models have been developed separately as methods to learn low-order functions, where there are few interactions between variables, but neither offers computational efficiency. On the other hand, ℓ_2-based shrunk additive models are efficient but do not offer feature selection as the resulting coefficient vectors are dense. Inspired by the success of the iterative magnitude pruning technique in finding lottery tickets of neural networks, we propose a new method – Sparser Random Feature Models via IMP (ShRIMP) – to efficiently fit high-dimensional data with inherent low-dimensional structure in the form of sparse variable dependencies. Our method can be viewed as a combined process to construct and find sparse lottery tickets for two-layer dense networks. We explain the observed benefit of SHRIMP through a refined analysis on the generalization error for thresholded Basis Pursuit and resulting bounds on eigenvalues. From function approximation experiments on both synthetic data and real-world benchmark datasets, we show that SHRIMP obtains better than or competitive test accuracy compared to state-of-art sparse feature and additive methods such as SRFE-S, SSAM, and SALSA. Meanwhile, SHRIMP performs feature selection with low computational complexity and is robust to the pruning rate, indicating a robustness in the structure of the obtained subnetworks. We gain insight into the lottery ticket hypothesis through SHRIMP by noting a correspondence between our model and weight/neuron subnetworks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/10/2023

Supervised Feature Selection with Neuron Evolution in Sparse Neural Networks

Feature selection that selects an informative subset of variables from d...
research
06/06/2020

Chromatic Learning for Sparse Datasets

Learning over sparse, high-dimensional data frequently necessitates the ...
research
05/30/2023

Generalization Bounds for Magnitude-Based Pruning via Sparse Matrix Sketching

In this paper, we derive a novel bound on the generalization error of Ma...
research
12/12/2013

Sparse Matrix-based Random Projection for Classification

As a typical dimensionality reduction technique, random projection can b...
research
02/06/2022

HARFE: Hard-Ridge Random Feature Expansion

We propose a random feature model for approximating high-dimensional spa...
research
05/16/2022

Distributed Feature Selection for High-dimensional Additive Models

Distributed statistical learning is a common strategy for handling massi...

Please sign up or login with your details

Forgot password? Click here to reset