Understanding Sparse JL for Feature Hashing

03/08/2019
by   Meena Jagadeesan, et al.
0

Feature hashing and more general projection schemes are commonly used in machine learning to reduce the dimensionality of feature vectors. The goal is to efficiently project a high-dimensional feature vector living in R^n into a lower-dimensional space R^m, while approximately preserving Euclidean norm. These schemes can be constructed using sparse random projections, for example using a sparse Johnson-Lindenstrauss (JL) transform. In practice, feature vectors often have a low ℓ_∞-to-ℓ_2 norm ratio, and for this restricted set of vectors, many sparse JL-based schemes can achieve the norm-preserving objective with smaller dimension m than is necessary for the scheme on the full space R^n. A line of work introduced by Weinberger et. al (ICML '09) analyzes the sparse JL transform with one nonzero entry per column, which is a standard feature hashing scheme. Recently, Freksen, Kamma, and Larsen (NIPS '18) closed this line of work by proving an essentially tight tradeoff between ℓ_∞-to-ℓ_2 norm ratio, distortion, failure probability, and dimension m for this feature hashing scheme. We study more general projection schemes that are constructed using sparse JL transforms permitted to have more than one (but still a small fraction of) nonzero entries per column. Our main result is an essentially tight tradeoff between ℓ_∞-to-ℓ_2 norm ratio, distortion, failure probability, and dimension m for a general sparsity s, that generalizes the result of Freksen et. al. We also connect our result to the sparse JL literature by showing that it implies lower bounds on dimension-sparsity tradeoffs that essentially match upper bounds by Cohen (SODA '16). Moreover, our proof introduces a new perspective on bounding moments of certain random variables, that could be useful in other settings in theoretical computer science.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2018

Fully Understanding the Hashing Trick

Feature hashing, also known as the hashing trick, introduced by Weinber...
research
02/03/2021

CountSketches, Feature Hashing and the Median of Three

In this paper, we revisit the classic CountSketch method, which is a spa...
research
09/01/2022

Johnson-Lindenstrauss embeddings for noisy vectors – taking advantage of the noise

This paper investigates theoretical properties of subsampling and hashin...
research
05/03/2022

Understanding the Moments of Tabulation Hashing via Chaoses

Simple tabulation hashing dates back to Zobrist in 1970 and is defined a...
research
06/29/2020

Binary Random Projections with Controllable Sparsity Patterns

Random projection is often used to project higher-dimensional vectors on...
research
06/20/2023

Cuckoo Hashing in Cryptography: Optimal Parameters, Robustness and Applications

Cuckoo hashing is a powerful primitive that enables storing items using ...
research
12/09/2019

Grouped sparse projection

As evident from deep learning, very large models bring improvements in t...

Please sign up or login with your details

Forgot password? Click here to reset