Hashing-like Johnson–Lindenstrauss transforms and their extreme singular values

12/30/2022
by   Kwassi Joseph Dzahini, et al.
0

The Johnson–Lindenstrauss (JL) lemma is a powerful tool for dimensionality reduction in modern algorithm design. The lemma states that any set of high-dimensional points in a Euclidean space can be flattened to lower dimensions while approximately preserving pairwise Euclidean distances. Random matrices satisfying this lemma are called JL transforms (JLTs). Inspired by existing s-hashing JLTs with exactly s nonzero elements on each column, the present work introduces s-hashing-like matrices whose expected number of nonzero elements on each column is s. Unlike s-hashing matrices, the independence of the sub-Gaussian entries of s-hashing-like matrices and the knowledge of their exact distribution play an important role in their analyses. Using properties of independent sub-Gaussian random variables, these matrices are demonstrated to be JLTs, and their smallest and largest singular values are estimated non-asymptotically in terms of known quantities using a technique from geometric functional analysis, that is, without any unknown “absolute constant” as is often the case in random matrix theory. By using the universal Bai–Yin law, these singular values are proved to converge almost surely to fixed quantities as the dimensions of the matrix grow to infinity. Numerical experiments suggest that s-hashing-like matrices are more effective in dimensionality reduction and have better behaviors of their extreme singular values than is borne out by our non-asymptotic analyses.

READ FULL TEXT

page 18

page 19

page 20

page 21

page 22

research
09/06/2021

The asymptotic joint distribution of the largest and the smallest singular values for random circulant matrices

In this manuscript, we study the limiting distribution for the joint law...
research
09/03/2020

On the smoothed analysis of the smallest singular value with discrete noise

Let A be an n× n real matrix, and let M be an n× n random matrix whose e...
research
02/22/2023

nSimplex Zen: A Novel Dimensionality Reduction for Euclidean and Hilbert Spaces

Dimensionality reduction techniques map values from a high dimensional s...
research
03/28/2022

Random matrix analysis of deep neural network weight matrices

Neural networks have been used successfully in a variety of fields, whic...
research
09/01/2022

Johnson-Lindenstrauss embeddings for noisy vectors – taking advantage of the noise

This paper investigates theoretical properties of subsampling and hashin...
research
11/21/2010

Random Projections for k-means Clustering

This paper discusses the topic of dimensionality reduction for k-means c...

Please sign up or login with your details

Forgot password? Click here to reset