Precise expressions for random projections: Low-rank approximation and randomized Newton

06/18/2020
by   Michał Dereziński, et al.
0

It is often desirable to reduce the dimensionality of a large dataset by projecting it onto a low-dimensional subspace. Matrix sketching has emerged as a powerful technique for performing such dimensionality reduction very efficiently. Even though there is an extensive literature on the worst-case performance of sketching, existing guarantees are typically very different from what is observed in practice. We exploit recent developments in the spectral analysis of random matrices to develop novel techniques that provide provably accurate expressions for the expected value of random projection matrices obtained via sketching. These expressions can be used to characterize the performance of dimensionality reduction in a variety of common machine learning tasks, ranging from low-rank approximation to iterative stochastic optimization. Our results apply to several popular sketching methods, including Gaussian and Rademacher sketches, and they enable precise analysis of these methods in terms of spectral properties of the data. Empirical results show that the expressions we derive reflect the practical performance of these sketching methods, down to lower-order effects and even constant factors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/27/2023

Practical Sketching Algorithms for Low-Rank Tucker Approximation of Large Tensors

Low-rank approximation of tensors has been widely used in high-dimension...
research
01/28/2019

Stochastic Linear Bandits with Hidden Low Rank Structure

High-dimensional representations often have a lower dimensional underlyi...
research
05/23/2019

Johnson-Lindenstrauss Property Implies Subspace Restricted Isometry Property

Dimensionality reduction is a popular approach to tackle high-dimensiona...
research
06/17/2018

On Sketching the q to p norms

We initiate the study of data dimensionality reduction, or sketching, fo...
research
05/14/2019

Dimensionality Reduction for Tukey Regression

We give the first dimensionality reduction methods for the overconstrain...
research
02/23/2018

Harnessing Structures in Big Data via Guaranteed Low-Rank Matrix Estimation

Low-rank modeling plays a pivotal role in signal processing and machine ...
research
03/03/2022

Uniform Approximations for Randomized Hadamard Transforms with Applications

Randomized Hadamard Transforms (RHTs) have emerged as a computationally ...

Please sign up or login with your details

Forgot password? Click here to reset