Random Projections and Dimension Reduction

by   Rishi Advani, et al.

This paper, broadly speaking, covers the use of randomness in two main areas: low-rank approximation and kernel methods. Low-rank approximation is very important in numerical linear algebra. Many applications depend on matrix decomposition algorithms that provide accurate low-rank representations of data. In modern problems, however, various factors make this hard to accomplish. One solution to these problems is the use of random projections. Instead of directly computing the matrix factorization, we randomly project the matrix onto a lower-dimensional subspace and then compute the factorization. Often, we are able to do this without significant loss of accuracy. We describe how randomization can be used to create more efficient algorithms to perform low-rank matrix approximation, as well as introducing a novel randomized algorithm for matrix decomposition. Compared to standard approaches, random algorithms are often faster and more robust. With these randomized algorithms, analyzing massive data sets becomes tractable. Kernel methods are almost diametrically opposite from low-rank approximation. The idea is to project low-dimensional data into a higher-dimensional 'feature space,' such that it is linear separable in the feature space. This enables the model to learn a nonlinear separation of the data. As before, with large data matrices, computing the kernel matrix can be expensive, so we use randomized methods to approximate the matrix. In addition, we propose an extension of the random Fourier features kernel in which hyperparameter values are randomly sampled from an interval or Borel set. The experiments discussed in this paper can be found on our website at https://rishi1999.github.io/random-projections.



page 1

page 2

page 3

page 4


Pass-Efficient Randomized LU Algorithms for Computing Low-Rank Matrix Approximation

Low-rank matrix approximation is extremely useful in the analysis of dat...

Randomized Clustered Nystrom for Large-Scale Kernel Machines

The Nystrom method has been popular for generating the low-rank approxim...

Fast Approximate CoSimRanks via Random Projections

Given a graph G with n nodes, and two nodes u,v in G, the CoSim-Rank val...

Completing a joint PMF from projections: a low-rank coupled tensor factorization approach

There has recently been considerable interest in completing a low-rank m...

Randomized Numerical Linear Algebra: Foundations Algorithms

This survey describes probabilistic algorithms for linear algebra comput...

Theory of Dual-sparse Regularized Randomized Reduction

In this paper, we study randomized reduction methods, which reduce high-...

Deep matrix factorizations

Constrained low-rank matrix approximations have been known for decades a...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.