Near-Optimal Algorithms for Linear Algebra in the Current Matrix Multiplication Time

by   Nadiia Chepurko, et al.

Currently, in the numerical linear algebra community, it is thought that to obtain nearly-optimal bounds for various problems such as rank computation, finding a maximal linearly independent subset of columns, regression, low rank approximation, maximum matching on general graphs and linear matroid union, one would need to resolve the main open question of Nelson and Nguyen (FOCS, 2013) regarding the logarithmic factors in the sketching dimension for existing constant factor approximation oblivious subspace embeddings. We show how to bypass this question using a refined sketching technique, and obtain optimal or nearly optimal bounds for these problems. A key technique we use is an explicit mapping of Indyk based on uncertainty principles and extractors, which after first applying known oblivious subspace embeddings, allows us to quickly spread out the mass of the vector so that sampling is now effective, and we avoid a logarithmic factor that is standard in the sketching dimension resulting from matrix Chernoff bounds. For the fundamental problems of rank computation and finding a linearly independent subset of columns, our algorithms improve Cheung, Kwok, and Lau (JACM, 2013) and are optimal to within a constant factor and a loglog(n)-factor, respectively. Further, for constant factor regression and low rank approximation we give the first optimal algorithms, for the current matrix multiplication exponent.



page 1

page 2

page 3

page 4


Quantum-Inspired Algorithms from Randomized Numerical Linear Algebra

We create classical (non-quantum) dynamic data structures supporting que...

Leveraging Well-Conditioned Bases: Streaming & Distributed Summaries in Minkowski p-Norms

Work on approximate linear algebra has led to efficient distributed and ...

Tight Bounds for ℓ_p Oblivious Subspace Embeddings

An ℓ_p oblivious subspace embedding is a distribution over r × n matrice...

Average Case Column Subset Selection for Entrywise ℓ_1-Norm Loss

We study the column subset selection problem with respect to the entrywi...

Regularized Weighted Low Rank Approximation

The classical low rank approximation problem is to find a rank k matrix ...

Optimal approximate matrix product in terms of stable rank

We prove, using the subspace embedding guarantee in a black box way, tha...

A Bootstrap Method for Error Estimation in Randomized Matrix Multiplication

In recent years, randomized methods for numerical linear algebra have re...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.