Total Least Squares Regression in Input Sparsity Time

09/27/2019
by   Huaian Diao, et al.
0

In the total least squares problem, one is given an m × n matrix A, and an m × d matrix B, and one seeks to "correct" both A and B, obtaining matrices  and B̂, so that there exists an X satisfying the equation ÂX = B̂. Typically the problem is overconstrained, meaning that m ≫max(n,d). The cost of the solution Â, B̂ is given by A-Â_F^2 + B - B̂_F^2. We give an algorithm for finding a solution X to the linear system ÂX=B̂ for which the cost A-Â_F^2 + B-B̂_F^2 is at most a multiplicative (1+ϵ) factor times the optimal cost, up to an additive error η that may be an arbitrarily small function of n. Importantly, our running time is Õ( nnz(A) + nnz(B) ) + poly(n/ϵ) · d, where for a matrix C, nnz(C) denotes its number of non-zero entries. Importantly, our running time does not directly depend on the large parameter m. As total least squares regression is known to be solvable via low rank approximation, a natural approach is to invoke fast algorithms for approximate low rank approximation, obtaining matrices  and B̂ from this low rank approximation, and then solving for X so that ÂX = B̂. However, existing algorithms do not apply since in total least squares the rank of the low rank approximation needs to be n, and so the running time of known methods would be at least mn^2. In contrast, we are able to achieve a much faster running time for finding X by never explicitly forming the equation  X = B̂, but instead solving for an X which is a solution to an implicit such equation. Finally, we generalize our algorithm to the total least squares problem with regularization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/29/2019

Optimal Sketching for Kronecker Product Regression and Low Rank Approximation

We study the Kronecker product regression problem, in which the design m...
research
11/01/2021

Improved Algorithms for Low Rank Approximation from Sparsity

We overcome two major bottlenecks in the study of low rank approximation...
research
07/12/2021

In-Database Regression in Input Sparsity Time

Sketching is a powerful dimensionality reduction technique for accelerat...
research
12/08/2019

The AZ algorithm for least squares systems with a known incomplete generalized inverse

We introduce an algorithm for the least squares solution of a rectangula...
research
02/18/2019

Improved Convergence for ℓ_∞ and ℓ_1 Regression via Iteratively Reweighted Least Squares

The iteratively reweighted least squares method (IRLS) is a popular tech...
research
08/17/2018

Randomized Least Squares Regression: Combining Model- and Algorithm-Induced Uncertainties

We analyze the uncertainties in the minimum norm solution of full-rank r...
research
12/12/2019

Sublinear Time Numerical Linear Algebra for Structured Matrices

We show how to solve a number of problems in numerical linear algebra, s...

Please sign up or login with your details

Forgot password? Click here to reset