Efficient Symmetric Norm Regression via Linear Sketching

10/04/2019 ∙ by Zhao Song, et al. ∙ 0

We provide efficient algorithms for overconstrained linear regression problems with size n × d when the loss function is a symmetric norm (a norm invariant under sign-flips and coordinate-permutations). An important class of symmetric norms are Orlicz norms, where for a function G and a vector y ∈R^n, the corresponding Orlicz norm y_G is defined as the unique value α such that ∑_i=1^n G(|y_i|/α) = 1. When the loss function is an Orlicz norm, our algorithm produces a (1 + ε)-approximate solution for an arbitrarily small constant ε > 0 in input-sparsity time, improving over the previously best-known algorithm which produces a d ·polylog n-approximate solution. When the loss function is a general symmetric norm, our algorithm produces a √(d)·polylog n ·mmc(ℓ)-approximate solution in input-sparsity time, where mmc(ℓ) is a quantity related to the symmetric norm under consideration. To the best of our knowledge, this is the first input-sparsity time algorithm with provable guarantees for the general class of symmetric norm regression problem. Our results shed light on resolving the universal sketching problem for linear regression, and the techniques might be of independent interest to numerical linear algebra problems more broadly.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.