Reduced Label Complexity For Tight ℓ_2 Regression

05/12/2023
by   Alex Gittens, et al.
0

Given data X∈ℝ^n× d and labels 𝐲∈ℝ^n the goal is find 𝐰∈ℝ^d to minimize ‖ X𝐰-𝐲‖^2. We give a polynomial algorithm that, oblivious to 𝐲, throws out n/(d+√(n)) data points and is a (1+d/n)-approximation to optimal in expectation. The motivation is tight approximation with reduced label complexity (number of labels revealed). We reduce label complexity by Ω(√(n)). Open question: Can label complexity be reduced by Ω(n) with tight (1+d/n)-approximation?

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/04/2017

Supporting Ruled Polygons

We explore several problems related to ruled polygons. Given a ruling of...
research
05/15/2019

Tight Kernel Query Complexity of Kernel Ridge Regression and Kernel k-means Clustering

We present tight lower bounds on the number of kernel evaluations requir...
research
07/13/2022

Online Active Regression

Active regression considers a linear regression problem where the learne...
research
06/08/2019

Doubly Robust Crowdsourcing

Large-scale labeled datasets are the indispensable fuel that ignites the...
research
06/01/2021

L_0 Isotonic Regression With Secondary Objectives

We provide algorithms for isotonic regression minimizing L_0 error (Hamm...
research
11/07/2018

A Tight Analysis of Bethe Approximation for Permanent

We prove that the permanent of nonnegative matrices can be deterministic...
research
11/16/2018

A tight kernel for computing the tree bisection and reconnection distance between two phylogenetic trees

In 2001 Allen and Steel showed that, if subtree and chain reduction rule...

Please sign up or login with your details

Forgot password? Click here to reset