Forster Decomposition and Learning Halfspaces with Noise

07/12/2021
by   Ilias Diakonikolas, et al.
10

A Forster transform is an operation that turns a distribution into one with good anti-concentration properties. While a Forster transform does not always exist, we show that any distribution can be efficiently decomposed as a disjoint mixture of few distributions for which a Forster transform exists and can be computed efficiently. As the main application of this result, we obtain the first polynomial-time algorithm for distribution-independent PAC learning of halfspaces in the Massart noise model with strongly polynomial sample complexity, i.e., independent of the bit complexity of the examples. Previous algorithms for this learning problem incurred sample complexity scaling polynomially with the bit complexity, even though such a dependence is not information-theoretically necessary.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/06/2022

A Strongly Polynomial Algorithm for Approximate Forster Transforms and its Application to Halfspace Learning

The Forster transform is a method of regularizing a dataset by placing i...
research
12/13/2022

How Does Independence Help Generalization? Sample Complexity of ERM on Product Distributions

While many classical notions of learnability (e.g., PAC learnability) ar...
research
12/09/2014

Max vs Min: Tensor Decomposition and ICA with nearly Linear Sample Complexity

We present a simple, general technique for reducing the sample complexit...
research
02/11/2021

Sample-Optimal PAC Learning of Halfspaces with Malicious Noise

We study efficient PAC learning of homogeneous halfspaces in ℝ^d in the ...
research
05/14/2019

List-Decodable Linear Regression

We give the first polynomial-time algorithm for robust regression in the...
research
05/28/2023

On the Role of Noise in the Sample Complexity of Learning Recurrent Neural Networks: Exponential Gaps for Long Sequences

We consider the class of noisy multi-layered sigmoid recurrent neural ne...
research
02/24/2020

Learning Structured Distributions From Untrusted Batches: Faster and Simpler

We revisit the problem of learning from untrusted batches introduced by ...

Please sign up or login with your details

Forgot password? Click here to reset