An Efficient Tester-Learner for Halfspaces

02/28/2023
by   Aravind Gollakota, et al.
0

We give the first efficient algorithm for learning halfspaces in the testable learning model recently defined by Rubinfeld and Vasilyan (2023). In this model, a learner certifies that the accuracy of its output hypothesis is near optimal whenever the training set passes an associated test, and training sets drawn from some target distribution – e.g., the Gaussian – must pass the test. This model is more challenging than distribution-specific agnostic or Massart noise models where the learner is allowed to fail arbitrarily if the distributional assumption does not hold. We consider the setting where the target distribution is Gaussian (or more generally any strongly log-concave distribution) in d dimensions and the noise model is either Massart or adversarial (agnostic). For Massart noise, our tester-learner runs in polynomial time and outputs a hypothesis with (information-theoretically optimal) error 𝗈𝗉𝗍 + ϵ for any strongly log-concave target distribution. For adversarial noise, our tester-learner obtains error O(𝗈𝗉𝗍) + ϵ in polynomial time when the target distribution is Gaussian; for strongly log-concave distributions, we obtain Õ(𝗈𝗉𝗍) + ϵ in quasipolynomial time. Prior work on testable learning ignores the labels in the training set and checks that the empirical moments of the covariates are close to the moments of the base distribution. Here we develop new tests of independent interest that make critical use of the labels and combine them with the moment-matching approach of Gollakota et al. (2023). This enables us to simulate a variant of the algorithm of Diakonikolas et al. (2020) for learning noisy halfspaces using nonconvex SGD but in the testable learning setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2023

Tester-Learners for Halfspaces: Universal Algorithms

We give the first tester-learner for halfspaces that succeeds universall...
research
03/09/2023

Efficient Testable Learning of Halfspaces with Adversarial Label Noise

We give the first polynomial-time algorithm for the testable learning of...
research
02/15/2021

Efficient Learning with Arbitrary Covariate Shift

We give an efficient algorithm for learning a binary function in a given...
research
03/08/2018

Efficient Algorithms for Outlier-Robust Regression

We give the first polynomial-time algorithm for performing linear or pol...
research
02/14/2020

Local continuity of log-concave projection, with applications to estimation under model misspecification

The log-concave projection is an operator that maps a d-dimensional dist...
research
04/14/2022

Testing distributional assumptions of learning algorithms

There are many important high dimensional function classes that have fas...
research
12/06/2022

A Strongly Polynomial Algorithm for Approximate Forster Transforms and its Application to Halfspace Learning

The Forster transform is a method of regularizing a dataset by placing i...

Please sign up or login with your details

Forgot password? Click here to reset