A faster and simpler algorithm for learning shallow networks

07/24/2023
by   Sitan Chen, et al.
0

We revisit the well-studied problem of learning a linear combination of k ReLU activations given labeled examples drawn from the standard d-dimensional Gaussian measure. Chen et al. [CDG+23] recently gave the first algorithm for this problem to run in poly(d,1/ε) time when k = O(1), where ε is the target error. More precisely, their algorithm runs in time (d/ε)^quasipoly(k) and learns over multiple stages. Here we show that a much simpler one-stage version of their algorithm suffices, and moreover its runtime is only (d/ε)^O(k^2).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/25/2023

Most Neural Networks Are Almost Learnable

We present a PTAS for learning random constant-depth networks. We show t...
research
06/23/2020

Robust Gaussian Covariance Estimation in Nearly-Matrix Multiplication Time

Robust covariance estimation is the following, well-studied problem in h...
research
02/24/2020

Learning Structured Distributions From Untrusted Batches: Faster and Simpler

We revisit the problem of learning from untrusted batches introduced by ...
research
03/02/2020

ETH-Tight Algorithms for Long Path and Cycle on Unit Disk Graphs

We present an algorithm for the extensively studied Long Path and Long C...
research
08/10/2021

A D-competitive algorithm for the Multilevel Aggregation Problem with Deadlines

In this paper, we consider the multi-level aggregation problem with dead...
research
03/02/2020

Connected Components on a PRAM in Log Diameter Time

We present an O(log d + loglog_m/n n)-time randomized PRAM algorithm for...
research
10/15/2018

An Illuminating Algorithm for the Light Bulb Problem

The Light Bulb Problem is one of the most basic problems in data analysi...

Please sign up or login with your details

Forgot password? Click here to reset