Learning a Single Neuron with Gradient Methods

01/15/2020
by   Gilad Yehudai, et al.
0

We consider the fundamental problem of learning a single neuron x σ(w^ x) using standard gradient methods. As opposed to previous works, which considered specific (and not always realistic) input distributions and activation functions σ(·), we ask whether a more general result is attainable, under milder assumptions. On the one hand, we show that some assumptions on the distribution and the activation function are necessary. On the other hand, we prove positive guarantees under mild assumptions, which go beyond those studied in the literature so far. We also point out and study the challenges in further strengthening and generalizing our results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/16/2022

Learning a Single Neuron for Non-monotonic Activation Functions

We study the problem of learning a single neuron 𝐱↦σ(𝐰^T𝐱) with gradient...
research
06/02/2021

Learning a Single Neuron with Bias Using Gradient Descent

We theoretically study the fundamental problem of learning a single neur...
research
05/29/2020

Agnostic Learning of a Single Neuron with Gradient Descent

We consider the problem of learning the best-fitting single neuron as me...
research
11/21/2020

Efficient Neural Network Implementation with Quadratic Neuron

Previous works proved that the combination of the linear neuron network ...
research
01/15/2021

A New Artificial Neuron Proposal with Trainable Simultaneous Local and Global Activation Function

The activation function plays a fundamental role in the artificial neura...
research
05/09/2021

Directional Convergence Analysis under Spherically Symmetric Distribution

We consider the fundamental problem of learning linear predictors (i.e.,...
research
06/13/2023

Robustly Learning a Single Neuron via Sharpness

We study the problem of learning a single neuron with respect to the L_2...

Please sign up or login with your details

Forgot password? Click here to reset