The NTK approximation is valid for longer than you think

05/22/2023
by   Enric Boix-Adserà, et al.
0

We study when the neural tangent kernel (NTK) approximation is valid for training a model with the square loss. In the lazy training setting of Chizat et al. 2019, we show that rescaling the model by a factor of α = O(T) suffices for the NTK approximation to be valid until training time T. Our bound is tight and improves on the previous bound of Chizat et al. 2019, which required a larger rescaling factor of α = O(T^2).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/29/2021

Local planar domination revisited

We show how to compute a 20-approximation of a minimum dominating set in...
research
11/20/2019

Online Spectral Approximation in Random Order Streams

This paper studies spectral approximation for a positive semidefinite ma...
research
04/23/2020

Directed Girth

It is known that a better than 2-approximation algorithm for the girth i...
research
11/22/2018

The Price of Uncertain Priors in Source Coding

We consider the problem of one-way communication when the recipient does...
research
12/16/2021

A model sufficiency test using permutation entropy

Using the ordinal pattern concept in permutation entropy, we propose a m...
research
06/23/2023

Valid inference after prediction

Recent work has focused on the very common practice of prediction-based ...
research
11/18/2022

Improved Approximations for Unrelated Machine Scheduling

We revisit two well-studied scheduling problems in the unrelated machine...

Please sign up or login with your details

Forgot password? Click here to reset