Turing-Universal Learners with Optimal Scaling Laws

11/09/2021
by   Preetum Nakkiran, et al.
0

For a given distribution, learning algorithm, and performance metric, the rate of convergence (or data-scaling law) is the asymptotic behavior of the algorithm's test performance as a function of number of train samples. Many learning methods in both theory and practice have power-law rates, i.e. performance scales as n^-α for some α > 0. Moreover, both theoreticians and practitioners are concerned with improving the rates of their learning algorithms under settings of interest. We observe the existence of a "universal learner", which achieves the best possible distribution-dependent asymptotic rate among all learning algorithms within a specified runtime (e.g. O(n^2)), while incurring only polylogarithmic slowdown over this runtime. This algorithm is uniform, and does not depend on the distribution, and yet achieves best-possible rates for all distributions. The construction itself is a simple extension of Levin's universal search (Levin, 1973). And much like universal search, the universal learner is not at all practical, and is primarily of theoretical and philosophical interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2021

Learning Curve Theory

Recently a number of empirical "universal" scaling law papers have been ...
research
11/09/2020

A Theory of Universal Learning

How quickly can a given class of concepts be learned from examples? It i...
research
12/25/2017

Uniform Rates of Convergence of Some Representations of Extremes : a first approach

Uniform convergence rates are provided for asymptotic representations of...
research
05/19/2023

Tester-Learners for Halfspaces: Universal Algorithms

We give the first tester-learner for halfspaces that succeeds universall...
research
05/17/2017

AI, Native Supercomputing and The Revival of Moore's Law

Based on Alan Turing's proposition on AI and computing machinery, which ...
research
07/14/2021

Mapping Learning Algorithms on Data, a useful step for optimizing performances and their comparison

In the paper, we propose a novel methodology to map learning algorithms ...
research
02/14/2023

Cliff-Learning

We study the data-scaling of transfer learning from foundation models in...

Please sign up or login with your details

Forgot password? Click here to reset