Cliff-Learning

02/14/2023
by   Tony T. Wang, et al.
0

We study the data-scaling of transfer learning from foundation models in the low-downstream-data regime. We observe an intriguing phenomenon which we call cliff-learning. Cliff-learning refers to regions of data-scaling laws where performance improves at a faster than power law rate (i.e. regions of concavity on a log-log scaling plot). We conduct an in-depth investigation of foundation-model cliff-learning and study toy models of the phenomenon. We observe that the degree of cliff-learning reflects the degree of compatibility between the priors of a learning algorithm and the task being learned.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2021

Learning Curve Theory

Recently a number of empirical "universal" scaling law papers have been ...
research
08/17/2022

Understanding Scaling Laws for Recommendation Models

Scale has been a major driving force in improving machine learning perfo...
research
11/01/2022

Transfer Learning with Kernel Methods

Transfer learning refers to the process of adapting a model trained on a...
research
06/22/2023

TaCA: Upgrading Your Visual Foundation Model with Task-agnostic Compatible Adapter

Visual foundation models like CLIP excel in learning feature representat...
research
06/29/2022

Beyond neural scaling laws: beating power law scaling via data pruning

Widely observed neural scaling laws, in which error falls off as a power...
research
11/09/2021

Turing-Universal Learners with Optimal Scaling Laws

For a given distribution, learning algorithm, and performance metric, th...
research
04/21/2023

The simplicity bubble effect as a zemblanitous phenomenon in learning systems

The ubiquity of Big Data and machine learning in society evinces the nee...

Please sign up or login with your details

Forgot password? Click here to reset