Relative stability toward diffeomorphisms in deep nets indicates performance

05/06/2021
by   Leonardo Petrini, et al.
9

Understanding why deep nets can classify data in large dimensions remains a challenge. It has been proposed that they do so by becoming stable to diffeomorphisms, yet existing empirical measurements support that it is often not the case. We revisit this question by defining a maximum-entropy distribution on diffeomorphisms, that allows to study typical diffeomorphisms of a given norm. We confirm that stability toward diffeomorphisms does not strongly correlate to performance on four benchmark data sets of images. By contrast, we find that the stability toward diffeomorphisms relative to that of generic transformations R_f correlates remarkably with the test error ϵ_t. It is of order unity at initialization but decreases by several decades during training for state-of-the-art architectures. For CIFAR10 and 15 known architectures, we find ϵ_t≈ 0.2√(R_f), suggesting that obtaining a small R_f is important to achieve good performance. We study how R_f depends on the size of the training set and compare it to a simple model of invariant learning.

READ FULL TEXT

page 7

page 10

page 11

page 12

page 14

page 15

page 19

page 22

research
10/04/2022

How deep convolutional neural networks lose spatial information with training

A central question of machine learning is how deep nets manage to learn ...
research
11/20/2022

Instability in clinical risk stratification models using deep learning

While it has been well known in the ML community that deep learning mode...
research
07/22/2020

Compressing invariant manifolds in neural nets

We study how neural networks compress uninformative input space in model...
research
06/16/2021

Locality defeats the curse of dimensionality in convolutional teacher-student scenarios

Convolutional neural networks perform a local and translationally-invari...
research
04/12/2018

Learned Deformation Stability in Convolutional Neural Networks

Conventional wisdom holds that interleaved pooling layers in convolution...
research
06/14/2019

Explaining Landscape Connectivity of Low-cost Solutions for Multilayer Nets

Mode connectivity is a surprising phenomenon in the loss landscape of de...
research
03/09/2023

Inversion dynamics of class manifolds in deep learning reveals tradeoffs underlying generalisation

To achieve near-zero training error in a classification problem, the lay...

Please sign up or login with your details

Forgot password? Click here to reset