Memorization Through the Lens of Curvature of Loss Function Around Samples

07/11/2023
by   Isha Garg, et al.
0

Neural networks are overparametrized and easily overfit the datasets they train on. In the extreme case, it is shown that they can memorize a training set with fully randomized labels. We propose using the curvature of loss function around the training sample as a measure of its memorization, averaged over all training epochs. We use this to study the generalization versus memorization properties of different samples in popular image datasets. We visualize samples with the highest curvature of loss around them, and show that these visually correspond to long-tailed, mislabeled or conflicting samples. This analysis helps us find a, to the best of our knowledge, novel failure model on the CIFAR100 dataset, that of duplicated images with different labels. We also synthetically mislabel a proportion of the dataset by randomly corrupting the labels of a few samples, and show that sorting by curvature yields high AUROC values for identifying the mislabeled samples.

READ FULL TEXT

page 2

page 3

page 5

page 7

page 11

page 12

page 13

research
11/01/2020

Learning Euler's Elastica Model for Medical Image Segmentation

Image segmentation is a fundamental topic in image processing and has be...
research
07/08/2020

RicciNets: Curvature-guided Pruning of High-performance Neural Networks Using Ricci Flow

A novel method to identify salient computational paths within randomly w...
research
03/22/2023

Curvature-Balanced Feature Manifold Learning for Long-Tailed Classification

To address the challenges of long-tailed classification, researchers hav...
research
08/02/2022

Curvature-informed multi-task learning for graph networks

Properties of interest for crystals and molecules, such as band gap, ela...
research
06/07/2018

Dimensionality-Driven Learning with Noisy Labels

Datasets with significant proportions of noisy (incorrect) class labels ...
research
03/27/2023

Probing optimisation in physics-informed neural networks

A novel comparison is presented of the effect of optimiser choice on the...
research
07/20/2020

GarNet++: Improving Fast and Accurate Static3D Cloth Draping by Curvature Loss

In this paper, we tackle the problem of static 3D cloth draping on virtu...

Please sign up or login with your details

Forgot password? Click here to reset