Loss-Curvature Matching for Dataset Selection and Condensation

03/08/2023
by   Seungjae Shin, et al.
0

Training neural networks on a large dataset requires substantial computational costs. Dataset reduction selects or synthesizes data instances based on the large dataset, while minimizing the degradation in generalization performance from the full dataset. Existing methods utilize the neural network during the dataset reduction procedure, so the model parameter becomes important factor in preserving the performance after reduction. By depending upon the importance of parameters, this paper introduces a new reduction objective, coined LCMat, which Matches the Loss Curvatures of the original dataset and reduced dataset over the model parameter space, more than the parameter point. This new objective induces a better adaptation of the reduced dataset on the perturbed parameter region than the exact point matching. Particularly, we identify the worst case of the loss curvature gap from the local parameter region, and we derive the implementable upper bound of such worst-case with theoretical analyses. Our experiments on both coreset selection and condensation benchmarks illustrate that LCMat shows better generalization performances than existing baselines.

READ FULL TEXT

page 8

page 23

research
02/24/2011

New Worst-Case Upper Bound for X3SAT

The rigorous theoretical analyses of algorithms for exact 3-satisfiabili...
research
12/16/2020

Study on the Large Batch Size Training of Neural Networks Based on the Second Order Gradient

Large batch size training in deep neural networks (DNNs) possesses a wel...
research
10/10/2020

Regularizing Neural Networks via Adversarial Model Perturbation

Recent research has suggested that when training neural networks, flat l...
research
11/09/2020

Numerical Exploration of Training Loss Level-Sets in Deep Neural Networks

We present a computational method for empirically characterizing the tra...
research
07/08/2020

RicciNets: Curvature-guided Pruning of High-performance Neural Networks Using Ricci Flow

A novel method to identify salient computational paths within randomly w...
research
06/10/2020

Exploring the Vulnerability of Deep Neural Networks: A Study of Parameter Corruption

We argue that the vulnerability of model parameters is of crucial value ...
research
11/22/2021

Two step clustering for data reduction combining DBSCAN and k-means clustering

A novel combination of two widely-used clustering algorithms is proposed...

Please sign up or login with your details

Forgot password? Click here to reset