Rethinking Data Distillation: Do Not Overlook Calibration

07/24/2023
by   Dongyao Zhu, et al.
0

Neural networks trained on distilled data often produce over-confident output and require correction by calibration methods. Existing calibration methods such as temperature scaling and mixup work well for networks trained on original large-scale data. However, we find that these methods fail to calibrate networks trained on data distilled from large source datasets. In this paper, we show that distilled data lead to networks that are not calibratable due to (i) a more concentrated distribution of the maximum logits and (ii) the loss of information that is semantically meaningful but unrelated to classification tasks. To address this problem, we propose Masked Temperature Scaling (MTS) and Masked Distillation Training (MDT) which mitigate the limitations of distilled data and achieve better calibration results while maintaining the efficiency of dataset distillation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/23/2014

Committees of deep feedforward networks trained with few data

Deep convolutional neural networks are known to give good results on ima...
research
03/04/2023

IKD+: Reliable Low Complexity Deep Models For Retinopathy Classification

Deep neural network (DNN) models for retinopathy have estimated predicti...
research
03/22/2022

Dataset Distillation by Matching Training Trajectories

Dataset distillation is the task of synthesizing a small dataset such th...
research
07/31/2022

Adaptive Temperature Scaling for Robust Calibration of Deep Neural Networks

In this paper, we study the post-hoc calibration of modern neural networ...
research
08/16/2023

Dual-Branch Temperature Scaling Calibration for Long-Tailed Recognition

The calibration for deep neural networks is currently receiving widespre...
research
01/03/2021

Neural network algorithm and its application in temperature control of distillation tower

Distillation process is a complex process of conduction, mass transfer a...
research
11/18/2022

Layer-Stack Temperature Scaling

Recent works demonstrate that early layers in a neural network contain u...

Please sign up or login with your details

Forgot password? Click here to reset