Calibrate and Prune: Improving Reliability of Lottery Tickets Through Prediction Calibration

02/10/2020
by   Bindya Venkatesh, et al.
5

The hypothesis that sub-network initializations (lottery) exist within the initializations of over-parameterized networks, which when trained in isolation produce highly generalizable models, has led to crucial insights into network initialization and has enabled computationally efficient inferencing. In order to realize the full potential of these pruning strategies, particularly when utilized in transfer learning scenarios, it is necessary to understand the behavior of winning tickets when they might overfit to the dataset characteristics. In supervised and semi-supervised learning, prediction calibration is a commonly adopted strategy to handle such inductive biases in models. In this paper, we study the impact of incorporating calibration strategies during model training on the quality of the resulting lottery tickets, using several evaluation metrics. More specifically, we incorporate a suite of calibration strategies to different combinations of architectures and datasets, and evaluate the fidelity of sub-networks retrained based on winning tickets. Furthermore, we report the generalization performance of tickets across distributional shifts, when the inductive biases are explicitly controlled using calibration mechanisms. Finally, we provide key insights and recommendations for obtaining reliable lottery tickets, which we demonstrate to achieve improved generalization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/28/2019

Evaluating Lottery Tickets Under Distributional Shifts

The Lottery Ticket Hypothesis suggests large, over-parameterized neural ...
research
02/11/2021

When and How Mixup Improves Calibration

In many machine learning applications, it is important for the model to ...
research
05/12/2018

Active Semi-supervised Transfer Learning (ASTL) for Offline BCI Calibration

Single-trial classification of event-related potentials in electroenceph...
research
10/30/2019

Learn-By-Calibrating: Using Calibration as a Training Objective

Calibration error is commonly adopted for evaluating the quality of unce...
research
12/23/2022

Benchmark for Uncertainty Robustness in Self-Supervised Learning

Self-Supervised Learning (SSL) is crucial for real-world applications, e...
research
04/27/2020

Calibrating Healthcare AI: Towards Reliable and Interpretable Deep Predictive Models

The wide-spread adoption of representation learning technologies in clin...
research
07/19/2022

Assaying Out-Of-Distribution Generalization in Transfer Learning

Since out-of-distribution generalization is a generally ill-posed proble...

Please sign up or login with your details

Forgot password? Click here to reset