Deep Learning Through the Lens of Example Difficulty

06/17/2021
by   Robert J. N. Baldock, et al.
0

Existing work on understanding deep learning often employs measures that compress all data-dependent information into a few numbers. In this work, we adopt a perspective based on the role of individual examples. We introduce a measure of the computational difficulty of making a prediction for a given input: the (effective) prediction depth. Our extensive investigation reveals surprising yet simple relationships between the prediction depth of a given input and the model's uncertainty, confidence, accuracy and speed of learning for that data point. We further categorize difficult examples into three interpretable groups, demonstrate how these groups are processed differently inside deep models and showcase how this understanding allows us to improve prediction accuracy. Insights from our study lead to a coherent view of a number of separately reported phenomena in the literature: early layers generalize while later layers memorize; early layers converge faster and networks learn easy data and simple functions first.

READ FULL TEXT

page 29

page 31

page 32

page 33

page 34

page 40

page 41

page 42

research
06/05/2020

Uncertainty-Aware CNNs for Depth Completion: Uncertainty from Beginning to End

The focus in deep learning research has been mostly to push the limits o...
research
02/21/2020

An Investigation of Interpretability Techniques for Deep Learning in Predictive Process Analytics

This paper explores interpretability techniques for two of the most succ...
research
01/12/2023

Understanding Difficulty-based Sample Weighting with a Universal Difficulty Measure

Sample weighting is widely used in deep learning. A large number of weig...
research
09/25/2019

Attention-based Deep Tropical Cyclone Rapid Intensification Prediction

Rapid intensification (RI) is when a sudden and considerable increase in...
research
01/15/2023

EENet: Learning to Early Exit for Adaptive Inference

Budgeted adaptive inference with early exits is an emerging technique to...
research
06/26/2021

Interflow: Aggregating Multi-layer Feature Mappings with Attention Mechanism

Traditionally, CNN models possess hierarchical structures and utilize th...
research
04/10/2023

Simulated Annealing in Early Layers Leads to Better Generalization

Recently, a number of iterative learning methods have been introduced to...

Please sign up or login with your details

Forgot password? Click here to reset