Improving Calibration in Mixup-trained Deep Neural Networks through Confidence-Based Loss Functions

03/22/2020
by   Juan Maroñas, et al.
5

Deep Neural Networks (DNN) represent the state of the art in many tasks. However, due to their overparameterization, their generalization capabilities are in doubt and are still under study. Consequently, DNN can overfit and assign overconfident predictions, as they tend to learn highly oscillating decision thresholds. This has been shown to affect the calibration of the confidences assigned to unseen data. Data Augmentation (DA) strategies have been proposed to overcome some of these limitations. One of the most popular is Mixup, which has shown a great ability to improve the accuracy of these models. Recent work has provided evidence that Mixup also improves the uncertainty quantification and calibration of DNN. In this work, we argue and provide empirical evidence that, due to its fundamentals, Mixup does not necessarily improve calibration. Based on our observations we propose a new loss function that improves the calibration, and also sometimes the accuracy. Our loss is inspired by Bayes decision theory and introduces a new training framework for designing losses for probabilistic modelling. We provide state-of-the-art accuracy with consistent improvements in calibration performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/30/2021

Soft Calibration Objectives for Neural Networks

Optimal decision making requires that classifiers produce uncertainty es...
research
08/23/2019

Calibration of Deep Probabilistic Models with Decoupled Bayesian Neural Networks

Deep Neural Networks (DNNs) have achieved state-of-the-art accuracy perf...
research
03/25/2022

A Stitch in Time Saves Nine: A Train-Time Regularizing Loss for Improved Neural Network Calibration

Deep Neural Networks ( DNN s) are known to make overconfident mistakes, ...
research
06/17/2021

On the Dark Side of Calibration for Modern Neural Networks

Modern neural networks are highly uncalibrated. It poses a significant c...
research
06/30/2023

Impact of Noise on Calibration and Generalisation of Neural Networks

Noise injection and data augmentation strategies have been effective for...
research
05/05/2020

Designing Accurate Emulators for Scientific Processes using Calibration-Driven Deep Models

Predictive models that accurately emulate complex scientific processes c...
research
11/04/2020

Extracting Chemical-Protein Interactions via Calibrated Deep Neural Network and Self-training

The extraction of interactions between chemicals and proteins from sever...

Please sign up or login with your details

Forgot password? Click here to reset