Leveraging Simple Model Predictions for Enhancing its Performance

05/30/2019
by   Amit Dhurandhar, et al.
0

There has been recent interest in improving performance of simple models for multiple reasons such as interpretability, robust learning from small data, and deployment in memory constrained environments. In this paper, we propose a novel method SRatio that can utilize information from high performing complex models (viz. deep neural networks, boosted trees, random forests) to reweight a training dataset for a potentially low performing simple model such as a decision tree or a shallow network enhancing its performance. Our method also leverages the per sample hardness estimate of the simple model which is not the case with the prior works which primarily consider the complex model's confidences/predictions and is thus conceptually novel. Moreover, we generalize and formalize the concept of attaching probes to intermediate layers of a neural network, which was one of the main ideas in previous work, to other commonly used classifiers and incorporate this into our method. The benefit of these contributions is witnessed in the experiments where on 6 UCI datasets and CIFAR-10 we outperform competitors in a majority (16 out of 27) of the cases and tie for best performance in the remaining cases. In fact, in a couple of cases, we even approach the complex model's performance. We also show for popular loss functions such as cross-entropy loss, least squares loss, and hinge loss that the weighted loss minimized by simple models using our weighting is an upper bound on the loss of the complex model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2023

Reevaluating Loss Functions: Enhancing Robustness to Label Noise in Deep Learning Models

Large annotated datasets inevitably contain incorrect labels, which pose...
research
07/19/2018

Improving Simple Models with Confidence Profiles

In this paper, we propose a new method called ProfWeight for transferrin...
research
05/20/2018

Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels

Deep neural networks (DNNs) have achieved tremendous success in a variet...
research
02/27/2018

Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs

The loss functions of deep neural networks are complex and their geometr...
research
10/09/2022

Prediction intervals for neural network models using weighted asymmetric loss functions

We develop a novel and simple method to produce prediction intervals (PI...
research
12/13/2020

Demysifying Deep Neural Networks Through Interpretation: A Survey

Modern deep learning algorithms tend to optimize an objective metric, su...
research
07/07/2021

Mitigating Performance Saturation in Neural Marked Point Processes: Architectures and Loss Functions

Attributed event sequences are commonly encountered in practice. A recen...

Please sign up or login with your details

Forgot password? Click here to reset