Why M Heads are Better than One: Training a Diverse Ensemble of Deep Networks

11/19/2015
by   Stefan Lee, et al.
0

Convolutional Neural Networks have achieved state-of-the-art performance on a wide range of tasks. Most benchmarks are led by ensembles of these powerful learners, but ensembling is typically treated as a post-hoc procedure implemented by averaging independently trained models with model variation induced by bagging or random initialization. In this paper, we rigorously treat ensembling as a first-class problem to explicitly address the question: what are the best strategies to create an ensemble? We first compare a large number of ensembling strategies, and then propose and evaluate novel strategies, such as parameter sharing (through a new family of models we call TreeNets) as well as training under ensemble-aware and diversity-encouraging losses. We demonstrate that TreeNets can improve ensemble performance and that diverse ensembles can be trained end-to-end under a unified loss, achieving significantly higher "oracle" accuracies than classical ensembles.

READ FULL TEXT

page 7

page 13

page 14

page 15

research
05/24/2019

EnsembleNet: End-to-End Optimization of Multi-headed Models

Ensembling is a universally useful approach to boost the performance of ...
research
06/20/2020

Collective Learning by Ensembles of Altruistic Diversifying Neural Networks

Combining the predictions of collections of neural networks often outper...
research
01/26/2023

Joint Training of Deep Ensembles Fails Due to Learner Collusion

Ensembles of machine learning models have been well established as a pow...
research
09/07/2020

ECOC as a Method of Constructing Deep Convolutional Neural Network Ensembles

Deep neural networks have enhanced the performance of decision making sy...
research
03/24/2016

Deep Extreme Feature Extraction: New MVA Method for Searching Particles in High Energy Physics

In this paper, we present Deep Extreme Feature Extraction (DEFE), a new ...
research
06/24/2016

Stochastic Multiple Choice Learning for Training Diverse Deep Ensembles

Many practical perception systems exist within larger processes that inc...
research
01/23/2023

DeepFEL: Deep Fastfood Ensemble Learning for Histopathology Image Analysis

Computational pathology tasks have some unique characterises such as mul...

Please sign up or login with your details

Forgot password? Click here to reset