Rapid Training of Very Large Ensembles of Diverse Neural Networks

09/12/2018
by   Abdul Wasay, et al.
0

Ensembles of deep neural networks with diverse architectures significantly improve generalization accuracy. However, training such ensembles requires a large amount of computational resources and time as every network in the ensemble has to be separately trained. In practice, this restricts the number of different deep neural network architectures that can be included within an ensemble. We propose a new approach to address this problem. Our approach captures the structural similarity between members of a neural network ensemble and train it only once. Subsequently, this knowledge is transferred to all members of the ensemble using function-preserving transformations. Then, these ensemble networks converge significantly faster as compared to training from scratch. We show through experiments on CIFAR-10, CIFAR-100, and SVHN data sets that our approach can train large and diverse ensembles of deep neural networks achieving comparable accuracy to existing approaches in a fraction of their training time. In particular, our approach trains an ensemble of 100 variants of deep neural networks with diverse architectures up to 6 × faster as compared to existing approaches. This improvement in training cost grows linearly with the size of the ensemble.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/23/2022

Prune and Tune Ensembles: Low-Cost Ensemble Learning With Sparse Independent Subnetworks

Ensemble Learning is an effective method for improving generalization in...
research
02/17/2020

BatchEnsemble: an Alternative Approach to Efficient Ensemble and Lifelong Learning

Ensembles, where multiple neural networks are trained individually and t...
research
06/27/2022

Effective training-time stacking for ensembling of deep neural networks

Ensembling is a popular and effective method for improving machine learn...
research
04/09/2019

Intra-Ensemble in Neural Networks

Improving model performance is always the key problem in machine learnin...
research
07/25/2020

Economical ensembles with hypernetworks

Averaging the predictions of many independently trained neural networks ...
research
12/03/2020

Multiple Networks are More Efficient than One: Fast and Accurate Models via Ensembles and Cascades

Recent work on efficient neural network architectures focuses on discove...
research
03/15/2023

Bayesian Quadrature for Neural Ensemble Search

Ensembling can improve the performance of Neural Networks, but existing ...

Please sign up or login with your details

Forgot password? Click here to reset