Sparse Mutation Decompositions: Fine Tuning Deep Neural Networks with Subspace Evolution

02/12/2023
by   Tim Whitaker, et al.
0

Neuroevolution is a promising area of research that combines evolutionary algorithms with neural networks. A popular subclass of neuroevolutionary methods, called evolution strategies, relies on dense noise perturbations to mutate networks, which can be sample inefficient and challenging for large models with millions of parameters. We introduce an approach to alleviating this problem by decomposing dense mutations into low-dimensional subspaces. Restricting mutations in this way can significantly reduce variance as networks can handle stronger perturbations while maintaining performance, which enables a more controlled and targeted evolution of deep networks. This approach is uniquely effective for the task of fine tuning pre-trained models, which is an increasingly valuable area of research as networks continue to scale in size and open source models become more widely available. Furthermore, we show how this work naturally connects to ensemble learning where sparse mutations encourage diversity among children such that their combined predictions can reliably improve performance. We conduct the first large scale exploration of neuroevolutionary fine tuning and ensembling on the notoriously difficult ImageNet dataset, where we see small generalization improvements with only a single evolutionary generation using nearly a dozen different deep neural network architectures.

READ FULL TEXT

page 1

page 2

page 5

research
02/06/2018

Digital Watermarking for Deep Neural Networks

Although deep neural networks have made tremendous progress in the area ...
research
07/02/2018

Make (Nearly) Every Neural Network Better: Generating Neural Network Ensembles by Weight Parameter Resampling

Deep Neural Networks (DNNs) have become increasingly popular in computer...
research
07/20/2022

Pretraining a Neural Network before Knowing Its Architecture

Training large neural networks is possible by training a smaller hyperne...
research
01/15/2017

Embedding Watermarks into Deep Neural Networks

Deep neural networks have recently achieved significant progress. Sharin...
research
03/03/2017

Large-Scale Evolution of Image Classifiers

Neural networks have proven effective at solving difficult problems but ...
research
11/05/2020

Identifying and interpreting tuning dimensions in deep networks

In neuroscience, a tuning dimension is a stimulus attribute that account...
research
08/08/2023

Asynchronous Evolution of Deep Neural Network Architectures

Many evolutionary algorithms (EAs) take advantage of parallel evaluation...

Please sign up or login with your details

Forgot password? Click here to reset