A Generative Model for Sampling High-Performance and Diverse Weights for Neural Networks

05/07/2019
by   Lior Deutsch, et al.
1

Recent work on mode connectivity in the loss landscape of deep neural networks has demonstrated that the locus of (sub-)optimal weight vectors lies on continuous paths. In this work, we train a neural network that serves as a hypernetwork, mapping a latent vector into high-performance (low-loss) weight vectors, generalizing recent findings of mode connectivity to higher dimensional manifolds. We formulate the training objective as a compromise between accuracy and diversity, where the diversity takes into account trivial symmetry transformations of the target network. We demonstrate how to reduce the number of parameters in the hypernetwork by parameter sharing. Once learned, the hypernetwork allows for a computationally efficient, ancestral sampling of neural network weights, which we recruit to form large ensembles. The improvement in classification accuracy obtained by this ensembling indicates that the generated manifold extends in dimensions other than directions implied by trivial symmetries. For computational efficiency, we distill an ensemble into a single classifier while retaining generalization.

READ FULL TEXT

page 4

page 5

page 22

page 23

page 24

research
01/06/2018

Generating Neural Networks with Neural Networks

Hypernetworks are neural networks that transform a random input vector i...
research
06/20/2023

Traversing Between Modes in Function Space for Fast Ensembling

Deep ensemble is a simple yet powerful way to improve the performance of...
research
12/05/2019

Deep Ensembles: A Loss Landscape Perspective

Deep ensembles have been empirically shown to be a promising approach fo...
research
01/30/2019

HyperGAN: A Generative Model for Diverse, Performant Neural Networks

We introduce HyperGAN, a generative network that learns to generate all ...
research
05/23/2023

Transferring Learning Trajectories of Neural Networks

Training deep neural networks (DNNs) is computationally expensive, which...
research
08/22/2023

Mode Combinability: Exploring Convex Combinations of Permutation Aligned Models

We explore element-wise convex combinations of two permutation-aligned n...
research
02/20/2021

Learning Neural Network Subspaces

Recent observations have advanced our understanding of the neural networ...

Please sign up or login with your details

Forgot password? Click here to reset