HyperGAN: A Generative Model for Diverse, Performant Neural Networks

01/30/2019
by   Neale Ratzlaff, et al.
0

We introduce HyperGAN, a generative network that learns to generate all the weights within a deep neural network. HyperGAN employs a novel mixer to transform independent Gaussian noise into a latent space where dimensions are correlated, which is then transformed to generate weights in each layer of a deep neural network. We utilize an architecture that bears resemblance to generative adversarial networks, but we evaluate the likelihood of samples with a classification loss. This is equivalent to minimizing the KL-divergence between the generated network parameter distribution and an unknown true parameter distribution. We apply HyperGAN to classification, showing that HyperGAN can learn to generate parameters which solve the MNIST and CIFAR-10 datasets with competitive performance to fully supervised learning, while learning a rich distribution of effective parameters. We also show that HyperGAN can also provide better uncertainty than standard ensembles. This is evaluated by the ability of HyperGAN generated ensembles to detect out of distribution data as well as adversarial examples. We see that in addition to being highly accurate on inlier data, HyperGAN can provide reasonable uncertainty estimates.

READ FULL TEXT
research
06/07/2020

Uncertainty-Aware Deep Classifiers using Generative Models

Deep neural networks are often ignorant about what they do not know and ...
research
10/25/2017

mixup: Beyond Empirical Risk Minimization

Large deep neural networks are powerful, but exhibit undesirable behavio...
research
12/11/2017

Training Ensembles to Detect Adversarial Examples

We propose a new ensemble method for detecting and classifying adversari...
research
01/08/2019

A Comprehensive guide to Bayesian Convolutional Neural Network with Variational Inference

Artificial Neural Networks are connectionist systems that perform a give...
research
05/07/2019

A Generative Model for Sampling High-Performance and Diverse Weights for Neural Networks

Recent work on mode connectivity in the loss landscape of deep neural ne...
research
07/15/2023

Generative Meta-Learning Robust Quality-Diversity Portfolio

This paper proposes a novel meta-learning approach to optimize a robust ...
research
09/26/2022

Learning to Learn with Generative Models of Neural Network Checkpoints

We explore a data-driven approach for learning to optimize neural networ...

Please sign up or login with your details

Forgot password? Click here to reset