Winning Lottery Tickets in Deep Generative Models

10/05/2020
by   Neha Mukund Kalibhat, et al.
0

The lottery ticket hypothesis suggests that sparse, sub-networks of a given neural network, if initialized properly, can be trained to reach comparable or even better performance to that of the original network. Prior works in lottery tickets have primarily focused on the supervised learning setup, with several papers proposing effective ways of finding "winning tickets" in classification problems. In this paper, we confirm the existence of winning tickets in deep generative models such as GANs and VAEs. We show that the popular iterative magnitude pruning approach (with late rewinding) can be used with generative losses to find the winning tickets. This approach effectively yields tickets with sparsity up to 99 CIFAR and Celeb-A datasets. We also demonstrate the transferability of winning tickets across different generative models (GANs and VAEs) sharing the same architecture, suggesting that winning tickets have inductive biases that could help train a wide range of deep generative models. Furthermore, we show the practical benefits of lottery tickets in generative models by detecting tickets at very early stages in training called "early-bird tickets". Through early-bird tickets, we can achieve up to 88 operations (FLOPs) and 54 train large-scale generative models over tight resource constraints. These results out-perform existing early pruning methods like SNIP (Lee, Ajanthan, and Torr 2019) and GraSP (Wang, Zhang, and Grosse 2020). Our findings shed light towards existence of proper network initializations that could improve convergence and stability of generative models.

READ FULL TEXT

page 2

page 12

page 13

page 14

research
02/17/2016

Auxiliary Deep Generative Models

Deep generative models parameterized by neural networks have recently ac...
research
06/04/2022

Learning Robust Representations Of Generative Models Using Set-Based Artificial Fingerprints

With recent progress in deep generative models, the problem of identifyi...
research
09/06/2022

Unifying Generative Models with GFlowNets

There are many frameworks for deep generative modeling, each often prese...
research
11/08/2018

Bias and Generalization in Deep Generative Models: An Empirical Study

In high dimensional settings, density estimation algorithms rely crucial...
research
10/26/2020

Reducing the Computational Cost of Deep Generative Models with Binary Neural Networks

Deep generative models provide a powerful set of tools to understand rea...
research
06/22/2023

CEMSSL: A Unified Framework for Multi-Solution Inverse Kinematic Model Learning of Robot Arms with High-Precision Manipulation

Multiple solutions mainly originate from the existence of redundant degr...
research
05/17/2021

Finding an Unsupervised Image Segmenter in Each of Your Deep Generative Models

Recent research has shown that numerous human-interpretable directions e...

Please sign up or login with your details

Forgot password? Click here to reset