The Geometry of Deep Generative Image Models and its Applications

01/15/2021
by   Binxu Wang, et al.
16

Generative adversarial networks (GANs) have emerged as a powerful unsupervised method to model the statistical patterns of real-world data sets, such as natural images. These networks are trained to map random inputs in their latent space to new samples representative of the learned data. However, the structure of the latent space is hard to intuit due to its high dimensionality and the non-linearity of the generator, which limits the usefulness of the models. Understanding the latent space requires a way to identify input codes for existing real-world images (inversion), and a way to identify directions with known image transformations (interpretability). Here, we use a geometric framework to address both issues simultaneously. We develop an architecture-agnostic method to compute the Riemannian metric of the image manifold created by GANs. The eigen-decomposition of the metric isolates axes that account for different levels of image variability. An empirical analysis of several pretrained GANs shows that image variation around each position is concentrated along surprisingly few major axes (the space is highly anisotropic) and the directions that create this large variation are similar at different positions in the space (the space is homogeneous). We show that many of the top eigenvectors correspond to interpretable transforms in the image space, with a substantial part of eigenspace corresponding to minor transforms which could be compressed out. This geometric understanding unifies key previous results related to GAN interpretability. We show that the use of this metric allows for more efficient optimization in the latent space (e.g. GAN inversion) and facilitates unsupervised discovery of interpretable axes. Our results illustrate that defining the geometry of the GAN image manifold can serve as a general framework for understanding GANs.

READ FULL TEXT

page 1

page 10

page 11

page 12

page 13

page 14

page 16

page 18

research
02/15/2018

Inverting The Generator Of A Generative Adversarial Network (II)

Generative adversarial networks (GANs) learn a deep generative model tha...
research
11/23/2021

Tensor Component Analysis for Interpreting the Latent Space of GANs

This paper addresses the problem of finding interpretable directions in ...
research
12/23/2019

RPGAN: GANs Interpretability via Random Routing

In this paper, we introduce Random Path Generative Adversarial Network (...
research
04/06/2020

GANSpace: Discovering Interpretable GAN Controls

This paper describes a simple technique to analyze Generative Adversaria...
research
05/17/2021

Finding an Unsupervised Image Segmenter in Each of Your Deep Generative Models

Recent research has shown that numerous human-interpretable directions e...
research
07/20/2022

Interpreting Latent Spaces of Generative Models for Medical Images using Unsupervised Methods

Generative models such as Generative Adversarial Networks (GANs) and Var...
research
12/09/2020

GAN Steerability without optimization

Recent research has shown remarkable success in revealing "steering" dir...

Please sign up or login with your details

Forgot password? Click here to reset