Galaxy Image Simulation Using Progressive GANs

09/26/2019
by   Mohamad Dia, et al.
EPFL
FHNW
0

In this work, we provide an efficient and realistic data-driven approach to simulate astronomical images using deep generative models from machine learning. Our solution is based on a variant of the generative adversarial network (GAN) with progressive training methodology and Wasserstein cost function. The proposed solution generates naturalistic images of galaxies that show complex structures and high diversity, which suggests that data-driven simulations using machine learning can replace many of the expensive model-driven methods used in astronomical data processing.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 4

02/12/2019

Progressively Growing Generative Adversarial Networks for High Resolution Semantic Segmentation of Satellite Images

Machine learning has proven to be useful in classification and segmentat...
05/02/2018

Text to Image Synthesis Using Generative Adversarial Networks

Generating images from natural language is one of the primary applicatio...
11/20/2018

Attributing Fake Images to GANs: Analyzing Fingerprints in Generated Images

Research in computer graphics has been in pursuit of realistic image gen...
03/08/2022

On generative models as the basis for digital twins

A framework is proposed for generative models as a basis for digital twi...
04/07/2022

Risk-based regulation for all: The need and a method for a wide adoption solution for data-driven inspection targeting

Access to data and data processing, including the use of machine learnin...
05/08/2020

On the use of Data-Driven Cost Function Identification in Parametrized NMPC

In this paper, a framework with complete numerical investigation is prop...
11/17/2017

3D Reconstruction of Incomplete Archaeological Objects Using a Generative Adversary Network

We introduce a data-driven approach to aid the repairing and conservatio...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Investigating the reasons behind the accelerated expansion of the universe is one of the main challenges in astronomy and modern cosmology. Future space missions, such as Euclid, will provide images of billions of galaxies in order to investigate the so-called dark matter and probe the geometry of the universe through the gravitational lensing effect. Due to the very large-scale of data provided by such missions, automated algorithms are needed for measurement and detection purposes. The training and calibration of such algorithms require simulated, or synthetic, images of galaxies that mimic the real observations and exhibit real morphologies.

In the case of weak lensing for instance, the accuracy of the shape measurement algorithms is very sensitive to any statistical bias induced by the Point Spread Function (PSF). Therefore, simulated images of galaxies with known ground-truth lensing are required to calibrate and detect any potential bias in the ensemble averages. Moreover, the training of automated strong lensing detector, such as deep learning architectures

(lensFinding2019), requires simulated images in order to mitigate class imbalance and avoid false-positive type of error in the current datasets.

2 Model-Driven v.s. Data-Driven Galaxy Image Simulation

The current approaches to simulate images of galaxies in the cosmology literature are mostly model-driven, or rule-based, approaches. These might involve the fitting of parametric analytic profiles (size, ellipticity, brightness, etc.) to the observed galaxies. This approach is usually unable to reproduce all the complex morphologies. An alternative, more expensive and often infeasible, model-driven approach is to start with high-quality galaxy images as the input of the simulation pipeline followed by a model that reproduces all the data acquisition effects (galsim2015).

Recently, several data-driven approaches have been investigated in order to generate synthetic images of galaxies via generative models used in machine learning (celeste2015; EnablingDE2016)

, mainly variational autoencoder (VAE)

(Kingma2013AutoEncodingVB) and generative adversarial network (GAN) (GAN2014). Such approaches have shown some promising preliminary results in generating galaxy images. Following this data-driven approach, and motivated by the success and recent impressive improvements in GANs, we have further investigated the use of such architecture in generating galaxy images.

3 Generative Adversarial Network

Unlike most of the generative models used in machine learning, GAN represents a novel approach that learns how to sample from the data distribution without explicitly tracking the parameters of the probability distribution function via traditional maximum likelihood estimation. The GAN architecture consists of two neural networks that compete against each other in a two-player minimax game. The first network is the “generator” that is responsible of generating the data, while the second network is the “discriminator” that represents the adversarial loss function. Despite its elegant mathematical formulation and the theoretical guarantees provided by a non-parametric analysis, the initial GAN architecture suffered from some practical implementation problems.

After the invention of GAN in 2014, a plethora of work have been done to improve the training (in terms of convergence and stability) and to obtain more realistic generated data (in terms of quality and diversity). Most of this effort was made towards improving the cost function and stabilizing the training methodology, which has recently lead to unprecedented results in generating synthetic images. Based on these recent advances, we have investigated variants of GAN that use the Wasserstein distance (Wasserstein2017) and the progressive training (karras2018progressive) on galaxy images provided by the Galaxy-Zoo dataset (galaxyZoo).

4 Proposed Architecture

Following (karras2018progressive), we employ blocks of convolutional layers to progressively build the generator and the discriminator as mirror images of each other (see Table 1). Intuitively speaking, training a small network to generate low-resolution images that capture the large-scale structure of the galaxies is an easier task than directly training a full network to generate high-resolution images with fine details. Hence, we start by training the network to generate low-resolution images (), we then progressively increase the resolution, in steps until resolution, by smoothly and synchronously adding blocks of convolutional layers to both the generator and discriminator. For the generator, each progress block is preceded by an up-sampling operation while a down-sampling operation follows each progress block in the discriminator.111

One can also use fractionally-strided and strided convolution respectively.

Such methodology leads to a more stable and faster training.

Generator Output Dimensions Latent Space

Input latent vector

Conv Conv Progress Conv Conv   Progress Conv Conv   Progress Conv Conv   Progress Conv Conv  RGB Extraction Conv
Discriminator Output Dimensions RGB Reading Input image Conv Progress Conv Conv   Progress Conv Conv   Progress Conv Conv   Progress Conv Conv  Cost Calculation Conv Conv Conv
Table 1: Blocks of convolutional layers added progressively for both the generator and the discriminator.

Moreover, the Wasserstein distance with gradient penalty (Wasserstein2017) is used as a cost function to mitigate the gradient problems. Furthermore, various normalization techniques are used to avoid the unhealthy competition between the generator and discriminator. In particular, we use “weight scaling” and “pixelwise feature normalization” as done in (He2015; AlexNet2012)

. In addition to that, the “mini-batch standard deviation

(karras2018progressive) is computed and incorporated in the cost function in order to favor diversity in the synthetic data.

4.1 Results

Our architecture is implemented in Python using PyTorch library and trained on a GPU system. The dataset is made of 6157 images of galaxies in RGB format. The images were centered at

resolution, normalized, and augmented using standard data augmentation techniques. A batch size was used with data loading workers.

The training was performed over a total of epochs and lasted less than hours. During the first epochs of training, the generator and discriminator were competing to reach the minimax equilibrium and the performance was fluctuating (in terms of their loss functions). The performance stabilized after that while the image quality continued to improve. After training, the discriminator, which plays the role of an adaptive loss function, is detached from the architecture and dismissed. The generator is then able to generate galaxy images starting from a latent space made of

standard Gaussian i.i.d. random variables.

Figure 1: Left: Progressive increase of resolution in four steps. Right: Diverse set of simulated images obtained by the proposed GAN architecture.

By changing the latent vector, we were able to obtain very diverse and high quality images of galaxies showing complex structures and morphologies (e.g. arm and disk features). Furthermore, the simulated images exhibited realistic effects (e.g. companion stars) as shown in Figure 1.

5 Future Work

We are planning to investigate the latent space of our GAN model in order to gain insight on the effect of each latent variable on the galaxies morphology. This will provide us with more control on the generation task and will permit to interpolate between the variables and perform latent space arithmetics. Furthermore, we are planning to incorporate the labels of the galaxies, when available, in a supervised or semi-supervised approach using variants of "Conditional GAN" architectures

(Odena2016ConditionalIS) in order to improve the quality of the generated images and guide the generator.

M. Dia and E. Savary would like to acknowledge funding from the SNSF (grant number 173716).

References