1 Introduction
Finding new chemical compounds with desired properties is a challenging task with important applications such as de novo drug design (Schneider & Fechner, 2005). The space of synthesizable molecules is vast and search in this space proves to be very difficult, mostly owing to its discrete nature.
Recent progress in the development of deep generative models has spawned a range of promising proposals to address this issue. Most works in this area (GómezBombarelli et al., 2016; Kusner et al., 2017; Guimaraes et al., 2017; Dai et al., 2018) make use of a socalled SMILES representation (Weininger, 1988)
of molecules: a stringbased representation derived from molecular graphs. Recurrent neural networks (RNNs) are ideal candidates for these representations and consequently, most recent works follow the recipe of applying RNNbased generative models on this type of encoding. Stringbased representations of molecules, however, have certain disadvantages: RNNs have to spend capacity on learning both the syntactic rules and the order ambiguity of the representation. Besides, this is approach not applicable to generic (nonmolecular) graphs.
SMILES strings are generated from a graphbased representation of molecules, thereby working in the original graph space has the benefit of removing additional overhead. With recent progress in the area of deep learning on graphs
(Bronstein et al., 2017; Hamilton et al., 2017), training deep generative models directly on graph representations becomes a feasible alternative that has been explored in a range of recent works (Kipf & Welling, 2016b; Johnson, 2017; Grover et al., 2017; Li et al., 2018b; Simonovsky & Komodakis, 2018; You et al., 2018).Likelihoodbased methods for molecular graph generation (Li et al., 2018b; Simonovsky & Komodakis, 2018) however, either require providing a fixed (or randomly chosen) ordered representation of the graph or an expensive graph matching procedure to evaluate the likelihood of a generated molecule, as the evaluation of all possible node orderings is prohibitive already for graphs of small size.
In this work, we sidestep this issue by utilizing implicit, likelihoodfree methods, in particular, a generative adversarial network (GAN) (Goodfellow et al., 2014) that we adapt to work directly on graph representations. We further utilize a reinforcement learning (RL) objective similar to ORGAN (Guimaraes et al., 2017) to encourage the generation of molecules with particular properties.
Our molecular GAN (MolGAN) model (outlined in Figure 1) is the first to address the generation of graphstructured data in the context of molecular synthesis using GANs (Goodfellow et al., 2014). The generative model of MolGAN predicts discrete graph structure at once (i.e., nonsequentially) for computational efficiency, although sequential variants are possible in general. MolGAN further utilizes a permutationinvariant discriminator and reward network (for RLbased optimization towards desired chemical properties) based on graph convolution layers (Bruna et al., 2013; Duvenaud et al., 2015; Kipf & Welling, 2016a; Schlichtkrull et al., 2017) that both operate directly on graphstructured representations.
2 Background
2.1 Molecules as graphs
Most previous deep generative models for molecular data (GómezBombarelli et al., 2016; Kusner et al., 2017; Guimaraes et al., 2017; Dai et al., 2018) resort to generating SMILES representations of molecules. The SMILES syntax, however, is not robust to small changes or mistakes, which can result in the generation of invalid or drastically different structures. Grammar VAEs (Kusner et al., 2017) alleviate this problem by constraining the generative process to follow a particular grammar.
Operating directly in the space of graphs has recently been shown to be a viable alternative for generative modeling of molecular data (Li et al., 2018b; Simonovsky & Komodakis, 2018) with the added benefit that all generated outputs are valid graphs (but not necessarily valid molecules).
We consider that each molecule can be represented by an undirected graph with a set of edges and nodes . Each atom corresponds to a node that is associated with a dimensional onehot vector , indicating the type of the atom. We further represent each atomic bond as an edge associated with a bond type . For a molecular graph with nodes, we can summarize this representation in a node feature matrix
and an adjacency tensor
where is a onehot vector indicating the type of the edge between and .2.2 Implicit vs. likelihoodbased methods
Likelihoodbased methods such as the variational autoencoder (VAE) (Kingma & Welling, 2013; Rezende et al., 2014) typically allow for easier and more stable optimization than implicit generative models such as a GAN (Goodfellow et al., 2014). When generating graphstructured data, however, we wish to be invariant to reordering of nodes in the (ordered) matrix representation of the graph, which requires us to either perform a prohibitively expensive graph matching procedure (Simonovsky & Komodakis, 2018) or to evaluate the likelihood for all possible node permutations explicitly.
By resorting to implicit generative models, in particular to the GAN framework, we circumvent the need for an explicit likelihood. While the discriminator of the GAN can be made invariant to node ordering by utilizing graph convolutions (Bruna et al., 2013; Duvenaud et al., 2015; Kipf & Welling, 2016a) and a node aggregation operator (Li et al., 2016), the generator still has to decide on a specific node ordering when generating a graph. Since we do not provide a likelihood, however, the generator is free to choose any suitable ordering for the task at hand. We provide a brief introduction to GANs in the following.
Generative adversarial networks
GANs (Goodfellow et al., 2014) are implicit generative models in the sense that they allow for inference of model parameters without requiring one to specify a likelihood.
A GAN consist of two main components: a generative model , that learns a map from a prior to the data distribution to sample new datapoints, and a discriminative model , that learns to classify whether samples came from the data distribution rather than from
. Those two models are implemented as neural networks and trained simultaneously with stochastic gradient descent (SGD).
and have different objectives, and they can be seen as two players in a minimax game(1) 
where tries to generate samples to fool the discriminator and tries to differentiate samples correctly. To prevent undesired behaviour such as mode collapse (Salimans et al., 2016) and to stabilize learning, we use minibatch discrimination (Salimans et al., 2016) and improved WGAN (Gulrajani et al., 2017), an alternative and more stable GAN model that minimizes a better suited divergence.
Improved WGAN
WGANs (Arjovsky et al., 2017)
minimize an approximation of the Earth Mover (EM) distance (also know as Wasserstein1 distance) defined between two probability distributions. Formally, the Wasserstein distance between
and , using the KantorovichRubinstein duality is(2) 
where in the case of WGAN, is the empirical distribution and is the generator distribution. Note that the supremum is over all the KLipschitz functions for some .
Gulrajani et al. (2017) introduce a gradient penalty as an alternative soft constraint on the 1Lipschitz continuity as an improvement upon the gradient clipping
scheme from the original WGAN. The loss with respect to the generator remains the same as in WGAN, but the loss function with respect to the discriminator is modified to be
(3) 
where
is a hyperparameter (we use
as in the original paper), is a sampled linear combination between and with , thus with .2.3 Deterministic policy gradients
A GAN generator learns a transformation from a prior distribution to the data distribution. Thus, generated samples resemble data samples. However, in de novo drug design methods, we are not only interested in generating chemically valid compounds, but we want them to have some useful property (e.g., to be easily synthesizable). Therefore, we also optimize the generation process towards some nondifferentiable metrics using reinforcement learning.
In reinforcement learning, a stochastic policy is represented by which is a parametric probability distribution in that selects a categorical action conditioned on an environmental state . Conversely, a deterministic policy is represented by which deterministically outputs an action.
In initial experiments, we explored using REINFORCE (Williams, 1992) in combination with a stochastic policy that models graph generation as a set of categorical choices (actions). However, we found that it converged poorly due to the high dimensional action space when generating graphs at once. We instead base our method on a deterministic policy gradient algorithm which is known to perform well in highdimensional action spaces (Silver et al., 2014). In particular, we employ a version of deep deterministic policy gradient (DDPG) introduced by Lillicrap et al. (2015), an offpolicy actorcritic algorithm that uses deterministic policy gradients to maximize an approximation of the expected future reward.
In our case, the policy is the GAN generator which takes a sample for the prior as input, instead of an environmental state , and it outputs a molecular graph as an action (). Moreover, we do not model episodes, so there is no need to assess the quality of a stateaction combination since it does only depend on the graph . Therefore, we introduce a learnable and differentiable approximation of the reward function that predicts the immediate reward, and we train it via a mean squared error objective based on the real reward provided by an external system (e.g., the synthesizability score of a molecule). Then, we train the generator maximizing the predicted reward via which, being differentiable, provides a gradient to the policy towards the desired metric. Notice that, differently form DDPG, we do not use experience replay or target networks (see original work).
3 Model
The MolGAN architecture (Figure 2) consists of three main components: a generator , a discriminator and a reward network .
The generator takes a sample from a prior distribution and generates an annotated graph representing a molecule. Nodes and edges of are associated with annotations denoting atom type and bond type respectively. The discriminator takes both samples from the dataset and the generator and learns to distinguish them. Both and are trained using improved WGAN such that the generator learns to match the empirical distribution and eventually outputs valid molecules.
The reward network is used to approximate the reward function of a sample and optimize molecule generation towards nondifferentiable metrics using reinforcement learning. Dataset and generated samples are inputs of , but, differently from the discriminator, it assigns scores to them (e.g., how likely the generated molecule is to be soluble in water). The reward network learns to assign a reward to each molecule to match a score provided by an external software^{1}^{1}1We used the RDKit OpenSource Cheminformatics Software: http://www.rdkit.org.. Notice that, when MolGAN outputs a nonvalid molecule, it is not possible to assign a reward since the graph is not even a compound. Thus, for invalid molecular graphs, we assign zero rewards.
The discriminator is then trained using the WGAN objective while the generator uses a linear combination of the WGAN loss and the RL loss:
(4) 
where is a hyperparameter that regulates the tradeoff between the two components.
3.1 Generator
takes Ddimensional vectors
sampled from a standard normal distribution
and outputs graphs. While recent works have shown that it is feasible to generate graphs of small size by using an RNNbased generative model (Johnson, 2017; You et al., 2018; Li et al., 2018a, b)we, for simplicity, utilize a generative model that predicts the entire graph at once using a simple multilayer perceptron (MLP), as similarly done in
Simonovsky & Komodakis (2018). While this limits our study to graphs of a prechosen maximum size, we find that it is significantly faster and easier to optimize.We restrict the domain to graphs of a limited number of nodes and, for each , outputs two continuous and dense objects: that defines atom types and that defines bonds types (see Section 2.1). Both and have a probabilistic interpretation since each node and edge type is represented with probabilities of categorical distributions over types. To generate a molecule we obtain discrete, sparse objects and via categorical sampling from and , respectively. We overload notation and also represent samples from the dataset with binary and .
As this discretization process is nondifferentiable, we explore three model variations to allow for gradientbased training: We can i) use the continuous objects and directly during the forward pass (i.e., and ), ii) add Gumbel noise to and before passing them to and in order to make the generation stochastic while still forwarding continuous objects (i.e., and ), or iii) use a straightthrough gradient based on categorical reparameterization with the GumbelSoftmax (Jang et al., 2016; Maddison et al., 2016), taht is we use a sample form a categorical distribution during the forward pass (i.e., and ) and the continuous relaxed values (i.e., the original and ) in the backward pass.
3.2 Discriminator and reward network
Both the discriminator and the reward network receive a graph as input, and they output a scalar value each. We choose the same architecture for both networks but do not share parameters between them. A series of graph convolution layers convolve node signals using the graph adjacency tensor . We base our model on RelationalGCN (Schlichtkrull et al., 2017), a convolutional network for graphs with support for multiple edge types. At every layer, feature representations of nodes are convolved/propagated according to:
(5) 
where is the signal of the node at layer and
is a linear transformation function that acts as a selfconnection between layers. We further utilize an edge typespecific affine function
for each layer. denotes the set of neighbors for node . The normalization factor ensures that activations are on a similar scale irrespective of the number of neighbors.After several layers of propagation via graph convolutions, following Li et al. (2016) we aggregate node embeddings into a graph level representation vector as
(6) 
where
is the logistic sigmoid function,
and are MLPs with a linear output layer and denotes elementwise multiplication. Then, is a vector representation of the graph and it is further processed by an MLP to produce a graph level scalar output for the discriminator and for the reward network.4 Related work
ObjectiveReinforced Generative Adversarial Networks (ORGAN) by Guimaraes et al. (2017) is the closest related work to ours. Their model relies on SeqGAN (Yu et al., 2017) to adversarially learn to output sequences while optimizing towards chemical metrics with REINFORCE (Williams, 1992). The main differences from our approach is that they model sequences of SMILES as molecular representations instead of graphs, and their RL component uses REINFORCE while we use DDPG. Segler et al. (2017) also employs RL for drug discovery by searching retrosynthetic routes using Monte Carlo Tree Search (MCTS) in combination with an expansion policy network.
Several other works have explored training generative models on SMILES representations of molecules: CharacterVAE (GómezBombarelli et al., 2016) is the first such model that is based on a VAE with recurrent encoder and decoder networks. GrammarVAE (Kusner et al., 2017) and SDVAE (Dai et al., 2018) constrain the decoding process to follow particular syntactic and semantic rules.
A related line of research considers training deep generative models to output graphstructured data directly. Several works explored autoencoder architectures utilizing graph convolutions for link prediction within graphs (Kipf & Welling, 2016b; Grover et al., 2017; Davidson et al., 2018). Johnson (2017); Li et al. (2018b); You et al. (2018); Li et al. (2018a) on the other hand developed likelihoodbased methods to directly output graphs of arbitrary size in a sequential manner. Several related works have explored extending VAEs to generate graphs directly, examples include the GraphVAE (Simonovsky & Komodakis, 2018), Junction Tree VAE (Jin et al., 2018) and the NeVAE (Samanta et al., 2018) model.
For link prediction within graphs, a range of adversarial methods have been introduced in the literature (Minervini et al., 2017; Wang et al., 2018; Bojchevski et al., 2018). This class of models, however, is not suitable to generate molecular graphs from scratch, which makes direct comparison infeasible.
5 Experiments
We compare MolGAN against recent neural networkbased drug generation models in a range of experiments on established benchmarks using the QM9 (Ramakrishnan et al., 2014) chemical database. We first focus on studying the effect of the parameter to find the best tradeoff between the GAN and RL objective (see Section 5.1). We then compare MolGAN with ORGAN (Guimaraes et al., 2017) since it is the most related work to ours: ORGAN is a sequential generative model operating on SMILES representations, optimizing towards several chemical properties with an RL objective (see Section 5.2
). We also compare our model against variational autoencoding methods (Section
5.3) such as CharacterVAE (GómezBombarelli et al., 2016), GrammarVAE (Kusner et al., 2017), as well as a recent graphbased generative model: GraphVAE (Simonovsky & Komodakis, 2018).Dataset
Generator architecture
The generator architecture is fixed for all experiments. We use as the maximum number of nodes,
as the number of atom types (C, O, N, F, and one padding symbol), and
as the number of bond types (single, double, triple and no bond). These dimensionalities are enough to cover all molecules in QM9. The generator takes a 32dimensional vector sampled from a standard normal distribution and process it with a 3layer MLP of hidden units respectively, withas activation functions. Eventually, the last layer is linearly projected to match
and dimensions and normalized in their last dimension with a operation ().Discriminator and reward network architecture
Both networks use a RelationalGCN encoder (see Eq. 3.2) with two layers and hidden units, respectively, to process the input graphs. Subsequently, we compute a 128dimensional graphlevel representation (see Eq. 3.2) further processed by a 2layer MLP of dimensions and with as hidden layer activation function. In the reward network, we further use a sigmoid activation function on the output.
Evaluation measures
We measure the following statistics as defined in Samanta et al. (2018): validity, novelty, and uniqueness. Validity is defined as the ratio between the number of valid and all generated molecules. Novelty measures the ratio between the set of valid samples that are not in the dataset and the total number of valid samples. Finally, uniqueness is defined as the ratio between the number of unique samples and valid samples and it measures the degree of variety during sampling.
Training
In all experiments, we use a batch size of 32 and train using the Adam (Kingma & Ba, 2014) optimizer with a learning rate of . For each setting, we employ a grid search over dropout rates (Srivastava et al., 2014) as well as over discretization variations (as described in Section 3.1). We always report the results of the best model depending on what we are optimizing for (e.g., when optimizing solubility we report the model with the highest solubility score – when no metric is optimized we report the model with the highest sum of individual scores). Although the use of WGAN should prevent, to some extent, undesired behaviors like mode collapse (Salimans et al., 2016), we notice that our models suffer from that problem. We leave addressing this issue for future work. As a simple countermeasure, we employ early stopping to avoid completely collapsed modes. In particular, we use the unique score to measure the degree of collapse of our models since it intrinsically indicates how much variety there is in the generation process. We set an arbitrary threshold of 2% under which we consider a model to be collapsed and stop training.
During early stages of our work, we noticed that the reward network needs several epochs of pretraining before being used to propagate the gradient to the generator, otherwise the generator easily diverges. We think this happens because at the beginning of the training,
does not predict the reward accurately and then it does not optimize the generator well. Therefore, in each experiment, we train the generator for the first half of the epochs without the RL component, but using the WGAN objective only. We train the reward network during these epochs, but no RL loss is used to train the generator. For the second half of the epochs we use the combined loss in Equation 4.5.1 Effect of
As in Guimaraes et al. (2017), the hyperparameter regulates a tradeoff between maximizing the desired objective and regulating the generator distribution to match the data distribution. We study the effects of on the solubility metric (see Section 5.2 for more details).
We train for 300 epochs on the 5k subset of QM9 used in Guimaraes et al. (2017). We use the best parameter – determined via the model with the maximum sum of valid, unique, novel, and solubility scores – on all other experiments (Section 5.2 and 5.3) without doing any further search.
Results
We report results in Table 1. We observe a clear trend towards higher validity scores for lower values of . This is likely due to the implicit optimization of valid molecules since invalid ones receive zero reward during training. Therefore, if the RL loss component is strong, the generator is optimized to generate mostly valid molecular graphs. Conversely, it appears that does not mainly affect the unique and novel scores. Notice that these scores are not optimized, neither directly nor indirectly, and therefore they are a result of model architecture, hyperparameters, and training procedure. Indeed, the unique score is always close to 2% (which is our threshold) indicating that models appear to collapse (even in the RL only case) if we do not apply early stopping.
Since controls the tradeoff between the WGAN and RL losses, it is not surprising that (i.e., only RL in the second half of training) results in the highest valid and solubility scores compared to other values. The value with the highest sum of scores is . We use this value for subsequent experiments.
Algorithm  Valid  Unique  Novel  Solubility 

(full RL)  99.8  2.3  97.9  0.86 
98.2  2.2  98.1  0.74  
92.2  2.7  95.0  0.67  
87.3  3.2  87.2  0.56  
88.2  2.1  88.2  0.65  
86.6  2.1  87.5  0.48  
89.6  2.8  89.6  0.57  
(no RL)  87.7  2.9  97.7  0.54 
5.2 Objectives optimization
Objective  Algorithm  Valid (%)  Unique (%)  Time (h)  Diversity  Druglikeliness  Synthesizability  Solubility 

Druglikeliness  ORGAN  88.2  69.4*  9.63*  0.55  0.52  0.32  0.35 
OR(W)GAN  85.0  8.2*  10.06*  0.95  0.60  0.54  0.47  
Naive RL  97.1  54.0*  9.39*  0.80  0.57  0.53  0.50  
MolGAN  99.9  2.0  1.66  0.95  0.61  0.68  0.52  
MolGAN (QM9)  100.0  2.2  4.12  0.97  0.62  0.59  0.53  
Synthesizability  ORGAN  96.5  45.9*  8.66*  0.92  0.51  0.83  0.45 
OR(W)GAN  97.6  30.7*  9.60*  1.00  0.20  0.75  0.84  
Naive RL  97.7  13.6*  10.60*  0.96  0.52  0.83  0.46  
MolGAN  99.4  2.1  1.04  0.75  0.52  0.90  0.67  
MolGAN (QM9)  100.0  2.1  2.49  0.95  0.53  0.95  0.68  
Solubility  ORGAN  94.7  54.3*  8.65*  0.76  0.50  0.63  0.55 
OR(W)GAN  94.1  20.8*  9.21*  0.90  0.42  0.66  0.54  
Naive RL  92.7  100.0*  10.51*  0.75  0.49  0.70  0.78  
MolGAN  99.8  2.3  0.58  0.97  0.45  0.42  0.86  
MolGAN (QM9)  99.8  2.0  1.62  0.99  0.44  0.22  0.89  
All/Alternated  ORGAN  96.1  97.2*  10.2*  0.92  0.52  0.71  0.53 
All/Simultaneously  MolGAN  97.4  2.4  2.12  0.91  0.47  0.84  0.65 
All/Simultaneously  MolGAN (QM9)  98.0  2.3  5.83  0.93  0.51  0.82  0.69 
Similarly to the previous experiment, we train our model for 300 epochs on the 5k QM9 subset while optimizing the same objectives as Guimaraes et al. (2017) to compare against their work. Moreover, we also report results on the full dataset trained for 30 epochs (note that the full dataset is 20 times larger than the subset). We choose to optimize the following objectives which represent qualities typically desired for drug discovery:
 Druglikeness:

how likely a compound is to be a drug. The Quantitative Estimate of Druglikeness (QED) score quantifies compound quality with a weighted geometric mean of desirability scores capturing the underlying data distribution of several drug properties
(Bickerton et al., 2012).  Solubility:

the degree to which a molecule is hydrophilic. The log octanolwater partition coefficient (logP), is defined as the logarithm of the ratio of the concentrations between two solvents of a solute (Comer & Tam, 2001).
 Synthetizability:

this measure quantifies how easy a molecule is to synthesize. The Synthetic Accessibility score (Ertl & Schuffenhauer, 2009) is a method to estimate the ease of synthesis in a probabilistic way.
All scores are normalized to lie within . We assign a score of zero to invalid compounds (i.e., implicitly we are also optimizing a validity score). We also measure, without optimizing for it, a diversity score which indicates how likely a molecule is to be diverse with respect to the dataset. This measure compares substructures between samples and a random subset from the dataset indicating how many repetitions there are.
For evaluation, we report average scores for 6400 sampled compounds as in (Guimaraes et al., 2017). Additionally, we rerun experiments from (Guimaraes et al., 2017) to report unique scores and execution time since they are not provided in the original work. Differently from ORGAN, to optimize for all objectives, we do not alternate between optimizing them individually during training which in our case is not possible since the reward network is specific to a single type of reward. Thus, we instead optimize a joint reward which we define as the product (to lie within ) of all objectives.
Results
Results are reported in Table 2. Qualitative samples are provided in the Appendix (Figure 3). We observe that MolGAN models always converge to very high validity outputs at the end of the training. This is coherent as observed in the previous experiment, since also here there is an implicit optimization of validity. Moreover, in all single metrics settings, our models beat ORGAN models in terms of valid scores as well as all the three objective scores we optimize for.
We argue that this should be mainly due to two factors: i) intuitively, it should be easier to optimize a molecular graph predicted as a single sample than to optimize an RNN model that outputs a sequence of characters, and ii) using the deterministic policy gradient instead of REINFORCE effectively provides a better gradient and it improves the sampling procedure towards metrics while penalizing invalid graphs.
Training on the full QM9 dataset for 10 times fewer epochs further improves results in almost all scores. During training, our algorithm observes more different samples, and therefore it can learn well with much fewer iterations. Moreover, it can observe molecules with more diverse structures and properties.
As previously observed in Section 5.1, also in this experiment the unique score is always close to 2% confirming our hypothesis that our models are susceptible to mode collapse. This is not the case for the ORGAN baseline. During sampling, ORGAN generates sequences of maximum 51 characters which allows it to generate larger molecules whereas our model is (by choice) constrained to generate up to 9 atoms. This can explain the difference in unique score since the chance of generating different molecules in a smaller space is much lower. Notice that in ORGAN, the RL component relies on REINFORCE, and the unique score is optimized penalizing nonunique outputs which we do not.
In terms of training time, our model outperforms ORGAN by a large margin when training on the 5k dataset (at least 5 times faster in each setting), as we do not rely on sequential generation or discrimination. Both ORGAN and MolGAN have a comparable number of parameters, with the latter being approximately larger.
5.3 VAE Baselines
Algorithm  Valid  Unique  Novel 

CharacterVAE  10.3  67.5  90.0 
GrammarVAE  60.2  9.3  80.9 
GraphVAE  55.7  76.0  61.6 
GraphVAE/imp  56.2  42.0  75.8 
GraphVAE NoGM  81.0  24.1  61.0 
MolGAN  98.1  10.4  94.2 
In this experiment, we compare MolGAN against recent likelihoodbased methods that utilize VAEs. We report a comparison with CharacterVAE (GómezBombarelli et al., 2016), GrammarVAE (Kusner et al., 2017), and GraphVAE (Simonovsky & Komodakis, 2018). Here we train using the complete QM9 dataset. Naturally, we compare only with metrics that measure the quality of the generative process since the likelihood is not computed directly in MolGAN. Moreover, we do not optimize any particular chemical property except validity (i.e., we do not optimize any metric described above, but we optimize towards chemically valid compounds). The final evaluation scores are an average from random samples. The number of samples differs from the previous experiment to be in line with the setting in Simonovsky & Komodakis (2018).
Results
Results are reported in Table 3. Training on the full QM9 dataset (without optimizing any metric except validity) results in a model with a higher unique score compared to the ones in Section 5.2.
Though the unique score of MolGAN is slightly higher compared to GrammarVAE, the other baselines are superior in terms of this score. Even though here we do not consider our model to be collapsed, such a low score confirms our hypothesis that our model is prone to mode collapse. On the other hand, we observe significantly higher validity scores compared to the VAEbased baselines.
Notice that, differently from our approach, VAEs optimize the evidence lower bound (ELBO) and there is no explicit nor implicit optimization of output validity. Moreover, since a part of the ELBO maximizes reconstruction of the observations, the novelty in the sampling process is not expected to be high since it is not optimized. However, in all reported methods novelty is and, in the case of CharacterVAE, . Though CharacterVAE can achieve a high novelty score, it underperforms in terms of validity. MolGAN, on the other hand, achieves both high validity and novelty scores.
6 Conclusions
In this work, we have introduced MolGAN: an implicit generative model for molecular graphs of small size. Through joint training with a GAN and an RL objective, our model is capable of generating molecular graphs with both higher validity and novelty than previous comparable VAEbased generative models, while not requiring a permutationdependent likelihood function. Compared to a recent SMILESbased sequential GAN model for molecular generation, MolGAN can achieve higher chemical property scores (such as solubility) while allowing for at least 5x faster training time.
A central limitation of our current formulation of MolGANs is their susceptibility to mode collapse: both the GAN and the RL objective do not encourage generation of diverse and nonunique outputs whereby the model tends to be pulled towards a solution that only involves little sample variability. This ultimately results in the generation of only a handful of different molecules if training is not stopped early.
We think that this issue can be addressed in future work, for example via careful design of reward functions or some form of pretraining. The MolGAN framework taken together with established benchmark datasets for chemical synthesis offer a new test bed for improvements on GAN stability with respect to the issue of mode collapse. We believe that insights gained from such evaluations will be valuable to the community even outside of the scope of generating molecular graphs. Lastly, it will be promising to explore alternative generative architectures within the MolGAN framework, such as recurrent graphbased generative models (Johnson, 2017; Li et al., 2018b; You et al., 2018), as our current oneshot prediction of the adjacency tensor is most likely feasible only for graphs of small size.
Acknowledgements
The authors would like to thank Luca Falorsi, Tim R. Davidson, Herke van Hoof and Max Welling for helpful discussions and feedback. T.K. is supported by the SAP Innovation Center Network.
References
 Arjovsky et al. (2017) Arjovsky, Martin, Chintala, Soumith, and Bottou, Léon. Wasserstein generative adversarial networks. In International Conference on Machine Learning, pp. 214–223, 2017.
 Bickerton et al. (2012) Bickerton, G Richard, Paolini, Gaia V, Besnard, Jérémy, Muresan, Sorel, and Hopkins, Andrew L. Quantifying the chemical beauty of drugs. Nature chemistry, 4(2):90, 2012.
 Bojchevski et al. (2018) Bojchevski, Aleksandar, Shchur, Oleksandr, Zügner, Daniel, and Günnemann, Stephan. Netgan: Generating graphs via random walks. In International Conference on Machine Learning, 2018.
 Bronstein et al. (2017) Bronstein, Michael M, Bruna, Joan, LeCun, Yann, Szlam, Arthur, and Vandergheynst, Pierre. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4):18–42, 2017.
 Bruna et al. (2013) Bruna, Joan, Zaremba, Wojciech, Szlam, Arthur, and LeCun, Yann. Spectral networks and locally connected networks on graphs. arXiv preprint arXiv:1312.6203, 2013.
 Comer & Tam (2001) Comer, John and Tam, Kin. Lipophilicity profiles: theory and measurement. WileyVCH: Zürich, Switzerland, 2001.
 Dai et al. (2018) Dai, Hanjun, Tian, Yingtao, Dai, Bo, Skiena, Steven, and Song, Le. Syntaxdirected variational autoencoder for molecule generation. In International Conference on Machine Learning, 2018.
 Davidson et al. (2018) Davidson, Tim R, Falorsi, Luca, De Cao, Nicola, Kipf, Thomas, and Tomczak, Jakub M. Hyperspherical variational autoencoders. In UAI, 2018.
 Duvenaud et al. (2015) Duvenaud, David K, Maclaurin, Dougal, Iparraguirre, Jorge, Bombarell, Rafael, Hirzel, Timothy, AspuruGuzik, Alán, and Adams, Ryan P. Convolutional networks on graphs for learning molecular fingerprints. In Advances in neural information processing systems, pp. 2224–2232, 2015.
 Ertl & Schuffenhauer (2009) Ertl, Peter and Schuffenhauer, Ansgar. Estimation of synthetic accessibility score of druglike molecules based on molecular complexity and fragment contributions. Journal of cheminformatics, 1(1):8, 2009.
 GómezBombarelli et al. (2016) GómezBombarelli, Rafael, Wei, Jennifer N, Duvenaud, David, HernándezLobato, José Miguel, SánchezLengeling, Benjamín, Sheberla, Dennis, AguileraIparraguirre, Jorge, Hirzel, Timothy D, Adams, Ryan P, and AspuruGuzik, Alán. Automatic chemical design using a datadriven continuous representation of molecules. ACS Central Science, 2016.
 Goodfellow et al. (2014) Goodfellow, Ian, PougetAbadie, Jean, Mirza, Mehdi, Xu, Bing, WardeFarley, David, Ozair, Sherjil, Courville, Aaron, and Bengio, Yoshua. Generative adversarial nets. In Advances in neural information processing systems, pp. 2672–2680, 2014.
 Grover et al. (2017) Grover, Aditya, Zweig, Aaron, and Ermon, Stefano. Graphite: Iterative generative modeling of graphs. In NIPS Bayesian Deep Learning Workshop, 2017.
 Guimaraes et al. (2017) Guimaraes, Gabriel Lima, SanchezLengeling, Benjamin, Farias, Pedro Luis Cunha, and AspuruGuzik, Alán. Objectivereinforced generative adversarial networks (organ) for sequence generation models. arXiv preprint arXiv:1705.10843, 2017.
 Gulrajani et al. (2017) Gulrajani, Ishaan, Ahmed, Faruk, Arjovsky, Martin, Dumoulin, Vincent, and Courville, Aaron C. Improved training of wasserstein gans. In Advances in Neural Information Processing Systems, pp. 5769–5779, 2017.
 Hamilton et al. (2017) Hamilton, William L, Ying, Rex, and Leskovec, Jure. Representation learning on graphs: Methods and applications. arXiv preprint arXiv:1709.05584, 2017.
 Jang et al. (2016) Jang, Eric, Gu, Shixiang, and Poole, Ben. Categorical reparameterization with gumbelsoftmax. arXiv preprint arXiv:1611.01144, 2016.
 Jin et al. (2018) Jin, Wengong, Barzilay, Regina, and Jaakkola, Tommi. Junction tree variational autoencoder for molecular graph generation. arXiv preprint arXiv:1802.04364, 2018.
 Johnson (2017) Johnson, Daniel D. Learning graphical state transitions. ICLR, 2017.
 Kingma & Ba (2014) Kingma, Diederik P and Ba, Jimmy. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
 Kingma & Welling (2013) Kingma, Diederik P and Welling, Max. Autoencoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
 Kipf & Welling (2016a) Kipf, Thomas N and Welling, Max. Semisupervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016a.
 Kipf & Welling (2016b) Kipf, Thomas N and Welling, Max. Variational graph autoencoders. In NIPS Bayesian Deep Learning Workshop, 2016b.
 Kusner et al. (2017) Kusner, Matt J, Paige, Brooks, and HernándezLobato, José Miguel. Grammar variational autoencoder. In International Conference on Machine Learning, pp. 1945–1954, 2017.
 Li et al. (2018a) Li, Yibo, Zhang, Liangren, and Liu, Zhenming. Multiobjective de novo drug design with conditional graph generative model. arXiv preprint arXiv:1801.07299, 2018a.
 Li et al. (2016) Li, Yujia, Tarlow, Daniel, Brockschmidt, Marc, and Zemel, Richard. Gated graph sequence neural networks. International Conference on Learning Representations, 2016.
 Li et al. (2018b) Li, Yujia, Vinyals, Oriol, Dyer, Chris, Pascanu, Razvan, and Battaglia, Peter. Learning deep generative models of graphs. arXiv preprint arXiv:1803.03324, 2018b.
 Lillicrap et al. (2015) Lillicrap, Timothy P, Hunt, Jonathan J, Pritzel, Alexander, Heess, Nicolas, Erez, Tom, Tassa, Yuval, Silver, David, and Wierstra, Daan. Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971, 2015.
 Maddison et al. (2016) Maddison, Chris J, Mnih, Andriy, and Teh, Yee Whye. The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:1611.00712, 2016.
 Minervini et al. (2017) Minervini, Pasquale, Demeester, Thomas, Rocktäschel, Tim, and Riedel, Sebastian. Adversarial sets for regularising neural link predictors. arXiv preprint arXiv:1707.07596, 2017.
 Ramakrishnan et al. (2014) Ramakrishnan, Raghunathan, Dral, Pavlo O, Rupp, Matthias, and Von Lilienfeld, O Anatole. Quantum chemistry structures and properties of 134 kilo molecules. Scientific data, 1:140022, 2014.
 Rezende et al. (2014) Rezende, Danilo J., Mohamed, Shakir, and Wierstra, Daan. Stochastic backpropagation and approximate inference in deep generative models. arXiv preprint arXiv:1401.4082, 2014.
 Ruddigkeit et al. (2012) Ruddigkeit, Lars, Van Deursen, Ruud, Blum, Lorenz C, and Reymond, JeanLouis. Enumeration of 166 billion organic small molecules in the chemical universe database gdb17. Journal of chemical information and modeling, 52(11):2864–2875, 2012.
 Salimans et al. (2016) Salimans, Tim, Goodfellow, Ian, Zaremba, Wojciech, Cheung, Vicki, Radford, Alec, and Chen, Xi. Improved techniques for training gans. In Advances in Neural Information Processing Systems, pp. 2234–2242, 2016.
 Samanta et al. (2018) Samanta, Bidisha, De, Abir, Ganguly, Niloy, and GomezRodriguez, Manuel. Designing random graph models using variational autoencoders with applications to chemical design. arXiv preprint arXiv:1802.05283, 2018.
 Schlichtkrull et al. (2017) Schlichtkrull, Michael, Kipf, Thomas N, Bloem, Peter, Berg, Rianne van den, Titov, Ivan, and Welling, Max. Modeling relational data with graph convolutional networks. arXiv preprint arXiv:1703.06103, 2017.
 Schneider & Fechner (2005) Schneider, Gisbert and Fechner, Uli. Computerbased de novo design of druglike molecules. Nature Reviews Drug Discovery, 4(8):649, 2005.
 Segler et al. (2017) Segler, Marwin HS, Preuss, Mike, and Waller, Mark P. Learning to plan chemical syntheses. arXiv preprint arXiv:1708.04202, 2017.
 Silver et al. (2014) Silver, David, Lever, Guy, Heess, Nicolas, Degris, Thomas, Wierstra, Daan, and Riedmiller, Martin. Deterministic policy gradient algorithms. In International Conference on Machine Learning, 2014.
 Simonovsky & Komodakis (2018) Simonovsky, Martin and Komodakis, Nikos. Graphvae: Towards generation of small graphs using variational autoencoders. arXiv preprint arXiv:1802.03480, 2018.
 Srivastava et al. (2014) Srivastava, Nitish, Hinton, Geoffrey, Krizhevsky, Alex, Sutskever, Ilya, and Salakhutdinov, Ruslan. Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research, 15(1):1929–1958, 2014.
 Wang et al. (2018) Wang, Hongwei, Wang, Jia, Wang, Jialin, Zhao, Miao, Zhang, Weinan, Zhang, Fuzheng, Xie, Xing, and Guo, Minyi. Graphgan: Graph representation learning with generative adversarial nets. In AAAI, 2018.
 Weininger (1988) Weininger, David. Smiles, a chemical language and information system. 1. introduction to methodology and encoding rules. Journal of chemical information and computer sciences, 28(1):31–36, 1988.
 Williams (1992) Williams, Ronald J. Simple statistical gradientfollowing algorithms for connectionist reinforcement learning. In Reinforcement Learning, pp. 5–32. Springer, 1992.
 You et al. (2018) You, Jiaxuan, Ying, Rex, Ren, Xiang, Hamilton, William L, and Leskovec, Jure. Graphrnn: A deep generative model for graphs. In International Conference on Machine Learning, 2018.
 Yu et al. (2017) Yu, Lantao, Zhang, Weinan, Wang, Jun, and Yu, Yong. Seqgan: Sequence generative adversarial nets with policy gradient. In AAAI, pp. 2852–2858, 2017.
Comments
There are no comments yet.