IV-VAE
Code for "Semi-supervised Disentanglement with Independent Vector Variational Autoencoders," arXiv preprint arXiv:2003.06581 (2020)
view repo
We aim to separate the generative factors of data into two latent vectors in a variational autoencoder. One vector captures class factors relevant to target classification tasks, while the other vector captures style factors relevant to the remaining information. To learn the discrete class features, we introduce supervision using a small amount of labeled data, which can simply yet effectively reduce the effort required for hyperparameter tuning performed in existing unsupervised methods. Furthermore, we introduce a learning objective to encourage statistical independence between the vectors. We show that (i) this vector independence term exists within the result obtained on decomposing the evidence lower bound with multiple latent vectors, and (ii) encouraging such independence along with reducing the total correlation within the vectors enhances disentanglement performance. Experiments conducted on several image datasets demonstrate that the disentanglement achieved via our method can improve classification performance and generation controllability.
READ FULL TEXT VIEW PDF
We present a new flavor of Variational Autoencoder (VAE) that interpolat...
read it
We investigate the problem of learning representations that are invarian...
read it
In recent years, extending variational autoencoder's framework to learn
...
read it
We develop a novel probabilistic generative model based on the variation...
read it
We decompose the evidence lower bound to show the existence of a term
me...
read it
Automated discovery of early visual concepts from raw image data is a ma...
read it
We present a unified invariance framework for supervised neural networks...
read it
Code for "Semi-supervised Disentanglement with Independent Vector Variational Autoencoders," arXiv preprint arXiv:2003.06581 (2020)
A desirably disentangled representation contains individual units, each corresponding to a single generative factor of data while being invariant to changes in other units (tpami13_bengio_reprLearn). Such interpretable and invariant properties lead to benefits in downstream tasks including image classification and generation (nips14_kingmaDgm; iclrw16_aae; nips17_disent_semiSupDgm; cvpr19_ssl_genEx).
Variational autoencoders (VAEs) (iclr14_vae) have been actively utilized for unsupervised disentanglement learning (icml18_factorVae; nips18_betaTcVae; aistats19_structDisent; iclr18_dipVae; aistats19_gaoAutoEncTc; nipsw17_bVae_anneal). To capture the generative factors that are assumed to be statistically independent, many studies have encouraged the independence of latent variables within a representation (iclr17_betaVae; icml18_factorVae; nips18_betaTcVae). Despite their promising results, the usage of only continuous variables frequently causes difficulty in the discovery of discrete factors (e.g., object categories).
To address this issue, researchers have utilized discrete variables together with continuous variables to separately capture discrete and continuous factors (nips18_jointVae; nips14_kingmaDgm) and trained their models by maximizing the evidence lower bound (ELBO). In this paper, we show that the ability of disentanglement in their models is derived from not only using the two types of variables but also encouraging several sources of disentanglement. We expose these sources by decomposing the ELBO.
Unsupervised learning with discrete and continuous variables is difficult, because continuous units with a large informational capacity often store all information, causing discrete units to store nothing and be neglected by models (nips18_jointVae). Previously, this issue was mostly solved using hyperparameter-sensitive capacity controls (nips18_jointVae) or additional steps for inferring discrete features during training (icml19_cascadeVae). In contrast, we simply inject weak supervision with a few class labels and effectively resolve the difficulty in learning of discrete variables. icml18_commAssump also suggested the exploitation of available supervision for improving disentanglement.
We introduce a semi-supervised disentanglement learning method, which is shown in Figure 1. A VAE was used to extract two feature vectors: one containing discrete variables to capture class factors and the other vector containing continuous variables to capture style factors. The contributions of this work to the relevant field of study are as follows:
[noitemsep]
We introduce the vector independence objective that measures the statistical independence between latent vectors in VAEs and enables different vectors to store different information. We name our model an Independent Vector VAE (IV-VAE).
We decompose the ELBO containing multiple latent vectors to reveal the vector independence term along with well-known total correlation (TC) terms that measure independence between the variables within the vectors. We show that these terms are the sources of disentanglement in jointly learning discrete and continuous units.
We introduce supervision with a small number of class labels, significantly reducing the difficulty in learning discrete class features.
We empirically show that our method enhances disentanglement learning on several image datasets, positively affecting the classification and controllable generation.
The supplementary material includes derivation details and additional results, and the sections are indicated with S (e.g., Section S1).
A -dimensional random vector,
, consists of scalar-valued random variables,
. Previous methods have encouraged independence between variables, whereas our method encourages independence between vectors (by introducing another vector, y) along with independence between the variables within the vectors.Given dataset containing i.i.d. samples of random vector x, a VAE learns latent vector z involved in the data generation process. The VAE consists of encoder and decoder trained by maximizing the ELBO on , where the empirical distribution is represented as . The training objective consists of the reconstruction term and the KL divergence from the prior to individual posteriors:
(1) |
(2) |
where and in (1) represent the vanilla VAE and -VAE (iclr17_betaVae) objectives, respectively. Under factorized prior (e.g., ), the -VAE enhances the independence of latent variables, leading to disentanglement.
To understand the disentangling mechanism in VAEs, the KL term (2) was decomposed into (3) in (nipsw16_elboSurgery) and further into (4) under the factorized prior in (nips18_betaTcVae):
(3) |
(4) |
where is the aggregate posterior (iclrw16_aae) that describes the latent structure for all data samples, is the mutual information (MI) between the data and latent vectors under empirical distribution , and is the regularization term for penalizing individual latent dimensions that excessively differ from their priors.
The second term in (4) represents the TC (ibm1960_tc) that measures the statistical dependency of more than two random variables. icml18_factorVae and nips18_betaTcVae argued that (i) -VAEs heavily penalize the MI term, , along with the other terms, causing z to be less informative about x, and (ii) the TC term is the source of disentanglement and thus should be strongly penalized. To address these issues, icml18_factorVae proposed the FactorVAE by adding the TC term to the vanilla VAE objective, and nips18_betaTcVae proposed the -TCVAE by separately controlling the three terms in (4) using individual weights.
The previous methods (iclr17_betaVae; icml18_factorVae; nips18_betaTcVae)
learned disentangled representations by encouraging the independence of continuous variables. They primarily employed Gaussian distributions for variational posteriors and consequently focused on modeling continuous generative factors. However, most datasets naturally contain discrete factors of variation (e.g., object classes), which these methods frequently fail to capture. We address this issue by incorporating discrete random variables and further enhance the disentanglement by promoting the independence between one set of discrete variables and another set of continuous variables.
To separately capture discrete and continuous factors, researchers have proposed the simultaneous utilization of discrete and continuous variables, which are stored in two latent vectors, y and z, respectively. By introducing joint prior , approximate posterior , and likelihood , the ELBO containing the reconstruction and KL terms becomes (see Section S1 for the derivation)
(5) |
(6) |
By assuming factorized prior , the KL can be decomposed depending on the factorized form of ; nips14_kingmaDgm assumed ^{1}^{1}1The unsupervised objective in Eqn. 7 of nips14_kingmaDgm can be reformulated as . Eqn. 7 was extended to reveal the objective for labeled data and the entropy for discrete y., and nips18_jointVae assumed to derive
(7) |
In contrast, we decompose the KL term (6) to reveal the existence of the vector independence term between y and z as well as the TC terms within the vectors (see our decomposition in (11)). Moreover, our method explicitly encourages the vector independence term while not penalizing the latent-data MI terms, and , to obtain disentangled informative features. We show that our method outperforms the -VAE (iclr17_betaVae) and its variant (nips18_jointVae), which strongly penalize the KL term (7) and consequently minimize the latent-data MI terms.
In addition, unsupervised learning with discrete and continuous variables often causes the discrete units to capture less information and be discarded by the model (nips18_jointVae) because of a larger informational capacity of continuous units than that of discrete units. To address this issue, existing unsupervised methods involve sophisticated hyperparameter tunings or additional computations. For example, nips18_jointVae modified (7) using capacity control terms separately for discrete and continuous units, and icml19_cascadeVae
proposed an alternating optimization between inferring probable discrete features and updating the encoder. In contrast, we introduce weak classification supervision to guide the encoder to store class factors in discrete units; this method simply but significantly reduces the effort needed for designing inference steps.
The problem scenario is identical to that described in Section 2.2, with discrete random vector for capturing class factors and continuous random vector for capturing style factors. First, we show that the ELBO can be decomposed into the proposed vector independence term and the others. Then, we present our semi-supervised learning (SSL) strategy.
Assuming the conditional independence of , an encoder produces the parameters for variational posteriors and , for the n-th sample, . The aggregate posteriors that capture the entire latent space under data distribution are defined as , , and
(8) |
where is computed using its decomposition form.^{2}^{2}2In this paper, we assumed the conditional independence by following (nips18_jointVae). However, can be computed as by incorporating architectural dependency from y to z. In this case, only is revised, whereas the computation of (8) and (9) remains unchanged. Then, we define our vector independence objective as the MI between the two vectors that measures their statistical dependency:
(9) |
Thr reduction of this term can enforce y and z to capture different semantics. Here, we emphasize that our method is applicable to cases with multiple L latent vectors by extending (9) to , which has a form similar to the TC computed over the variables (i.e.,
) but is computed over the vectors. The relationship between the TC and the vector independence term is similar to that between the objectives of independent component analysis (ICA;
sp91_ica_1; nips96_ica_2_amari) and independent vector analysis (IVA; iva_1_conf; iva_2_jour).nips18_betaTcVae showed that the TC term measuring the dependency between latent variables (i) exists in the decomposion of the ELBO containing a single latent vector, z, and (ii) is a source of disentanglement in VAEs. Similarly, we reveal that the vector independence term (i) exists in the decomposition of the ELBO containing two latent vectors, y and z, and (ii) is another source of disentanglement.
Concretely, we decompose the KL term (6) of the ELBO into (10) under and further into (11) under and :
(10) |
(11) |
where is the MI between x and latent vectors y and z, and and are the dimension-wise regularization terms. The derivation from (10) to (11) is motivated by that from (3) to (4). See Section S2 for the derivation details.
As suggested by icml18_factorVae and nips18_betaTcVae, data-latent MI is not penalized during training so as to allow the latent vectors to capture data information. The second term in the RHS is our vector independence objective, and the third and fifth terms are the TC terms for the variables in y and z, respectively. We empirically show that simultaneously reducing these three terms provides better disentanglement compared to penalizing only the TC terms without considering the vector independence. The regularization terms, and , forbid individual latent variables from deviating largely from the priors.
A concurrent work (aistats19_structDisent) introduced a decomposition similar to our result in (11). Their derivation was initiated by augmenting the ELBO with a data entropy term (i.e., ). In contrast to our method, their method with discrete and continuous units uses purely unsupervised learning, which often causes the discrete units to be ignored by the model.
Here, we investigate the relationship between vector independence objective (9) and the following TC terms:
(12) |
(13) |
(12) measures the independence of the variables within each vector (hereafter called the “separate” TC). In addition, (13) simultaneously considers the variables in y and z (hereafter called the “collective” TC), and it can be viewed as the TC on concatenated vector yz (i.e., ).
We introduce two relationships. First, perfectly penalized collective TC indicates perfect vector independence:
(14) |
by letting and . In this case, the vector independence objective would be naturally satisfied, resulting in unnecessary optimization. However, this case is rare because of the existence of other loss terms (e.g., a reconstruction term) that often prevent the collective TC from being zero. Furthermore, under the factorized prior, the perfectly penalized TC may be undesirable because it could imply the occurrence of posterior collapse (i.e., learning a trivial posterior that collapses to the prior and fails to capture data features). We present the experimental setup and results regarding this relationship in Sections S5 and S6.
Second, perfect vector independence does not ensure that all variables within and between the vectors are perfectly independent, i.e., . However, perfect vector independence ensures that the collective TC is the sum of the two separate TCs, i.e., (see Section S3 for the derivation).
A weak classification supervision guides y to suitably represent discrete class factors. In addition, our vector independence objective further enforces y and z to capture different types of information. In our experiments, we simplify the problem setup by assuming that a given dataset involves a single classification task with classes. This enabled us to design
as a single categorical variable,
, where .^{3}^{3}3For multiple classification tasks (e.g., identity recognition and eye-glass detection tasks for face images), our method can be applied with multidimensional y, where each dimension, , corresponds to one categorical variable for each task. We represent as a -dimensional one-hot vector. For , we assume the existence of multiple style factors and expect each factor to be captured by each variable, , within z.The training image dataset consists of labeled set , where the -th image, , is paired with the corresponding class label, , and unlabeled set . Here, and are the numbers of samples in datasets and , respectively, and . The empirical data distributions over and are denoted by and , respectively.
To update encoder parameter and decoder parameter , the objectives for the labeled and unlabeled sets are given as
(15) |
(16) |
For the following reconstruction terms, we use true label for and inferred feature for . This strategy helps the decoder accurately recognize one-hot class vectors.
(17) |
(18) |
We compute the classification term for as
(19) |
where hyper-parameter controls the effect of discriminative learning. With scaling constant , we set (nips14_kingmaDgm; icml16_adgm_auxiliary).
Next, we introduce the following commonly used objective for and to learn disentangled features and regularize the encoder:
(20) |
This is identical to applying individual loss weights to the vector independence, TC, and dimension-wise regularization terms in our KL decomposition (11). Here, the expectation over the empirical distribution, i.e., or , is included in computing the aggregated posteriors, z and . Note that disappears for a single class variable, (i.e., in (12)), and data-latent MI is removed so as to allow and z to properly store the information about x.
The final optimization function is given as
(21) |
We used the dSprites (dsprites17), Fashion-MNIST (xiao2017fashionmnist), and MNIST (lecun2010mnist) datasets. For SSL, the labeled data were selected to be distributed evenly across classes, and the entire training set was used for the unlabeled set. We prevented overfitting to training data in classification tasks by introducing validation data. For the dSprites dataset, we divided the images in a ratio of 10:1:1 for training, validation, and testing and tested two SSL setups with 2% and 0.25% labeled training data. For the Fashion-MNIST and MNIST datasets, we divided the training set in a ratio of 5:1 for training and validation while maintaining the original test set and tested the SSL setup with 2% labeled training data.
The architectures of the encoder and decoder were the same as the convolutional networks used in (nips18_jointVae; icml19_cascadeVae). The priors were set as and , where denotes evenly distributed class probabilities. We employed the Gumbel-Softmax distribution (iclr17_gumbel; iclr17_concrete) for reparametrizing categorical . We trained networks with minibatch weighted sampling (nips18_betaTcVae). Further details of experimental settings are described in Section S4.
We considered the vanilla VAE (iclr14_vae), -VAE (iclr17_betaVae), -TCVAE (nips18_betaTcVae), and jointVAE (nips18_jointVae) as the baselines. For fair comparison, we augmented their original unsupervised objectives with the classification term in (19) and the same loss weight, . For the VAE, -VAE, and -TCVAE, we augmented continuous vector z in their original objectives with discrete variable . Note that the main difference between the -TCVAE and our IV-VAE is the existence of the vector independence term in (20). We also removed the data-latent MI term from the -TCVAE objective, as applied in our objective. See Section S5 for the baseline details.
We measured the classification error with to assess the amount of class information in and the ELBO to examine the generative modeling ability. As the disentanglement score, we computed the MI gap (nips18_betaTcVae) based on the empirical MI between latent variables and known generative factors: , where is the number of known factors and is the score for each factor, , for quantifying the gap between the top two variables with the highest MI. Here, is the normalized MI between latent variable and factor , and it is theoretically bounded between 0 and 1. Additionally, . This vanilla MIG is denoted by . Figure 2 shows the example of the normalized MI computed on dSprites with different models, where the top two MI values for each factor are indicated.
The entangled results for the case where one variable captures both style and class factors (e.g., in Figure 2(a)) can obtain fairly good scores. To alleviate this issue, we separately computed the MIG for the class and style factors as (i.e., the MIG of the first row in Figure 2) and (i.e., the mean MIG averaged from the second to last rows), where denotes ground truth class labels and is the set of style factors. For the cases without known generative factors but with available labels, , we also computed the MI that assesses how much represents the class information and z does not as and , respectively.
All results presented in this section were obtained using test data. See our supplementary material for additional results.
Figure 2 depicts the benefit of SSL and vector independence. Using purely unsupervised learning failed to capture the class factor with discrete variable , but employing SSL with 2% class labels easily resolved this issue. Encouraging vector independence helped z better capture the style factors by forcing and z to store different information.
Figure 3 shows the results on dSprites obtained under various SSL setups. The unsupervised setup yielded good vanilla scores, which were similar to those of the 2%-labeled setup. However, the unsupervised setup caused the continuous vector z to mostly captured the class information and the discrete to fail to store it. This is evidenced by higher and lower than those of the SSL setups. Furthermore, as indicated by lower scores, more than two variables undesirably captured the class factor under the unsupervised setup.
Injecting weak class supervision to the training process simply yet effectively alleviated this difficulty, as shown in the decreased and increased in Figure 3. In addition, our IV-VAEs with proper weights outperformed -VAEs and -TCVAEs for most scores. In particular, the score improvements with the 0.25% labels were larger than those with the 2% labels, indicating the benefit of vector independence under a few class labels. In terms of the ELBO, our IV-VAEs did not outperform -TCVAEs but showed less trade-off between density modeling and disentanglement than -VAE (i.e., higher ELBO and MIG scores).
Figure 4 shows the relationships between the vector independence objective and evaluation metrics. In Figure 4-1 (left), we depicted scatter plots, where each circle represents the median score of seven networks trained with the same loss weights but initiated from different random seeds. In Figure 4-2 (right), the different values were further integrated to analyze the general result trends. Notice that in the figures, a larger led to a smaller KL of vector independence, i.e., a stronger independence between and z. The analyses are summarized below. See Section S6 for the additional analyses with extended baselines.
[noitemsep]
ELBO in Figure 4-1(a). The ELBO values of IV-VAEs were similar to those of -TCVAEs for lower values of 0.5 and 1. Increasing caused slightly decreased ELBO values, because the models focused more on learning disentangled features than maximizing the reconstruction term. Nevertheless, under higher values of 4 and 8, IV-VAEs yielded better ELBO than -VAEs^{4}^{4}4The -VAEs with of 4 and 8 under the 0.25% SSL setup yielded the median ELBO of -102.2 and -120.8, respectively. Those with of 4 and 8 under the 2% SSL setup yielded the median ELBO of -53.9 and -72.5. because we did not penalize the latent-data MI term, which led to an eased trade-off between reconstruction and disentanglement.
, , and in Figure 4-1(b), (c), and (d). For most of the setups, IV-VAEs with the values of 0.5, 1, or 4 achieved better MIG scores than -TCVAEs, showing that reducing the vector independence objective along with the TC helps disentanglement. Higher that heavily penalized the TC often led to higher MIG scores (i.e., the blue and green lines showed better scores than the red lines). The optimal yielding the highest MIG differed depending on the value of , but the of 0.5, 1, and 4 generally worked well.
Classification error and I(y; t) in Figure 4-1(e) and (f). Given the value of , the lowest classification error and highest score were often obtained with IV-VAE, in comparison to -TCVAE. This result implies that vector independence encourages to better capture the class factor. In the SSL setup with 0.25% labels, excessively increasing values harmed the classification performance.
I(z; t) in Figure 4-1(g). IV-VAEs often achieved lower than that of -TCVAEs. This result implies that vector independence prevents z from storing the class factor by enforcing z and to represent different information, leading to better disentanglement.
Overall result trends in Figure 4-2. Our IV-VAEs outperformed -TCVAEs for most scores, demonstrating the benefit of vector independence. In general, the of 1 for the 0.25% setup and the of 4 for the 2% setup worked well.
Figure 5 shows the results on Fashion-MNIST with 2% labeled data. Because of the absence of ground truth style factors, we only measured the MI between the latent variables and class labels. For most values of , promoting vector independence with greater allowed and z to capture different factors, causing more class information to be stored in . Our IV-VAEs achieved better classification errors and as well as lower than all the baselines, indicating improved disentanglement. Figure 6 shows the results on MNIST with 2% labeled data. We also observed the benefit of our vector independence under most of the weight settings.
Figures 7 and 8 depict the qualitative results of the two networks trained without and with the vector independence objective. The networks were obtained from the same random seed and loss weights except the weight.^{5}^{5}5In Figure 7, the of 4 was used for training the IV-VAE, while the of 96, of 1, and of 2 were commonly used for both networks. In Figure 8, the of 8 was used, while the of 32, of 1, and of 2 were the same for both networks. See Section S7 and S8 for additional examples.
To visually show the corrected class labels by encouraging vector independence, Figures 7(a) and 8
(a) depict reconstruction examples with estimated class information, i.e., inferred class probabilities or one-hot labels via the argmax operation. The results obtained from the class probabilities were often blurry, indicating that the inputs were confusing to be classified. Employing vector independence often corrects the classification results by enforcing
to better store the class factor.We have proposed an approach for semi-supervised disentanglement learning with a variational autoencoder. In our method, two latent vectors separately capture class and style factors. To boost disentanglement, we have proposed the vector independence objective that enforces the vectors to be statistically independent. We have revealed that, along with the total correlation term, our vector independence term is another source of disentanglement in the evidence lower bound. Furthermore, the difficulty in the learning of discrete factors can be reduced by exploiting a small number of class labels. The experiments on the dSprites, Fashion-MNIST, and MNIST datasets have confirmed the effectiveness of our method for image classification and generation.
We thank Prof. Sung Ju Hwang (KAIST) for providing his valuable comments. This work was supported by an Institute for Information and Communications Technology Promotion (IITP) grant funded by the Korean Government (MSIT) under grant no. 2016-0-00562 (R0124-16-0002).
Suppose that the generation process of the data involves two latent vectors, y and z, with a joint prior, . By introducing an approximate posterior, , and a likelihood, , the ELBO for a single data sample becomes
(22) |
Figure S1 verifies our decomposition of the KL term, , in the ELBO on .
Figure S2 verifies the following relationship: perfect vector independence ensures that the collective TC becomes the sum of the two separate TCs (i.e., ).
Moreover, in this document, we provide the experimental results regarding the relationship described in (14) of the main paper: . We show that (i) a perfectly penalized collective TC (i.e., ) rarely occurs because of the existence of other loss terms (e.g., a reconstruction term) and (ii) encouraging vector independence under penalizing either the separate TC or the collective TC can improve disentanglement performance. See Section S5 for the experimental settings and Section S6 for the results.
We used the dSprites (dsprites17), MNIST (lecun2010mnist), and Fashion-MNIST (xiao2017fashionmnist) datasets. For semi-supervised learning, the labeled data were selected to be distributed evenly across classes. The size of labeled set was either 2% or 0.25% of the entire training data, which we used for unlabeled set . The classification loss weight, , was set as for the 2% and 0.25%-labeled dSprites setups. Here, for the 0.25% setup, we failed to apply scaling constant (which caused an extremely large ), and we instead used as applied in the 2% setup. We used (i.e., ) for the 2%-labeled MNIST. We used (i.e., ) for the 2%-labeled Fashion-MNIST.
Figure S3 shows the network architectures used in our experiments, which are identical to the convolutional architectures used in (nips18_jointVae; icml19_cascadeVae). The priors were set as and , where denotes evenly distributed class probabilities. We employed the Gumbel-Softmax distribution (iclr17_gumbel; iclr17_concrete) for reparametrizing categorical . We will release our experimental codes on GitHub.
The dSprites dataset contains 2D-shape binary images with a size of 6464, which were synthetically generated with five independent factors: shape (3 classes; heart, oval, and square), position X (32 values), position Y (32), scale (6), and rotation (40). We divided the 737,280 images into training, validation, and test sets in a ratio of 10:1:1. We tested two cases with the 2% and 0.25% labels available in the training data.
One 6-dimensional isotropic Gaussian vector was used for , and one categorical variable representing 3 classes was used for
. The Gumbel-Softmax temperature parameter was set as 0.75. The Adam optimizer was used with an initial learning rate of 0.001 and a minibatch size of 2048. Every batch contained 1024 labeled samples. We trained networks for 100 epochs and reported the results measured at the epoch of the best validation loss. We tested 7 different random weeds for network weight initialization.
The MNIST dataset contains 0–9 handwritten digit images with a size of 2828. We divided the original training set into 50,000 training and 10,000 validation images while maintaining the test set of 10,000 images. The images were normalized to have [0, 1] continuous values and resized to 3232 by following (nips18_jointVae). We tested the case with the 2% labels available in the training data.
One 10-dimensional isotropic Gaussian vector was used for , and one categorical variable representing 10 classes was used for . The Gumbel-Softmax temperature parameter was set as 0.67. The Adam optimizer was used with an initial learning rate of 0.001 and a minibatch size of 512. Every batch contained 256 labeled samples. We trained networks for 200 epochs and reported the results measured at the epoch of the best validation loss. We tested 10 different random seeds for network weight initialization.
The Fashion-MNIST dataset (xiao2017fashionmnist) contains grayscale images with a size of 2828 and 10 fashion categories (e.g., t-shirt, trouser, sandal, and bag). The experimental settings were the same as those used in the MNIST dataset, except that the Gumbel-Softmax temperature parameter was set as 0.75.
We assumed that discrete random variable and continuous random vector were involved in the data generation process. Under this scenario, we compared our IV-VAE with the vanilla VAE (iclr14_vae), -VAE (iclr17_betaVae), -TCVAE (nips18_betaTcVae), and JointVAE (nips18_jointVae). For fair comparison, we applied the following settings that were used in our method to the baselines.
[noitemsep]
We utilized the same dimensional latent units and network architectures. For the VAE, -VAE, and -TCVAE, we augmented the original continuous vector, z, in their objectives with discrete variable . The jointVAE was originally designed to incorporate both of the continuous z and discrete .
We assumed the conditional independence of by following (nips18_jointVae).
We modified their original unsupervised setups to use weak classification supervision for semi-supervised learning (SSL). Concretely, as applied in our method, their objectives for labeled set and unlabeled set became
and .
The reconstruction and classification terms were also identical to those of our method, as shown below. For the reconstruction terms, true label for and inferred feature for were used to enable the decoder to better recognize one-hot class vectors.
and
The methods differ in the definition of to learn disentangled features and regularize the encoder, as shown below. For notational brevity, we omit the dependence of objectives and on their parameters . The terms related to independence of latent units are indicated by colors (i.e., , , and ). Note that we did not penalize the latent-data MI terms by setting in order to help and z capture the information about x.
[noitemsep]
VAE: the case with in the below term of -VAE
-VAE:
-TCVAE-1:
-TCVAE-2:
JointVAE:
IV-VAE-1 (ours):
IV-VAE-2 (ours):
where is the separate TC on z, is the collective TC that simultaneously considers and the variables in z, is the vector independence term between and z, and and are the dimension-wise regularization terms.
The term of -TCVAE-1 is obtained by applying the KL decomposition of (4) in the main paper separately to and in the -VAE objective and further assigning individual loss weights to the decomposed terms. Notice that does not exist for the single-dimensional variable (i.e., with is zero). The term of -TCVAE-2 is obtained by applying the KL decomposition of (4) in the main paper to in the -VAE objective, where the concatenation is considered as a single latent vector, and further assigning individual loss weights to the decomposed terms. For fair comparison, we assigned and separately to
Comments
There are no comments yet.