1 Introduction
The highlevel goal of this paper is to train a neural model to learn a semantic embedding space into which speech audio and image inputs can be mapped. This goal is inspired by the fact that children are able to learn through associating stimuli during their early years. For example, a child hearing his or her mother pronounce “seven” or write a “7” might learn to think of the same concept upon hearing or seeing either. In fact, Man et al. showed that the temporoparietal cortex of the human brain produces contentspecific and modalityinvariant neural responses to audio and visual stimuli [1].
In our method, we map image and audio inputs to the parameterizations of diagonal Gaussians representing the posterior distribution over semantic embeddings. We then sample embeddings from this distribution and use a loss function which encourages samples from paired audio and image inputs to be more similar than mismatched pairs of audio and images. Although this objective has been shown to encourage the embedding space to be rich semantically [2], in this paper we explore methods of better encouraging modalityinvariance. That is, not only should semantically relevant content be clustered in the embedding space, but the distributions of embeddings for semantically equivalent audio and images should be the same. This goal is based on the assumption that information concerning modality is noise for tasks requiring only the semantic content of the sensory input.
To drive the posterior distributions over embeddings to be the same for semantically equivalent inputs across modalities, we introduce a term to the objective which regularizes the amount of information encoded in the semantic embedding. The term, borrowed from variational autoencoders (VAEs), is the sum of the KL divergences of the posterior distributions from the unit Gaussian. Our results suggest that when this regularization term is increased from zero during hyperparameter tuning, modalityinformation tends to be filtered out prior to semanticinformation. We believe this technique has the potential to be useful for modalityinvariant and domaininvariant applications.
2 Previous Work
(1) 
The unsupervised problem of learning semantic relations through the cooccurrence and lack of cooccurrence of sensory inputs is an increasingly attractive pursuit for researchers [2, 3, 4, 5]. This attraction is primarily due to the expense of attaining labels for data. The ability to learn semantic relevance with input pairings alone unlocks the potential of training models using inexpensivelycollected data with the only supervisory signal being the cooccurrence of sensory inputs [3]. In addition, the learned semantic space has direct practical applications. One particular application of a semantic space is crossmodality transfer learning: using paired inputs from two modalities and labels for one modality to learn how to predict labels for the unlabeled modality. Aytar, Vondrick, and Torralba [5]
use a teacherstudent model on videos to transfer knowledge from pretrained ImageNet and Places convolutional neural networks (CNNs) identifying object and scene information in images to train a CNN run on the raw audio waveform from the video to recognize the same information. In Aytar et al.’s model, the shared semantic space consists of the two distributions over objects and scenes as opposed to being an arbitrary (yet highly linearly correlated) semantic space, as is the case in our model. Wang et al.
[3] gave a comprehensive overview of existing approaches to another practical application of shared semantic spaces: crossmodality information retrieval. The task is formulated as follows: given an input of one modality, find related instances of another modality. In 2016, Harwath et al. [2] presented a method to learn a semantic embedding space into which images and spoken audio recordings of captions of the images could be mapped. They evaluated their method by looking at the crossmodality retrieval recall scores: e.g., given an image and audio captions, which of the audio captions describes the image? K. Saito et al. developed an adversarial neural architecture to learn modalityinvariant representations of paired images and text [4]. Modalityinvariance was encouraged using an adversarial setup in which the discriminator was given one of the two representations or a sample drawn from the unit Gaussian. The discriminator was tasked with determining which modality the input originated from or whether it was drawn from the unit Gaussian. The encoders were trained through gradient reversal, as used previously in adversarial domain adaptation and generative adversarial networks [4, 6, 7, 8, 9]. Kashyap [10] also applied Harwath et al.’s [2] approach to the MNIST and TIDIGITs dataset, focusing primarily on using the embeddings for crossmodality transfer learning. Our work focuses more on the embeddings themselves, and methods to promote modalityinvariance. Hsu et al. [11] designed a convolutional variational autoencoder (CVAE) for log melfilterbanks of speech drawn from the TIMIT dataset. In our work, we use the same convolutional network architecture for our audio encoder.Our network architecture and loss function is based on Harwath et al., but instead of deterministically mapping inputs to embeddings, we map inputs to the parameterization of a diagonal Gaussian, and sample embeddings from it. In addition, we add a regularization term for the posterior distributions. In this regard, our method takes a similar approach to achieving modalityinvariance as K. Saito et al. insofar as we both drive the distribution of embeddings to have minimal deviation from a unit Gaussian prior distribution of embeddings [4], but we found that our encoders can deceive a discriminator without using gradient reversal. In addition, we believe the problem of modalityinvariant embeddings using speech as one of the modalities has yet to be explored, so our research makes a novel contribution in this area.
3 Methods
We first formalize the problem. Given a set of cooccurring images and captions, where (image space) and (audio caption space), functions and are chosen to optimize some objective that promotes the encoding of semantic information contained in the inputs and into and , respectively. For example, if is a picture of a handwritten “7” and is an audio recording of someone saying “seven”, and should be considered highly semantically related by some similarity metric. As in [2], we aim to increase the margin between the similarity of representations of cooccurring inputs and the similarity of representations of noncooccurring inputs. The similarity loss function is given in Equation 1.
In contrast to [2], our encoders, and , are nondeterministic. We learn the deterministic functions and . Then we use and to parameterize a diagonal Gaussian representing the posterior distribution over embeddings:
(2) 
Embeddings are then sampled from the posterior:
(3) 
and likewise for . We illustrate this process in Figure 1.
In addition to , defined in Equation 1, we average the KL divergence of the predicted posteriors over embeddings from the prior over embeddings (the unit Gaussian) as a regularization term we call information gain (IG) loss:
(4) 
Our total loss function is then:
(5) 
where is the sum of all Frobenius norms of weight matrices and convolutional kernels, and and are tunable hyperparameters.
4 Datasets
For images, we used the MNIST dataset of handwritten digits [12]. The dataset contains 60K training images and 10K test images. The images are 28x28 8bit grayscale images, and we preprocess each image to have pixel values between 0 and 1. For audio, we used the TIDIGITS dataset of spoken utterances sampled at 20 KHz [13]. We only used digit strings containing a single number, and we used utterances from men, women, and children. After filtering out utterances which contain more than one number, we have 6,456 training utterances, 1,076 test utterances, and 1,076 validation utterances. Using the Kaldi speech recognition toolkit [14]
, we generated 80 dimensional log melfilterbank features with a 25ms window size and a 10ms frame shift, multiplied by a Povey window. To create inputs of the same size, we pad or crop each spectrogram to 100 frames (i.e., one second of speech) which is one frame longer than the mean frame length of the available utterances. We preprocessed each filterbank to have zero mean and unit variance. Longer utterances were center cropped. Shorter utterances were zero padded at the end after adjusting the filterbank to have zero mean. For TIDIGITS, we also combined the utterances labeled “oh” and “zero” into one class for the purpose of labeling clusters in our analysis
^{1}^{1}1Training does not depend on explicit class labels except insofar as pairing audio and image inputs based on their ground truth digit labels..5 Experiments
We used convolutional neural networks to predict the parameterizations of and (Equation 2). We trained the networks to minimize Equation 5 for the MNIST and TIDIGITS datasets described in Section 4. We compared the embedding spaces produced when and when to gauge the effect of regularizing information gain in the posterior.
We set the embedding dimension to be , which is consistent with the latent embedding dimensionality used by [11] for their variational autoencoder for 58 phones. We did not explore other values of . The encoders for both images and audio are convolutional networks which produce the parameterization (the mean and log variance vectors) of the posterior distribution over embeddings. The audio encoder uses the same architecture as the encoder portion of Hsu et al.’s variational autoencoder for 80 dimensional log melfilterbank speech [11].
The image encoder is also convolutional, taking the following form:

[topsep=0pt,itemsep=1ex,partopsep=1ex,parsep=1ex]

conv., 64 filters, same padding

conv., strides, 128 filters, same padding

conv., strides, 256 filters, same padding

unit fully connected

unit linear output ( for , for )
ReLu activations were used for each layer except the final linear layer. A weight decay () of is used for all convolutional and fully connected layers. The initial learning rate was
which was decayed by a factor of 0.9 every 10 epochs. The Adam optimization algorithm was used with
, , and . 128 distinct imageaudio pairs were used for each batch. After processing each image or audio input through the respective encoder to produce a posterior distribution, 16 embeddings were sampled per input.^{2}^{2}2Positive imageaudio embedding pairings were established by matching corresponding sampled embeddings for each input. This produced a total of 2,048 imageaudio embedding pairs in each batch.Negative sampling was performed by selecting one of the other 2,047 sample pairs in the batch. While it would at first seem reasonable to disallow negative samples for a training pair to be drawn from the same underlying digit class, such a mechanism implies a ground truth digit labelling of all examples within a batch. In other words, the knowledge of which negative example pairs not to sample is equivalent to the network possessing an oracle that knows which audio/visual sample pairs within a batch were drawn from the same underlying digit class. This oracle would allow the network to trivially recover the ground truth digit labelling of all examples within a batch. In an effort to avoid this, we allow negative samples to be chosen from any digit class regardless of the initial example’s digit class. Empirically, we found that the weight of the positive examples can easily overcome the “contradictory” signals introduced by this sampling scheme, allowing the model to produce a semantically rich embedding space.
The model was trained for 100 epochs. An epoch was defined as the number of batches required to cover all training examples in the larger of the two datasets (MNIST) exactly once. Training required about 35 minutes on an NVIDIA TitanX GPU.
6 Results and Analysis
To analyze the learned semantic space, we sampled embeddings for inputs from the unseen test set, sampling 16 samples per input point. We ran Kmeans clustering with
and calculated the cluster purity of the resulting clusters, defined as:(6) 
where is the set of all points in cluster and is the set of all points of class
(their ground truth digit label). This metric represents the accuracy of a classifier which classifies a point,
, according to the majority class of the cluster whose mean is closest to using euclidean distance.We then used a subset of 2,152 sample embeddings (1,076 from images, 1,076 from audio) and performed a classification task to predict the original input point’s modality from the embeddings using an SVM with a Gaussian RBF kernel. 1600 examples were used for the training set and the remaining were used for the test set. We used 3fold validation to select a value for the SVM. Comparing the modality classification test accuracy to the prior on modality () allows us to gauge the extent to which the embeddings are modalityinvariant. Perfectly modalityinvariant embeddings would result in a modality classification test accuracy of .
We evaluated the effect of on the cluster purity and modality invariance of the embeddings learned by our model. Results from using the modality classifier and cluster purity analysis are shown in Figure 3 and Table 1.
Cluster Purity  Modality SVM Acc.  

0.00e+00  0.525  1.000 
1.00e05  0.542  1.000 
6.81e05  0.516  1.000 
4.64e04  0.707  1.000 
3.16e03  0.980  0.859 
2.15e02  0.984  0.554 
1.47e01  0.975  0.520 
1.00e+00  0.679  0.516 
In addition, we used 200 samples per modality to compute a two dimensional tSNE projection^{3}^{3}3tSNE was selected over PCA for its ability to show relative pairwise distances [15] of the embeddings produced by each hyperparameter setting. We plotted these samples in Figure 2 and colored them according to class label. For both cells in a row, the same TSNE model was used, so the embeddings for both modalities were projected into the same twodimensional space.
The additional term resulted in greater cluster purity, as shown in Figure 3. The lower cluster purity for alone () is visually evident in the first row of Figure 2: though there are clear semantic clusterings of samples from the same digit, there are typically two clusters per digit—one for images and one for audio. One possible explanation for why the cluster purity is low () for is that when KMeans is performed with , is about half the number of digit clusters in the embedding space (one for each digitmodality pair), resulting in KMeans clusters with members nearly evenly split between two digits. This finding shows that while using alone, embeddings originating from the same modality may still be significantly closer together than embeddings of different modalities, regardless of the similarity of semantic content. The 100% accuracy of the SVM in predicting the modality of embeddings when , as shown in Figure 3(a), further supports the finding that the embedding space produced from using alone is not modality invariant. This trend is not conducive to modality invariance of the embeddings.
In contrast, the embeddings produced when using for training were only able to be classified by an SVM with 55.4% accuracy, as shown in Table 1. Although this metric is not the ideal 50% accuracy of truly modalityinvariant embeddings, the embedding space produced using is much closer to being modalityinvariant than the space produced by alone.
Figure 3 shows that minimizing the divergence of the posterior over embeddings from the prior improves modality invariance. This could be due to the fact that the KL divergence represents the amount of information about an embedding conveyed by an input, and by limiting the amount of information, we force the encoders to filter out information. This is the same reason why variational autoencoders exhibit denoising behavior [16]. Since semantic information is important for minimizing , modality information tends to be filtered out before semantic information. For our applications, Figure 3(ii) shows that is the empirically observed ideal cutoff point at which increasing to further limit the total information conveyed in the posterior begins to also to overly restrict the semantic information conveyed, resulting in a drop in cluster purity. Figure 2 shows this trend qualitatively. Row 1 shows sampled embeddings resulting from an underregularized model; row 3, well regularized; and row 5, over regularized.
7 Conclusions
In this work, our goal was to learn a joint modalityinvariant semantic embedding space for speech and images in an unsupervised manner. We focused on spoken utterances and images of handwritten digits. We found that by sampling encodings rather than predicting them directly, and by regularizing the posterior distribution over embeddings, we were able to learn a more modalityinvariant semantic embedding space. From an adversarial perspective, we were able to deceive an adversarial discriminator (the modalityclassifying SVM) without the use of gradient reversal or any adversarial setup during training. This leads us to suspect may be a useful regularization term in other generative adversarial approaches to learning domain or modality invariant embeddings.
Further research could be done to attempt to combine the techniques used by VAEs and GANs. One potential direction in the vein of multimodal learning of a semantic space is to replace the dot product similarity with a symmetric divergence of the variational distributions of matched and mismatched inputs. This would allow for a more probabilistically theoretical formulation of the loss function which could have more general implications for other areas of research.
Acknowledgements
The authors would like to thank WeiNing Hsu for his help with the audio encoder architecture, and advice on variational autoencoders.
References
 [1] Kingson Man, Jonas T Kaplan, Antonio Damasio, and Kaspar Meyer, “Sight and sound converge to form modalityinvariant representations in temporoparietal cortex,” Journal of Neuroscience, vol. 32, no. 47, pp. 16629–16636, 2012.
 [2] David Harwath, Antonio Torralba, and James Glass, “Unsupervised learning of spoken language with visual context,” in Advances in Neural Information Processing Systems, 2016, pp. 1858–1866.
 [3] Kaiye Wang, Qiyue Yin, Wei Wang, Shu Wu, and Liang Wang, “A comprehensive survey on crossmodal retrieval,” arXiv preprint arXiv:1607.06215, 2016.
 [4] Kuniaki Saito, Yusuke Mukuta, Yoshitaka Ushiku, and Tatsuya Harada, “Demian: Deep modality invariant adversarial network,” arXiv preprint arXiv:1612.07976, 2016.
 [5] Yusuf Aytar, Carl Vondrick, and Antonio Torralba, “Soundnet: Learning sound representations from unlabeled video,” in Advances in Neural Information Processing Systems, 2016, pp. 892–900.

[6]
Yaroslav Ganin and Victor Lempitsky,
“Unsupervised domain adaptation by backpropagation,”
inInternational Conference on Machine Learning
, 2015, pp. 1180–1189.  [7] Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky, “Domainadversarial training of neural networks,” Journal of Machine Learning Research, vol. 17, no. 59, pp. 1–35, 2016.
 [8] Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell, “Adversarial discriminative domain adaptation,” arXiv preprint arXiv:1702.05464, 2017.
 [9] Ian Goodfellow, Jean PougetAbadie, Mehdi Mirza, Bing Xu, David WardeFarley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio, “Generative adversarial nets,” in Advances in neural information processing systems, 2014, pp. 2672–2680.
 [10] Karan Kashyap, “Learning digits via joint audiovisual representations,” M.S. thesis, Massachusetts Institute of Technology, 2017.
 [11] WeiNing Hsu, Yu Zhang, and James Glass, “Learning latent representations for speech generation and transformation,” arXiv preprint arXiv:1704.04222, 2017.
 [12] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradientbased learning applied to document recognition.,” in Proceedings of the IEEE, 86(11):2272324, November 1998.
 [13] R Gary Leonard and George Doddington, “Tidigits speech corpus,” Texas Instruments, Inc, 1993.
 [14] Daniel Povey, Arnab Ghoshal, Gilles Boulianne, Lukas Burget, Ondrej Glembek, Nagendra Goel, Mirko Hannemann, Petr Motlicek, Yanmin Qian, Petr Schwarz, et al., “The kaldi speech recognition toolkit,” in IEEE 2011 workshop on automatic speech recognition and understanding. IEEE Signal Processing Society, 2011, number EPFLCONF192584.
 [15] Laurens van der Maaten and Geoffrey Hinton, “Visualizing data using tsne,” Journal of Machine Learning Research, vol. 9, no. Nov, pp. 2579–2605, 2008.
 [16] Diederik P Kingma and Max Welling, “Autoencoding variational bayes,” arXiv preprint arXiv:1312.6114, 2013.
Comments
There are no comments yet.