Relating by Contrasting: A Data-efficient Framework for Multimodal Generative Models

07/02/2020
by   Yuge Shi, et al.
0

Multimodal learning for generative models often refers to the learning of abstract concepts from the commonality of information in multiple modalities, such as vision and language. While it has proven effective for learning generalisable representations, the training of such models often requires a large amount of "related" multimodal data that shares commonality, which can be expensive to come by. To mitigate this, we develop a novel contrastive framework for generative model learning, allowing us to train the model not just by the commonality between modalities, but by the distinction between "related" and "unrelated" multimodal data. We show in experiments that our method enables data-efficient multimodal learning on challenging datasets for various multimodal VAE models. We also show that under our proposed framework, the generative model can accurately identify related samples from unrelated ones, making it possible to make use of the plentiful unlabeled, unpaired multimodal data.

READ FULL TEXT
research
07/05/2022

A survey of multimodal deep generative models

Multimodal learning is a framework for building models that make predict...
research
10/08/2021

On the Limitations of Multimodal VAEs

Multimodal variational autoencoders (VAEs) have shown promise as efficie...
research
12/20/2021

Multimodal Adversarially Learned Inference with Factorized Discriminators

Learning from multimodal data is an important research topic in machine ...
research
10/26/2022

Using multimodal learning and deep generative models for corporate bankruptcy prediction

This research introduces for the first time the concept of multimodal le...
research
06/15/2020

Multimodal Generative Learning Utilizing Jensen-Shannon-Divergence

Learning from different data types is a long-standing goal in machine le...
research
07/01/2023

S-Omninet: Structured Data Enhanced Universal Multimodal Learning Architecture

Multimodal multitask learning has attracted an increasing interest in re...
research
09/07/2022

Benchmarking Multimodal Variational Autoencoders: GeBiD Dataset and Toolkit

Multimodal Variational Autoencoders (VAEs) have been a subject of intens...

Please sign up or login with your details

Forgot password? Click here to reset