Generalized Multimodal ELBO

05/06/2021
by   Thomas M. Sutter, et al.
9

Multiple data types naturally co-occur when describing real-world phenomena and learning from them is a long-standing goal in machine learning research. However, existing self-supervised generative models approximating an ELBO are not able to fulfill all desired requirements of multimodal models: their posterior approximation functions lead to a trade-off between the semantic coherence and the ability to learn the joint data distribution. We propose a new, generalized ELBO formulation for multimodal data that overcomes these limitations. The new objective encompasses two previous methods as special cases and combines their benefits without compromises. In extensive experiments, we demonstrate the advantage of the proposed method compared to state-of-the-art models in self-supervised, generative learning tasks.

READ FULL TEXT

page 7

page 9

page 18

page 19

page 20

page 23

research
06/15/2020

Multimodal Generative Learning Utilizing Jensen-Shannon-Divergence

Learning from different data types is a long-standing goal in machine le...
research
12/27/2022

GEDI: GEnerative and DIscriminative Training for Self-Supervised Learning

Self-supervised learning is a popular and powerful method for utilizing ...
research
03/06/2020

Noise Estimation Using Density Estimation for Self-Supervised Multimodal Learning

One of the key factors of enabling machine learning models to comprehend...
research
03/31/2023

Self-Supervised Multimodal Learning: A Survey

Multimodal learning, which aims to understand and analyze information fr...
research
04/22/2023

Learning Symbolic Representations Through Joint GEnerative and DIscriminative Training

We introduce GEDI, a Bayesian framework that combines existing self-supe...
research
10/08/2021

On the Limitations of Multimodal VAEs

Multimodal variational autoencoders (VAEs) have shown promise as efficie...

Please sign up or login with your details

Forgot password? Click here to reset