Learning multimodal VAEs through mutual supervision
Multimodal VAEs seek to model the joint distribution over heterogeneous data (e.g.\ vision, language), whilst also capturing a shared representation across such modalities. Prior work has typically combined information from the modalities by reconciling idiosyncratic representations directly in the...
Main Authors: | Joy, T, Shi, Y, Torr, PHS, Rainforth, T, Schmon, SM, Siddharth, N |
---|---|
Format: | Conference item |
Language: | English |
Published: |
OpenReview
2022
|
Similar Items
-
Capturing label characteristics in VAEs
by: Joy, T, et al.
Published: (2021) -
Adversarial masking for self-supervised learning
by: Shi, Y, et al.
Published: (2022) -
Relating by contrasting: A data-efficient framework for multimodal DGMs
by: Shi, Y, et al.
Published: (2021) -
Improving generative modelling in VAEs using Multimodal Prior
by: Abrol, V, et al.
Published: (2020) -
iVAE-GAN: Identifiable VAE-GAN Models for Latent Representation Learning
by: Bjorn Uttrup Dideriksen, et al.
Published: (2022-01-01)