arXiv Analytics

Sign in

arXiv:2110.04616 [cs.LG]AbstractReferencesReviewsResources

Discriminative Multimodal Learning via Conditional Priors in Generative Models

Rogelio A. Mancisidor, Michael Kampffmeyer, Kjersti Aas, Robert Jenssen

Published 2021-10-09, updated 2022-07-26Version 2

Deep generative models with latent variables have been used lately to learn joint representations and generative processes from multi-modal data. These two learning mechanisms can, however, conflict with each other and representations can fail to embed information on the data modalities. This research studies the realistic scenario in which all modalities and class labels are available for model training, but where some modalities and labels required for downstream tasks are missing. We show, in this scenario, that the variational lower bound limits mutual information between joint representations and missing modalities. We, to counteract these problems, introduce a novel conditional multi-modal discriminative model that uses an informative prior distribution and optimizes a likelihood-free objective function that maximizes mutual information between joint representations and missing modalities. Extensive experimentation shows the benefits of the model we propose, the empirical results showing that our model achieves state-of-the-art results in representative problems such as downstream classification, acoustic inversion and annotation generation.

Related articles: Most relevant | Search more
arXiv:2006.10236 [cs.LG] (Published 2020-06-18)
Unsupervised Meta-Learning through Latent-Space Interpolation in Generative Models
arXiv:1903.09030 [cs.LG] (Published 2019-03-21)
Generative Models For Deep Learning with Very Scarce Data
arXiv:1911.06679 [cs.LG] (Published 2019-11-15)
Generative Models for Effective ML on Private, Decentralized Datasets