arXiv Analytics

Sign in

arXiv:2010.03459 [stat.ML]AbstractReferencesReviewsResources

Learning disentangled representations with the Wasserstein Autoencoder

Benoit Gaujac, Ilya Feige, David Barber

Published 2020-10-07Version 1

Disentangled representation learning has undoubtedly benefited from objective function surgery. However, a delicate balancing act of tuning is still required in order to trade off reconstruction fidelity versus disentanglement. Building on previous successes of penalizing the total correlation in the latent variables, we propose TCWAE (Total Correlation Wasserstein Autoencoder). Working in the WAE paradigm naturally enables the separation of the total-correlation term, thus providing disentanglement control over the learned representation, while offering more flexibility in the choice of reconstruction cost. We propose two variants using different KL estimators and perform extensive quantitative comparisons on data sets with known generative factors, showing competitive results relative to state-of-the-art techniques. We further study the trade off between disentanglement and reconstruction on more-difficult data sets with unknown generative factors, where the flexibility of the WAE paradigm in the reconstruction term improves reconstructions.

Related articles: Most relevant | Search more
arXiv:1811.00007 [stat.ML] (Published 2018-10-31)
Interventional Robustness of Deep Latent Variable Models
arXiv:1912.03915 [stat.ML] (Published 2019-12-09)
Learning Disentangled Representations via Mutual Information Estimation
arXiv:1706.00400 [stat.ML] (Published 2017-06-01)
Learning Disentangled Representations with Semi-Supervised Deep Generative Models
N. Siddharth et al.