arXiv Analytics

Sign in

arXiv:2004.03623 [cs.CV]AbstractReferencesReviewsResources

PatchVAE: Learning Local Latent Codes for Recognition

Kamal Gupta, Saurabh Singh, Abhinav Shrivastava

Published 2020-04-07Version 1

Unsupervised representation learning holds the promise of exploiting large amounts of unlabeled data to learn general representations. A promising technique for unsupervised learning is the framework of Variational Auto-encoders (VAEs). However, unsupervised representations learned by VAEs are significantly outperformed by those learned by supervised learning for recognition. Our hypothesis is that to learn useful representations for recognition the model needs to be encouraged to learn about repeating and consistent patterns in data. Drawing inspiration from the mid-level representation discovery work, we propose PatchVAE, that reasons about images at patch level. Our key contribution is a bottleneck formulation that encourages mid-level style representations in the VAE framework. Our experiments demonstrate that representations learned by our method perform much better on the recognition tasks compared to those learned by vanilla VAEs.

Related articles: Most relevant | Search more
arXiv:1910.01853 [cs.CV] (Published 2019-10-04)
DELP-DAR System for License Plate Detection and Recognition
arXiv:0903.4513 [cs.CV] (Published 2009-03-26, updated 2011-10-13)
Building the information kernel and the problem of recognition
arXiv:1506.01398 [cs.CV] (Published 2015-06-03)
Recognition of Changes in SAR Images Based on Gauss-Log Ratio and MRFFCM