arXiv Analytics

Sign in

arXiv:1912.08265 [cs.CV]AbstractReferencesReviewsResources

Learning from Synthetic Animals

Jiteng Mu, Weichao Qiu, Gregory Hager, Alan Yuille

Published 2019-12-17Version 1

Despite great success in human parsing, progress for parsing other deformable articulated objects, like animals, is still limited by the lack of labeled data. In this paper, we use synthetic images and ground truth generated from CAD animal models to address this challenge. To bridge the gap between real and synthetic images, we propose a novel consistency-constrained semi-supervised learning method (CC-SSL). Our method leverages both spatial and temporal consistencies, to bootstrap weak models trained on synthetic data with unlabeled real images. We demonstrate the effectiveness of our method on highly deformable animals, such as horses and tigers. Without using any real image label, our method allows for accurate keypoints prediction on real images. Moreover, we quantitatively show that models using synthetic data achieve better generalization performance than models trained on real images across different domains in the Visual Domain Adaptation Challenge dataset. Our synthetic dataset contains 10+ animals with diverse poses and rich ground truth, which enables us to use the multi-task learning strategy to further boost models' performance.

Related articles: Most relevant | Search more
arXiv:2403.10004 [cs.CV] (Published 2024-03-15)
ST-LDM: A Universal Framework for Text-Grounded Object Generation in Real Images
arXiv:1711.11556 [cs.CV] (Published 2017-11-30)
ROAD: Reality Oriented Adaptation for Semantic Segmentation of Urban Scenes
arXiv:1903.05820 [cs.CV] (Published 2019-03-14)
Purifying Naturalistic Images through a Real-time Style Transfer Semantics Network