arXiv Analytics

Sign in

arXiv:1812.11166 [cs.CV]AbstractReferencesReviewsResources

Learning to Reconstruct Shapes from Unseen Classes

Xiuming Zhang, Zhoutong Zhang, Chengkai Zhang, Joshua B. Tenenbaum, William T. Freeman, Jiajun Wu

Published 2018-12-28Version 1

From a single image, humans are able to perceive the full 3D shape of an object by exploiting learned shape priors from everyday life. Contemporary single-image 3D reconstruction algorithms aim to solve this task in a similar fashion, but often end up with priors that are highly biased by training classes. Here we present an algorithm, Generalizable Reconstruction (GenRe), designed to capture more generic, class-agnostic shape priors. We achieve this with an inference network and training procedure that combine 2.5D representations of visible surfaces (depth and silhouette), spherical shape representations of both visible and non-visible surfaces, and 3D voxel-based representations, in a principled manner that exploits the causal structure of how 3D shapes give rise to 2D images. Experiments demonstrate that GenRe performs well on single-view shape reconstruction, and generalizes to diverse novel objects from categories not seen during training.

Comments: NeurIPS 2018 (Oral). The first two authors contributed equally to this paper. Project page: http://genre.csail.mit.edu/
Categories: cs.CV, cs.AI
Related articles: Most relevant | Search more
arXiv:2210.15200 [cs.CV] (Published 2022-10-27)
Deep-MDS Framework for Recovering the 3D Shape of 2D Landmarks from a Single Image
arXiv:2111.14465 [cs.CV] (Published 2021-11-29, updated 2022-04-07)
Motion-from-Blur: 3D Shape and Motion Estimation of Motion-blurred Objects in Videos
arXiv:1809.07917 [cs.CV] (Published 2018-09-21)
Adaptive O-CNN: A Patch-based Deep Representation of 3D Shapes