arXiv Analytics

Sign in

arXiv:2410.16499 [cs.CV]AbstractReferencesReviewsResources

SINGAPO: Single Image Controlled Generation of Articulated Parts in Object

Jiayi Liu, Denys Iliash, Angel X. Chang, Manolis Savva, Ali Mahdavi-Amiri

Published 2024-10-21Version 1

We address the challenge of creating 3D assets for household articulated objects from a single image. Prior work on articulated object creation either requires multi-view multi-state input, or only allows coarse control over the generation process. These limitations hinder the scalability and practicality for articulated object modeling. In this work, we propose a method to generate articulated objects from a single image. Observing the object in resting state from an arbitrary view, our method generates an articulated object that is visually consistent with the input image. To capture the ambiguity in part shape and motion posed by a single view of the object, we design a diffusion model that learns the plausible variations of objects in terms of geometry and kinematics. To tackle the complexity of generating structured data with attributes in multiple domains, we design a pipeline that produces articulated objects from high-level structure to geometric details in a coarse-to-fine manner, where we use a part connectivity graph and part abstraction as proxies. Our experiments show that our method outperforms the state-of-the-art in articulated object creation by a large margin in terms of the generated object realism, resemblance to the input image, and reconstruction quality.

Comments: Project page: https://3dlg-hcvc.github.io/singapo
Categories: cs.CV
Related articles: Most relevant | Search more
arXiv:1211.2881 [cs.CV] (Published 2012-11-13, updated 2012-11-28)
Deep Attribute Networks
arXiv:1905.03556 [cs.CV] (Published 2019-05-09)
Cycle-IR: Deep Cyclic Image Retargeting
arXiv:1807.07044 [cs.CV] (Published 2018-07-18)
Location Augmentation for CNN