arXiv Analytics

Sign in

arXiv:2505.06219 [cs.CV]AbstractReferencesReviewsResources

VIN-NBV: A View Introspection Network for Next-Best-View Selection for Resource-Efficient 3D Reconstruction

Noah Frahm, Dongxu Zhao, Andrea Dunn Beltran, Ron Alterovitz, Jan-Michael Frahm, Junier Oliva, Roni Sengupta

Published 2025-05-09Version 1

Next Best View (NBV) algorithms aim to acquire an optimal set of images using minimal resources, time, or number of captures to enable efficient 3D reconstruction of a scene. Existing approaches often rely on prior scene knowledge or additional image captures and often develop policies that maximize coverage. Yet, for many real scenes with complex geometry and self-occlusions, coverage maximization does not lead to better reconstruction quality directly. In this paper, we propose the View Introspection Network (VIN), which is trained to predict the reconstruction quality improvement of views directly, and the VIN-NBV policy. A greedy sequential sampling-based policy, where at each acquisition step, we sample multiple query views and choose the one with the highest VIN predicted improvement score. We design the VIN to perform 3D-aware featurization of the reconstruction built from prior acquisitions, and for each query view create a feature that can be decoded into an improvement score. We then train the VIN using imitation learning to predict the reconstruction improvement score. We show that VIN-NBV improves reconstruction quality by ~30% over a coverage maximization baseline when operating with constraints on the number of acquisitions or the time in motion.

Related articles: Most relevant | Search more
arXiv:2105.14230 [cs.CV] (Published 2021-05-29)
Transforming the Latent Space of StyleGAN for Real Face Editing
arXiv:2307.04187 [cs.CV] (Published 2023-07-09)
Predictive Coding For Animation-Based Video Compression
arXiv:2012.09036 [cs.CV] (Published 2020-12-13)
Improved StyleGAN Embedding: Where are the Good Latents?