arXiv Analytics

Sign in

arXiv:2301.00411 [cs.CV]AbstractReferencesReviewsResources

Detachable Novel Views Synthesis of Dynamic Scenes Using Distribution-Driven Neural Radiance Fields

Boyu Zhang, Wenbo Xu, Zheng Zhu, Guan Huang

Published 2023-01-01Version 1

Representing and synthesizing novel views in real-world dynamic scenes from casual monocular videos is a long-standing problem. Existing solutions typically approach dynamic scenes by applying geometry techniques or utilizing temporal information between several adjacent frames without considering the underlying background distribution in the entire scene or the transmittance over the ray dimension, limiting their performance on static and occlusion areas. Our approach $\textbf{D}$istribution-$\textbf{D}$riven neural radiance fields offers high-quality view synthesis and a 3D solution to $\textbf{D}$etach the background from the entire $\textbf{D}$ynamic scene, which is called $\text{D}^4$NeRF. Specifically, it employs a neural representation to capture the scene distribution in the static background and a 6D-input NeRF to represent dynamic objects, respectively. Each ray sample is given an additional occlusion weight to indicate the transmittance lying in the static and dynamic components. We evaluate $\text{D}^4$NeRF on public dynamic scenes and our urban driving scenes acquired from an autonomous-driving dataset. Extensive experiments demonstrate that our approach outperforms previous methods in rendering texture details and motion areas while also producing a clean static background. Our code will be released at https://github.com/Luciferbobo/D4NeRF.

Related articles: Most relevant | Search more
arXiv:1507.02438 [cs.CV] (Published 2015-07-09)
Generalized Video Deblurring for Dynamic Scenes
arXiv:1201.4895 [cs.CV] (Published 2012-01-23, updated 2013-06-26)
Compressive Acquisition of Dynamic Scenes
arXiv:1806.05620 [cs.CV] (Published 2018-06-14)
DynSLAM: Tracking, Mapping and Inpainting in Dynamic Scenes