arXiv Analytics

Sign in

arXiv:1706.06629 [cs.CV]AbstractReferencesReviewsResources

Co-Fusion: Real-time Segmentation, Tracking and Fusion of Multiple Objects

Martin Rünz, Lourdes Agapito

Published 2017-06-20Version 1

In this paper we introduce Co-Fusion, a dense SLAM system that takes a live stream of RGB-D images as input and segments the scene into different objects (using either motion or semantic cues) while simultaneously tracking and reconstructing their 3D shape in real time. We use a multiple model fitting approach where each object can move independently from the background and still be effectively tracked and its shape fused over time using only the information from pixels associated with that object label. Previous attempts to deal with dynamic scenes have typically considered moving regions as outliers, and consequently do not model their shape or track their motion over time. In contrast, we enable the robot to maintain 3D models for each of the segmented objects and to improve them over time through fusion. As a result, our system can enable a robot to maintain a scene description at the object level which has the potential to allow interactions with its working environment; even in the case of dynamic scenes.

Comments: International Conference on Robotics and Automation (ICRA) 2017, http://visual.cs.ucl.ac.uk/pubs/cofusion, https://github.com/martinruenz/co-fusion
Categories: cs.CV
Related articles: Most relevant | Search more
arXiv:1201.4895 [cs.CV] (Published 2012-01-23, updated 2013-06-26)
Compressive Acquisition of Dynamic Scenes
arXiv:1910.06407 [cs.CV] (Published 2019-10-14)
FireNet: Real-time Segmentation of Fire Perimeter from Aerial Video
Jigar Doshi et al.
arXiv:1507.02438 [cs.CV] (Published 2015-07-09)
Generalized Video Deblurring for Dynamic Scenes