arXiv Analytics

Sign in

arXiv:1707.05564 [cs.CV]AbstractReferencesReviewsResources

Batch based Monocular SLAM for Egocentric Videos

Suvam Patra, Kartikeya Gupta, Faran Ahmad, Chetan Arora, Subhashis Banerjee

Published 2017-07-18Version 1

Simultaneous Localization and Mapping (SLAM) from a monocular camera has been a well researched area. However, estimating camera pose and 3d geometry reliably for egocentric videos still remain a challenge. Some of the common causes of failures are dominant 3D rotations and low parallax between successive frames, resulting in unreliable pose and 3d estimates. For forward moving cameras, with no opportunities for loop closures, the drift leads to eventual failures for traditional feature based and direct SLAM techniques. We propose a novel batch mode structure from motion based technique for robust SLAM in such scenarios. In contrast to most of the existing techniques, we process frames in short batches, wherein we exploit short loop closures arising out of to-and-fro motion of wearer's head, and stabilize the egomotion estimates by 2D batch mode techniques such as motion averaging on pairwise epipolar results. Once pose estimates are obtained reliably over a batch, we refine the 3d estimate by triangulation and batch mode Bundle Adjustment (BA). Finally, we merge the batches using 3D correspondences and carry out a BA refinement post merging. We present both qualitative and quantitative comparison of our method on various public first and third person video datasets, to establish the robustness and accuracy of our algorithm over the state of the art.

Related articles: Most relevant | Search more
arXiv:1603.09200 [cs.CV] (Published 2016-03-30)
Unsupervised Understanding of Location and Illumination Changes in Egocentric Videos
arXiv:2403.08125 [cs.CV] (Published 2024-03-12)
Q-SLAM: Quadric Representations for Monocular SLAM
arXiv:1909.00713 [cs.CV] (Published 2019-09-02)
Estimation of Absolute Scale in Monocular SLAM Using Synthetic Data