arXiv Analytics

Sign in

arXiv:2211.09771 [cs.CV]AbstractReferencesReviewsResources

Boosting Object Representation Learning via Motion and Object Continuity

Quentin Delfosse, Wolfgang Stammer, Thomas Rothenbacher, Dwarak Vittal, Kristian Kersting

Published 2022-11-16Version 1

Recent unsupervised multi-object detection models have shown impressive performance improvements, largely attributed to novel architectural inductive biases. Unfortunately, they may produce suboptimal object encodings for downstream tasks. To overcome this, we propose to exploit object motion and continuity, i.e., objects do not pop in and out of existence. This is accomplished through two mechanisms: (i) providing priors on the location of objects through integration of optical flow, and (ii) a contrastive object continuity loss across consecutive image frames. Rather than developing an explicit deep architecture, the resulting Motion and Object Continuity (MOC) scheme can be instantiated using any baseline object detection model. Our results show large improvements in the performances of a SOTA model in terms of object discovery, convergence speed and overall latent object representations, particularly for playing Atari games. Overall, we show clear benefits of integrating motion and object continuity for downstream tasks, moving beyond object representation learning based only on reconstruction.

Related articles: Most relevant | Search more
arXiv:2204.03934 [cs.CV] (Published 2022-04-08)
Does Robustness on ImageNet Transfer to Downstream Tasks?
arXiv:2109.01134 [cs.CV] (Published 2021-09-02)
Learning to Prompt for Vision-Language Models
arXiv:2301.04101 [cs.CV] (Published 2023-01-10)
Neural Radiance Field Codebooks