arXiv Analytics

Sign in

arXiv:2105.14077 [cs.CV]AbstractReferencesReviewsResources

On the Bias Against Inductive Biases

George Cazenavette, Simon Lucey

Published 2021-05-28Version 1

Borrowing from the transformer models that revolutionized the field of natural language processing, self-supervised feature learning for visual tasks has also seen state-of-the-art success using these extremely deep, isotropic networks. However, the typical AI researcher does not have the resources to evaluate, let alone train, a model with several billion parameters and quadratic self-attention activations. To facilitate further research, it is necessary to understand the features of these huge transformer models that can be adequately studied by the typical researcher. One interesting characteristic of these transformer models is that they remove most of the inductive biases present in classical convolutional networks. In this work, we analyze the effect of these and more inductive biases on small to moderately-sized isotropic networks used for unsupervised visual feature learning and show that their removal is not always ideal.

Related articles: Most relevant | Search more
arXiv:2411.05001 [cs.CV] (Published 2024-11-07)
Analyzing The Language of Visual Tokens
arXiv:2210.11795 [cs.CV] (Published 2022-10-21)
PoseScript: 3D Human Poses from Natural Language
arXiv:2010.11701 [cs.CV] (Published 2020-09-29)
Spatial Attention as an Interface for Image Captioning Models