arXiv Analytics

Sign in

arXiv:2107.04735 [cs.CV]AbstractReferencesReviewsResources

Local-to-Global Self-Attention in Vision Transformers

Jinpeng Li, Yichao Yan, Shengcai Liao, Xiaokang Yang, Ling Shao

Published 2021-07-10Version 1

Transformers have demonstrated great potential in computer vision tasks. To avoid dense computations of self-attentions in high-resolution visual data, some recent Transformer models adopt a hierarchical design, where self-attentions are only computed within local windows. This design significantly improves the efficiency but lacks global feature reasoning in early stages. In this work, we design a multi-path structure of the Transformer, which enables local-to-global reasoning at multiple granularities in each stage. The proposed framework is computationally efficient and highly effective. With a marginal increasement in computational overhead, our model achieves notable improvements in both image classification and semantic segmentation. Code is available at https://github.com/ljpadam/LG-Transformer

Related articles: Most relevant | Search more
arXiv:2204.12451 [cs.CV] (Published 2022-04-26)
Understanding The Robustness in Vision Transformers
arXiv:2307.04066 [cs.CV] (Published 2023-07-09)
Random Position Adversarial Patch for Vision Transformers
arXiv:2106.03714 [cs.CV] (Published 2021-06-07)
Refiner: Refining Self-attention for Vision Transformers
Daquan Zhou et al.