arXiv Analytics

Sign in

arXiv:2208.13975 [cs.CV]AbstractReferencesReviewsResources

MRL: Learning to Mix with Attention and Convolutions

Shlok Mohta, Hisahiro Suganuma, Yoshiki Tanaka

Published 2022-08-30Version 1

In this paper, we present a new neural architectural block for the vision domain, named Mixing Regionally and Locally (MRL), developed with the aim of effectively and efficiently mixing the provided input features. We bifurcate the input feature mixing task as mixing at a regional and local scale. To achieve an efficient mix, we exploit the domain-wide receptive field provided by self-attention for regional-scale mixing and convolutional kernels restricted to local scale for local-scale mixing. More specifically, our proposed method mixes regional features associated with local features within a defined region, followed by a local-scale features mix augmented by regional features. Experiments show that this hybridization of self-attention and convolution brings improved capacity, generalization (right inductive bias), and efficiency. Under similar network settings, MRL outperforms or is at par with its counterparts in classification, object detection, and segmentation tasks. We also show that our MRL-based network architecture achieves state-of-the-art performance for H&E histology datasets. We achieved DICE of 0.843, 0.855, and 0.892 for Kumar, CoNSep, and CPM-17 datasets, respectively, while highlighting the versatility offered by the MRL framework by incorporating layers like group convolutions to improve dataset-specific generalization.

Related articles: Most relevant | Search more
arXiv:1603.06759 [cs.CV] (Published 2016-03-22)
Convolution in Convolution for Network in Network
arXiv:2303.10999 [cs.CV] (Published 2023-03-20)
Induced Feature Selection by Structured Pruning
arXiv:2209.06953 [cs.CV] (Published 2022-09-14)
On the interplay of adversarial robustness and architecture components: patches, convolution and attention