arXiv Analytics

Sign in

arXiv:2209.11189 [cs.CV]AbstractReferencesReviewsResources

Learning Visual Explanations for DCNN-Based Image Classifiers Using an Attention Mechanism

Ioanna Gkartzonika, Nikolaos Gkalelis, Vasileios Mezaris

Published 2022-09-22Version 1

In this paper two new learning-based eXplainable AI (XAI) methods for deep convolutional neural network (DCNN) image classifiers, called L-CAM-Fm and L-CAM-Img, are proposed. Both methods use an attention mechanism that is inserted in the original (frozen) DCNN and is trained to derive class activation maps (CAMs) from the last convolutional layer's feature maps. During training, CAMs are applied to the feature maps (L-CAM-Fm) or the input image (L-CAM-Img) forcing the attention mechanism to learn the image regions explaining the DCNN's outcome. Experimental evaluation on ImageNet shows that the proposed methods achieve competitive results while requiring a single forward pass at the inference stage. Moreover, based on the derived explanations a comprehensive qualitative analysis is performed providing valuable insight for understanding the reasons behind classification errors, including possible dataset biases affecting the trained classifier.

Comments: Accepted for publication; to be included in Proc. ECCV Workshops 2022. The version posted here is the "submitted manuscript" version
Categories: cs.CV
Related articles: Most relevant | Search more
arXiv:1312.6082 [cs.CV] (Published 2013-12-20, updated 2014-04-14)
Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks
arXiv:2309.00331 [cs.CV] (Published 2023-09-01)
Human trajectory prediction using LSTM with Attention mechanism
arXiv:1812.10025 [cs.CV] (Published 2018-12-25)
Attention Branch Network: Learning of Attention Mechanism for Visual Explanation