arXiv Analytics

Sign in

arXiv:1910.07416 [cs.CV]AbstractReferencesReviewsResources

Understanding Misclassifications by Attributes

Sadaf Gulshad, Zeynep Akata, Jan Hendrik Metzen, Arnold Smeulders

Published 2019-10-15Version 1

In this paper, we aim to understand and explain the decisions of deep neural networks by studying the behavior of predicted attributes when adversarial examples are introduced. We study the changes in attributes for clean as well as adversarial images in both standard and adversarially robust networks. We propose a metric to quantify the robustness of an adversarially robust network against adversarial attacks. In a standard network, attributes predicted for adversarial images are consistent with the wrong class, while attributes predicted for the clean images are consistent with the true class. In an adversarially robust network, the attributes predicted for adversarial images classified correctly are consistent with the true class. Finally, we show that the ability to robustify a network varies for different datasets. For the fine grained dataset, it is higher as compared to the coarse-grained dataset. Additionally, the ability to robustify a network increases with the increase in adversarial noise.

Comments: arXiv admin note: substantial text overlap with arXiv:1904.08279
Categories: cs.CV
Related articles: Most relevant | Search more
arXiv:2305.08439 [cs.CV] (Published 2023-05-15)
Exploiting Frequency Spectrum of Adversarial Images for General Robustness
arXiv:1611.01646 [cs.CV] (Published 2016-11-05)
Boosting Image Captioning with Attributes
arXiv:2102.13262 [cs.CV] (Published 2021-02-26)
Improving Robustness of Learning-based Autonomous Steering Using Adversarial Images