arXiv Analytics

Sign in

arXiv:1905.13284 [cs.LG]AbstractReferencesReviewsResources

Identifying Classes Susceptible to Adversarial Attacks

Rangeet Pan, Md Johirul Islam, Shibbir Ahmed, Hridesh Rajan

Published 2019-05-30Version 1

Despite numerous attempts to defend deep learning based image classifiers, they remain susceptible to the adversarial attacks. This paper proposes a technique to identify susceptible classes, those classes that are more easily subverted. To identify the susceptible classes we use distance-based measures and apply them on a trained model. Based on the distance among original classes, we create mapping among original classes and adversarial classes that helps to reduce the randomness of a model to a significant amount in an adversarial setting. We analyze the high dimensional geometry among the feature classes and identify the k most susceptible target classes in an adversarial attack. We conduct experiments using MNIST, Fashion MNIST, CIFAR-10 (ImageNet and ResNet-32) datasets. Finally, we evaluate our techniques in order to determine which distance-based measure works best and how the randomness of a model changes with perturbation.

Related articles: Most relevant | Search more
arXiv:1802.05193 [cs.LG] (Published 2018-02-14)
Security Analysis and Enhancement of Model Compressed Deep Learning Systems under Adversarial Attacks
arXiv:2001.07769 [cs.LG] (Published 2020-01-21)
Massif: Interactive Interpretation of Adversarial Attacks on Deep Learning
Nilaksh Das et al.
arXiv:1906.03563 [cs.LG] (Published 2019-06-09)
Beyond Adversarial Training: Min-Max Optimization in Adversarial Attack and Defense