arXiv Analytics

Sign in

arXiv:1905.07672 [cs.LG]AbstractReferencesReviewsResources

Things You May Not Know About Adversarial Example: A Black-box Adversarial Image Attack

Yuchao Duan, Zhe Zhao, Lei Bu, Fu Song

Published 2019-05-19Version 1

Numerous methods for crafting adversarial examples were proposed recently with high attack success rate. Most of the existing works normalize images into a continuous vector, domain firstly, and then craft adversarial examples in the continuous vector space. However, "adversarial" examples may become benign after de-normalizing them back into discrete integer domain, known as the discretization problem. The discretization problem was mentioned in some work, but was despised and have received relatively little attention. In this work, we conduct the first comprehensive study of this discretization problem. We theoretically analyzed 34 representative methods and empirically studied 20 representative open source tools for crafting discretization images. Our findings reveal that almost all of existing works suffer from the discretization problem and the problem is far more serious than we thought. This suggests that the discretization problem should be taken into account when crafting adversarial examples. As a first step towards addressing the discretization problem, we propose a black-box attack method to encode the adversarial example searching problem as a derivative-free optimization problem. Our method is able to craft "real'' adversarial images by derivative-free search on the discrete integer domain. Experimental results show that our method achieves significantly higher attack success rates on the discrete integer domain than most of the other tools, no matter white-box or black-box. Moreover, our method is able to handle any model that is not differentiable and we successfully break the winner of NIPS 17 competition on defense with a 95\% success rate.

Related articles: Most relevant | Search more
arXiv:2303.14173 [cs.LG] (Published 2023-03-24)
How many dimensions are required to find an adversarial example?
arXiv:1905.00441 [cs.LG] (Published 2019-05-01)
NATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks
arXiv:1901.08360 [cs.LG] (Published 2019-01-24)
Cross-Entropy Loss and Low-Rank Features Have Responsibility for Adversarial Examples