arXiv:1807.10272 [stat.ML]AbstractReferencesReviewsResources
Evaluating and Understanding the Robustness of Adversarial Logit Pairing
Logan Engstrom, Andrew Ilyas, Anish Athalye
Published 2018-07-26Version 1
We evaluate the robustness of Adversarial Logit Pairing, a recently proposed defense against adversarial examples. We find that a network trained with Adversarial Logit Pairing achieves 0.6% accuracy in the threat model in which the defense is considered. We provide a brief overview of the defense and the threat models/claims considered, as well as a discussion of the methodology and results of our attack, which may offer insights into the reasons underlying the vulnerability of ALP to adversarial attack.
Comments: Source code at https://github.com/labsix/adversarial-logit-pairing-analysis
Keywords: robustness, adversarial logit pairing achieves, evaluating, adversarial examples, threat models/claims
Tags: github project
Related articles: Most relevant | Search more
arXiv:1706.03922 [stat.ML] (Published 2017-06-13)
Analyzing the Robustness of Nearest Neighbors to Adversarial Examples
arXiv:1905.02175 [stat.ML] (Published 2019-05-06)
Adversarial Examples Are Not Bugs, They Are Features
arXiv:1902.08265 [stat.ML] (Published 2019-02-21)
Quantifying Perceptual Distortion of Adversarial Examples