arXiv Analytics

Sign in

arXiv:2105.14710 [cs.LG]AbstractReferencesReviewsResources

Robustifying $\ell_\infty$ Adversarial Training to the Union of Perturbation Models

Ameya D. Patil, Michael Tuttle, Alexander G. Schwing, Naresh R. Shanbhag

Published 2021-05-31Version 1

Classical adversarial training (AT) frameworks are designed to achieve high adversarial accuracy against a single attack type, typically $\ell_\infty$ norm-bounded perturbations. Recent extensions in AT have focused on defending against the union of multiple perturbations but this benefit is obtained at the expense of a significant (up to $10\times$) increase in training complexity over single-attack $\ell_\infty$ AT. In this work, we expand the capabilities of widely popular single-attack $\ell_\infty$ AT frameworks to provide robustness to the union of ($\ell_\infty, \ell_2, \ell_1$) perturbations while preserving their training efficiency. Our technique, referred to as Shaped Noise Augmented Processing (SNAP), exploits a well-established byproduct of single-attack AT frameworks -- the reduction in the curvature of the decision boundary of networks. SNAP prepends a given deep net with a shaped noise augmentation layer whose distribution is learned along with network parameters using any standard single-attack AT. As a result, SNAP enhances adversarial accuracy of ResNet-18 on CIFAR-10 against the union of ($\ell_\infty, \ell_2, \ell_1$) perturbations by 14%-to-20% for four state-of-the-art (SOTA) single-attack $\ell_\infty$ AT frameworks, and, for the first time, establishes a benchmark for ResNet-50 and ResNet-101 on ImageNet.

Related articles: Most relevant | Search more
arXiv:1910.04279 [cs.LG] (Published 2019-10-09)
Adversarial Training: embedding adversarial perturbations into the parameter space of a neural network to build a robust system
arXiv:1611.03383 [cs.LG] (Published 2016-11-10)
Disentangling factors of variation in deep representations using adversarial training
arXiv:2006.00387 [cs.LG] (Published 2020-05-30)
Exploring Model Robustness with Adaptive Networks and Improved Adversarial Training