arXiv Analytics

Sign in

arXiv:2008.06631 [stat.ML]AbstractReferencesReviewsResources

On the Generalization Properties of Adversarial Training

Yue Xing, Qifan Song, Guang Cheng

Published 2020-08-15Version 1

Modern machine learning and deep learning models are shown to be vulnerable when testing data are slightly perturbed. Theoretical studies of adversarial training algorithms mostly focus on their adversarial training losses or local convergence properties. In contrast, this paper studies the generalization performance of a generic adversarial training algorithm. Specifically, we consider linear regression models and two-layer neural networks (with lazy training) using squared loss under both low-dimensional and high-dimensional regimes. In the former regime, the adversarial risk of the trained models will converge to the minimal adversarial risk. In the latter regime, we discover that data interpolation prevents the adversarial robust estimator from being consistent (i.e. converge in probability). Therefore, inspired by successes of the least absolute shrinkage and selection operator (LASSO), we incorporate the L1 penalty in the high dimensional adversarial learning, and show that it leads to consistent adversarial robust estimation in both theory and numerical trials.

Related articles: Most relevant | Search more
arXiv:2305.05642 [stat.ML] (Published 2023-05-09)
A duality framework for generalization analysis of random feature models and two-layer neural networks
arXiv:2404.18769 [stat.ML] (Published 2024-04-29)
Learning with Norm Constrained, Over-parameterized, Two-layer Neural Networks
arXiv:2310.07891 [stat.ML] (Published 2023-10-11)
A Theory of Non-Linear Feature Learning with One Gradient Step in Two-Layer Neural Networks