arXiv Analytics

Sign in

arXiv:1805.09370 [cs.LG]AbstractReferencesReviewsResources

Towards Robust Training of Neural Networks by Regularizing Adversarial Gradients

Fuxun Yu, Zirui Xu, Yanzhi Wang, Chenchen Liu, Xiang Chen

Published 2018-05-23Version 1

In recent years, neural networks have demonstrated outstanding effectiveness in a large amount of applications.However, recent works have shown that neural networks are susceptible to adversarial examples, indicating possible flaws intrinsic to the network structures. To address this problem and improve the robustness of neural networks, we investigate the fundamental mechanisms behind adversarial examples and propose a novel robust training method via regulating adversarial gradients. The regulation effectively squeezes the adversarial gradients of neural networks and significantly increases the difficulty of adversarial example generation.Without any adversarial example involved, the robust training method could generate naturally robust networks, which are near-immune to various types of adversarial examples. Experiments show the naturally robust networks can achieve optimal accuracy against Fast Gradient Sign Method (FGSM) and C\&W attacks on MNIST, Cifar10, and Google Speech Command dataset. Moreover, our proposed method also provides neural networks with consistent robustness against transferable attacks.

Related articles: Most relevant | Search more
arXiv:2306.09844 [cs.LG] (Published 2023-06-16)
Wasserstein distributional robustness of neural networks
arXiv:1706.02690 [cs.LG] (Published 2017-06-08)
Principled Detection of Out-of-Distribution Examples in Neural Networks
arXiv:1901.08360 [cs.LG] (Published 2019-01-24)
Cross-Entropy Loss and Low-Rank Features Have Responsibility for Adversarial Examples