arXiv Analytics

Sign in

arXiv:1812.07385 [cs.LG]AbstractReferencesReviewsResources

Perturbation Analysis of Learning Algorithms: A Unifying Perspective on Generation of Adversarial Examples

Emilio Rafael Balda, Arash Behboodi, Rudolf Mathar

Published 2018-12-15Version 1

Despite the tremendous success of deep neural networks in various learning problems, it has been observed that adding an intentionally designed adversarial perturbation to inputs of these architectures leads to erroneous classification with high confidence in the prediction. In this work, we propose a general framework based on the perturbation analysis of learning algorithms which consists of convex programming and is able to recover many current adversarial attacks as special cases. The framework can be used to propose novel attacks against learning algorithms for classification and regression tasks under various new constraints with closed form solutions in many instances. In particular we derive new attacks against classification algorithms which are shown to achieve comparable performances to notable existing attacks. The framework is then used to generate adversarial perturbations for regression tasks which include single pixel and single subset attacks. By applying this method to autoencoding and image colorization tasks, it is shown that adversarial perturbations can effectively perturb the output of regression tasks as well.

Comments: arXiv admin note: text overlap with arXiv:1803.03607
Categories: cs.LG, cs.AI, cs.IT, math.IT, stat.ML
Related articles: Most relevant | Search more
arXiv:1812.01804 [cs.LG] (Published 2018-12-05)
Random Spiking and Systematic Evaluation of Defenses Against Adversarial Examples
arXiv:1911.05268 [cs.LG] (Published 2019-11-13)
Adversarial Examples in Modern Machine Learning: A Review
arXiv:1807.09705 [cs.LG] (Published 2018-07-25)
Limitations of the Lipschitz constant as a defense against adversarial examples