arXiv Analytics

Sign in

arXiv:2011.05074 [cs.LG]AbstractReferencesReviewsResources

Efficient and Transferable Adversarial Examples from Bayesian Neural Networks

Martin Gubri, Maxime Cordy, Mike Papadakis, Yves Le Traon

Published 2020-11-10Version 1

Deep neural networks are vulnerable to evasion attacks, i.e., carefully crafted examples designed to fool a model at test time. Attacks that successfully evade an ensemble of models can transfer to other independently trained models, which proves useful in black-box settings. Unfortunately, these methods involve heavy computation costs to train the models forming the ensemble. To overcome this, we propose a new method to generate transferable adversarial examples efficiently. Inspired by Bayesian deep learning, our method builds such ensembles by sampling from the posterior distribution of neural network weights during a single training process. Experiments on CIFAR-10 show that our approach improves the transfer rates significantly at equal or even lower computation costs. Intra-architecture transfer rate is increased by 23% compared to classical ensemble-based attacks, while requiring 4 times less training epochs. In the inter-architecture case, we show that we can combine our method with ensemble-based attacks to increase their transfer rate by up to 15% with constant training computational cost.

Related articles: Most relevant | Search more
arXiv:2201.02478 [cs.LG] (Published 2022-01-07)
Bayesian Neural Networks for Reversible Steganography
arXiv:2206.06779 [cs.LG] (Published 2022-06-08)
Quantitative performance evaluation of Bayesian neural networks
arXiv:2410.10395 [cs.LG] (Published 2024-10-14)
Improved Depth Estimation of Bayesian Neural Networks