arXiv Analytics

Sign in

arXiv:2102.07304 [cs.LG]AbstractReferencesReviewsResources

CAP-GAN: Towards_Adversarial_Robustness_with_Cycle-consistent_Attentional_Purification

Mingu Kang, Trung Quang Tran, Seungju Cho, Daeyoung Kim

Published 2021-02-15Version 1

Adversarial attack is aimed at fooling the target classifier with imperceptible perturbation. Adversarial examples, which are carefully crafted with a malicious purpose, can lead to erroneous predictions, resulting in catastrophic accidents. To mitigate the effects of adversarial attacks, we propose a novel purification model called CAP-GAN. CAP-GAN takes account of the idea of pixel-level and feature-level consistency to achieve reasonable purification under cycle-consistent learning. Specifically, we utilize the guided attention module and knowledge distillation to convey meaningful information to the purification model. Once a model is fully trained, inputs would be projected into the purification model and transformed into clean-like images. We vary the capacity of the adversary to argue the robustness against various types of attack strategies. On the CIFAR-10 dataset, CAP-GAN outperforms other pre-processing based defenses under both black-box and white-box settings.

Related articles: Most relevant | Search more
arXiv:1905.12797 [cs.LG] (Published 2019-05-30)
Bandlimiting Neural Networks Against Adversarial Attacks
arXiv:1906.03563 [cs.LG] (Published 2019-06-09)
Beyond Adversarial Training: Min-Max Optimization in Adversarial Attack and Defense
arXiv:2307.07916 [cs.LG] (Published 2023-07-16)
On the Robustness of Split Learning against Adversarial Attacks