arXiv Analytics

Sign in

arXiv:1806.09186 [cs.CV]AbstractReferencesReviewsResources

Detecting Adversarial Examples Based on Steganalysis

Jiayang Liu, Weiming Zhang, Yiwei Zhang, Dongdong Hou, Yujia Liu, Nenghai Yu

Published 2018-06-21Version 1

Deep Neural Networks (DNNs) have recently led to significant improvement in many fields, such as image classification. However, these machine learning models are vulnerable to adversarial examples which can mislead machine learning classifiers to give incorrect classifications. Adversarial examples pose security concerns in areas where privacy requirements are strict, such as face recognition, autonomous cars and malware detection. What's more, they could be used to perform an attack on machine learning systems, even if the adversary has no access to the underlying model. In this paper, we focus on detecting adversarial examples. We propose to augment deep neural networks with a detector. The detector is constructed by modeling the differences between adjacent pixels in natural images. And then we identify deviations from this model and assume that such deviations are due to adversarial attack. We construct the detector based on steganalysis which can detect minor modifications to an image because the adversarial attack can be treated as a sort of accidental steganography.

Related articles: Most relevant | Search more
arXiv:1909.09263 [cs.CV] (Published 2019-09-19)
Propagated Perturbation of Adversarial Attack for well-known CNNs: Empirical Study and its Explanation
arXiv:2501.07044 [cs.CV] (Published 2025-01-13)
Protego: Detecting Adversarial Examples for Vision Transformers via Intrinsic Capabilities
arXiv:2003.01895 [cs.CV] (Published 2020-03-04)
Double Backpropagation for Training Autoencoders against Adversarial Attack