arXiv Analytics

Sign in

arXiv:2206.08255 [cs.LG]AbstractReferencesReviewsResources

Gradient-Based Adversarial and Out-of-Distribution Detection

Jinsol Lee, Mohit Prabhushankar, Ghassan AlRegib

Published 2022-06-16Version 1

We propose to utilize gradients for detecting adversarial and out-of-distribution samples. We introduce confounding labels -- labels that differ from normal labels seen during training -- in gradient generation to probe the effective expressivity of neural networks. Gradients depict the amount of change required for a model to properly represent given inputs, providing insight into the representational power of the model established by network architectural properties as well as training data. By introducing a label of different design, we remove the dependency on ground truth labels for gradient generation during inference. We show that our gradient-based approach allows for capturing the anomaly in inputs based on the effective expressivity of the models with no hyperparameter tuning or additional processing, and outperforms state-of-the-art methods for adversarial and out-of-distribution detection.

Comments: International Conference on Machine Learning (ICML) Workshop on New Frontiers in Adversarial Machine Learning, July 2022
Categories: cs.LG, cs.CV
Related articles: Most relevant | Search more
arXiv:1912.03133 [cs.LG] (Published 2019-12-05)
Why Should we Combine Training and Post-Training Methods for Out-of-Distribution Detection?
arXiv:2301.10454 [cs.LG] (Published 2023-01-25)
A Data-Centric Approach for Improving Adversarial Training Through the Lens of Out-of-Distribution Detection
arXiv:2205.03493 [cs.LG] (Published 2022-05-06)
Norm-Scaling for Out-of-Distribution Detection