arXiv Analytics

Sign in

arXiv:1912.03133 [cs.LG]AbstractReferencesReviewsResources

Why Should we Combine Training and Post-Training Methods for Out-of-Distribution Detection?

Aristotelis-Angelos Papadopoulos, Nazim Shaikh, Mohammad Reza Rajati

Published 2019-12-05Version 1

Deep neural networks are known to achieve superior results in classification tasks. However, it has been recently shown that they are incapable to detect examples that are generated by a distribution which is different than the one they have been trained on since they are making overconfident prediction for Out-Of-Distribution (OOD) examples. OOD detection has attracted a lot of attention recently. In this paper, we review some of the most seminal recent algorithms in the OOD detection field, we divide those methods into training and post-training and we experimentally show how the combination of the former with the latter can achieve state-of-the-art results in the OOD detection task.

Comments: Preprint, 9 pages. arXiv admin note: text overlap with arXiv:1906.03509
Categories: cs.LG, cs.CV, stat.ML
Related articles: Most relevant | Search more
arXiv:1904.12220 [cs.LG] (Published 2019-04-27)
Analysis of Confident-Classifiers for Out-of-distribution Detection
arXiv:1811.07308 [cs.LG] (Published 2018-11-18, updated 2019-02-02)
Enhancing the Robustness of Prior Network in Out-of-Distribution Detection
arXiv:2205.03493 [cs.LG] (Published 2022-05-06)
Norm-Scaling for Out-of-Distribution Detection