arXiv Analytics

Sign in

arXiv:2211.13445 [cs.CV]AbstractReferencesReviewsResources

Delving into Out-of-Distribution Detection with Vision-Language Representations

Yifei Ming, Ziyang Cai, Jiuxiang Gu, Yiyou Sun, Wei Li, Yixuan Li

Published 2022-11-24Version 1

Recognizing out-of-distribution (OOD) samples is critical for machine learning systems deployed in the open world. The vast majority of OOD detection methods are driven by a single modality (e.g., either vision or language), leaving the rich information in multi-modal representations untapped. Inspired by the recent success of vision-language pre-training, this paper enriches the landscape of OOD detection from a single-modal to a multi-modal regime. Particularly, we propose Maximum Concept Matching (MCM), a simple yet effective zero-shot OOD detection method based on aligning visual features with textual concepts. We contribute in-depth analysis and theoretical insights to understand the effectiveness of MCM. Extensive experiments demonstrate that MCM achieves superior performance on a wide variety of real-world tasks. MCM with vision-language features outperforms a common baseline with pure visual features on a hard OOD task with semantically similar classes by 13.1% (AUROC). Code is available at https://github.com/deeplearning-wisc/MCM.

Related articles: Most relevant | Search more
arXiv:2306.10485 [cs.CV] (Published 2023-06-18)
Balanced Energy Regularization Loss for Out-of-distribution Detection
arXiv:2405.16766 [cs.CV] (Published 2024-05-27, updated 2025-01-07)
Concept Matching with Agent for Out-of-Distribution Detection
arXiv:2407.15739 [cs.CV] (Published 2024-07-22)
Diffusion for Out-of-Distribution Detection on Road Scenes and Beyond