arXiv Analytics

Sign in

arXiv:1807.00493 [cs.CV]AbstractReferencesReviewsResources

Active Testing: An Efficient and Robust Framework for Estimating Accuracy

Phuc Nguyen, Deva Ramanan, Charless Fowlkes

Published 2018-07-02Version 1

Much recent work on visual recognition aims to scale up learning to massive, noisily-annotated datasets. We address the problem of scaling- up the evaluation of such models to large-scale datasets with noisy labels. Current protocols for doing so require a human user to either vet (re-annotate) a small fraction of the test set and ignore the rest, or else correct errors in annotation as they are found through manual inspection of results. In this work, we re-formulate the problem as one of active testing, and examine strategies for efficiently querying a user so as to obtain an accu- rate performance estimate with minimal vetting. We demonstrate the effectiveness of our proposed active testing framework on estimating two performance metrics, Precision@K and mean Average Precision, for two popular computer vision tasks, multi-label classification and instance segmentation. We further show that our approach is able to save significant human annotation effort and is more robust than alternative evaluation protocols.

Related articles: Most relevant | Search more
arXiv:1003.5249 [cs.CV] (Published 2010-03-27)
Active Testing for Face Detection and Localization
arXiv:1610.09590 [cs.CV] (Published 2016-10-30)
A Scalable and Robust Framework for Intelligent Real-time Video Surveillance
arXiv:2411.10951 [cs.CV] (Published 2024-11-17)
TSFormer: A Robust Framework for Efficient UHD Image Restoration