arXiv Analytics

Sign in

arXiv:2409.09078 [stat.ML]AbstractReferencesReviewsResources

Bounds on the Generalization Error in Active Learning

Vincent Menden, Yahya Saleh, Armin Iske

Published 2024-09-10Version 1

We establish empirical risk minimization principles for active learning by deriving a family of upper bounds on the generalization error. Aligning with empirical observations, the bounds suggest that superior query algorithms can be obtained by combining both informativeness and representativeness query strategies, where the latter is assessed using integral probability metrics. To facilitate the use of these bounds in application, we systematically link diverse active learning scenarios, characterized by their loss functions and hypothesis classes to their corresponding upper bounds. Our results show that regularization techniques used to constraint the complexity of various hypothesis classes are sufficient conditions to ensure the validity of the bounds. The present work enables principled construction and empirical quality-evaluation of query algorithms in active learning.

Related articles: Most relevant | Search more
arXiv:2410.08361 [stat.ML] (Published 2024-10-10)
Upper Bounds for Learning in Reproducing Kernel Hilbert Spaces for Orbits of an Iterated Function System
arXiv:2409.18836 [stat.ML] (Published 2024-09-27)
Constructing Confidence Intervals for 'the' Generalization Error -- a Comprehensive Benchmark Study
arXiv:1810.06397 [stat.ML] (Published 2018-10-15)
A Priori Estimates of the Generalization Error for Two-layer Neural Networks