arXiv Analytics

Sign in

arXiv:2505.16713 [stat.ML]AbstractReferencesReviewsResources

Sharp concentration of uniform generalization errors in binary linear classification

Shogo Nakakita

Published 2025-05-22, updated 2025-06-26Version 2

We examine the concentration of uniform generalization errors around their expectation in binary linear classification problems via an isoperimetric argument. In particular, we establish Poincar\'{e} and log-Sobolev inequalities for the joint distribution of the output labels and the label-weighted input vectors, which we apply to derive concentration bounds. The derived concentration bounds are sharp up to moderate multiplicative constants by those under well-balanced labels. In asymptotic analysis, we also show that almost sure convergence of uniform generalization errors to their expectation occurs in very broad settings, such as proportionally high-dimensional regimes. Using this convergence, we establish uniform laws of large numbers under dimension-free conditions.

Comments: 26 pages, 1 figure; minor edits to improve readability
Categories: stat.ML, cs.LG, math.ST, stat.TH
Related articles: Most relevant | Search more
arXiv:2006.05240 [stat.ML] (Published 2020-06-09)
How Robust is the Median-of-Means? Concentration Bounds in Presence of Outliers
arXiv:1905.10155 [stat.ML] (Published 2019-05-24)
Concentration bounds for linear Monge mapping estimation and optimal transport domain adaptation
arXiv:2008.02464 [stat.ML] (Published 2020-08-06)
Concentration Bounds for Co-occurrence Matrices of Markov Chains