arXiv Analytics

Sign in

arXiv:2402.08508 [stat.ML]AbstractReferencesReviewsResources

A PAC-Bayesian Link Between Generalisation and Flat Minima

Maxime Haddouche, Paul Viallard, Umut Simsekli, Benjamin Guedj

Published 2024-02-13, updated 2025-02-11Version 2

Modern machine learning usually involves predictors in the overparameterised setting (number of trained parameters greater than dataset size), and their training yields not only good performance on training data, but also good generalisation capacity. This phenomenon challenges many theoretical results, and remains an open problem. To reach a better understanding, we provide novel generalisation bounds involving gradient terms. To do so, we combine the PAC-Bayes toolbox with Poincar\'e and Log-Sobolev inequalities, avoiding an explicit dependency on the dimension of the predictor space. Our results highlight the positive influence of flat minima (being minima with a neighbourhood nearly minimising the learning problem as well) on generalisation performance, involving directly the benefits of the optimisation phase.

Comments: Published at International Conference on Algorithmic Learning Theory 2025
Categories: stat.ML, cs.LG
Related articles: Most relevant | Search more
arXiv:1111.5648 [stat.ML] (Published 2011-11-23)
Falsification and future performance
arXiv:2204.12868 [stat.ML] (Published 2022-04-27)
Performance and Interpretability Comparisons of Supervised Machine Learning Algorithms: An Empirical Study
arXiv:2405.13456 [stat.ML] (Published 2024-05-22)
Deep linear networks for regression are implicitly regularized towards flat minima