arXiv Analytics

Sign in

arXiv:1802.04784 [stat.ML]AbstractReferencesReviewsResources

MONK -- Outlier-Robust Mean Embedding Estimation by Median-of-Means

Matthieu Lerasle, Zoltan Szabo, Gaspar Massiot, Eric Moulines

Published 2018-02-13Version 1

Mean embeddings provide an extremely flexible and powerful tool in machine learning and statistics to represent probability distributions and define a semi-metric (MMD, maximum mean discrepancy; also called N-distance or energy distance), with numerous successful applications. The representation is constructed as the expectation of the feature map defined by a kernel. As a mean, its classical empirical estimator, however, can be arbitrary severely affected even by a single outlier in case of unbounded features. To the best of our knowledge, unfortunately even the consistency of the existing few techniques trying to alleviate this serious sensitivity bottleneck is unknown. In this paper, we show how the recently emerged principle of median-of-means can be used to design minimax-optimal estimators for kernel mean embedding and MMD, with finite-sample strong outlier-robustness guarantees.

Related articles: Most relevant | Search more
arXiv:2302.09930 [stat.ML] (Published 2023-02-20)
Nyström $M$-Hilbert-Schmidt Independence Criterion
arXiv:2006.05240 [stat.ML] (Published 2020-06-09)
How Robust is the Median-of-Means? Concentration Bounds in Presence of Outliers
arXiv:2105.14035 [stat.ML] (Published 2021-05-28)
DeepMoM: Robust Deep Learning With Median-of-Means