arXiv Analytics

Sign in

arXiv:1902.03840 [math.NA]AbstractReferencesReviewsResources

On the rotational invariant $L_1$-norm PCA

Sebastian Neumayer, Max Nimmer, Simon Setzer, Gabriele Steidl

Published 2019-02-11Version 1

Principal component analysis (PCA) is a powerful tool for dimensionality reduction. Unfortunately, it is sensitive to outliers, so that various robust PCA variants were proposed in the literature. Among them the so-called rotational invariant $L_1$-norm PCA is rather popular. In this paper, we reinterpret this robust method as conditional gradient algorithm and show moreover that it coincides with a gradient descent algorithm on Grassmannian manifolds. Based on this point of view, we prove for the first time convergence of the whole series of iterates to a critical point using the Kurdyka-{\L}ojasiewicz property of the energy functional.

Related articles: Most relevant | Search more
arXiv:2005.05449 [math.NA] (Published 2020-05-11)
Robust PCA via Regularized REAPER with a Matrix-Free Proximal Algorithm
arXiv:1906.12085 [math.NA] (Published 2019-06-28)
FameSVD: Fast and Memory-efficient Singular Value Decomposition
arXiv:1903.09576 [math.NA] (Published 2019-03-22)
Data-Space Inversion with Ensemble Smoother