arXiv Analytics

Sign in

arXiv:1510.04822 [stat.ML]AbstractReferencesReviewsResources

SGD with Variance Reduction beyond Empirical Risk Minimization

Massil Achab, Agathe Guilloux, Stéphane Ga\"\iffas, Emmanuel Bacry

Published 2015-10-16Version 1

We introduce a doubly stochastic proximal gradient algorithm for optimizing a finite average of smooth convex functions, whose gradients depend on numerically expensive expectations. Our main motivation is the acceleration of the optimization of the regularized Cox partial-likelihood (the core model used in survival analysis), but our algorithm can be used in different settings as well. The proposed algorithm is doubly stochastic in the sense that gradient steps are done using stochastic gradient descent (SGD) with variance reduction, where the inner expectations are approximated by a Monte-Carlo Markov-Chain (MCMC) algorithm. We derive conditions on the MCMC number of iterations guaranteeing convergence, and obtain a linear rate of convergence under strong convexity and a sublinear rate without this assumption. We illustrate the fact that our algorithm improves the state-of-the-art solver for regularized Cox partial-likelihood on several datasets from survival analysis.

Related articles: Most relevant | Search more
arXiv:2305.14961 [stat.ML] (Published 2023-05-24)
Deep Learning for Survival Analysis: A Review
arXiv:1802.08626 [stat.ML] (Published 2018-02-23)
Empirical Risk Minimization under Fairness Constraints
arXiv:1806.10701 [stat.ML] (Published 2018-06-27)
Empirical Risk Minimization and Stochastic Gradient Descent for Relational Data