arXiv Analytics

Sign in

arXiv:1405.4980 [math.OC]AbstractReferencesReviewsResources

Convex Optimization: Algorithms and Complexity

Sébastien Bubeck

Published 2014-05-20, updated 2015-11-16Version 2

This monograph presents the main complexity theorems in convex optimization and their corresponding algorithms. Starting from the fundamental theory of black-box optimization, the material progresses towards recent advances in structural optimization and stochastic optimization. Our presentation of black-box optimization, strongly influenced by Nesterov's seminal book and Nemirovski's lecture notes, includes the analysis of cutting plane methods, as well as (accelerated) gradient descent schemes. We also pay special attention to non-Euclidean settings (relevant algorithms include Frank-Wolfe, mirror descent, and dual averaging) and discuss their relevance in machine learning. We provide a gentle introduction to structural optimization with FISTA (to optimize a sum of a smooth and a simple non-smooth term), saddle-point mirror prox (Nemirovski's alternative to Nesterov's smoothing), and a concise description of interior point methods. In stochastic optimization we discuss stochastic gradient descent, mini-batches, random coordinate descent, and sublinear algorithms. We also briefly touch upon convex relaxation of combinatorial problems and the use of randomness to round solutions, as well as random walks based methods.

Comments: A previous version of the manuscript was titled "Theory of Convex Optimization for Machine Learning"
Journal: In Foundations and Trends in Machine Learning, Vol. 8: No. 3-4, pp 231-357, 2015
Categories: math.OC, cs.CC, cs.LG, cs.NA, stat.ML
Related articles: Most relevant | Search more
arXiv:1509.05647 [math.OC] (Published 2015-09-18)
Fast and Simple PCA via Convex Optimization
arXiv:1602.04847 [math.OC] (Published 2016-02-15)
Black-box optimization with a politician
arXiv:2305.00438 [math.OC] (Published 2023-04-30)
META-SMGO-$Δ$: similarity as a prior in black-box optimization