arXiv Analytics

Sign in

arXiv:1506.07613 [cs.CV]AbstractReferencesReviewsResources

Generalized Majorization-Minimization

Sobhan Naderi Parizi, Kun He, Stan Sclaroff, Pedro Felzenszwalb

Published 2015-06-25Version 1

Non-convex optimization is ubiquitous in machine learning. The Majorization-Minimization (MM) procedure systematically optimizes non-convex functions through an iterative construction and optimization of upper bounds on the objective function. The bound at each iteration is required to \emph{touch} the objective function at the optimizer of the previous bound. We show that this touching constraint is unnecessary and overly restrictive. We generalize MM by relaxing this constraint, and propose a new framework for designing optimization algorithms, named Generalized Majorization-Minimization (G-MM). Compared to MM, G-MM is much more flexible. For instance, it can incorporate application-specific biases into the optimization procedure without changing the objective function. We derive G-MM algorithms for several latent variable models and show that they consistently outperform their MM counterparts in optimizing non-convex objectives. In particular, G-MM algorithms appear to be less sensitive to initialization.

Related articles: Most relevant | Search more
arXiv:2012.04837 [cs.CV] (Published 2020-12-09)
Deep Unsupervised Image Anomaly Detection: An Information Theoretic Framework
arXiv:2211.03646 [cs.CV] (Published 2022-11-07)
Contrastive Classification and Representation Learning with Probabilistic Interpretation
arXiv:2012.08950 [cs.CV] (Published 2020-12-16)
Deep Reinforcement Learning of Graph Matching