arXiv Analytics

Sign in

arXiv:1702.05122 [math.OC]AbstractReferencesReviewsResources

Exact Diffusion for Distributed Optimization and Learning --- Part I: Algorithm Development

Kun Yuan, Bicheng Ying, Xiaochuan Zhao, Ali H. Sayed

Published 2017-02-16Version 1

This work develops a distributed optimization strategy with guaranteed exact convergence for a broad class of left-stochastic combination policies. The resulting exact diffusion strategy is shown in Part II to have a wider stability range and superior convergence performance than the EXTRA strategy. The exact diffusion solution is applicable to non-symmetric left-stochastic combination matrices, while most earlier developments on exact consensus implementations are limited to doubly-stochastic matrices; these latter matrices impose stringent constraints on the network topology. Similar difficulties arise for implementations with right-stochastic policies, which are common in push-sum consensus solutions. The derivation of the exact diffusion strategy in this work relies on reformulating the aggregate optimization problem as a penalized problem and resorting to a diagonally-weighted incremental construction. Detailed stability and convergence analyses are pursued in Part II and are facilitated by examining the evolution of the error dynamics in a transformed domain. Numerical simulations illustrate the theoretical conclusions.

Related articles: Most relevant | Search more
arXiv:1803.07143 [math.OC] (Published 2018-03-19)
Communication reduction in distributed optimization via estimation of the proximal operator
arXiv:2106.07703 [math.OC] (Published 2021-06-14)
Distributed Optimization with Global Constraints Using Noisy Measurements
arXiv:2404.03946 [math.OC] (Published 2024-04-05)
Distributed Optimization for Energy Grids: A Tutorial on ADMM and ALADIN