arXiv Analytics

Sign in

arXiv:2002.05516 [cs.LG]AbstractReferencesReviewsResources

Federated Learning of a Mixture of Global and Local Models

Filip Hanzely, Peter Richtárik

Published 2020-02-10Version 1

We propose a new optimization formulation for training federated learning models. The standard formulation has the form of an empirical risk minimization problem constructed to find a single global model trained from the private data stored across all participating devices. In contrast, our formulation seeks an explicit trade-off between this traditional global model and the local models, which can be learned by each device from its own private data without any communication. Further, we develop several efficient variants of SGD (with and without partial participation and with and without variance reduction) for solving the new formulation and prove communication complexity guarantees. Notably, our methods are similar but not identical to federated averaging / local SGD, thus shedding some light on the essence of the elusive method. In particular, our methods do not perform full averaging steps and instead merely take steps towards averaging. We argue for the benefits of this new paradigm for federated learning.

Comments: 39 pages, 8 algorithms, 6 figures, 1 table
Categories: cs.LG, cs.DC, math.OC, stat.ML
Related articles: Most relevant | Search more
arXiv:1911.07652 [cs.LG] (Published 2019-11-15)
Information-Theoretic Perspective of Federated Learning
arXiv:2009.06303 [cs.LG] (Published 2020-09-14)
Fed+: A Family of Fusion Algorithms for Federated Learning
arXiv:2001.01523 [cs.LG] (Published 2020-01-06)
Think Locally, Act Globally: Federated Learning with Local and Global Representations