arXiv Analytics

Sign in

arXiv:1904.03293 [cs.LG]AbstractReferencesReviewsResources

Collaborative Learning with Limited Interaction: Tight Bounds for Distributed Exploration in Multi-Armed Bandits

Chao Tao, Qin Zhang, Yuan Zhou

Published 2019-04-05Version 1

Best arm identification (or, pure exploration) in multi-armed bandits is a fundamental problem in machine learning. In this paper we study the distributed version of this problem where we have multiple agents, and they want to learn the best arm collaboratively. We want to quantify the power of collaboration under limited interaction (or, communication steps), as interaction is expensive in many settings. We measure the running time of a distributed algorithm as the speedup over the best centralized algorithm where there is only one agent. We give almost tight round-speedup tradeoffs for this problem, along which we develop several new techniques for proving lower bounds on the number of communication steps under time or confidence constraints.

Related articles: Most relevant | Search more
arXiv:1311.0800 [cs.LG] (Published 2013-11-04)
Distributed Exploration in Multi-Armed Bandits
arXiv:2006.00082 [cs.LG] (Published 2020-05-29)
Meta Clustering for Collaborative Learning
arXiv:2402.01845 [cs.LG] (Published 2024-02-02, updated 2024-07-15)
Multi-Armed Bandits with Interference