arXiv Analytics

Sign in

arXiv:2308.10722 [cs.LG]AbstractReferencesReviewsResources

Clustered Linear Contextual Bandits with Knapsacks

Yichuan Deng, Michalis Mamakos, Zhao Song

Published 2023-08-21Version 1

In this work, we study clustered contextual bandits where rewards and resource consumption are the outcomes of cluster-specific linear models. The arms are divided in clusters, with the cluster memberships being unknown to an algorithm. Pulling an arm in a time period results in a reward and in consumption for each one of multiple resources, and with the total consumption of any resource exceeding a constraint implying the termination of the algorithm. Thus, maximizing the total reward requires learning not only models about the reward and the resource consumption, but also cluster memberships. We provide an algorithm that achieves regret sublinear in the number of time periods, without requiring access to all of the arms. In particular, we show that it suffices to perform clustering only once to a randomly selected subset of the arms. To achieve this result, we provide a sophisticated combination of techniques from the literature of econometrics and of bandits with constraints.

Related articles: Most relevant | Search more
arXiv:2303.10181 [cs.LG] (Published 2023-03-17)
Operating critical machine learning models in resource constrained regimes
arXiv:2003.06308 [cs.LG] (Published 2020-03-11)
Compressing deep neural networks on FPGAs to binary and ternary precision with HLS4ML
arXiv:2210.03204 [cs.LG] (Published 2022-10-06)
Enabling Deep Learning on Edge Devices