arXiv Analytics

Sign in

arXiv:1310.1404 [stat.ML]AbstractReferencesReviewsResources

Sequential Monte Carlo Bandits

Michael Cherkassky, Luke Bornn

Published 2013-10-04Version 1

In this paper we propose a flexible and efficient framework for handling multi-armed bandits, combining sequential Monte Carlo algorithms with hierarchical Bayesian modeling techniques. The framework naturally encompasses restless bandits, contextual bandits, and other bandit variants under a single inferential model. Despite the model's generality, we propose efficient Monte Carlo algorithms to make inference scalable, based on recent developments in sequential Monte Carlo methods. Through two simulation studies, the framework is shown to outperform other empirical methods, while also naturally scaling to more complex problems for which existing approaches can not cope. Additionally, we successfully apply our framework to online video-based advertising recommendation, and show its increased efficacy as compared to current state of the art bandit algorithms.

Related articles: Most relevant | Search more
arXiv:2308.07983 [stat.ML] (Published 2023-08-15)
Monte Carlo guided Diffusion for Bayesian linear inverse problems
arXiv:1907.10477 [stat.ML] (Published 2019-07-24)
On the relationship between variational inference and adaptive importance sampling
arXiv:1602.06701 [stat.ML] (Published 2016-02-22)
Inference Networks for Sequential Monte Carlo in Graphical Models