arXiv Analytics

Sign in

arXiv:1811.01926 [cs.LG]AbstractReferencesReviewsResources

contextual: Evaluating Contextual Multi-Armed Bandit Problems in R

Robin van Emden, Maurits Kaptein

Published 2018-11-06Version 1

Over the past decade, contextual bandit algorithms have been gaining in popularity due to their effectiveness and flexibility in solving sequential decision problems---from online advertising and finance to clinical trial design and personalized medicine. At the same time, there are, as of yet, surprisingly few options that enable researchers and practitioners to simulate and compare the wealth of new and existing bandit algorithms in a standardized way. To help close this gap between analytical research and empirical evaluation the current paper introduces the object-oriented R package "contextual": a user-friendly and, through its object-oriented structure, easily extensible framework that facilitates parallelized comparison of contextual and context-free bandit policies through both simulation and offline analysis.

Related articles: Most relevant | Search more
arXiv:2008.07146 [cs.LG] (Published 2020-08-17)
A Large-scale Open Dataset for Bandit Algorithms
arXiv:2009.06606 [cs.LG] (Published 2020-09-14)
Hellinger KL-UCB based Bandit Algorithms for Markovian and i.i.d. Settings
arXiv:2006.12038 [cs.LG] (Published 2020-06-22)
Bandit algorithms: Letting go of logarithmic regret for statistical robustness