arXiv:1506.07291 [math.OC]AbstractReferencesReviewsResources
Two-Armed Restless Bandits with Imperfect Information: Stochastic Control and Indexability
Published 2015-06-24Version 1
We present a two-armed bandit model of decision making under uncertainty where the expected return to investing in the "risky arm" increases when choosing that arm and decreases when choosing the "safe" arm. These dynamics are natural in applications such as human capital development, job search, and occupational choice. Using new insights from stochastic control, along with a monotonicity condition on the payoff dynamics, we show that optimal strategies in our model are stopping rules that can be characterized by an index which formally coincides with Gittins' index. Our result implies the indexability of a new class of restless bandit models.
Comments: 46 pages
Categories: math.OC
Related articles: Most relevant | Search more
arXiv:2101.06205 [math.OC] (Published 2021-01-15)
Maximum principle for stochastic control of SDEs with measurable drifts
The Separation Principle in Stochastic Control, Redux
arXiv:2007.09978 [math.OC] (Published 2020-07-20)
Mathematical and computational approaches for stochastic control of river environment and ecology: from fisheries viewpoint