arXiv:1606.03203 [stat.ML]AbstractReferencesReviewsResources
Causal Bandits: Learning Good Interventions via Causal Inference
Finnian Lattimore, Tor Lattimore, Mark D. Reid
Published 2016-06-10Version 1
We study the problem of using causal models to improve the rate at which good interventions can be learned online in a stochastic environment. Our formalism combines multi-arm bandits and causal inference to model a novel type of bandit feedback that is not exploited by existing approaches. We propose a new algorithm that exploits the causal feedback and prove a bound on its simple regret that is strictly better (in all quantities) than algorithms that do not use the additional causal information.
Related articles: Most relevant | Search more
arXiv:1906.00442 [stat.ML] (Published 2019-06-02)
An Evaluation Toolkit to Guide Model Selection and Cohort Definition in Causal Inference
Yishai Shimoni et al.
arXiv:2306.10551 [stat.ML] (Published 2023-06-18)
Can predictive models be used for causal inference?
arXiv:2302.00370 [stat.ML] (Published 2023-02-01)
How to select predictive models for causal inference?