{ "id": "1507.00664", "version": "v1", "published": "2015-07-02T17:21:00.000Z", "updated": "2015-07-02T17:21:00.000Z", "title": "On the Reduction of Total-Cost and Average-Cost MDPs to Discounted MDPs", "authors": [ "Eugene A. Feinberg", "Jefferson Huang" ], "categories": [ "math.OC" ], "abstract": "This paper provides conditions under which countable-state total-cost and average-cost Markov decision processes (MDPs) can be reduced to discounted ones. Results are given for transient total-cost MDPs with transition rates that are not necessarily probabilities, as well as for average-cost MDPs with transition probabilities satisfying the condition that there is a state such that the expected time to reach it is uniformly bounded for all initial states and stationary policies. When the state and action sets are finite, these reductions lead to linear programming formulations and complexity estimates for MDPs under the aforementioned criteria.", "revisions": [ { "version": "v1", "updated": "2015-07-02T17:21:00.000Z" } ], "analyses": { "keywords": [ "average-cost mdps", "discounted mdps", "average-cost markov decision processes", "transient total-cost mdps", "transition probabilities" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable", "adsabs": "2015arXiv150700664F" } } }