{ "id": "1702.05921", "version": "v1", "published": "2017-02-20T10:20:11.000Z", "updated": "2017-02-20T10:20:11.000Z", "title": "A Mean-field Stochastic Control Problem with Partial Observations", "authors": [ "Rainer Buckdahn", "Juan Li", "Jin Ma" ], "comment": "The paper was accepted by Annals of Applied Probability", "categories": [ "math.PR" ], "abstract": "In this paper we are interested in a new type of {\\it mean-field}, non-Markovian stochastic control problems with partial observations. More precisely, we assume that the coefficients of the controlled dynamics depend not only on the paths of the state, but also on the conditional law of the state, given the observation to date. Our problem is strongly motivated by the recent study of the mean field games and the related McKean-Vlasov stochastic control problem, but with added aspects of path-dependence and partial observation. We shall first investigate the well-posedness of the state-observation dynamics, with combined reference probability measure arguments in nonlinear filtering theory and the Schauder fixed point theorem. We then study the stochastic control problem with a partially observable system in which the conditional law appears nonlinearly in both the coefficients of the system and cost function. As a consequence the control problem is intrinsically \"time-inconsistent\", and we prove that the Pontryagin Stochastic Maximum Principle holds in this case and characterize the adjoint equations, which turn out to be a new form of mean-field type BSDEs.", "revisions": [ { "version": "v1", "updated": "2017-02-20T10:20:11.000Z" } ], "analyses": { "keywords": [ "mean-field stochastic control problem", "partial observation", "mckean-vlasov stochastic control problem", "pontryagin stochastic maximum principle holds", "reference probability measure arguments" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }