arXiv Analytics

Sign in

arXiv:2109.10854 [math.OC]AbstractReferencesReviewsResources

Imitation Learning of Stabilizing Policies for Nonlinear Systems

Sebastian East

Published 2021-09-22Version 1

There has been a recent interest in imitation learning methods that are guaranteed to produce a stabilizing control law with respect to a known system. Work in this area has generally considered linear systems and controllers, for which stabilizing imitation learning takes the form of a biconvex optimization problem. In this paper it is demonstrated that the same methods developed for linear systems and controllers can be readily extended to polynomial systems and controllers using sum of squares techniques. A projected gradient descent algorithm and an alternating direction method of multipliers algorithm are proposed as heuristics for solving the stabilizing imitation learning problem, and their performance is illustrated through numerical experiments.

Related articles: Most relevant | Search more
arXiv:2004.01815 [math.OC] (Published 2020-04-03)
Near optimal tracking control of a class of nonlinear systems and an experimental comparison
arXiv:1803.08386 [math.OC] (Published 2018-03-22, updated 2018-04-16)
Observability and State Estimation for a Class of Nonlinear Systems
arXiv:2303.08707 [math.OC] (Published 2023-03-15)
On the design of persistently exciting inputs for data-driven control of linear and nonlinear systems