Taylor expansions of the value function associated with a bilinear optimal control problem |
| |
Institution: | 1. Institute of Mathematics, University of Graz, Austria;2. RICAM Institute, Austrian Academy of Sciences, Linz, Austria |
| |
Abstract: | A general bilinear optimal control problem subject to an infinite-dimensional state equation is considered. Polynomial approximations of the associated value function are derived around the steady state by repeated formal differentiation of the Hamilton–Jacobi–Bellman equation. The terms of the approximations are described by multilinear forms, which can be obtained as solutions to generalized Lyapunov equations with recursively defined right-hand sides. They form the basis for defining a suboptimal feedback law. The approximation properties of this feedback law are investigated. An application to the optimal control of a Fokker–Planck equation is also provided. |
| |
Keywords: | Value function Hamilton–Jacobi–Bellman equation Bilinear control systems Riccati equation Generalized Lyapunov equations Fokker–Planck equation |
本文献已被 ScienceDirect 等数据库收录! |