Policy Optimization for Markovian Jump Linear Quadratic Control: Gradient Method and Global Convergence
成果类型:
Article
署名作者:
Jansch-Porto, Joao Paulo; Hu, Bin; Dullerud, Geir E.
署名单位:
University of Illinois System; University of Illinois Urbana-Champaign; University of Illinois System; University of Illinois Urbana-Champaign
刊物名称:
IEEE TRANSACTIONS ON AUTOMATIC CONTROL
ISSN/ISSBN:
0018-9286
DOI:
10.1109/TAC.2022.3176439
发表日期:
2023
页码:
2475-2482
关键词:
Optimization
Linear systems
gradient methods
CONVERGENCE
State feedback
COSTS
Markov processes
Markovian jump linear systems (MJLS)
optimal control
policy gradient methods
reinforcement learning (RL)
摘要:
Recently, policy optimization has received renewed attention from the control community due to various applications in reinforcement learning tasks. In this article, we investigate the global convergence of the gradient method for quadratic optimal control of discrete-time Markovian jump linear systems (MJLS). First, we study the optimization landscape of direct policy optimization for MJLS, with static-state feedback controllers and quadratic performance costs. Despite the nonconvexity of the resultant problem, we are still able to identify several useful properties such as coercivity, gradient dominance, and smoothness. Based on these properties, we prove that the gradient method converges to the optimal-state feedback controller for MJLS at a linear rate if initialized at a controller, which is mean-square stabilizing. This article brings new insights for understanding the performance of the policy gradient method on the Markovian jump linear quadratic control problem.