Optimal Output Feedback Learning Control for Continuous-Time Linear Quadratic Regulation

成果类型:
Article
署名作者:
Xie, Kedi; Guay, Martin; Lu, Maobin; Wang, Shimin; Deng, Fang
署名单位:
Beijing Institute of Technology; Beijing Institute of Technology; Queens University - Canada
刊物名称:
IEEE TRANSACTIONS ON AUTOMATIC CONTROL
ISSN/ISSBN:
0018-9286
DOI:
10.1109/TAC.2025.3532182
发表日期:
2025
页码:
4124-4131
关键词:
Output feedback observers Symmetric matrices CONVERGENCE State feedback Heuristic algorithms regulation Adaptation models vectors transient analysis Linear quadratic regulation (LQR) observer error optimal output feedback control reinforcement learning (RL)
摘要:
The classical linear quadratic regulation (LQR) problem of linear systems by state feedback has been widely addressed. However, the LQR problem by dynamic output feedback with optimal transient performance remains open. The main reason is that the observer error inevitably leads to suboptimal transient performance of the closed-loop system. In this article, we propose an optimal dynamic output feedback learning control approach to solve the LQR problem of linear continuous-time systems with unknown dynamics. In particular, we propose a novel internal dynamics called the internal model. Unlike the classical $p$-copy internal model, it is driven by the input and output of the system, and the role of the proposed internal model is to compensate for the transient error of the observer such that the output feedback LQR problem is solved with guaranteed optimality. A model-free learning algorithm is developed to estimate the optimal control gain of the dynamic output feedback controller. The algorithm does not require any prior knowledge of the system matrices or the system's initial state, thus leading to an optimal solution to the model-free LQR problem. The effectiveness of the proposed method is illustrated using an aircraft control system.
来源URL: