Optimal Control of Probability on a Target Set for Continuous-Time Markov Chains

成果类型:
Article
署名作者:
Ma, Chenglin; Zhao, Huaizhong
署名单位:
Shandong University; Durham University; Shandong University
刊物名称:
IEEE TRANSACTIONS ON AUTOMATIC CONTROL
ISSN/ISSBN:
0018-9286
DOI:
10.1109/TAC.2023.3278789
发表日期:
2024
页码:
1202-1209
关键词:
Markov processes optimal control dynamic programming PROCESS CONTROL games Aerospace electronics safety controlled Markov chains dynamic programming principle (DPP) Hamilton-Jacobi-Bellman (HJB) equation optimal controls risk probability criteria
摘要:
In this article, a stochastic optimal control problem is considered for a continuous-time Markov chain taking values in a denumerable state space over a fixed finite horizon. The optimality criterion is the probability that the process remains in a target set before and at a certain time. The optimal value is a superadditive capacity of target sets. Under some minor assumptions for the controlled Markov process, we establish the dynamic programming principle, based on which we prove that the value function is a classical solution of the Hamilton-Jacobi-Bellman (HJB) equation on a discrete lattice space. We then prove that there exists an optimal deterministic Markov control under the compactness assumption of control domain. We further prove that the value function is the unique solution of the HJB equation. We also consider the case starting from the outside of the target set and give the corresponding results. Finally, we apply our results to two examples.