Variance minimization and the overtaking optimality approach to continuous-time controlled Markov chains |
| |
Authors: | Tomás Prieto-Rumeau Onésimo Hernández-Lerma |
| |
Affiliation: | (1) Department of Mathematics, South China Normal University, Guangzhou, 510631, People’s Republic of China |
| |
Abstract: | ![]() This paper deals with denumerable-state continuous-time controlled Markov chains with possibly unbounded transition and reward rates. It concerns optimality criteria that improve the usual expected average reward criterion. First, we show the existence of average reward optimal policies with minimal average variance. Then we compare the variance minimization criterion with overtaking optimality. We present an example showing that they are opposite criteria, and therefore we cannot optimize them simultaneously. This leads to a multiobjective problem for which we identify the set of Pareto optimal policies (also known as nondominated policies). |
| |
Keywords: | |
本文献已被 SpringerLink 等数据库收录! |
|