Approximation and adaptive control of Markov processes: Average reward criterion
Onésimo Hernández-Lerma (1987)
Kybernetika
Similarity:
Onésimo Hernández-Lerma (1987)
Kybernetika
Similarity:
Evgueni Gordienko, Onésimo Hernández-Lerma (1995)
Applicationes Mathematicae
Similarity:
This paper considers discrete-time Markov control processes on Borel spaces, with possibly unbounded costs, and the long run average cost (AC) criterion. Under appropriate hypotheses on weighted norms for the cost function and the transition law, the existence of solutions to the average cost optimality inequality and the average cost optimality equation are shown, which in turn yield the existence of AC-optimal and AC-canonical policies respectively.
Armando F. Mendoza-Pérez, Onésimo Hernández-Lerma (2012)
Applicationes Mathematicae
Similarity:
This paper deals with discrete-time Markov control processes in Borel spaces with unbounded rewards. Under suitable hypotheses, we show that a randomized stationary policy is optimal for a certain expected constrained problem (ECP) if and only if it is optimal for the corresponding pathwise constrained problem (pathwise CP). Moreover, we show that a certain parametric family of unconstrained optimality equations yields convergence properties that lead to an approximation scheme which...
Zhu, Quanxin, Guo, Xianping (2006)
Journal of Applied Mathematics and Stochastic Analysis
Similarity:
Onésimo Hernández-Lerma (1993)
Kybernetika
Similarity:
Łukasz Stettner (1993)
Applicationes Mathematicae
Similarity:
Optimal control with long run average cost functional of a partially observed Markov process is considered. Under the assumption that the transition probabilities are equivalent, the existence of the solution to the Bellman equation is shown, with the use of which optimal strategies are constructed.
Nico M. van Dijk, Arie Hordijk (1996)
Kybernetika
Similarity: