Exact decomposition approaches for Markov decision processes: a survey.
Daoui, Cherki, Abbad, Mohamed, Tkiouat, Mohamed (2010)
Advances in Operations Research
Similarity:
Daoui, Cherki, Abbad, Mohamed, Tkiouat, Mohamed (2010)
Advances in Operations Research
Similarity:
Xu, Qing, Batabyal, Amitrajeet A. (2002)
Discrete Dynamics in Nature and Society
Similarity:
Brianzoni, Serena, Mammana, Cristiana, Michetti, Elisabetta, Zirilli, Francesco (2008)
Discrete Dynamics in Nature and Society
Similarity:
Ulrich Herkenrath, Radu Theodorescu (1981)
Trabajos de Estadística e Investigación Operativa
Similarity:
The aim of this paper is to show that the theory of (generalized) random systems with complete connection may serve as a mathematical framework for learning and adaption. Chapter 1 is of an introductory nature and gives a general description of the problems with which one is faced. In Chapter 2 the mathematical model and some results about it are explained. Chapter 3 deals with special learning and adaption models.
Roberta Jungblut-Hessel, Brigitte Plateau, William J. Stewart, Bernard Ycart (2001)
RAIRO - Operations Research - Recherche Opérationnelle
Similarity:
In this paper we present a method to perform fast simulation of large markovian systems. This method is based on the use of three concepts: Markov chain uniformization, event-driven dynamics, and modularity. An application of urban traffic simulation is presented to illustrate the performance of our approach.
Nico M. van Dijk, Arie Hordijk (1996)
Kybernetika
Similarity:
Buckley, F.M., Pollett, P.K. (2010)
Probability Surveys [electronic only]
Similarity:
Liu, R.H., Zhang, Q., Yin, G. (2006)
Journal of Applied Mathematics and Stochastic Analysis
Similarity:
Raúl Montes-de-Oca, Francisco Salem-Silva (2005)
Kybernetika
Similarity:
This paper deals with Markov decision processes (MDPs) with real state space for which its minimum is attained, and that are upper bounded by (uncontrolled) stochastically ordered (SO) Markov chains. We consider MDPs with (possibly) unbounded costs, and to evaluate the quality of each policy, we use the objective function known as the average cost. For this objective function we consider two Markov control models and . and have the same components except for the transition laws....