Discrete-time Markov control processes with recursive discount rates
Yofre H. García; Juan González-Hernández
Kybernetika (2016)
- Volume: 52, Issue: 3, page 403-426
- ISSN: 0023-5954
Access Full Article
topAbstract
topHow to cite
topH. García, Yofre, and González-Hernández, Juan. "Discrete-time Markov control processes with recursive discount rates." Kybernetika 52.3 (2016): 403-426. <http://eudml.org/doc/281537>.
@article{H2016,
abstract = {This work analyzes a discrete-time Markov Control Model (MCM) on Borel spaces when the performance index is the expected total discounted cost. This criterion admits unbounded costs. It is assumed that the discount rate in any period is obtained by using recursive functions and a known initial discount rate. The classic dynamic programming method for finite-horizon case is verified. Under slight conditions, the existence of deterministic non-stationary optimal policies for infinite-horizon case is proven. Also, to find deterministic non-stationary $\epsilon -$optimal policies, the value-iteration method is used. To illustrate an example of recursive functions that generate discount rates, we consider the expected values of stochastic processes, which are solutions of certain class of Stochastic Differential Equations (SDE) between consecutive periods, when the initial condition is the previous discount rate. Finally, the consumption-investment problem and the discount linear-quadratic problem are presented as examples; in both cases, the discount rates are obtained using a SDE, similar to the Vasicek short-rate model.},
author = {H. García, Yofre, González-Hernández, Juan},
journal = {Kybernetika},
keywords = {dynamic programming method; optimal stochastic control},
language = {eng},
number = {3},
pages = {403-426},
publisher = {Institute of Information Theory and Automation AS CR},
title = {Discrete-time Markov control processes with recursive discount rates},
url = {http://eudml.org/doc/281537},
volume = {52},
year = {2016},
}
TY - JOUR
AU - H. García, Yofre
AU - González-Hernández, Juan
TI - Discrete-time Markov control processes with recursive discount rates
JO - Kybernetika
PY - 2016
PB - Institute of Information Theory and Automation AS CR
VL - 52
IS - 3
SP - 403
EP - 426
AB - This work analyzes a discrete-time Markov Control Model (MCM) on Borel spaces when the performance index is the expected total discounted cost. This criterion admits unbounded costs. It is assumed that the discount rate in any period is obtained by using recursive functions and a known initial discount rate. The classic dynamic programming method for finite-horizon case is verified. Under slight conditions, the existence of deterministic non-stationary optimal policies for infinite-horizon case is proven. Also, to find deterministic non-stationary $\epsilon -$optimal policies, the value-iteration method is used. To illustrate an example of recursive functions that generate discount rates, we consider the expected values of stochastic processes, which are solutions of certain class of Stochastic Differential Equations (SDE) between consecutive periods, when the initial condition is the previous discount rate. Finally, the consumption-investment problem and the discount linear-quadratic problem are presented as examples; in both cases, the discount rates are obtained using a SDE, similar to the Vasicek short-rate model.
LA - eng
KW - dynamic programming method; optimal stochastic control
UR - http://eudml.org/doc/281537
ER -
References
top- Arnold, L., 10.1002/zamm.19770570413, John Wiley and Sons, New York 1973. Zbl0278.60039MR0443083DOI10.1002/zamm.19770570413
- Ash, R., Doléans-Dade, C., Probability and Measure Theory., Academic Press, San Diego 2000. Zbl0944.60004MR1810041
- Bellman, R., Dynamic Programming., Princeton Univ. Press, New Jersey 1957. Zbl1205.90002MR0090477
- Bertsekas, D., Shreve, S., Stochastic Optimal Control: The Discrete Time Case., Athena Scientific, Massachusetts 1996. Zbl0633.93001MR0511544
- Brigo, D., Mercurio, F., 10.1007/978-3-662-04553-4, Springer-Verlag, New York 2001. Zbl1109.91023MR1846525DOI10.1007/978-3-662-04553-4
- Black, F., Karasinski, P., 10.2469/faj.v47.n4.52, Financ. Anal. J. 47 (1991), 4, 52-59. DOI10.2469/faj.v47.n4.52
- Carmon, Y., Shwartz, A., 10.1016/j.orl.2008.10.005, Oper. Res. Lett. 37 (2009), 51-55. Zbl1154.90610MR2488083DOI10.1016/j.orl.2008.10.005
- Vecchia, E. Della, Marco, S. Di, Vidal, F., Dynamic programming for variable discounted Markov decision problems., In: Jornadas Argentinas de Informática e Investigación Operativa (43JAIIO) - XII Simposio Argentino de Investigación Operativa (SIO), Buenos Aires, 2014, pp. 50-62.
- Cox, J., Ingersoll, J., Ross, S., 10.2307/1911242, Econometrica 53 (1985), 385-407. Zbl1274.91447MR0785475DOI10.2307/1911242
- Dothan, U., 10.1016/0304-405x(78)90020-x, J. Financ. Econ. 6 (1978), 59-69. DOI10.1016/0304-405x(78)90020-x
- Feinberg, E., Shwartz, A., 10.1287/moor.19.1.152, J. Finan. Econ. 19 (1994), 152-168. Zbl0803.90123MR1290017DOI10.1287/moor.19.1.152
- González-Hernández, J., López-Martínez, R., Pérez-Hernández, J., 10.1007/s00186-006-0092-2, Math. Method Oper. Res. 65 (2006), 1, 27-44. Zbl1126.90075MR2302022DOI10.1007/s00186-006-0092-2
- González-Hernández, J., López-Martínez, R., Minjarez-Sosa, A., Adaptive policies for stochastic systems under a randomized discounted cost criterion., Bol. Soc. Mat. Mex. 14 (2008), 3, 149-163. Zbl1201.93130MR2667162
- González-Hernández, J., López-Martínez, R., Minjarez-Sosa, A., Approximation, estimation and control of stochastic systems under a randomized discounted cost criterion., Kybernetika 45 (2008), 5, 737-754. Zbl1190.93105MR2599109
- Guo, X., Hernández-Del-Valle, A., Hernández-Lerma, O., 10.3166/ejc.18.528-538, Eur. J. Control 15 (2012), 7, 528-538. MR3086896DOI10.3166/ejc.18.528-538
- Hernández-Lerma, O., Lasserre, J. B., 10.1007/978-1-4612-0729-0, Springer-Verlag, New York 1996. Zbl0840.93001MR1363487DOI10.1007/978-1-4612-0729-0
- Minjarez-Sosa, J., 10.1007/s11750-015-0360-5, TOP 23 (2015), 3, 743-772. MR3407674DOI10.1007/s11750-015-0360-5
- Hinderer, K., 10.1007/978-3-642-46229-0, In: Lecture Notes Operations Research (M. Bechmann and H. Künzi, eds.), Springer-Verlag 33, Zürich 1970. MR0267890DOI10.1007/978-3-642-46229-0
- Ho, T., Lee, S., 10.1111/j.1540-6261.1986.tb02528.x, J. Financ. 41 (1986), 1011-1029. DOI10.1111/j.1540-6261.1986.tb02528.x
- Hull, J., Options, Futures and other Derivatives. Sixth edition., Prentice Hall, New Jersey 2006.
- Hull, J., White, A., 10.1093/rfs/3.4.573, Rev. Financ. Stud. 3 (1990), 573-592. DOI10.1093/rfs/3.4.573
- Mercurio, F., Moraleda, J., 10.1080/13518470151141440, Eur. J. Finance 7 (2001), 93-116. DOI10.1080/13518470151141440
- Rendleman, R., Bartter, B., 10.2307/2979016, J. Financ. Quant. Anal. 15 (1980), 11-24. DOI10.2307/2979016
- Rieder, U., 10.1007/bf01168566, Manuscripta Math. 24 (1978), 115-131. Zbl0385.28005MR0493590DOI10.1007/bf01168566
- Schäl, M., 10.1007/bf00532612, Z. Wahrscheinlichkeit 32 (1975), 179-196. Zbl0316.90080MR0378841DOI10.1007/bf00532612
- Vasicek, O., 10.1016/0304-405x(77)90016-2, J. Financ. Econ. 5 (1977), 177-188. DOI10.1016/0304-405x(77)90016-2
- Wei, Q., X, X. Guo, 10.1016/j.orl.2011.06.014, Oper. Res. Lett. 39 (2011), 369-374. Zbl1235.90178MR2835530DOI10.1016/j.orl.2011.06.014
- Ye, L., Guo, X., 10.1007/s10440-012-9669-3, Acta Appl. Math. 121 (2012), 1, 5-27. Zbl1281.90082MR2966962DOI10.1007/s10440-012-9669-3
- Zhang, Y., 10.1007/s11750-011-0186-8, TOP 21 (2013), 2, 378-408. Zbl1273.90235MR3068494DOI10.1007/s11750-011-0186-8
Citations in EuDML Documents
top- Fu Xingjian, Li Zizheng, Neural network optimal control for nonlinear system based on zero-sum differential game
- Rocio Ilhuicatzi-Roldán, Hugo Cruz-Suárez, Selene Chávez-Rodríguez, Markov decision processes with time-varying discount factors and random horizon
- Yofre H. García, Juan González-Hernández, Solutions of semi-Markov control models with recursive discount rates and approximation by -optimal policies
NotesEmbed ?
topTo embed these notes on your page include the following JavaScript code on your page where you want the notes to appear.