On variance conditions for Markov chain CLTs.
Häggström, Olle, Rosenthal, Jeffrey S. (2007)
Electronic Communications in Probability [electronic only]
Similarity:
Häggström, Olle, Rosenthal, Jeffrey S. (2007)
Electronic Communications in Probability [electronic only]
Similarity:
Kalashnikov, Vladimir V. (1994)
Journal of Applied Mathematics and Stochastic Analysis
Similarity:
Roberts, Gareth O., Rosenthal, Jeffrey S. (1997)
Electronic Communications in Probability [electronic only]
Similarity:
Marius Losifescu (1979)
Banach Center Publications
Similarity:
Cohn, Harry (1979)
International Journal of Mathematics and Mathematical Sciences
Similarity:
E. Nummelin, R. L. Tweedie (1976)
Annales scientifiques de l'Université de Clermont. Mathématiques
Similarity:
Dawid Czapla (2012)
Annales Polonici Mathematici
Similarity:
Stettner [Bull. Polish Acad. Sci. Math. 42 (1994)] considered the asymptotic stability of Markov-Feller chains, provided the sequence of transition probabilities of the chain converges to an invariant probability measure in the weak sense and converges uniformly with respect to the initial state variable on compact sets. We extend those results to the setting of Polish spaces and relax the original assumptions. Finally, we present a class of Markov-Feller chains with a linear state space...
Witold Bednorz (2013)
Applicationes Mathematicae
Similarity:
We give an improved quantitative version of the Kendall theorem. The Kendall theorem states that under mild conditions imposed on a probability distribution on the positive integers (i.e. a probability sequence) one can prove convergence of its renewal sequence. Due to the well-known property (the first entrance last exit decomposition) such results are of interest in the stability theory of time-homogeneous Markov chains. In particular this approach may be used to measure rates of convergence...
Jones, Galin L. (2004)
Probability Surveys [electronic only]
Similarity:
Mariusz Górajski (2009)
Annales UMCS, Mathematica
Similarity:
In this paper we consider an absorbing Markov chain with finite number of states. We focus especially on random walk on transient states. We present a graph reduction method and prove its validity. Using this method we build algorithms which allow us to determine the distribution of time to absorption, in particular we compute its moments and the probability of absorption. The main idea used in the proofs consists in observing a nondecreasing sequence of stopping times. Random walk on...
Rosenthal, Jeffrey S. (2002)
Electronic Communications in Probability [electronic only]
Similarity:
Thomas Kaijser
Similarity:
Consider a Hidden Markov Model (HMM) such that both the state space and the observation space are complete, separable, metric spaces and for which both the transition probability function (tr.pr.f.) determining the hidden Markov chain of the HMM and the tr.pr.f. determining the observation sequence of the HMM have densities. Such HMMs are called fully dominated. In this paper we consider a subclass of fully dominated HMMs which we call regular. A fully dominated,...
Kengo Kamatani (2014)
ESAIM: Probability and Statistics
Similarity:
We study asymptotic behavior of Markov chain Monte Carlo (MCMC) procedures. Sometimes the performances of MCMC procedures are poor and there are great importance for the study of such behavior. In this paper we call degeneracy for a particular type of poor performances. We show some equivalent conditions for degeneracy. As an application, we consider the cumulative probit model. It is well known that the natural data augmentation (DA) procedure does not work well for this model and the...