The definition of Markov chains has evolved during the 20th century. In 1953 the term Markov chain was used for stochastic processes with discrete or continuous index set, living on a countable or finite state space, see Doob.1 or Chung.2 Since the late 20th century it became more popular to consider a Markov chain as a stochastic process with discrete index set, living on a measurable state space.345
Denote with ( E , Σ ) {\displaystyle (E,\Sigma )} a measurable space and with p {\displaystyle p} a Markov kernel with source and target ( E , Σ ) {\displaystyle (E,\Sigma )} . A stochastic process ( X n ) n ∈ N {\displaystyle (X_{n})_{n\in \mathbb {N} }} on ( Ω , F , P ) {\displaystyle (\Omega ,{\mathcal {F}},\mathbb {P} )} is called a time homogeneous Markov chain with Markov kernel p {\displaystyle p} and start distribution μ {\displaystyle \mu } if
is satisfied for any n ∈ N , A 0 , … , A n ∈ Σ {\displaystyle n\in \mathbb {N} ,\,A_{0},\dots ,A_{n}\in \Sigma } . One can construct for any Markov kernel and any probability measure an associated Markov chain.6
For any measure μ : Σ → [ 0 , ∞ ] {\displaystyle \mu \colon \Sigma \to [0,\infty ]} we denote for μ {\displaystyle \mu } -integrable function f : E → R ∪ { ∞ , − ∞ } {\displaystyle f\colon E\to \mathbb {R} \cup \{\infty ,-\infty \}} the Lebesgue integral as ∫ E f ( x ) μ ( d x ) {\displaystyle \int _{E}f(x)\,\mu (dx)} . For the measure ν x : Σ → [ 0 , ∞ ] {\displaystyle \nu _{x}\colon \Sigma \to [0,\infty ]} defined by ν x ( A ) := p ( x , A ) {\displaystyle \nu _{x}(A):=p(x,A)} we used the following notation:
If μ {\displaystyle \mu } is a Dirac measure in x {\displaystyle x} , we denote for a Markov kernel p {\displaystyle p} with starting distribution μ {\displaystyle \mu } the associated Markov chain as ( X n ) n ∈ N {\displaystyle (X_{n})_{n\in \mathbb {N} }} on ( Ω , F , P x ) {\displaystyle (\Omega ,{\mathcal {F}},\mathbb {P} _{x})} and the expectation value
for a P x {\displaystyle \mathbb {P} _{x}} -integrable function X {\displaystyle X} . By definition, we have then P x [ X 0 = x ] = 1 {\displaystyle \mathbb {P} _{x}[X_{0}=x]=1} .
We have for any measurable function f : E → [ 0 , ∞ ] {\displaystyle f\colon E\to [0,\infty ]} the following relation:7
For a Markov kernel p {\displaystyle p} with starting distribution μ {\displaystyle \mu } one can introduce a family of Markov kernels ( p n ) n ∈ N {\displaystyle (p_{n})_{n\in \mathbb {N} }} by
for n ∈ N , n ≥ 1 {\displaystyle n\in \mathbb {N} ,\,n\geq 1} and p 1 := p {\displaystyle p_{1}:=p} . For the associated Markov chain ( X n ) n ∈ N {\displaystyle (X_{n})_{n\in \mathbb {N} }} according to p {\displaystyle p} and μ {\displaystyle \mu } one obtains
A probability measure μ {\displaystyle \mu } is called stationary measure of a Markov kernel p {\displaystyle p} if
holds for any A ∈ Σ {\displaystyle A\in \Sigma } . If ( X n ) n ∈ N {\displaystyle (X_{n})_{n\in \mathbb {N} }} on ( Ω , F , P ) {\displaystyle (\Omega ,{\mathcal {F}},\mathbb {P} )} denotes the Markov chain according to a Markov kernel p {\displaystyle p} with stationary measure μ {\displaystyle \mu } , and the distribution of X 0 {\displaystyle X_{0}} is μ {\displaystyle \mu } , then all X n {\displaystyle X_{n}} have the same probability distribution, namely:
for any A ∈ Σ {\displaystyle A\in \Sigma } .
A Markov kernel p {\displaystyle p} is called reversible according to a probability measure μ {\displaystyle \mu } if
holds for any A , B ∈ Σ {\displaystyle A,B\in \Sigma } . Replacing A = E {\displaystyle A=E} shows that if p {\displaystyle p} is reversible according to μ {\displaystyle \mu } , then μ {\displaystyle \mu } must be a stationary measure of p {\displaystyle p} .
Joseph L. Doob: Stochastic Processes. New York: John Wiley & Sons, 1953. ↩
Kai L. Chung: Markov Chains with Stationary Transition Probabilities. Second edition. Berlin: Springer-Verlag, 1974. ↩
Sean Meyn and Richard L. Tweedie: Markov Chains and Stochastic Stability. 2nd edition, 2009. ↩
Daniel Revuz: Markov Chains. 2nd edition, 1984. ↩
Rick Durrett: Probability: Theory and Examples. Fourth edition, 2005. ↩