Markov Chains

Discrete Stochastic Processes

A discrete stochastic process is a discrete system in which transitions occur randomly according to some probability distribution.

P[Xt+1=jXt=st,,X0=s0]=P[Xt+1=jXt=st].

Finite Markov Chains

A finite Markov chain is a memoryless homogeneous stochastic process with a finite number of states S={1n}. It is characterized by a transition matrix P=[pij]i,jS and an initial distribution q0, where (q0)i=P[X0=i].

State Probabilities

The probability distribution of states at time step t can be represented as a vector qt. By marginalizing out the distribution of the previous time step:

(qt+1)j=P[Xt+1=j]=iP[Xt+1=jXt=i]P[Xt=i]=ipij(qt)i=P:,jTqt

Therefore,

qt+1=PTqt=(PT)2qt1==(PT)tq0,

which indicates that the distribution at each time step is well-defined as long as the initial distribution q0 is specified.

We denote the probability P[Xt=jX0=i] as pij(t) for simplicity.

Stationary Distribution

A stationary distribution is a state vector π, such that

π=PTπ.

In other words, π is an eigenvector of PT to eigenvalue 1.

Irreducible Markov Chains

A Markov chain is irreducible if for all i,jS, n, s.t. pij(n)>0, i.e. State j is accessible from State i. A irreducible Markov chain has a strongly connected underlying transition graph.

An irreducible Markov chain has a unique stationary state π.

πj=hjj1,

where hij, the hitting time, is the average number of steps needed to go from i to j:

hij:=E[min{h1X0=i,Xh=j}].

In the following state diagram, State 0 and State 2 are not accessible to each other. Therefore, the Markov chain is not irreducible, and any [a,0,1a]T can be a stationary state.

stateDiagram-v2
	direction LR
    state "0" as zero
    state "1" as one
    state "2" as two

    zero --> zero : 1

    one --> zero: q
    one --> two : p

    two --> two: 1

Aperiodic Markov Chains

The periodicity of a state jS is gcd{n0pjj(n)>0}.

A Markov chain is aperiodic if all states have periodicity 1. i.e., if for all iS,

gcd{length of the walks from i to i}=1.
theorem

For every irreducible, aperiodic (ergodic) Markov chain, independently of q0,
limtqt=π.

Time-reversible Markov Chains

Consider an ergodic Markov chain {X0,X1,,Xt,}. Given a (large) time step n, we trace the states going back in time. It turns out that {Yk=Xnkk=0,1,} is a Markov chain with transition probabilities:

pjiR=P[Yk+1=jYk=i]=P[Xnk1=jXnk=i]=πjpijπi.
proof

Using Bayes' theorem,
pjiR=P[Xnk1=jXnk=i]=P[Xnk=iXnk1=j]P[Xnk1=j]P[Xnk=i]=πjpijπi.

Continuous-time Markov Chains

Rather than a transition matrix, we have a transition probability density p(xx), where x,xRd.

A stationary distribution π(x) satisfies

π(x)=Rdp(xx)π(x)dx.

The reverse transition probabilities is defined as

p(xx)=π(x)p(xx)π(x).