Jethro's Braindump

Markov Chains

tags
Stochastic Processes

Introduction

Consider a process that has a value in ecah time period. Let Xn denote its value in time period n, and suppose we want to make a probability model for the sequence of successive values $X_0, X_1, …, $. The simplest model would probably assume that the Xn are independent random variables, but often such an assumption is clearly unjustified. However, it may be reasonable to assume that the conditional distribution of Xn+1 given all the past observations, only depends on Xn. Such an assumption defines a Markov chain, a type of stochastic process.

Let {Xn,n=0,1,2,} be a stochastic process that takes on a finite or countable number of possible values. Unless otherwise mentioned, this set of possible values of the process will be denoted by the set of nonnegative integers {0,1,2,}. If Xn=i, then the process is said to be in state i at time n. We suppose that whenever the process is in state i, there is a fixed probabibility Pij that it will be next be in state j. That is, we suppose that:

P{Xn+1=j|Xn=i,Xn1=in1,}=Pij

for all states i0,i1,,in1,i,j and all n0. The value Pij represents the probability that the proces will, when in state i, next make a transition into state j.

Chapman-Kolmogorov Equations

Having defined the one-step transition probabilities Pij, we can now define the n -step transition probaiblities Pijn to be the probability that the process in state i will be in state j after n additional transitions. That is:

Pijn=P{Xn+k=j|Xk=i}

The Chapman-Kolmogorov equations provide a method for computing these n-step transition probabilities. These equations are:

Pijn+m=k=0PijnPkjm for all n,m0, all i,j

and are most easily understood by noting that PiknPkjm represents the probability that starting in i the process will go to state j in n+m transitions through a path which takes it into state k at the nth transition. Summing over all intermediate states yields the total probability that the process will be in state j after n+m transitions.

If we let P(n) denote the matrix of $n$-step transition probabilities Pijn then the Chapman-Kolmogorov equation asserts that:

P(n+m)=PnPm

The $n$-step transition matrix can be obtained by multiplying the matrix P by itself n times.

Classification of States

State j is said to be accessible from state i if Pijn>0 for some n0. Two states i and j that are accessible to each other are said to communicate, and we write ij.

Any state communicates with itself, since by definition Pii0=1.

The relation of communication satisfies the following 3 properties:

  1. State i communicates with state i, for all i0
  2. If state i communicates with state j, then state j communicates with state i.
  3. If state i communicates with state j, and state j communicates with state k, then state i communicates with state k.

Two states that communicate are said to be in the same class. The concept of communication divides the state space into several separate classes. The Markov chain is said to be irreducible if there is only 1 class, that is all states communicate with each other.

Limiting Probabilities

Suppose we have a P matrix such as:

P(4)=[0.57490.5241 0.56680.4332]

We can compute P(8):

P(8)=[0.5720.428 0.5700.430]

Notice that P(8) is almost identical to P(4), and each of the rows of P(8) has almost identical entries. It seems that Pijn is converging to some value that is the same for all i. There seems to exist a limiting probability that the process will be in state j after a large number of transitions, and this value is independent of the initial state.

State i is said to have period d if Piin=0 whenever n is not divisible by d, and d is the largest integer with this property. For instance, it may be possible for the process to enter state i only at times 2,4,6,,8, in which case state i has period. A state with period 1 is said to be aperiodic.

In a finite-state Markov chain, all recurrent states are positive recurrent. Positive recurrent, aperiodic states are called ergodic.

For an irreducible ergodic Markov chain limnPijn exists and is independent of i. Furthermore, letting

πj=limnPijn,j0

then πj is the unique nonnegative solution of:

πj=i=0πiPij,j0,i=0πj==1

We can obtain an expression for P(Xn+1=j) by conditioning on the state at time n:

P(Xn+1=j)=i=0P(Xn+1=j|Xn=i)P(Xn=i) =i=0PijP(Xn=i)

These long run proportions πj,j0 are often called stationary probabilities. The reason being that if the initial state is chosen according to probabilities πj,j0, then the probability of being in state j at any time n is also equal to πj.