Markov Chains
- tags
- Stochastic Processes
Introduction
Consider a process that has a value in ecah time period. Let
denote its value in time period , and suppose we want to make a
probability model for the sequence of successive values $X_0, X_1,
…, $. The simplest model would probably assume that the are
independent random variables, but often such an assumption is clearly
unjustified. However, it may be reasonable to assume that the
conditional distribution of given all the past observations,
only depends on . Such an assumption defines a Markov chain, a
type of stochastic process.
Let be a stochastic process
that takes on a finite or countable number of possible values. Unless
otherwise mentioned, this set of possible values of the process will
be denoted by the set of nonnegative integers . If , then the process is said to be in state
at time . We suppose that whenever the process is in state ,
there is a fixed probabibility that it will be next be in
state . That is, we suppose that:
for all states and all . The
value represents the probability that the proces will, when
in state , next make a transition into state .
Chapman-Kolmogorov Equations
Having defined the one-step transition probabilities , we can
now define the -step transition probaiblities to be the
probability that the process in state will be in state after
additional transitions. That is:
The Chapman-Kolmogorov equations provide a method for computing these
n-step transition probabilities. These equations are:
and are most easily understood by noting that
represents the probability that starting in the process will go to
state in transitions through a path which takes it into
state at the nth transition. Summing over all intermediate states
yields the total probability that the process will be in state
after transitions.
If we let denote the matrix of $n$-step transition
probabilities then the Chapman-Kolmogorov equation asserts
that:
The $n$-step transition matrix can be obtained by multiplying the
matrix by itself times.
Classification of States
State is said to be accessible from state if
for some . Two states and that are accessible to each
other are said to communicate, and we write .
Any state communicates with itself, since by definition .
The relation of communication satisfies the following 3 properties:
- State communicates with state , for all
- If state communicates with state , then state
communicates with state .
- If state communicates with state , and state
communicates with state , then state communicates with state
.
Two states that communicate are said to be in the same class. The
concept of communication divides the state space into several separate
classes. The Markov chain is said to be irreducible if there is only 1
class, that is all states communicate with each other.
Limiting Probabilities
Suppose we have a matrix such as:
We can compute :
Notice that is almost identical to
, and each of the rows of has
almost identical entries. It seems that is converging to
some value that is the same for all . There seems to exist a
limiting probability that the process will be in state after a
large number of transitions, and this value is independent of the
initial state.
State is said to have period if whenever is
not divisible by , and is the largest integer with this
property. For instance, it may be possible for the process to enter
state only at times , in which case state
has period. A state with period 1 is said to be aperiodic.
In a finite-state Markov chain, all recurrent states are positive
recurrent. Positive recurrent, aperiodic states are called ergodic.
For an irreducible ergodic Markov chain exists and is independent of . Furthermore, letting
then is the unique nonnegative solution of:
We can obtain an expression for by conditioning on
the state at time :
These long run proportions are often called
stationary probabilities. The reason being that if the initial state
is chosen according to probabilities , then the
probability of being in state at any time is also equal to
.