Generalizing Long Run Performance Beyond Finite MCs
Recall this from the previous page:
Note that we’re only talking about irreducible MCs for now. We already know what happens in all the cases except when we have a recurrent aperiodic MC with infinite states.
We will now try to generalize the results we’ve obtained for a regular MC to a recurrent aperiodic MC with infinite states. That is, we want to answer the question: when there are infinite states, how do we derive the regularity?
Is it possible that we have and ? Then it still seems “impossible” to return (even though the probability is ). We want to be able to intrepret this corectly (and gain intuition about this).
Generalising Main Theorem
Consider a recurrent MC.
First Return Time:
In words, it is the first time that the process returns to .
Recall that we defined (first return probability) to be “starting with , what is the probability that the first return to happens at th step).
So, what is the relationship between and ?
Clearly, we can express it as:
In other words, we can say that is the “probability mass” of at , i.e., gives us the probability that .
For recurrent states, we know that , i.e., . Therefore, we can calculate , which can be interpreted as the average first return time or (more commonly known as) the mean duration between visits.
By definition of “mean” (expectation), we can write it as such:
Note that we can only define when . When we have , then the probability that there are infinitely many steps between 2 visits is non-zero, and equal to so the expectation will be infinity (which is not very meaningful).
Limit Theorem
For any recurrent irreducible MC, define:
Then,
For any ,
If , then
If , then
Note that the theorem applies for MCs with infinitely many states too! It also applies for periodic MCs.
Interpretation
Let’s try to interpret the theorem:
Consider a fixed and observe that for all , the “average transition probability” (in the long-run) is the same, and equal to . This seems intuitive because we can think of a transition to as a “success” and a transition to any other state as a “failure” and model it as a geometric distribution. Then, we know that if the probability of success is , the mean time for success is . In this case, transitioning to any other state preserves the probability of success (i.e., it does not change because every state has the same probability of transition to ) and so, it does indeed fit well as a geometric distribution. For an aperiodic MC, in the long-run, the probability of transitioning from converges to .
Note that this theorem applies to periodic MCs as well, but the analogy of a geometric distribution only holds for an aperiodic MC.
In the first part of the theorem, we’re considering the limit of the average transition probability from . But when the MC is aperiodic, we’re saying that the -th step probability itself (not the average) converges to a fixed value, equal to .
This looks very similar to the regular MC case - we know that in the long-run, when we have finitely many states (recall that for a regular MC with finite states, converges to a “fixed” matrix where every row is equal to the limiting distribution). This should give a hint that 😲
Note that the average of a convergent series is determined by the “long-run terms” and since there are infinitely many terms which are equal to the limit (they dominate the initial few terms which may be different from the limit), the average is also equal to the limit. This is how we can understand that (2) and (1) are compatible/consistent with each other.
In the third part, notice that we’re only considering the probability of returning to when we start from itself. If we start from any other state, we can’t make any claims since we don’t know how many steps it takes to reach . For any that is NOT a multiple of the period , obviously . So, we only consider the limit of the non-zero entries, i.e., entries of where is a multiple of . In this case, we claim that the limit exists, and is equal to .
It is natural to expect to appear on the RHS because if we look at the first result, the “average” is brought down (by a factor of ) by all the zero-entries between every 2 multiples of . But in this case, we’re removing all these zero-entries and only considering multiples of , and so we expect the “average over non-zero entries” to be times higher.
Basically, we expect the non-zero entries, to converge to as .
Until now, we didn’t know what would happen for a periodic MC in the long run, since it doesn’t converge. While that is still true, we’ve shown that the subsequence formed by the non-zero entries does indeed converge (and so, we can make more sense out of the process).
Remarks
The mean duration between visits, , can be finite or infinite:
- Infinite: When , the limiting probability at each state is , although it is recurrent. We call such a MC to be null recurrent. For example, consider the symmetric random walk with and no absorbing state. Note that it is still recurrent (there’s only one class so it must be recurrent).
- Finite: When , the limiting probability at each state is . In such a case, we call it a positive recurrent MC. e.g. Random walk with (process eventually reaches 0) and “reflection” at , i.e., When , we can only consider the steps d. When , the limiting probability is positive, which means that it is a regular MC.
The notion of “null recurrence” only applies to MCs with states. For a MC with finite states, a state can never be null-recurrent.
Basic Limit Theorem
We can generalize the results from the previous theorem by only considering those MCs that have . This gives us the Basic Limit Theorem.
Basic Limit Theorem
For a positive recurrent (), irreducible, and aperiodic MC,
- exists for any and is given by:
- If is the solution to the equation , then we have:
Note:
- A positive recurrent, irreducible, aperiodic MC is called an ergodic MC. Hence, the basic limit theorem applies to all ergodic MCs.
- All regular MCs are also ergodic MCs. Why? Because if there are finite states (and it is irreducible so all of them must be recurrent), the limiting probability on each state is non-zero, hence it is positive recurrent.
- We do NOT require the MC to have finite/infinite states for the theorem to hold.
In general, it’s quite difficult to prove the “positive recurrent” property of a MC with inifinite states, so we normally deal with MCs with finite states (for which we can apply the theorem from the previous page for regular MCs)
How do we calculate ? Using the second part of the theorem, we can first solve to find the limiting distribution, and then take its reciprocal to obtain all the ’s. This gives us another interpretation for ⇒ knowing (and hence, ) also tells us the mean duration between visits.
Sometimes, it’s easier to find rather than (generally when we have infinite states). Exampe: A MC representing the number of consecutive successes of an binomial trials (with probability of success = ). To calculate in such a case, we need to find for all values of and then use the definition of .
Reducible Chains
Until now, we’ve been talking only about irreducible chains (and all the theorems only apply to irreducible chains). So, what do we do for reducible chains?
Intuitively we know that we can “reduce” (i.e., decompose) the reducible chain into irreducible sub-chains (each subchain consists of states in the same communication class).
Then, in the long-run, the limiting probability of the transient classes is zero. Moreover, we can find the probability of entering any recurrent class using FSA (by treating every recurrent class as an absorbing state since the process can never leave the recurrent class, i.e., no outoing arrows from the recurrent class).
Then, within each recurrent class of the reducible MC, we can check the period .
- If , then the limit of the subsequence is for every state in the class (only true if the process starts at state ).
- If
- If the class has finite states, then we can find out the limiting probability (since it is a regular markov “subchain”) using
- If the class has states, then the limiting probability at each state is either zero (if it is null recurrent) or can be obtained by if it is positive recurrent.
Now we know what happens whenever we have a recurrent class. Next, we’ll try to incorporate these results into a general MC so that we can understand the long-run probability of every state (note: not every class).
Long-Run Performance in General Markov Chains
As a quick recap, currently we know how to:
- Find the probability of a process entering any recurrent class (using FSA)
- Find the limiting probability of each state within the recurrent class (using ) if the period . (of course, if , then there is no limiting distribution since the limit does not converge)
If we can combine both, we can get the results for a general markov chain. Basically, we’re trying to find as for a general MC now (for a regular MC, we already know that will consist of all rows equal to but we don’t expect this to be the case for a general MC).
Intuitively (using conditional probability), we know that the limiting probabiilty (if it exists) at a state is equal to the probability of entering ’s recurrent class times the limiting probability in its own class.
Formally, consider where is a recurrent class. Then, we can set up a new MC and find the entering probability:
Once the process enters , it never leaves. Hence, we can restrict the process (i.e., consider the sub-chain) on only. Clearly, the new MC restricted on is irreducible and recurrent (since it is formed by states in the same class).
Then, we can find for state by solving .
Hence, if we repeat generating and making jumps for sufficiently large (i.e., ), then:
Procedure
For the sake of clarity and completeness, we outline the entire procedure to analyze the long-run performance of ANY general MC.
Find all the classes
Set up a new MC where every recurrent class is denoted by one state. Then, find denoted by → this gives the probability of entering any recurrent class, given the initial distribution.
We can ignore all transient classes 😲 because the process will eventually leave them in the long-run, i.e., their long-term probability is zero.
For every recurrent class , we find the period .
- Aperiodic (d=$1): find the corresponding limiting distribution of state $j in this class, denoted by , by considering the sub-MC restricted on
- Periodic (): there is NO limiting distribution, but we can still check the long-run proportion of time in each state by finding (i.e., we can still find but the interpretation is different in this case)
Consider the initial state
If is transient, then
If is recurrent, then:
Finally, given the initial distribution , then: