The equilibrium
Run a Markov chain long enough and the distribution over states stops changing. No matter where you started.
Stationary distribution
A probability vector is a stationary distribution of a Markov chain with transition matrix if: and . Once the chain reaches this distribution, it stays there forever.
Think of it as a balance condition: the probability flowing into each state equals the probability flowing out.
Finding the stationary distribution
In the long run, it's sunny 2/3 of the time and rainy 1/3 — regardless of today's weather.
Watch it converge
Pour "probability fluid" into the nodes. Watch it flow through the transitions until the levels stabilize at the stationary distribution:
The convergence theorem
If a Markov chain is irreducible and aperiodic, then:
- A unique stationary distribution exists.
- For any starting distribution, as .
- The fraction of time spent in state converges to .
Reducible chains may have multiple stationary distributions. Periodic chains cycle rather than converge. Both conditions (irreducible + aperiodic) are needed.
Detailed balance
A stronger condition than stationarity:
A distribution satisfies detailed balance (is "reversible") if: The flow from to equals the flow from to .
Detailed balance implies stationarity (but not vice versa). MCMC algorithms are built on it.
Applications
| Domain | States | Transitions |
|---|---|---|
| Page ranking | Web pages | Links between pages |
| Genetics | Alleles | Mutation + selection |
| Speech recognition | Phonemes | Language model |
| Finance | Market regimes | Bull ↔ Bear transitions |
| Board games | Board positions | Dice rolls + moves |
Practice problems
Summary
| Concept | Key Formula |
|---|---|
| Stationary distribution | , |
| Convergence | Irreducible + aperiodic → unique |
| Time averages | Fraction in state → |
| Detailed balance |
The stationary distribution is the Markov chain's version of the LLN. Just as sample averages converge to the true mean, state visit frequencies converge to .
What's next
The stationary distribution is a theoretical target. But what if we can't solve analytically? MCMC algorithms construct a Markov chain whose stationary distribution is exactly the distribution we want to sample from.