ConceptComplete

The Markov Property

The Markov property is the fundamental assumption that the future evolution of a stochastic process depends only on its present state, not on its past history. This "memoryless" property makes Markov chains tractable for analysis while remaining rich enough to model many real-world phenomena.


Discrete-time Markov chains

Definition1.1Markov chain

A discrete-time stochastic process (Xn)nβ‰₯0(X_n)_{n \geq 0} taking values in a countable state space SS is a Markov chain if for all nβ‰₯0n \geq 0 and all states i0,i1,…,in,j∈Si_0, i_1, \ldots, i_n, j \in S:

P(Xn+1=j∣Xn=in,Xnβˆ’1=inβˆ’1,…,X0=i0)=P(Xn+1=j∣Xn=in),\mathbb{P}(X_{n+1} = j \mid X_n = i_n, X_{n-1} = i_{n-1}, \ldots, X_0 = i_0) = \mathbb{P}(X_{n+1} = j \mid X_n = i_n),

whenever the conditional probabilities are well-defined.

The right-hand side P(Xn+1=j∣Xn=i)\mathbb{P}(X_{n+1} = j \mid X_n = i) is called the transition probability from state ii to state jj.

RemarkTime-homogeneous chains

A Markov chain is time-homogeneous (or stationary) if the transition probabilities do not depend on time: pij(n)=pijp_{ij}(n) = p_{ij} for all nn. Unless stated otherwise, we work with time-homogeneous chains and denote

pij=P(Xn+1=j∣Xn=i).p_{ij} = \mathbb{P}(X_{n+1} = j \mid X_n = i).

The matrix P=(pij)P = (p_{ij}) is called the transition matrix.

ExampleSimple random walk on Z

On S=ZS = \mathbb{Z}, let X0=0X_0 = 0 and

Xn+1=Xn+ΞΎn,X_{n+1} = X_n + \xi_n,

where ΞΎn\xi_n are i.i.d. with P(ΞΎn=1)=p\mathbb{P}(\xi_n = 1) = p and P(ΞΎn=βˆ’1)=q=1βˆ’p\mathbb{P}(\xi_n = -1) = q = 1-p. Then (Xn)(X_n) is a Markov chain with transition probabilities:

pi,i+1=p,pi,iβˆ’1=q,pij=0Β if ∣iβˆ’jβˆ£β‰ 1.p_{i,i+1} = p, \quad p_{i,i-1} = q, \quad p_{ij} = 0 \text{ if } |i-j| \neq 1.

This is the simple random walk on Z\mathbb{Z}.

ExampleTwo-state chain

Let S={0,1}S = \{0, 1\} with transition matrix

P=(1βˆ’Ξ±Ξ±Ξ²1βˆ’Ξ²).P = \begin{pmatrix} 1-\alpha & \alpha \\ \beta & 1-\beta \end{pmatrix}.

If Ξ±,β∈(0,1)\alpha, \beta \in (0,1), the chain moves between states with positive probability. The stationary distribution (if it exists) satisfies Ο€P=Ο€\pi P = \pi, giving

Ο€0=Ξ²Ξ±+Ξ²,Ο€1=Ξ±Ξ±+Ξ².\pi_0 = \frac{\beta}{\alpha + \beta}, \quad \pi_1 = \frac{\alpha}{\alpha + \beta}.


The Chapman-Kolmogorov equation

Definition1.2n-step transition probability

The nn-step transition probability is

pij(n)=P(Xm+n=j∣Xm=i),p_{ij}^{(n)} = \mathbb{P}(X_{m+n} = j \mid X_m = i),

the probability of being in state jj exactly nn steps after starting in state ii.

Theorem1.1Chapman-Kolmogorov equation

For all states i,j∈Si, j \in S and all integers m,nβ‰₯0m, n \geq 0:

pij(m+n)=βˆ‘k∈Spik(m)pkj(n).p_{ij}^{(m+n)} = \sum_{k \in S} p_{ik}^{(m)} p_{kj}^{(n)}.

In matrix form, P(m+n)=P(m)P(n)P^{(m+n)} = P^{(m)} P^{(n)}, i.e., P(n)=PnP^{(n)} = P^n.

The proof follows from the law of total probability and the Markov property: to go from ii to jj in m+nm+n steps, we must pass through some intermediate state kk at time mm.

ExampleComputing PΒ²

For the two-state chain with

P=(0.70.30.40.6),P = \begin{pmatrix} 0.7 & 0.3 \\ 0.4 & 0.6 \end{pmatrix},

we compute

P2=(0.70.30.40.6)2=(0.610.390.520.48).P^2 = \begin{pmatrix} 0.7 & 0.3 \\ 0.4 & 0.6 \end{pmatrix}^2 = \begin{pmatrix} 0.61 & 0.39 \\ 0.52 & 0.48 \end{pmatrix}.

Thus, p00(2)=0.61p_{00}^{(2)} = 0.61: starting in state 0, the probability of being in state 0 after 2 steps is 0.61.


Strong Markov property

Definition1.3Stopping time

A random variable Ο„:Ξ©β†’{0,1,2,…}βˆͺ{∞}\tau: \Omega \to \{0, 1, 2, \ldots\} \cup \{\infty\} is a stopping time with respect to (Xn)(X_n) if the event {Ο„=n}\{\tau = n\} depends only on X0,X1,…,XnX_0, X_1, \ldots, X_n for each nβ‰₯0n \geq 0.

Intuitively, the decision to stop at time nn can be made using only information available up to time nn.

ExampleHitting times are stopping times

For a set AβŠ†SA \subseteq S, the hitting time (or first passage time) is

Ο„A=inf⁑{nβ‰₯0:Xn∈A}.\tau_A = \inf\{n \geq 0 : X_n \in A\}.

Then Ο„A\tau_A is a stopping time: {Ο„A=n}\{\tau_A = n\} occurs iff X0,…,Xnβˆ’1βˆ‰AX_0, \ldots, X_{n-1} \notin A and Xn∈AX_n \in A, which depends only on X0,…,XnX_0, \ldots, X_n.

Theorem1.2Strong Markov property

Let Ο„\tau be a stopping time with P(Ο„<∞)=1\mathbb{P}(\tau < \infty) = 1. Then, conditionally on {Ο„<∞,XΟ„=i}\{\tau < \infty, X_\tau = i\}, the process (XΟ„+n)nβ‰₯0(X_{\tau+n})_{n \geq 0} is a Markov chain starting from ii, independent of X0,…,XΟ„X_0, \ldots, X_\tau.

The strong Markov property says that "restarting" the chain at a random (stopping) time preserves the Markov property. This is crucial for analyzing recurrence and transience.

RemarkApplication to recurrence

Using the strong Markov property, we can decompose long-run behavior into i.i.d. excursions between visits to a fixed state. This leads to the classification of states into recurrent and transient states (see Theorem 1.3 and related content).


Irreducibility and aperiodicity

Definition1.4Irreducibility

A Markov chain is irreducible if for every pair of states i,j∈Si, j \in S, there exists nβ‰₯0n \geq 0 such that pij(n)>0p_{ij}^{(n)} > 0. In other words, every state is accessible from every other state.

Definition1.5Period

The period of a state ii is

d(i)=gcd⁑{nβ‰₯1:pii(n)>0}.d(i) = \gcd\{n \geq 1 : p_{ii}^{(n)} > 0\}.

State ii is aperiodic if d(i)=1d(i) = 1. A chain is aperiodic if all states are aperiodic.

ExamplePeriodic random walk

The simple symmetric random walk on Z\mathbb{Z} has period 2: starting from the origin, XnX_n is even if nn is even and odd if nn is odd. Hence p00(n)=0p_{00}^{(n)} = 0 for all odd nn.

RemarkIrreducibility + aperiodicity

For an irreducible, aperiodic Markov chain on a finite state space, the distribution P(Xn=j∣X0=i)\mathbb{P}(X_n = j \mid X_0 = i) converges to a unique stationary distribution Ο€\pi as nβ†’βˆžn \to \infty, regardless of the initial state ii. This is the content of the convergence theorem (see Theorem 1.4).


Summary

The Markov property provides a mathematically tractable framework for modeling random evolution:

  • Memorylessness: Future depends only on present, not past.
  • Chapman-Kolmogorov: nn-step transitions multiply as matrices.
  • Strong Markov property: Restarting at a stopping time preserves the Markov property.
  • Irreducibility and aperiodicity: Ensure long-run convergence to equilibrium.

These foundational concepts underpin the entire theory of Markov chains, from finite-state models to general state space processes.