STAT 333: Applied Probability

Steve Drakich

Estimated study time: 1 hr 46 min

Table of contents

Introduction

STAT 333, Applied Probability, is a third-year course at the University of Waterloo taught in Spring 2020 by Professor Steve Drakich. Despite its title, the course is more theoretical than strictly applied—at its core, it is an introduction to stochastic processes, with particular emphasis on processes that possess a regenerative or memoryless character.

The prerequisite is STAT 230 (introductory probability) with a grade of at least 60%, together with solid backgrounds in calculus (differentiation, integration, limits, series, exponential functions) and linear algebra (solving systems of linear equations, matrix multiplication, determinants, inverses). The course builds cumulatively on all of this background.

The four chapters of the course are:

  1. Chapter 1 — Review of elementary probability theory.
  2. Chapter 2 — Conditional distributions and conditional expectations.
  3. Chapter 3 — Discrete-time Markov chains (the largest chapter).
  4. Chapter 4 — The exponential distribution and the Poisson process.

Chapter 1: Review of Elementary Probability

1.1 Probability Functions and Axioms

A probability function (or probability measure) \(P\) assigns a number to each event in a sample space \(\Omega\) and must satisfy three defining conditions:

  1. \(P(A) \geq 0\) for every event \(A\).
  2. \(P(\Omega) = 1\).
  3. For mutually exclusive events \(A_1, A_2, \ldots\), \(P\!\left(\bigcup_{i} A_i\right) = \sum_i P(A_i)\).
\[ P(A^c) = 1 - P(A). \]\[ P(A \mid B) = \frac{P(A \cap B)}{P(B)}. \]\[ P(A \cap B) = P(A \mid B)\, P(B). \]\[ P(A_1 \cap A_2 \cap \cdots \cap A_n) = P(A_1)\, P(A_2 \mid A_1)\, P(A_3 \mid A_1 \cap A_2) \cdots \]

Law of Total Probability

\[ P(A) = \sum_{i=1}^{n} P(A \mid B_i)\, P(B_i). \]

This expresses \(P(A)\) as a weighted average of conditional probabilities.

Bayes’ Theorem

\[ P(B_k \mid A) = \frac{P(A \mid B_k)\, P(B_k)}{\sum_{i=1}^{n} P(A \mid B_i)\, P(B_i)}. \]

1.2 Random Variables and Their Distributions

A random variable (RV) \(X\) is a function from the sample space to the real numbers. The two main types are:

  • Discrete — \(X\) takes on a finite or countably infinite set of values.
  • Continuous — \(X\) takes on values in a continuum.

Discrete Random Variables

\[ p(x) = P(X = x). \]\[ F(a) = P(X \leq a) = \sum_{x \leq a} p(x), \]\[ \bar{F}(a) = P(X > a) = 1 - F(a). \]

Special Discrete Distributions

The course reviews the following standard discrete distributions:

DistributionPMF \(p(x)\)MeanVariance
Bernoulli\((p)\)\(p^x(1-p)^{1-x}\), \(x\in\{0,1\}\)\(p\)\(p(1-p)\)
Binomial\((n,p)\)\(\binom{n}{x}p^x(1-p)^{n-x}\), \(x=0,\ldots,n\)\(np\)\(np(1-p)\)
Negative Binomial\((k,p)\)\(\binom{x-1}{k-1}p^k(1-p)^{x-k}\), \(x=k,k+1,\ldots\)\(k/p\)\(k(1-p)/p^2\)
Geometric\((p)\)\((1-p)^{x-1}p\), \(x=1,2,\ldots\)\(1/p\)\((1-p)/p^2\)
Poisson\((\lambda)\)\(e^{-\lambda}\lambda^x/x!\), \(x=0,1,2,\ldots\)\(\lambda\)\(\lambda\)
Hypergeometric\((N,r,n)\)\(\binom{r}{x}\binom{N-r}{n-x}/\binom{N}{n}\)\(nr/N\)

The table above shows the PMF, mean, and variance for each distribution. Two additional notes from the slides:

  • Discrete Uniform \(\text{DU}(a, b)\): \(p(x) = \frac{1}{b-a+1}\) for \(x = a, a+1, \ldots, b\), where \(a \leq b\) are integers.
  • Recovering the PMF from the CDF: If \(X\) takes values \(a_1 < a_2 < a_3 < \cdots\) with \(p(a_i) > 0\), then \(p(a_1) = F(a_1)\) and \(p(a_i) = F(a_i) - F(a_{i-1})\) for \(i \geq 2\).

The Poisson as a Limiting Binomial

\[ P(X = x) = \binom{n}{x}\left(\frac{\lambda}{n}\right)^x\!\left(1 - \frac{\lambda}{n}\right)^{n-x} \;\xrightarrow{n\to\infty}\; \frac{e^{-\lambda}\lambda^x}{x!}, \]

which is the Poisson PMF. The key step uses the limit definition of the exponential function, \(e^z = \lim_{n\to\infty}(1 + z/n)^n\).

Continuous Random Variables

\[ F(a) = \int_{-\infty}^{a} f(x)\,dx, \]

and the PDF is recovered by differentiation: \(f(x) = F'(x)\).

Common continuous distributions include the Uniform\((a,b)\), Beta\((m,n)\), Gamma (Erlang), Exponential, and Normal distributions. The Exponential\((\lambda)\) is obtained from the Erlang by taking \(n=1\), giving PDF \(f(x) = \lambda e^{-\lambda x}\) for \(x > 0\).

Special Continuous Distributions (summary)

Uniform \(U(a,b)\): PDF \(f(x) = \frac{1}{b-a}\) for \(a < x < b\).

\[ f(x) = \frac{(m+n-1)!}{(m-1)!(n-1)!}\, x^{m-1}(1-x)^{n-1}, \quad 0 < x < 1. \]\[ f(x) = \frac{\lambda^n x^{n-1} e^{-\lambda x}}{(n-1)!}, \quad x > 0, \]\[ F(x) = 1 - e^{-\lambda x} \sum_{j=0}^{n-1} \frac{(\lambda x)^j}{j!}, \quad x \geq 0, \]

mean \(n/\lambda\), variance \(n/\lambda^2\), and MGF \(\phi_X(t) = \bigl(\frac{\lambda}{\lambda - t}\bigr)^n\) for \(t < \lambda\). The Erlang\((n,\lambda)\) distribution equals the distribution of the sum of \(n\) iid \(\text{Exp}(\lambda)\) random variables, which follows immediately from the product form of the MGF.

1.3 Expectation and Moments

\[ E[g(X)] = \begin{cases} \sum_x g(x)\,p(x) & \text{discrete},\\[4pt] \int_{-\infty}^{\infty} g(x)\,f(x)\,dx & \text{continuous}. \end{cases} \]

Special cases:

  • Setting \(g(x) = x^n\) gives the \(n\)-th moment \(E[X^n]\). When \(n=1\), this is the mean \(\mu = E[X]\).
  • Setting \(g(x) = (x - \mu)^2\) gives the variance: \(\text{Var}(X) = E[X^2] - (E[X])^2\).
  • For a linear function \(aX + b\): \(E[aX+b] = aE[X]+b\) and \(\text{Var}(aX+b) = a^2\text{Var}(X)\).
  • For a linear combination \(aX + bY\): \(E[aX+bY] = aE[X] + bE[Y]\) and \[ \text{Var}(aX + bY) = a^2\text{Var}(X) + b^2\text{Var}(Y) + 2ab\,\text{Cov}(X,Y). \] The covariance satisfies \(\text{Cov}(X,Y) = E[XY] - E[X]E[Y]\) and \(\text{Cov}(X,X) = \text{Var}(X)\).

Moment-Generating Functions

The moment-generating function (MGF) of an RV \(X\) is defined as \[ \phi_X(t) = E\!\left[e^{tX}\right], \] provided this expectation is finite in some open interval containing \(t = 0\).

Key properties of the MGF:

  1. Moment extraction: \(E[X^n] = \phi_X^{(n)}(0)\), the \(n\)-th derivative of \(\phi_X\) evaluated at \(t=0\).
  2. Uniqueness: If two RVs have the same MGF, they have the same distribution.
  3. Taylor series representation: Expanding \(e^{tX}\) in a Taylor series and taking expectations gives \[ \phi_X(t) = \sum_{n=0}^{\infty} \frac{E[X^n]}{n!}\, t^n. \]
\[ \phi_X(t) = (pe^t + 1 - p)^n. \]

Differentiating and evaluating at \(t=0\) yields \(E[X] = np\) and \(\text{Var}(X) = np(1-p)\), confirming the known formulas.

1.4 Joint Distributions

\[ p_X(x) = \sum_y p(x,y), \qquad p_Y(y) = \sum_x p(x,y). \]\[ P(X_1 = x_1, \ldots, X_k = x_k) = \frac{n!}{x_1!\cdots x_k!}\, p_1^{x_1}\cdots p_k^{x_k}. \]

Covariance and independence: \(\text{Cov}(X,Y) = E[XY] - E[X]E[Y]\). If \(X\) and \(Y\) are independent, then \(\text{Cov}(X,Y) = 0\), but the converse is not true in general.

Formal definition of independence. Random variables \(X\) and \(Y\) are independent if and only if \[ F(a,b) = F_X(a)\,F_Y(b) \quad \forall\, a, b \in \mathbb{R}, \] or equivalently, \(p(x,y) = p_X(x)p_Y(y)\) in the jointly discrete case, or \(f(x,y) = f_X(x)f_Y(y)\) in the jointly continuous case. An important consequence: for arbitrary functions \(g\) and \(h\), independence implies \(E[g(X)h(Y)] = E[g(X)]\,E[h(Y)]\).

Caution. Zero covariance (\(X\) and \(Y\) are uncorrelated) does not imply independence. Independence implies zero covariance, but not conversely.

\[ \phi(s,t) = E\!\left[e^{sX + tY}\right], \]

and joint moments are recovered via \(E[X^m Y^n] = \frac{\partial^{m+n}}{\partial s^m \partial t^n}\phi(s,t)\big|_{s=0,\,t=0}\).

Theorem 1.1 — MGF of a Sum

Theorem 1.1. If \(X_1, X_2, \ldots, X_n\) are independent RVs with respective MGFs \(\phi_{X_i}(t)\), and \(T = X_1 + X_2 + \cdots + X_n\), then \[ \phi_T(t) = \prod_{i=1}^{n} \phi_{X_i}(t). \]
\[ \phi_T(t) = E\!\left[e^{tT}\right] = E\!\left[\prod_{i=1}^n e^{tX_i}\right] = \prod_{i=1}^n E\!\left[e^{tX_i}\right] = \prod_{i=1}^n \phi_{X_i}(t). \]

Corollary. If \(X_1, \ldots, X_n\) are also identically distributed (i.e., iid), then \(\phi_T(t) = [\phi_{X_1}(t)]^n\).

Example. If \(X_1, \ldots, X_m\) are independent Binomials with the same \(p\) but different \(n_i\), then \(T = X_1 + \cdots + X_m\) has MGF \((pe^t + 1-p)^{n_1+\cdots+n_m}\), which is the MGF of a \(\text{Binomial}(n_1+\cdots+n_m, p)\). By the uniqueness property, \(T\) is Binomial with parameters \(\sum n_i\) and \(p\).

1.5 The Strong Law of Large Numbers

Strong Law of Large Numbers. If \(X_1, X_2, \ldots\) are iid with common (finite) mean \(\mu\), then \[ \bar{X}_n = \frac{X_1 + X_2 + \cdots + X_n}{n} \;\xrightarrow{\text{a.s.}}\; \mu \quad \text{as } n \to \infty. \]

In plain English: the sample mean converges to the true mean as the number of independent trials grows. This theorem underpins much of applied probability and statistics.


Chapter 2: Conditional Distributions and Conditional Expectations

The purpose of this chapter is to develop deep fluency with conditioning, because Markov chains (Chapter 3) and the Poisson process (Chapter 4) both rely on it constantly.

2.1 Conditional Distributions

Jointly Discrete Case

Given two discrete RVs \(X_1\) and \(X_2\) with joint PMF \(p(x_1, x_2)\), the conditional PMF of \(X_1\) given \(X_2 = x_2\) is \[ p_{X_1 \mid X_2}(x_1 \mid x_2) = \frac{p(x_1, x_2)}{p_{X_2}(x_2)}, \qquad p_{X_2}(x_2) > 0. \]

Remark. If \(X_1\) and \(X_2\) are independent, then \(p_{X_1 \mid X_2}(x_1 \mid x_2) = p_{X_1}(x_1)\). Conditioning on an independent variable has no effect.

Linearity of conditional expectation. For constants \(a_1, \ldots, a_n \in \mathbb{R}\), \[ E\!\left[\sum_{i=1}^n a_i X_i \,\Big|\, Y = y\right] = \sum_{i=1}^n a_i\, E[X_i \mid Y = y]. \] Moreover, if \(h(\cdot)\) is a deterministic function and \(g(\cdot)\) is arbitrary, then \[ E[g(X_1)\,h(X_2) \mid X_2 = x_2] = h(x_2)\, E[g(X_1) \mid X_2 = x_2]. \]
\[ p_{X_1, X_2 \mid X_3}(x_1, x_2 \mid x_3) = \frac{p(x_1, x_2, x_3)}{p_{X_3}(x_3)}. \]

Key Examples

\[ P(X_1 = x_1 \mid Y = m) = \frac{\binom{n_1}{x_1}\binom{n_2}{m-x_1}}{\binom{n_1+n_2}{m}}, \]

for \(x_1 = \max(0, m-n_2), \ldots, \min(n_1, m)\). This is a Hypergeometric\((n_1+n_2, n_1, m)\) distribution, which has a natural combinatorial explanation.

\[ X_j \mid (Y = n) \;\sim\; \text{Binomial}\!\left(n,\, \frac{\lambda_j}{\sum_i \lambda_i}\right). \]

Example 2.4 (Reversed conditioning). If \(X \sim \text{Poisson}(\lambda)\) and \(Y \mid X \sim \text{Binomial}(X, p)\), one can show that the marginal distribution of \(Y\) is \(\text{Poisson}(\lambda p)\). The conditional distribution of \(X\) given \(Y = y\) is then a shifted Poisson: \(X \mid Y=y\) has the same distribution as \(W + y\), where \(W \sim \text{Poisson}(\lambda(1-p))\).

Jointly Continuous Case

\[ f_{X \mid Y}(x \mid y) = \frac{f(x, y)}{f_Y(y)}, \]

where \(f(x,y)\) is the joint PDF and \(f_Y(y)\) is the marginal PDF of \(Y\).

Example. If the joint PDF is \(f(x,y) = e^{-x-y}\) on the region \(y > 2x > 0\), the marginal of \(X\) is \(f_X(x) = e^{-5x}\) (Exponential with rate 5), and the conditional \(Y \mid X=x\) is a shifted Exponential starting at \(2x\).

Mixed Case (One Discrete, One Continuous)

\[ f_{X \mid Y}(x \mid y) = \frac{P(Y=y \mid X=x)\, f_X(x)}{P(Y=y)}, \]

where the marginal \(P(Y=y) = \int_{-\infty}^{\infty} P(Y=y \mid X=x)\, f_X(x)\,dx\) plays the role of the law of total probability.

This formula is derived by forming the ratio \(P(x \leq X \leq x+dx \mid Y=y)/dx\) and letting \(dx \to 0\), using Bayes’ theorem at the level of densities. Symmetrically, the conditional PMF of \(Y\) given \(X = x\) satisfies \(p(y \mid x) = f(x \mid y)\, p_Y(y) / f_X(x)\), and since it must sum to 1, \(f_X(x) = \sum_y f(x \mid y)\, p_Y(y)\).

Example 2.7 (Uniform–Bernoulli mixed case). Suppose \(X \sim U(0,1)\) and \(Y \mid (X = x) \sim \text{Bern}(x)\). The marginal PMF of \(Y\) is \[ P(Y = y) = \int_0^1 x^y(1-x)^{1-y}\,dx = \frac{1}{2}, \quad y \in \{0, 1\}. \] The conditional PDF of \(X \mid (Y = 0)\) is \(f(x \mid 0) = 2(1-x)\) on \((0,1)\), and the conditional PDF of \(X \mid (Y = 1)\) is \(f(x \mid 1) = 2x\) on \((0,1)\). Both are Beta distributions.

2.2 Conditional Expectation

The conditional mean of \(X_1\) given \(X_2 = x_2\) is \[ E[X_1 \mid X_2 = x_2] = \sum_{x_1} x_1\, p_{X_1 \mid X_2}(x_1 \mid x_2) \quad (\text{discrete}), \] or the corresponding integral in the continuous case.
\[ E[aX_1 + bX_2 \mid X_3 = x_3] = a\, E[X_1 \mid X_3 = x_3] + b\, E[X_2 \mid X_3 = x_3]. \]

Conditional Variance

The conditional variance of \(X_1\) given \(X_2 = x_2\) is \[ \text{Var}(X_1 \mid X_2 = x_2) = E[X_1^2 \mid X_2 = x_2] - \bigl(E[X_1 \mid X_2 = x_2]\bigr)^2. \]

This formula (Theorem 2.1) mirrors the ordinary variance formula with conditional moments replacing unconditional ones.

2.3 The Law of Total Expectation

A crucial insight is that \(E[g(X_1) \mid X_2 = x_2]\), which is a function of \(x_2\), becomes a random variable when \(x_2\) is replaced by the random \(X_2\). We write \(E[g(X_1) \mid X_2]\) for this random variable.

Theorem 2.2 (Law of Total Expectation). For any two RVs \(X\) and \(Y\), \[ E[g(X)] = E\!\bigl[E[g(X) \mid Y]\bigr]. \]
\[ E\!\bigl[E[g(X) \mid Y]\bigr] = \int_{-\infty}^{\infty} \left(\int_{-\infty}^{\infty} g(x)\, f_{X\mid Y}(x\mid y)\,dx\right) f_Y(y)\,dy. \]

Since \(f_{X \mid Y}(x \mid y)\, f_Y(y) = f(x,y)\), swapping integration order gives \(\int\!\int g(x) f(x,y)\,dy\,dx = \int g(x) f_X(x)\,dx = E[g(X)]\). \(\square\)

\[ E[X] = E[X \mid Y=0](1-p) + E[X \mid Y=1]\cdot p. \]\[ E[X] = (1 + E[X])(1-p) + p. \]

Solving gives \(E[X] = 1/p\). The same technique yields \(\text{Var}(X) = (1-p)/p^2\), both well-known results derived here without computing any sums or MGFs.

\[ P(A) = \int_{-\infty}^{\infty} P(A \mid Y=y)\, f_Y(y)\,dy. \]\[ P(X < Y) = \int_{-\infty}^{\infty} F_X(y)\, f_Y(y)\,dy. \]\[ P(X < Y) = \frac{\lambda_1}{\lambda_1 + \lambda_2}. \]

This elegant result will reappear prominently in Chapter 4.

2.4 The Conditional Variance Formula and Random Sums

Theorem 2.3 (Conditional Variance Formula). For any RVs \(X\) and \(Y\), \[ \text{Var}(X) = E[\text{Var}(X \mid Y)] + \text{Var}(E[X \mid Y]). \]

This theorem decomposes the total variance into two components: the average within-group variance and the variance of the group means.

Random Sums

\[ T = \sum_{i=1}^{N} X_i. \]

Using the law of total expectation conditioned on \(N\):

\[ E[T] = E[N\mu] = \mu\, E[N]. \]\[ \text{Var}(T) = E[N\sigma^2] + \text{Var}(N\mu) = \sigma^2 E[N] + \mu^2\text{Var}(N). \]

This formula for the variance of a random sum is used repeatedly throughout the remainder of the course.

Further Extensions: Iterated Conditioning

The law of total expectation can be extended to a conditional setting. For any three random variables \(X\), \(W\), and \(Y\):

Iterated conditioning. Conditional expectation satisfies the same properties as unconditional expectation, with all distributions replaced by their conditional counterparts. In particular, \[ E[g(X) \mid Y = y] = \begin{cases} \displaystyle\sum_w E[g(X) \mid W=w, Y=y]\, p_{W\mid Y}(w \mid y) & (W \text{ discrete}),\\[6pt] \displaystyle\int_{-\infty}^{\infty} E[g(X) \mid W=w, Y=y]\, f_{W\mid Y}(w \mid y)\,dw & (W \text{ continuous}). \end{cases} \] In random-variable notation, \(E[g(X) \mid Y] = E[E[g(X) \mid W, Y] \mid Y]\). Applying the law of total expectation once more gives \[ E[g(X)] = E\!\bigl[E[E[g(X) \mid W, Y] \mid Y]\bigr]. \]

Expected Number of Trials for \(k\) Consecutive Successes

\[ E[N_k] = \frac{1 + p\cdot E[N_{k-1}]}{p} + (1-p)\cdot E[N_k]. \]\[ E[N_k] = \sum_{j=1}^{k} \frac{1}{p^j} = \frac{1}{p} + \frac{1}{p^2} + \cdots + \frac{1}{p^k}. \]

This beautiful closed form — a geometric series — is obtained purely through the power of conditioning.


Chapter 3: Discrete-Time Markov Chains

3.1 Definitions and Fundamental Properties

Stochastic Processes

A stochastic process is a family of random variables \(\{X_t : t \in \mathcal{T}\}\) indexed by a set \(\mathcal{T}\). The set \(\mathcal{T}\) is the index set, usually interpreted as time. When \(\mathcal{T}\) is discrete (e.g., \(\mathcal{T} = \{0, 1, 2, \ldots\}\)), the process is a discrete-time stochastic process, denoted \(\{X_n\}_{n \geq 0}\).

The Markov Property

A discrete-time stochastic process \(\{X_n\}_{n \geq 0}\) is a discrete-time Markov chain (DTMC) if: 1. Each \(X_n\) is a discrete RV taking values in the state space \(S \subseteq \{0, 1, 2, \ldots\}\). 2. The Markov property holds: for all \(n \geq 0\) and states \(x_0, \ldots, x_{n+1} \in S\), \[ P(X_{n+1} = x_{n+1} \mid X_n = x_n, X_{n-1} = x_{n-1}, \ldots, X_0 = x_0) = P(X_{n+1} = x_{n+1} \mid X_n = x_n). \]

The Markov property says the conditional distribution of the future, given the present and the past, depends only on the present. The past states \(X_0, \ldots, X_{n-1}\) are irrelevant once the present \(X_n\) is known.

An important consequence is that the Markov property holds even when some past information is missing: you do not need the full history to invoke the memoryless structure.

Transition Probability Matrices

\[ p_{ij} = P(X_{n+1} = j \mid X_n = i) \]\[ P = \begin{pmatrix} p_{00} & p_{01} & p_{02} & \cdots \\ p_{10} & p_{11} & p_{12} & \cdots \\ \vdots & & \ddots \end{pmatrix}. \]

Properties of \(P\):

  • All entries satisfy \(p_{ij} \geq 0\).
  • Each row sums to 1: \(\sum_j p_{ij} = 1\) for every state \(i\).

A matrix with non-negative entries whose rows all sum to 1 is called a stochastic matrix.

Weather example. Suppose weather is either Clear (0), Overcast (1), or Rainy (2), and transitions depend only on today’s weather. If the one-day transition probabilities are given, they form a \(3 \times 3\) stochastic matrix \(P\).

\(n\)-Step Transition Probabilities and the Chapman–Kolmogorov Equations

The \(n\)-step transition probability is \[ p_{ij}^{(n)} = P(X_{n+m} = j \mid X_m = i), \] which by stationarity is independent of \(m\). The \(n\)-step transition probabilities form the matrix \(P^{(n)}\).
\[ p_{ij}^{(n)} = \sum_k p_{ik}^{(m)}\, p_{kj}^{(n-m)}, \]\[ P^{(n)} = P^n \quad (\text{the matrix } P \text{ multiplied by itself } n \text{ times}). \]

Initial Distribution and Marginal Probabilities

\[ \boldsymbol{\alpha}_n = \boldsymbol{\alpha}_0\, P^n, \]

so the probability of being in state \(k\) at time \(n\) is the \(k\)-th entry of \(\boldsymbol{\alpha}_0 P^n\).

A DTMC is completely characterized by its one-step TPM \(P\) and initial distribution \(\boldsymbol{\alpha}_0\).

Communication Classes and Irreducibility

State \(j\) is accessible from state \(i\) (written \(i \to j\)) if there exists \(n \geq 0\) such that \(p_{ij}^{(n)} > 0\).

States \(i\) and \(j\) communicate (written \(i \leftrightarrow j\)) if both \(i \to j\) and \(j \to i\).

A DTMC is irreducible if all states communicate (i.e., there is only one communication class).

Communication is an equivalence relation (reflexive, symmetric, transitive), so it partitions the state space into communication classes. A practical tool for finding classes is the state transition diagram, where a directed arrow connects \(i\) to \(j\) whenever \(p_{ij} > 0\).

1State 12State 23State 30.40.30.30.50.50.70.3
Proof that communication is an equivalence relation.
  • Reflexivity: \(i \leftrightarrow i\) since \(p_{ii}^{(0)} = 1 > 0\).
  • Symmetry: immediate from the definition of \(\leftrightarrow\).
  • Transitivity: if \(i \leftrightarrow j\) and \(j \leftrightarrow k\), then there exist \(m, n \geq 0\) such that \(p_{ij}^{(n)} > 0\) and \(p_{jk}^{(m)} > 0\). By the Chapman–Kolmogorov equations, \(p_{ik}^{(n+m)} \geq p_{ij}^{(n)} p_{jk}^{(m)} > 0\), so \(i \to k\). The same argument gives \(k \to i\), hence \(i \leftrightarrow k\).

A DTMC is completely characterised by its one-step TPM \(P\) and the initial conditions \(\boldsymbol{\alpha}_0\). All finite-dimensional distributions can be computed as \(P(X_0 = x_0, X_1 = x_1, \ldots, X_n = x_n) = \alpha_{0,x_0} P_{x_0,x_1} P_{x_1,x_2} \cdots P_{x_{n-1},x_n}\).

Periodicity

The period of state \(i\) is \[ d(i) = \gcd\!\bigl\{n \geq 1 : p_{ii}^{(n)} > 0\bigr\}. \] State \(i\) is aperiodic if \(d(i) = 1\), and the chain is aperiodic if all states are.
Theorem 3.1. If states \(i\) and \(j\) communicate, then \(d(i) = d(j)\). Consequently, the period is the same throughout a communication class.

Proof sketch. Since \(i \leftrightarrow j\), there exist \(m, n \geq 1\) such that \(p_{ij}^{(m)} > 0\) and \(p_{ji}^{(n)} > 0\). Therefore \(p_{ii}^{(m+n)} \geq p_{ij}^{(m)} p_{ji}^{(n)} > 0\), so \(d(i)\) divides \(m+n\). For any \(r\) with \(p_{jj}^{(r)} > 0\), we have \(p_{ii}^{(m+r+n)} \geq p_{ij}^{(m)} p_{jj}^{(r)} p_{ji}^{(n)} > 0\), so \(d(i)\) divides \(m+r+n\) and hence divides \(r\). This shows \(d(i) \leq d(j)\). By symmetry \(d(j) \leq d(i)\), so \(d(i) = d(j)\). \(\square\)

Aperiodicity does not require positive diagonal entries in \(P\). Even with all diagonal entries equal to zero, it is possible to have \(d(i) = 1\) for all states (as long as there exist cycles of coprime lengths).

If a state \(i\) has \(p_{ii} > 0\), then \(1 \in \{n : p_{ii}^{(n)} > 0\}\), so the GCD is 1 and \(i\) is aperiodic.

3.2 Transience and Recurrence

The first-passage probability from \(i\) to \(j\) at time \(n\) is \[ f_{ij}^{(n)} = P(X_n = j,\; X_k \neq j \text{ for } k=1,\ldots,n-1 \mid X_0 = i). \] The total first-passage probability (probability of ever visiting \(j\) from \(i\)) is \[ f_{ij} = \sum_{n=1}^{\infty} f_{ij}^{(n)}. \]
\[ p_{ij}^{(n)} = \sum_{k=1}^{n} f_{ij}^{(k)}\, p_{jj}^{(n-k)}, \quad n \geq 1. \]\[ f_{ij}^{(n)} = p_{ij}^{(n)} - \sum_{k=1}^{n-1} f_{ij}^{(k)}\, p_{jj}^{(n-k)}, \quad n \geq 2, \]

with \(f_{ij}^{(1)} = p_{ij}\).

State \(i\) is recurrent if \(f_{ii} = 1\) (the chain returns to \(i\) with probability 1).

State \(i\) is transient if \(f_{ii} < 1\) (there is positive probability of never returning).

Recurrent class (closed)R₁R₂Transient statesT₁T₂absorbedescape \[ E[M_i \mid X_0 = i] = \sum_{n=1}^{\infty} p_{ii}^{(n)}. \]

Since \(M_i \mid (X_0 = i)\) is geometrically distributed with success probability \(1 - f_{ii}\) when \(f_{ii} < 1\), or equals \(\infty\) with probability 1 when \(f_{ii} = 1\):

  • State \(i\) is recurrent if and only if \(\displaystyle\sum_{n=1}^{\infty} p_{ii}^{(n)} = \infty\).
  • State \(i\) is transient if and only if \(\displaystyle\sum_{n=1}^{\infty} p_{ii}^{(n)} < \infty\), and in this case \(E[M_i \mid X_0 = i] = \frac{f_{ii}}{1 - f_{ii}} < \infty\).
Theorem 3.2. If \(i \leftrightarrow j\) and state \(i\) is recurrent, then state \(j\) is recurrent. Hence recurrence and transience are class properties.
Theorem 3.3. If \(i \leftrightarrow j\) and state \(i\) is recurrent, then \(f_{ij} = P(\text{DTMC ever visits } j \mid X_0 = i) = 1\).
Theorem 3.4. A finite-state DTMC has at least one recurrent state. If the chain is also irreducible, then all states are recurrent.

Proof sketch. By irreducibility, all states communicate and thus share the same recurrence status. If any state were transient, the chain could eventually leave every state, but a finite state space has nowhere else to go — a contradiction.

Theorem 3.5. If state \(i\) is recurrent and state \(i\) does not communicate with state \(j\), then \(p_{ij}^{(k)} = 0\) for all \(k \in \mathbb{Z}^+\). In other words, a recurrent class is closed: once the chain enters it, it can never leave.

Remark. Theorem 3.5 also provides a test for transience: if \(p_{ij} > 0\) and states \(i\) and \(j\) do not communicate, then state \(i\) must be transient.

Examples

The simple random walk on \(\mathbb{Z}\) (moving right with probability \(p\) and left with \(1-p\) at each step) is recurrent when \(p = 1/2\) and transient otherwise. In higher dimensions, a symmetric random walk is recurrent in 1D and 2D but transient in 3D and above.

3.3 Limiting and Stationary Distributions

The central question of this section: does \(p_{ij}^{(n)}\) converge to a limit as \(n \to \infty\)?

Theorem 3.6 (Limiting behaviour of transient states). For any state \(i\) and any transient state \(j\), \[ \lim_{n \to \infty} p_{ij}^{(n)} = 0. \]

This follows because \(\sum_{n=1}^\infty p_{jj}^{(n)} < \infty\) for transient \(j\), so \(p_{jj}^{(n)} \to 0\), and consequently \(p_{ij}^{(n)} \to 0\) as well.

Mean Return Time and Positive/Null Recurrence

Suppose state \(i\) is recurrent. Define the first return time to state \(i\) as \[ N_i = \min\{n \in \mathbb{Z}^+ : X_n = i\}, \] whose conditional PMF is \(P(N_i = n \mid X_0 = i) = f_{ii}^{(n)}\). The mean recurrent time of state \(i\) is \[ m_i = E[N_i \mid X_0 = i] = \sum_{n=1}^{\infty} n\, f_{ii}^{(n)}. \] State \(i\) is positive recurrent if \(m_i < \infty\), and null recurrent if \(m_i = \infty\).

Example (null recurrence is possible). The random variable \(X\) with PMF \(P(X = x) = \frac{1}{x(x+1)}\) for \(x = 1, 2, 3, \ldots\) is a valid probability distribution (the sum telescopes to 1), yet \(E[X] = \sum_{x=1}^\infty \frac{1}{x+1} = \infty\). This shows that a recurrent state can have an infinite mean return time.

Facts about positive and null recurrence:

  1. Positive recurrence is a class property: if \(i \leftrightarrow j\) and \(i\) is positive recurrent, so is \(j\) (and null recurrence is likewise a class property).
  2. In any finite-state DTMC, all recurrent states are positive recurrent (null recurrence cannot occur).
  3. Positive recurrent, aperiodic states are called ergodic states.

Example: Periodic Chain

If the chain oscillates between two forms of the TPM (for even and odd \(n\)), the limit does not exist for individual \(p_{ij}^{(n)}\). The period is the root cause: periodicity of 2 prevents convergence.

Example: Irreducible Aperiodic Finite Chain

\[ \lim_{n \to \infty} p_{ij}^{(n)} = \pi_j \quad \text{for all } i, j, \]

and crucially, the limit \(\pi_j\) is independent of the starting state \(i\). The limiting matrix has all rows identical.

A stationary distribution (also called invariant or steady-state distribution) is a row vector \(\boldsymbol{\pi} = (\pi_0, \pi_1, \ldots)\) satisfying:
  1. \(\boldsymbol{\pi} P = \boldsymbol{\pi}\), i.e., \(\pi_j = \sum_{i=0}^\infty \pi_i P_{i,j}\) for all \(j\) (balance equations).
  2. \(\sum_j \pi_j = 1\) and \(\pi_j \geq 0\) for all \(j\).

The system \(\boldsymbol{\pi} P = \boldsymbol{\pi}\), together with the normalization \(\sum_j \pi_j = 1\), forms a linear system that can be solved to find \(\boldsymbol{\pi}\). The first set of equations is linearly dependent (one is redundant), so in a finite chain with \(N+1\) states one drops any one balance equation and replaces it with the normalization condition.

Remarks.

  • A stationary distribution does not exist if all states are null recurrent or transient.
  • An irreducible DTMC is positive recurrent if and only if a stationary distribution exists.
  • Stationary distributions need not be unique; uniqueness fails when there is more than one positive recurrent communication class.
Basic Limit Theorem (BLT). For an irreducible, recurrent, and aperiodic DTMC, \[ \lim_{n \to \infty} p_{ij}^{(n)} = \pi_j = \frac{1}{m_j} \quad \text{for all } i, j, \] where \(m_j\) is the mean recurrent time of state \(j\). If the chain is also positive recurrent, \(\{\pi_j\}\) is the unique positive solution to the balance equations \(\boldsymbol{\pi} = \boldsymbol{\pi} P\), \(\sum_j \pi_j = 1\).

The BLT derivation follows by taking \(n \to \infty\) in the Chapman–Kolmogorov equations \(p_{ij}^{(n)} = \sum_k p_{ik}^{(n-1)} P_{k,j}\) (assuming the limit may pass through the sum), which immediately gives the balance equations for \(\pi_j\).

\[ \frac{1}{n}\sum_{k=1}^n A_k \to \pi_j \quad \text{as } n \to \infty \]

in the sense that the expected fraction converges to \(\pi_j\). This follows because \(E\!\left[\frac{1}{n}\sum_{k=1}^n A_k \mid X_0 = i\right] = \frac{1}{n}\sum_{k=1}^n p_{ij}^{(k)} \to \pi_j\) by the Cesàro mean property. The Strong Law of Large Numbers gives the almost-sure version: starting from state \(j\), the inter-visit times \(\{N_j^{(n)}\}_{n \geq 1}\) are iid with common mean \(m_j\), so by the SLLN the long-run fraction equals \(1/m_j = \pi_j\).

Theorem 3.7 (Doubly stochastic TPM). Suppose a finite-state DTMC with \(N\) states is irreducible and aperiodic. If its TPM \(P\) is doubly stochastic (i.e., all column sums also equal 1: \(\sum_{i} P_{i,j} = 1\) for all \(j\)), then the unique stationary distribution is the uniform distribution: \[ \pi_j = \frac{1}{N}, \quad j = 0, 1, \ldots, N-1. \]

Proof. The uniform vector \(\pi_j = 1/N\) satisfies \(\sum_j \pi_j = 1\) and, using the doubly stochastic property, \(\sum_i \pi_i P_{i,j} = \frac{1}{N}\sum_i P_{i,j} = \frac{1}{N} \cdot 1 = \pi_j\). By the BLT, the stationary distribution is unique, so \(\pi_j = 1/N\). \(\square\)

3.4 Applications

3.4.1 The Galton–Watson Branching Process

The Galton–Watson branching process models population growth over discrete generations. Let \(X_n\) denote the size of generation \(n\). Each individual in generation \(n\) independently produces offspring according to a common distribution \(\{\alpha_m\}_{m \geq 0}\) (where \(\alpha_m = P(\text{number of offspring} = m)\)).

Gen 0Gen 1Gen 2Gen 3Z₀Z₁Z₂Z₃X₀=1X₁=3X₂=5 \[ X_{n+1} = \sum_{i=1}^{X_n} Z_i^{(n)}, \]

where \(Z_1^{(n)}, Z_2^{(n)}, \ldots\) are iid with PMF \(\{\alpha_m\}\). Since \(X_{n+1}\) depends only on \(X_n\), this is a DTMC. The state \(0\) (extinction) is absorbing and recurrent; all positive states are transient.

Let \(\mu = E[\text{offspring per individual}]\) and \(\sigma^2 = \text{Var}(\text{offspring per individual})\). Assuming \(X_0 = 1\):

Mean and variance of the branching process. For \(n \geq 1\), \[ E[X_n] = \mu^n, \] \[ \text{Var}(X_n) = \begin{cases} n\sigma^2 & \text{if } \mu = 1, \\[4pt] \sigma^2 \mu^{n-1} \dfrac{1 - \mu^n}{1 - \mu} & \text{if } \mu \neq 1. \end{cases} \]

These follow by the random sum formulas from Chapter 2: \(E[X_n] = \mu E[X_{n-1}]\) and \(\text{Var}(X_n) = \sigma^2 E[X_{n-1}] + \mu^2 \text{Var}(X_{n-1})\), applied recursively with \(E[X_0] = 1\) and \(\text{Var}(X_0) = 0\).

Let \(q = P(\text{population eventually becomes extinct})\). Then:

  • If \(\mu \leq 1\): \(q = 1\) (certain extinction).
  • If \(\mu > 1\): \(q < 1\) (positive probability of survival).
\[ \tilde{\alpha}(z) = \sum_{j=0}^{\infty} \alpha_j z^j \]

is the probability generating function (PGF) of the offspring distribution. The key properties of \(\tilde{\alpha}(z)\) on \([0,1]\) are: (i) \(\tilde{\alpha}(0) = \alpha_0 > 0\), (ii) \(\tilde{\alpha}(1) = 1\), (iii) \(\tilde{\alpha}'(z) > 0\) (increasing), and (iv) \(\tilde{\alpha}''(z) > 0\) (convex), since \(\alpha_0 + \alpha_1 < 1\). This means:

  • When \(\mu = \tilde{\alpha}'(1) = 1\) (critical case): \(z = 1\) is the only solution in \([0,1]\), so \(q = 1\).
  • When \(\mu > 1\) (supercritical case): there is a second solution \(z_0 \in (0,1)\). Since every non-negative solution \(z^*\) satisfies \(z^* \geq P(X_n = 0)\) for all \(n\), taking \(n \to \infty\) gives \(z^* \geq q\). Hence \(q = z_0\) is the smallest fixed point of \(\tilde{\alpha}\) in \([0,1]\).

Remark. If \(X_0 = n\), the extinction probability is \(q^n\), since each of the \(n\) initial individuals starts an independent family that must separately die out.

3.4.2 The Gambler’s Ruin Problem

\[ u_i = p\, u_{i+1} + q\, u_{i-1}, \quad i = 1, \ldots, N-1, \]\[ u_i = \begin{cases} 1 - \dfrac{i}{N} & \text{if } p = 1/2,\\[8pt] \dfrac{(q/p)^i - (q/p)^N}{1 - (q/p)^N} & \text{if } p \neq 1/2. \end{cases} \]\[ P(i) = \begin{cases} \dfrac{i}{N} & \text{if } p = 1/2,\\[8pt] \dfrac{1 - (q/p)^i}{1 - (q/p)^N} & \text{if } p \neq 1/2. \end{cases} \]

As \(N \to \infty\) (the gambler faces an infinitely rich adversary):

  • \(p = 1/2\): \(P(i) \to 0\) (certain ruin regardless of initial fortune).
  • \(p < 1/2\): \(P(i) \to 0\) (certain ruin, since \(q/p > 1\)).
  • \(p > 1/2\): \(P(i) \to 1 - (q/p)^i > 0\) (positive probability of unlimited wealth).

Only when the gambler has a genuine advantage (\(p > 1/2\)) is indefinite play viable.

3.5 Absorbing Markov Chains

A state is absorbing if \(p_{jj} = 1\) (once entered, never left). An absorbing Markov chain has at least one absorbing state and all non-absorbing (transient) states eventually reach an absorbing state with probability 1.

Structure of the TPM

\[ P = \begin{pmatrix} Q & R \\ \mathbf{0} & I \end{pmatrix}, \]

where:

  • \(Q\) (\(M \times M\)) — transitions among transient states.
  • \(R\) (\(M \times (N-M+1)\)) — transitions from transient to absorbing states.
  • \(\mathbf{0}\) — zero matrix (absorbing states cannot go transient).
  • \(I\) — identity matrix (absorbing states stay put).

Absorption Probabilities

\[ U_{i,k} = R_{i,k} + \sum_{j=0}^{M-1} Q_{i,j}\, U_{j,k}, \]\[ \mathbf{U} = (I - Q)^{-1} R. \]

The matrix \(I - Q\) is invertible (because \(Q\) governs only transient states, so \(Q^n \to 0\)).

The Fundamental Matrix

The fundamental matrix of an absorbing chain is \[ S = (I - Q)^{-1} = \sum_{n=0}^{\infty} Q^n. \] The \((i,l)\)-entry \(S_{i,l}\) equals the expected number of times the chain visits transient state \(l\) (including time 0 if \(i = l\)) before absorption, given \(X_0 = i\): \[ S_{i,l} = E\!\left[\sum_{n=0}^{T-1} \mathbf{1}[X_n = l] \,\Big|\, X_0 = i\right], \] where \(T = \min\{n \geq 1 : X_n \in \text{absorbing states}\}\) is the absorption time.

The matrix identity \(\sum_{n=0}^{N-1} Q^n = (I-Q)^{-1}(I - Q^N) \to (I-Q)^{-1}\) as \(N \to \infty\) (since \(Q^N \to 0\)) gives the geometric series for matrices: \(\sum_{n=0}^\infty Q^n = (I-Q)^{-1}\).

Expected Absorption Times and Visitation Probabilities

\[ \mathbf{w} = S\,\mathbf{1} = (I-Q)^{-1}\mathbf{1}, \]

where \(\mathbf{1}\) is a column vector of ones. This follows because \(w_i = \sum_l S_{i,l}\) (sum of expected visits across all transient states).

\[ f_{i,l} = \frac{S_{i,l}}{S_{l,l}}, \quad i \neq l, \]\[ f_{i,i} = 1 - \frac{1}{S_{i,i}}. \]

Derivation. By decomposing on whether state \(l\) is ever visited, \(S_{i,l} = \delta_{i,l} + f_{i,l}\, S_{l,l}\), rearranging gives the above formula. Note also that \(E[M_i \mid X_0 = i] = S_{i,i} - 1\) (expected number of future visits to transient state \(i\)), consistent with the geometric distribution of return visits.


Chapter 4: The Exponential Distribution and the Poisson Process

4.1 The Exponential Distribution — Key Properties

A random variable \(X\) has an exponential distribution with rate parameter \(\lambda > 0\), written \(X \sim \text{Exp}(\lambda)\), if its PDF is \[ f(x) = \lambda e^{-\lambda x}, \quad x > 0. \]

Standard facts:

QuantityFormula
CDF\(F(x) = 1 - e^{-\lambda x}\)
Tail probability\(\bar{F}(x) = e^{-\lambda x}\)
Mean\(E[X] = 1/\lambda\)
Variance\(\text{Var}(X) = 1/\lambda^2\)
MGF\(\phi_X(t) = \lambda/(\lambda - t)\), for \(t < \lambda\)

The parameter \(\lambda\) is often called the rate of the exponential distribution (it equals the reciprocal of the mean).

The Memoryless Property

The exponential distribution is the unique continuous distribution with the memoryless property: \[ P(X > s + t \mid X > s) = P(X > t) \quad \text{for all } s, t \geq 0. \]

Proof that Exp(\(\lambda\)) is memoryless. \(P(X > s+t \mid X > s) = e^{-\lambda(s+t)}/e^{-\lambda s} = e^{-\lambda t} = P(X > t)\). \(\square\)

Proof of uniqueness. An equivalent form of the memoryless property is \(\bar{F}(y+z) = \bar{F}(y)\bar{F}(z)\) for all \(y, z \geq 0\) (where \(\bar{F}(x) = P(X > x)\)). Starting from this, one shows \(\bar{F}(m/n) = \bar{F}(1)^{m/n}\) for all rationals \(m/n\), and by continuity \(\bar{F}(x) = \bar{F}(1)^x = e^{-\lambda x}\) for all \(x \geq 0\), where \(\lambda = -\ln \bar{F}(1) > 0\). This is precisely the exponential CDF.

This property means the past waiting time provides no information about the remaining waiting time. The geometric distribution is the unique discrete distribution with the memoryless property.

Generalized memoryless property. If \(X \sim \text{Exp}(\lambda)\) and \(Y, Z\) are independent non-negative random variables (also independent of \(X\)), then \[ P(X > Y + Z \mid X > Y) = P(X > Z). \] As a special case, \((X - Y) \mid (X > Y) \sim \text{Exp}(\lambda)\) regardless of the distribution of \(Y\).

Proof sketch. Write \(P(X > Y + Z, X > Y) = \int_0^\infty \int_0^\infty e^{-\lambda(y+z)} f_Z(z)\,dz\, f_Y(y)\,dy = P(X > Z) \cdot P(X > Y)\), so dividing by \(P(X > Y)\) gives the result.

Conditional independence after competition. Let \(X_1 \sim \text{Exp}(\lambda_1)\) and \(X_2 \sim \text{Exp}(\lambda_2)\) be independent. Given \(X_1 < X_2\), the random variables \(X_1\) and \(X_2 - X_1\) are conditionally independent, with \(X_1 \mid (X_1 < X_2) \sim \text{Exp}(\lambda_1 + \lambda_2)\) and \((X_2 - X_1) \mid (X_1 < X_2) \sim \text{Exp}(\lambda_2)\).

Minimum of Independent Exponentials

\[ Y = \min(X_1, \ldots, X_n) \sim \text{Exp}(\lambda_1 + \lambda_2 + \cdots + \lambda_n). \]\[ P(Y > y) = P(X_1 > y, \ldots, X_n > y) = \prod_{i=1}^n e^{-\lambda_i y} = e^{-(\sum \lambda_i)y}. \]

This is the tail probability of \(\text{Exp}(\sum_i \lambda_i)\).

Which Exponential is Smallest?

\[ P(X_j = \min_i X_i) = \frac{\lambda_j}{\lambda_1 + \lambda_2 + \cdots + \lambda_n}. \]\[ P(X_j < X_i \text{ for all } i \neq j \mid X_j = x) = \prod_{i \neq j} e^{-\lambda_i x}. \]

Integrating over \(x > 0\) against \(\lambda_j e^{-\lambda_j x}\,dx\) gives the stated ratio. This result generalizes the earlier formula \(P(X_1 < X_2) = \lambda_1/(\lambda_1+\lambda_2)\) from Chapter 2.

Competing Exponentials and the Future Minimum

A beautiful consequence: if \(X_j\) is the minimum, then the remaining minimum of the other \(n-1\) exponentials is \(\text{Exp}(\sum_{i \neq j} \lambda_i)\) and is independent of which \(X_j\) was the minimum. This allows very clean analysis of priority queues and similar systems.

Order Statistics and the Erlang Distribution

\[ f(x) = \frac{\lambda^n x^{n-1} e^{-\lambda x}}{(n-1)!}, \quad x > 0, \]\[ F(x) = 1 - e^{-\lambda x} \sum_{j=0}^{n-1} \frac{(\lambda x)^j}{j!}, \quad x \geq 0, \]

mean \(n/\lambda\), variance \(n/\lambda^2\), and MGF \(\phi_X(t) = \bigl(\frac{\lambda}{\lambda-t}\bigr)^n\) for \(t < \lambda\). The MGF is the \(n\)-fold product of Exp\((\lambda)\) MGFs, confirming the sum-of-exponentials representation.

Probability of an ordered sequence of independent exponentials. Let \(X_1, \ldots, X_n\) be independent with \(X_i \sim \text{Exp}(\lambda_i)\). Then \[ P(X_1 < X_2 < \cdots < X_n) = \prod_{i=1}^{n-1} \frac{\lambda_i}{\lambda_i + \lambda_{i+1} + \cdots + \lambda_n}. \]

4.2 The Poisson Process

Counting Processes

A counting process \(\{N(t) : t \geq 0\}\) satisfies: 1. \(N(0) = 0\). 2. \(N(t) \in \{0, 1, 2, \ldots\}\) for all \(t \geq 0\). 3. \(N(t)\) is non-decreasing: \(s < t \Rightarrow N(s) \leq N(t)\). 4. \(N(t) - N(s)\) counts the number of events in the interval \((s, t]\).

A counting process has independent increments if, for non-overlapping intervals, the counts are independent RVs. It has stationary increments if the distribution of \(N(s+t) - N(s)\) depends only on the length \(t\) and not on the starting time \(s\).

Definition of the Poisson Process

A counting process \(\{N(t)\}_{t \geq 0}\) is a Poisson process with rate \(\lambda > 0\) if: 1. \(N(0) = 0\). 2. The process has independent and stationary increments. 3. For small \(h > 0\): \(P(N(h) = 1) = \lambda h + o(h)\) and \(P(N(h) \geq 2) = o(h)\).

The third condition says that in a very short interval of length \(h\), the probability of exactly one event is approximately \(\lambda h\), and the probability of two or more events is negligible.

\[ N(t) \sim \text{Poisson}(\lambda t), \]

i.e., \(P(N(t) = n) = \frac{(\lambda t)^n e^{-\lambda t}}{n!}\). More generally, \(N(s+t) - N(s) \sim \text{Poisson}(\lambda t)\) for any \(s \geq 0\).

Proof sketch. Let \(\phi(t) = E[e^{u N(t)}]\) be the MGF of \(N(t)\) in the parameter \(u\). Using independent increments and the infinitesimal conditions, one derives the ODE \(\phi'(t) = \lambda(e^u - 1)\phi(t)\) with \(\phi(0) = 1\). The solution is \(\phi(t) = e^{\lambda t(e^u - 1)}\), which is the MGF of \(\text{Poisson}(\lambda t)\).

Poisson process: inter-arrival times and counting process N(t)

Inter-Arrival and Arrival Times

Define:

  • Inter-arrival times \(T_n = S_n - S_{n-1}\), where \(S_n\) is the time of the \(n\)-th event and \(S_0 = 0\).
  • Arrival time of the \(n\)-th event: \(S_n = T_1 + T_2 + \cdots + T_n\).

Theorem 4.4. The inter-arrival times \(T_1, T_2, \ldots\) of a Poisson process with rate \(\lambda\) are iid \(\text{Exp}(\lambda)\) random variables.

Proof. The distribution of \(N(t)\) gives \(P(T_1 > t) = P(N(t) = 0) = e^{-\lambda t}\), so \(T_1 \sim \text{Exp}(\lambda)\). By the stationary and independent increments properties, the same holds for all subsequent inter-arrival times, and they are mutually independent.

Corollary. The \(n\)-th arrival time \(S_n = T_1 + \cdots + T_n\) has an Erlang\((n, \lambda)\) distribution.

4.3 Further Properties of the Poisson Process

Conditional Distribution of Arrivals

\[ N(s) \mid N(t) = n \;\sim\; \text{Binomial}\!\left(n,\, \frac{s}{t}\right). \]\[ P(N(s) = m \mid N(t) = n) = \frac{P(N(s) = m)\,P(N(t) - N(s) = n-m)}{P(N(t) = n)}. \]

Using the Poisson PMFs and the independent increments property, after simplification (the exponential terms cancel), one obtains the Binomial PMF with parameters \(n\) and \(s/t\).

Conditional Distribution of Individual Arrival Times

Theorem 4.5 tells us about the count \(N(s)\) given \(N(t)=n\). A deeper question is: given that exactly one or \(n\) arrivals occurred in \([0,t]\), where in the interval did they occur? The next two theorems answer this.

Theorem 4.6. Let \(\{N(t)\}_{t \geq 0}\) be a Poisson process with rate \(\lambda\). Given \(N(t) = 1\), the first arrival time satisfies \[ S_1 \mid N(t) = 1 \;\sim\; \text{Uniform}(0, t). \]
\[ G(s) = \frac{P(S_1 \leq s,\; N(t) = 1)}{P(N(t) = 1)} = \frac{P(N(s) = 1)\,P(N(t) - N(s) = 0)}{P(N(t) = 1)}. \]\[ G(s) = \frac{(\lambda s\, e^{-\lambda s})(e^{-\lambda(t-s)})}{\lambda t\, e^{-\lambda t}} = \frac{\lambda s\, e^{-\lambda t}}{\lambda t\, e^{-\lambda t}} = \frac{s}{t}. \]

Since \(s/t\) is the CDF of \(\text{Uniform}(0,t)\), the result follows. \(\square\)

Intuition. The result says that, conditional on exactly one event occurring in \([0,t]\), the event time is uniformly scattered throughout the interval — the Poisson process has no preferred location.

Order Statistics Background

To state the generalisation, we need a brief introduction to order statistics.

Order statistics. Given \(n\) i.i.d. continuous RVs \(Y_1, Y_2, \ldots, Y_n\) with common CDF \(F\) and PDF \(f\), the order statistics \(Y_{(1)} \leq Y_{(2)} \leq \cdots \leq Y_{(n)}\) are defined by sorting: \(Y_{(k)}\) is the \(k\)-th smallest value among \(Y_1, \ldots, Y_n\).
\[ g(y_1, y_2, \ldots, y_n) = n!\, f(y_1)\, f(y_2) \cdots f(y_n), \qquad 0 < y_1 < y_2 < \cdots < y_n. \]

The factor \(n!\) accounts for the number of ways to assign \(n\) unordered observations to the \(n\) ordered slots.

\[ g(y_1, \ldots, y_n) = \frac{n!}{t^n}, \qquad 0 < y_1 < y_2 < \cdots < y_n < t. \]\[ g_{(i)}(y) = \frac{n!}{(i-1)!\,(n-i)!} \cdot \frac{y^{i-1}(t-y)^{n-i}}{t^n}, \qquad 0 < y < t, \]

which is a scaled Beta distribution.

Theorem 4.7 (Arrival times as uniform order statistics). For a Poisson process with rate \(\lambda\), given \(N(t) = n\), the conditional joint distribution of the \(n\) arrival times \((S_1, S_2, \ldots, S_n)\) equals the joint distribution of the \(n\) order statistics from \(n\) i.i.d. \(\text{Uniform}(0,t)\) random variables. In particular, the conditional joint PDF is \[ f(s_1, \ldots, s_n \mid N(t) = n) = \frac{n!}{t^n}, \qquad 0 < s_1 < s_2 < \cdots < s_n < t. \]
\[ P\!\left(\bigcap_{k=1}^n \{N(s_{k-1}, s_k) = 1\} \cap \{N(s_n, t) = 0\}\right) = \prod_{k=1}^n \Bigl[\lambda(s_k - s_{k-1})\, e^{-\lambda(s_k - s_{k-1})}\Bigr] \cdot e^{-\lambda(t-s_n)}, \]\[ = \lambda^n \prod_{k=1}^n (s_k - s_{k-1}) \cdot e^{-\lambda t}. \]\[ f(s_1, \ldots, s_n \mid N(t) = n) = \frac{\lambda^n\, e^{-\lambda t}}{(\lambda t)^n e^{-\lambda t}/n!} \cdot \frac{\partial^n}{\partial s_1 \cdots \partial s_n}\!\left[\prod_{k=1}^n (s_k - s_{k-1})\right] = \frac{n!}{t^n}. \quad \square \]

Remark. The remarkable conclusion is that, given \(n\) events have occurred by time \(t\), the event locations are completely uninformative about the Poisson rate \(\lambda\) — they are simply \(n\) i.i.d. uniform points placed on \([0,t]\) and then sorted. Theorem 4.6 is the special case \(n = 1\).

Applications of Theorem 4.7

\[ T = \sum_{i=1}^{N(t)} e^{-\alpha S_i}. \]\[ E[T \mid N(t) = n] = n\, E[e^{-\alpha U}], \qquad U \sim \text{Uniform}(0,t). \]\[ \boxed{E[T] = E[N(t)]\, E[e^{-\alpha U}] = \lambda t \cdot \frac{1 - e^{-\alpha t}}{\alpha t} = \frac{\lambda(1 - e^{-\alpha t})}{\alpha}.} \]\[ \text{Var}(T \mid N(t) = n) = n\, \text{Var}(e^{-\alpha U}) = n(D - B^2). \]\[ \text{Var}(T) = E[N(t)](D - B^2) + B^2\,\text{Var}(N(t)) = \lambda t\, D = \frac{\lambda(1-e^{-2\alpha t})}{2\alpha}. \]

Example 4.B (Satellite launches). Satellites are launched according to a Poisson process at rate \(\lambda = 3\) per year. Given that exactly 2 launches occurred in the past year (\(t = 1\)), find the probability that the first launch was within the first 5 months and the second launch was before the final 2 months.

In terms of years: we want \(P(S_1 \leq 5/12,\; S_2 \leq 5/6 \mid N(1) = 2)\).

\[ P(S_1 \leq {\textstyle\tfrac{5}{12}},\; S_2 \leq {\textstyle\tfrac{5}{6}} \mid N(1) = 2) = \int_0^{5/12} \int_{s_1}^{5/6} 2\, ds_2\, ds_1 = \int_0^{5/12} 2\!\left(\tfrac{5}{6} - s_1\right) ds_1. \]

Evaluating: \(= 2\!\left[\tfrac{5}{6}s_1 - \tfrac{s_1^2}{2}\right]_0^{5/12} = 2\!\left(\tfrac{25}{72} - \tfrac{25}{288}\right) = 2 \cdot \tfrac{75}{288} = \dfrac{25}{48} \approx 0.521\).

Superposition and Thinning

Superposition. If \(N_1(t) \sim \text{Poisson}(\lambda_1)\) and \(N_2(t) \sim \text{Poisson}(\lambda_2)\) are independent, then \(N_1(t) + N_2(t) \sim \text{Poisson}(\lambda_1 + \lambda_2)\).

Thinning. If each event of a Poisson\((\lambda)\) process is independently classified as Type1 (probability \(p\)) or Type2 (probability \(1-p\)), then the Type1 events form a Poisson\((\lambda p)\) process and the Type2 events form a Poisson\((\lambda(1-p))\) process, and the two are independent.

Comparison of Two Poisson Processes

Given two independent Poisson processes with arrival times \(S_n^{(1)}\) and \(S_n^{(2)}\) (Erlang-\(n\) distributed with rates \(\lambda_1\) and \(\lambda_2\) respectively), the probability that the \(n\)-th event of process 1 occurs before the \(n\)-th event of process 2 can be expressed as a sum of binomial probabilities. This connects back to the result from Chapter 2 about \(P(X < Y)\) for competing exponentials.

4.4 Generalizations of the Poisson Process

4.4.1 The Non-Homogeneous Poisson Process

The standard Poisson process assumes a constant arrival rate \(\lambda\). In many real-world scenarios (customer arrivals, traffic flow, etc.), the rate varies with time.

A non-homogeneous (non-stationary) Poisson process \(\{N(t)\}_{t \geq 0}\) with rate function \(\lambda(t)\) satisfies: 1. \(N(0) = 0\). 2. Independent increments (but not stationary increments). 3. For small \(h > 0\): \(P(N(t+h) - N(t) = 1) = \lambda(t)h + o(h)\) and \(P(N(t+h) - N(t) \geq 2) = o(h)\).
\[ \int_{s_1}^{s_1+s_2} \lambda(\tau)\,d\tau = m(s_1 + s_2) - m(s_1), \]

where \(m(t) = \int_0^t \lambda(\tau)\,d\tau\) is the mean value function.

\[ \frac{\partial}{\partial s_2}\phi_u(s_1, s_2) = \lambda(s_1 + s_2)(e^u - 1)\,\phi_u(s_1, s_2), \]

with initial condition \(\phi_u(s_1, 0) = 1\). The solution is \(\phi_u(s_1, s_2) = \exp\!\bigl[(e^u - 1)\int_{s_1}^{s_1+s_2}\lambda(\tau)\,d\tau\bigr]\), which is the MGF of the stated Poisson distribution.

4.4.2 The Compound Poisson Process

A compound Poisson process is a process of the form \[ X(t) = \sum_{i=1}^{N(t)} Y_i, \] where \(N(t)\) is a Poisson process with rate \(\lambda\), and \(Y_1, Y_2, \ldots\) are iid RVs independent of \(N(t)\).
\[ E[X(t)] = \lambda t\, E[Y_1], \qquad \text{Var}(X(t)) = \lambda t\, E[Y_1^2]. \]\[ \phi_{X(t)}(u) = \exp\!\bigl\{\lambda t\,(\phi_{Y_1}(u) - 1)\bigr\}. \]

Compound Poisson processes arise throughout insurance mathematics (aggregate claims models), queuing theory, and mathematical finance.


Summary and Key Theorems

The following table summarises the major results from each chapter.

ChapterKey ResultStatement
1Theorem 1.1MGF of a sum of independent RVs is the product of the individual MGFs.
2Theorem 2.2Law of Total Expectation: \(E[g(X)] = E[E[g(X)\mid Y]]\).
2Theorem 2.3Conditional Variance Formula: \(\text{Var}(X) = E[\text{Var}(X\mid Y)] + \text{Var}(E[X\mid Y])\).
3Chapman–Kolmogorov\(n\)-step TPM equals \(P^n\); intermediate conditioning gives \(P^{(n)} = P^{(m)} P^{(n-m)}\).
3Theorem 3.1Communicating states have the same period.
3Theorem 3.2Recurrence/transience is a class property.
3Theorem 3.3If \(i \leftrightarrow j\) and \(i\) is recurrent, then \(f_{ij} = 1\).
3Theorem 3.4Finite irreducible chain is recurrent.
3Theorem 3.5A recurrent class is closed: \(p_{ij}^{(k)} = 0\) for all \(k\) if \(i\) is recurrent and \(i \not\leftrightarrow j\).
3Basic Limit TheoremIrreducible, positive recurrent, aperiodic chain: \(\lim_{n\to\infty} p_{ij}^{(n)} = \pi_j = 1/m_j\), independent of \(i\).
3Theorem 3.7Finite irreducible aperiodic chain with doubly stochastic TPM: \(\pi_j = 1/N\).
3Absorbing chainsAbsorption probabilities \(\mathbf{U} = (I-Q)^{-1}R\); mean absorption times \(\mathbf{w} = (I-Q)^{-1}\mathbf{1}\).
4Theorem 4.1 (Min of exponentials)Min of independent exponentials is exponential with rate = sum of rates.
4Memoryless uniquenessExponential is the unique continuous memoryless distribution.
4Theorem 4.4Inter-arrival times of a Poisson process are iid \(\text{Exp}(\lambda)\).
4Theorem 4.5Given \(N(t)=n\), the earlier count \(N(s)\) is \(\text{Binomial}(n, s/t)\).
4Theorem 4.6Given \(N(t)=1\), the first arrival time \(S_1 \sim \text{Uniform}(0,t)\).
4Theorem 4.7Given \(N(t)=n\), the joint distribution of \((S_1,\ldots,S_n)\) equals that of \(n\) order statistics from i.i.d. \(\text{Uniform}(0,t)\); conditional joint PDF \(= n!/t^n\).
4Theorem 4.8Non-homogeneous Poisson process: events in \((s_1, s_1+s_2]\) are Poisson with mean \(\int_{s_1}^{s_1+s_2}\lambda(\tau)d\tau\).

Appendix: Notation Reference

SymbolMeaning
\(P(A)\)Probability of event \(A\)
\(P(A \mid B)\)Conditional probability of \(A\) given \(B\)
\(E[X]\), \(\mu\)Expected value (mean) of \(X\)
\(\text{Var}(X)\), \(\sigma^2\)Variance of \(X\)
\(\phi_X(t)\)MGF of \(X\): \(E[e^{tX}]\)
\(F(x)\)CDF of \(X\): \(P(X \leq x)\)
\(\bar{F}(x)\)Tail probability: \(P(X > x)\)
\(p_{ij}^{(n)}\)\(n\)-step transition probability from state \(i\) to state \(j\)
\(f_{ij}^{(n)}\)First-passage probability from \(i\) to \(j\) at step \(n\)
\(d(i)\)Period of state \(i\)
\(\boldsymbol{\pi}\)Stationary distribution of a Markov chain
\(N(t)\)Counting process at time \(t\)
\(\lambda\)Rate parameter (exponential / Poisson)
\(S_n\)Arrival time of the \(n\)-th event in a Poisson process
\(T_n\)\(n\)-th inter-arrival time
Back to top