Martingales in Discrete Time

All the stochastic processes that we will study in this class will be generated from a sequence of random variables. The specific sequence used to generate the stochastic process \{M_{n}: 0 \leq n < \infty\} determines the information set \mathcal{F}_{n} available at time n. As time passes, we learn more about the processes and what was used to generate the process.

We call the collection of information sets \{\mathcal{F}_{n}: 0 \leq n < \infty\} a filtration. Since information accumulates over time, we always have that \mathcal{F}_{m} \subset \mathcal{F}_{n} for m < n.

Martingales

Earlier, we defined the simple random walk as the process \{S_{n}: 0 \leq n < \infty\}, built from a sequence \{X_{n}: 1 \leq n < \infty\} of independent random variables, where S_{n} = S_{0} + \sum_{i = 1}^{n} X_{i} for n \geq 1. Each X_{i} is a fair coin toss, taking values -1 or 1 with probability 1/2. At time n, the outcomes of the first n coin tosses are known, so the information set is \mathcal{F}_{n} = \{X_{1}, X_{2}, \ldots, X_{n}\}, containing all information revealed by X_{1}, X_{2}, \ldots, X_{n}.

We can then see that \operatorname{E}(S_{n + 1} \mid \mathcal{F}_{n}) = \operatorname{E}(S_{n} + X_{n + 1} \mid \mathcal{F}_{n}) = S_{n} + \operatorname{E}(X_{n + 1} \mid \mathcal{F}_{n}). There is an implicit assumption that the coin tosses are independent from each other. Intuitively, we do not expect to gain any information from the first n tosses to determine the next one. Thus, \operatorname{E}(X_{n + 1} \mid \mathcal{F}_{n}) = \operatorname{E}(X_{n + 1} = 0), implying \operatorname{E}(S_{n + 1} \mid \mathcal{F}_{n}) = S_{n}.

The process just defined is what we call a martingale. A martingale is a stochastic process \{M_{n}\} adapted to a filtration \{\mathcal{F}_{n}\} such that for all n, \operatorname{E}(M_{n+1} \mid \mathcal{F}_{n}) = M_{n}. In other words, given all the information up to time n, the expected value of the next observation is equal to the current value. This property captures the idea that, on average, the process does not “drift” up or down given the available information.

Example 1 We can generalize the simple random walk by defining \{X_{n}: 1 \leq n < \infty\} as a sequence of independent random variables such that \operatorname{E}(X_{n}) = 0 for all n \geq 1. We do not assume any specific distribution nor we assume that the random variables are identically distributed. We can then define S_{n} = S_{0} + \sum_{i = 1}^{n} X_{i}. We can do as before to find \operatorname{E}(S_{n + 1} \mid \mathcal{F}_{n}) = \operatorname{E}(S_{n} + X_{n + 1} \mid \mathcal{F}_{n}) = S_{n} + \operatorname{E}(X_{n + 1} \mid \mathcal{F}_{n}) = S_{n} + \operatorname{E}(X_{n + 1}) = S_{n}, showing that \{S_{n}\} is a martingale with respect to \{\mathcal{F}_{n}\}.

Example 2 Let \{X_{n}: 1 \leq n < \infty\} be a sequence of independent random variables with \operatorname{E}(X_{n}) = 0 and \operatorname{V}(X_{n}) = \sigma^{2} for all n \geq 1. We do not require the random variables to be identically distributed, but each has mean zero and variance \sigma^{2}. Define S_{n} as in Example 1, and set M_{n} = S_{n}^{2} - n \sigma^{2}. Expanding M_{n+1}, M_{n+1} = (S_{n} + X_{n+1})^{2} - (n+1)\sigma^{2} = S_{n}^{2} + 2S_{n}X_{n+1} + X_{n+1}^{2} - (n+1)\sigma^{2}. Taking conditional expectation, \operatorname{E}(M_{n+1} \mid \mathcal{F}_{n}) = S_{n}^{2} + 2S_{n}\operatorname{E}(X_{n+1} \mid \mathcal{F}_{n}) + \operatorname{E}(X_{n+1}^{2} \mid \mathcal{F}_{n}) - (n+1)\sigma^{2}. Because X_{n+1} is independent of \mathcal{F}_{n}, we have \operatorname{E}(X_{n+1} \mid \mathcal{F}_{n}) = 0 and \operatorname{E}(X_{n+1}^{2} \mid \mathcal{F}_{n}) = \sigma^{2}. Therefore, \operatorname{E}(M_{n+1} \mid \mathcal{F}_{n}) = S_{n}^{2} + \sigma^{2} - (n+1)\sigma^{2} = S_{n}^{2} - n\sigma^{2} = M_{n}. Thus, \{M_{n}\} is a martingale with respect to \{\mathcal{F}_{n}\}.

Example 3 Let \{X_{n}: 1 \leq n < \infty\} be a sequence of independent random variables with X_{n} > 0 and \operatorname{E}(X_{n}) = 1 for all n \geq 1. Set M_{0} = 1 and define M_{n} = X_{1} X_{2} \cdots X_{n}. Since M_{n+1} = M_{n} X_{n+1}, we have \operatorname{E}(M_{n+1} \mid \mathcal{F}_{n}) = M_{n} \operatorname{E}(X_{n+1} \mid \mathcal{F}_{n}) = M_{n} \operatorname{E}(X_{n+1}) = M_{n}. Therefore, \{M_{n}\} is a martingale with respect to \{\mathcal{F}_{n}\}.

Stopped Processes

A random variable \tau taking values in \{0, 1, 2, \ldots\} \cup \{\infty\} is called a stopping time with respect to \{\mathcal{F}_{n}: 0 \leq n \leq \infty\} if \{\tau \leq n\} \in \mathcal{F}_{n} \text{ for all} 0 \leq n < \infty. If \tau < \infty almost surely, we can define the stopped process Y_{\tau} = \sum_{k = 0}^{\infty} \mathbf{1}_{\{\tau = k\}} Y_{k}. In cases where \operatorname{P}(\tau = \infty) > 0, we can always define the truncated stopping time n \land \tau = \min\{n, \tau\}. The stopped process Y_{n \land \tau} is always well defined.

WarningStopped Martingales are also Martingales

If \{M_{n}\} is a martingale with respect to \{\mathcal{F}_{n}\}, then the stopped process \{M_{n \land \tau}\} is also a martingale with respect to \{\mathcal{F}_{n}\}.

Proof. We can always write M_{n \land \tau} = M_{n + 1} \mathbf{1}_{\{\tau \geq n + 1\}} + M_{n \land \tau} \mathbf{1}_{\{\tau \leq n\}}. Since \mathbf{1}_{\{\tau \geq n + 1\}} = 1 - \mathbf{1}_{\{\tau \leq n\}}, we see that \mathbf{1}_{\{\tau \geq n + 1\}} is \mathcal{F}_{n} measurable. Thus \begin{aligned} \operatorname{E}(M_{n + 1} \mathbf{1}_{\{\tau \geq n + 1\}} \mid \mathcal{F}_{n}) & = \operatorname{E}(M_{n + 1} \mid \mathcal{F}_{n}) \mathbf{1}_{\{\tau \geq n + 1\}} \\ & = M_{n} \mathbf{1}_{\{\tau \geq n + 1\}} \\ & = M_{n \land \tau} \mathbf{1}_{\{\tau \geq n + 1\}}. \end{aligned} \begin{aligned} \operatorname{E}(M_{(n + 1) \land \tau} | \mathcal{F}_{n}) & = \operatorname{E}(M_{n + 1} \mathbf{1}_{\{\tau \geq n + 1\}} \mid \mathcal{F}_{n}) + M_{n \land \tau} \mathbf{1}_{\{\tau \leq n\}} \\ & = M_{n \land \tau} \mathbf{1}_{\{\tau \geq n + 1\}} + M_{n \land \tau} \mathbf{1}_{\{\tau \leq n\}} \\ & = M_{n \land \tau}, \end{aligned} proving that \{M_{n \land \tau}\} is a martingale with respect to \{\mathcal{F}_{n}\}.

Ruin Probabilities Again

Let us reconsider the simple symmetric random walk: S_{n} = \sum_{i = 1}^{n} X_{i} for n \geq 1, where each X_{i} is an independent fair coin toss, taking values -1 or 1 with probability 1/2. Define the stopping time \tau = \min\{n > 0 : S_{n} = A \text{ or } S_{n} = -B\}, which is measurable with respect to the filtration \mathcal{F}_{n}.

The Probability

Since \tau is almost surely finite, the stopped process S_{n \land \tau} converges to S_{\tau} as n \to \infty: \lim_{n \rightarrow \infty} S_{n \land \tau} = S_{\tau}.

Recall that S_{n} is a martingale, so by the stopped process property, S_{n \land \tau} is also a martingale. Moreover, since the process stops when it reaches either A or -B, we have |S_{n \land \tau}| \leq \max(A, B) for all n. This uniform bound allows us to apply the dominated convergence theorem, yielding \lim_{n \to \infty} \operatorname{E}(S_{n \land \tau}) = \operatorname{E}\left(\lim_{n \to \infty} S_{n \land \tau}\right) = \operatorname{E}(S_{\tau}). We can then conclude that \operatorname{E}(S_{\tau}) = \lim_{n \to \infty} \operatorname{E}(S_{n \land \tau}) = S_{0 \land \tau} = 0.

Since we can also write \operatorname{E}(S_{\tau}) = \operatorname{P}(S_{\tau} = A) A + (1 - \operatorname{P}(S_{\tau} = A)) (-B) = 0, we get that \operatorname{P}(S_{\tau} = A) = \frac{B}{A + B}.

The Expectation

Furthermore, define M_{n} = S_{n}^{2} - n as before. Since for the simple random walk \operatorname{E}(X_{n}) = 0 and \operatorname{E}(X_{n}^{2}) = 1, this implies that \{M_{n}\} is a martingale. Furthermore, |M_{n \land \tau}| \leq \max(A^{2}, B^{2}) + \tau. Because \operatorname{E}(\tau) \infty, we can apply again the dominated convergence theorem to show \lim_{n \to \infty} \operatorname{E}(M_{n \land \tau}) = \operatorname{E}\left(\lim_{n \to \infty} M_{n \land \tau}\right) = \operatorname{E}(M_{\tau}), but also \operatorname{E}(M_{\tau}) = \lim_{n \to \infty} \operatorname{E}(M_{n \land \tau}) = M_{0 \land \tau} = 0.

Since we also have that \operatorname{E}(M_{\tau}) = A^{2} \frac{B}{A + B} + B^{2} \frac{A}{A + B} - \operatorname{E}(\tau) = 0, we conclude that \operatorname{E}(\tau) = A B.