In contrast to optional processes, the class of predictable processes was used extensively in the development of stochastic integration in these notes. They appeared as integrands in stochastic integrals then, later on, as compensators and in the Doob-Meyer decomposition. Since they are also central to the theory of predictable section and projection, I will revisit the basic properties of predictable processes now. In particular, any of the collections of sets and processes in the following theorem can equivalently be used to define the predictable sigma-algebra. As usual, we work with respect to a complete filtered probability space . However, completeness is not actually required for the following result. All processes are assumed to be real valued, or take values in the extended reals .

Theorem 1The following collections of sets and processes each generate the same sigma-algebra on .

{:is a predictable stopping time}.as ranges over the predictable stopping times andZover the -measurable random variables..The elementary predictable processes.{:is a stopping time}{}.The left-continuous adapted processes.The continuous adapted processes.

Compare this with the analogous result for sets/processes generating the optional sigma-algebra given in the previous post. The proof of Theorem 1 is given further below. First, recall that the predictable sigma-algebra was previously defined to be generated by the left-continuous adapted processes. However, it can equivalently be defined by any of the collections stated in Theorem 1. To make this clear, I now restate the definition making use if this equivalence.

Definition 2The predictable sigma-algebra, , is the sigma-algebra on generated by any of the collections of sets/processes in Theorem 1.

A stochastic process is predictable iff it is -measurable.

We previously showed that the running maximum of a progressive process is optional. In the case that the process is predictable, then it is straightforward to show that its running maximum is also predictable.

Lemma 3IfXis a predictable process then so is .

*Proof:* As in the proof of Lemma 3 of the post on measurable projection, the process is left-continuous and adapted. So, *Y* is predictable and is the maximum of two predictable processes, so is predictable. ⬜

I proved, in the post on measurable projection, that the limit supremum, and left and right-limit supremum of a progressively measurable process is again progressive. Then, it was seen that most of these limits are in fact optional. I now show that the (strict) left limit-supremum is predictable.

Lemma 4IfXis a progressively measurable process thenis predictable. If

Xis predictable then so is

*Proof:* For each positive integer *n*, choose a sequence of times increasing to infinity, and such that tends to zero as *n* goes to infinity. For example, . Define the process by and

for . This is left-continuous and, by Lemma 4 of the post on measurable projection, is adapted. So, is predictable. Then,

is predictable. Finally, if *X* is predictable then

is the maximum of two predictable processes, so is predictable. ⬜

A map is measurable if and only if its graph, , is jointly measurable. Furthermore, is a stopping time if and only if is progressive, and also if and only if is optional. In the current context, it would seem natural to ask if being a predictable stopping time is equivalent to its graph being predictable. In fact, this is true. However, the proof of this requires the predictable section theorem or other advanced results. So, I will leave this until after introducing predictable section in the next post of these notes.

It can be shown that every optional process is equal to a predictable process outside of a thin set. Recall that a process *X* is thin if it is optional and the set is contained in the graphs of a countable sequence of stopping times.

Lemma 5For any optional processX, there exists a predictable processYsuch that is thin.

*Proof:* Consider the collection, , of all processes which are the sum of a predictable and a thin process. We need to show that every optional process is in , which will be done with the functional monotone class theorem.

First, letting be the collection of stochastic intervals for stopping times , then is a pi-system generating the optional sigma-algebra. For any in , we can decompose the indicator function

as the sum of a predictable and a thin process, so is in . Next, consider processes *X* and *Y* in . Then, and for predictable processes and thin processes . Writing

for any real numbers shows that is in .

Consider a sequence converging pointwise to a limit *X*. We can write for predictable and thin . Letting *S* be the collection of in for which converges to a finite limit, predictability of implies that . So, converges pointwise to a predictable process *Y*. On the set we have and, so, on this set. As is contained in for a sequence of stopping times ,

Therefore, . Then, the functional monotone class theorem says that every optional process is in as required. ⬜

#### Proof of Theorem 1

Finally for this post, I will give the proof of Theorem 1 stated above.

Letting , , be the sigma-algebra generated by each of the respective collections of sets/processes of the theorem, it needs to be shown that these are all equal. As is generated by the left-continuous adapted processes, it equals what we originally defined as the the predictable sigma-algebra.

We have already shown the following equalities.

- The predictable sigma-algebra is generated by the continuous adapted processes, . See Lemma 2 of the post on filtrations and adapted processes.
- The predictable sigma-algebra is generated by sets as in the third statement, . See Lemma 3 of the post on filtrations and adapted processes.
- The predictable sigma-algebra is generated by the stochastic intervals for predictable stopping times , . See Lemma 7 of the post on predictable stopping times.

I now prove the remaining equivalences.

*Proof of :* As elementary predictable processes are left-continuous and adapted, we have . Also, if is as in the third statement, then is an elementary predictable process for each and, letting *T* go to infinity, shows that . Hence, and, as we have already shown that , this completes the proof. ⬜

*Proof of :* If is a stopping time then is left-continuous and adapted, so we see that . Next, for any and ,

is a stopping time. Since is equal to , it follows that . As has already been shown, this completes the proof. ⬜

*Proof of :* It is clear by taking in the processes in the second statement that we have .

Now, letting be a predictable stopping time then it just remains to demonstrate that is -measurable for all -measurable *Z*. However, as is generated by as *X* ranges over the left-continuous and adapted processes, it is enough to consider for such a process *X*. In that case

where is the stopped process, which is left-continuous and adapted, hence -measurable. As we have already shown that , this shows that is -measurable as required. ⬜

Hi George, a process is predictable implies that it is progressive and is measurable for all . I am wondering if the converse is also true. It seems to me quite “reasonable” to define “predictable” by the requirement that is measurable. Thanks!

Comment by Yu Ding — 28 October 18 @ 6:38 PM |

Hi. No, the converse is not true. For one thing, being measurable is not enough to ensure that the same holds at all stopping times. Consider Poisson processes. These would be predictable by your definition, but they are not predictable (e.g., a compensated Poisson process is a non-continuous martingale, but predictable martingales are always continuous).

Comment by George Lowther — 6 December 18 @ 11:46 PM |

[…] we like. There is also a corresponding predictable section theorem, which says that if S is in the predictable sigma-algebra, its section can be chosen to be a predictable stopping […]

Pingback by Proof of Optional and Predictable Section | Almost Sure — 7 January 19 @ 12:03 PM |