In the previous post I started by introducing the concept of a stochastic process, and their modifications. It is necessary to introduce a further concept, to represent the information available at each time. A *filtration* on a probability space is a collection of sub-sigma-algebras of satisfying whenever . The idea is that represents the set of events observable by time . The probability space taken together with the filtration is called a *filtered probability space*.

Given a filtration, its right and left limits at any time and the limit at infinity are as follows

Here, denotes the sigma-algebra generated by a collection of sets. The left limit as defined here only really makes sense at positive times. Throughout these notes, I define the left limit at time zero as . The filtration is said to be *right-continuous* if .

A probability space is complete if contains all subsets of zero probability elements of . Any probability space can be extended to a complete probability space (its completion) in a unique way by enlarging the sigma-algebra to consist of all sets such that for satisfying . Similarly, a filtered probability space is said to be complete if the underlying probability space is complete and contains all zero probability sets.

Often, in stochastic process theory, filtered probability spaces are assumed to satisfy the *usual conditions*, meaning that it is complete and the filtration is right-continuous. Note that any filtered probability space can be completed simply by completing the underlying probability space and then adding all zero probability sets to each . Furthermore, replacing by , any filtration can be enlarged to a right-continuous one. By these constructions, any filtered probability space can be enlarged in a minimal way to one satisfying the usual conditions.

Throughout these notes I assume a complete filtered probability space, although many of the results can be extended to the non-complete case without much difficulty. However, for the sake of a bit more generality, I don’t assume that filtrations are right-continuous.

One reason for using filtrations is to define *adapted* processes. A stochastic process process is adapted if is an -measurable random variable for each time . This is just saying that the value is observable by time . Conversely, the filtration generated by any process is the smallest filtration with respect to which it is adapted. This is given by , and referred to as the *natural filtration* of .

As mentioned in the previous post, it is often necessary to impose measurability constraints on a process considered as a map . Right-continuous and left-continuous processes are automatically jointly measurability. When considering more general processes, it is useful to combine the measurability concept with adaptedness. This can be done in either of the following three ways, in order of increasing generality (see Lemma 4 below).

Definition 1

- The
predictablesigma-algebra on , denoted by , is generated by the left-continuous and adapted processes. A stochastic process is said to be predictable if it is -measurable. Alternatively, the predictable processes are sometimes calledprevisible.- The
optionalsigma-algebra on , denoted by , is generated by the right-continuous and adapted processes. A stochastic process is said to be optional if it is -measurable.- A process is
progressively measurable, or just progressive, if for each , the mapis -measurable.

The most important of these definitions, at least in these notes, is that of predictable processes. While adapted right-continuous processes will be used extensively, there is not much need to generalize to optional processes. Similarly, progressive measurability isn’t used a lot except in the context of adapted right-continuous (and therefore optional) processes. On the other hand, predictable processes are extensively used as integrands for stochastic integrals and in the Doob-Meyer decomposition, and are often not restricted to the adapted and left-continuous case.

Given any set of real-valued functions on a set, which is closed under multiplication, the set of functions measurable with respect to the generated sigma-algebra can be identitified as follows. They form the smallest set of real-valued functions containing the generating set and which is closed under taking linear combinations and increasing limits. So, for example, the predictable processes form the smallest set containing the adapted left-continuous processes which is closed under linear combinations and such that the limit of an increasing sequence of predictable processes is predictable.

Another way of defining predictable processes is in terms of continuous adapted processes.

Lemma 2The predictable sigma-algebra is generated by the continuous and adapted processes.

*Proof:* Clearly every continuous adapted process is left-continuous and, therefore, is predictable. Conversely, if is an adapted left-continuous process then it can be written as a limit of the continuous processes

Continuity of follows from the fact that is continuous for bounded and measurable . In fact, as is bounded by , it has Lipschitz constant . The limit follows from left-continuity of , which is therefore in the sigma-algebra generated by the continuous adapted processes . ⬜

A further method of defining the predictable sigma-algebra is in terms of simple sets generating it. The following is sometimes used.

Lemma 3The predictable sigma-algebra is generated by the sets of the form

(1)

*Proof:* If is any of the sets in the collection (1) then the process defined by is adapted and left-continuous, and therefore predictable. So, .

Conversely, let be left-continuous and adapted. Then it is the limit of the piecewise constant functions

as goes to infinity. Each of the summands on the right hand side is easily seen to be measurable with respect to the sigma-algebra generated by the collection (1). So, is also measurable. ⬜

In these notes, I refer to the collection of finite unions of sets in the collection (1) as the *elementary* or *elementary predictable* sets. Writing these as then .

Finally, the different forms of measurability can be listed in order of generality, starting with the predictable processes, up to the much larger class of jointly measurable adapted processes.

Lemma 4Each of the following properties of a stochastic process implies the next

- predictable.
- optional.
- progressive.
- adapted and jointly measurable.

*Proof:* As the predictable sigma-algebra is generated by the continuous adapted processes, which are also optional by definition, it follows that all predictable processes are optional.

Now, if is a right-continuous and adapted process and , then the process is right-continuous and -measurable at all times. By the joint measurability of right-continuous processes, is -measurable. As this holds for all times , is progressively measurable.

Finally, consider a progressively measurable process . From the definitions, it is jointly measurable. Furthermore, for any time , restricted to is -measurable. Therefore, is -measurable for all and, in particular, is adapted. ⬜

I have been trying to find clarification on a question relating to filtrations and measurability for awhile, but can’t find discussions of it. I figure someone knowledgeable about stochastic processes would know this. If you have a random variable X that is measurable with respect to a sigma algebra generated by random variable Y, then there exists a function g such that X=g(Y). But what if the sigma algebra is generated by an uncountable number of random variables, such as where F is the natural filtration of stochastic process Y. What can be said in this case for r.v. X measurable with respect to F_t? Is there any functional representation?

Thanks for any help.

Comment by Student — 29 September 10 @ 2:13 AM |

Yes there is! Assuming that X is a real-valued random variable, then it can be expressed as a function of Y.

If Y:Ω → E is a measurable map from the underlying probability space to a measurable space (E,ℰ) then any σ(Y)-measurable map X:Ω → ℝ can be written as X = g(Y) for a measurable g:E → ℝ. This is an application of the functional monotone class theorem. (i) Show that every X = 1

_{A}for A ∈ σ(Y) is of the form g(Y). (ii) Show that the set of random variables of the form g(Y) is closed under linear combinations and taking increasing limits. (iii) appeal to the monotone class theorem.If Y

_{t}is a (real-valued) process, it can be considered as a map to the space ℝ^{ℝ+}of real valued functions.Comment by George Lowther — 30 September 10 @ 10:58 PM |

Hi,

According to Lemma 2, will it be true to say that all brownian motions and martingales that are continuous are predictable processes?

Comment by Su — 4 February 11 @ 5:00 AM |

Yes, that’s correct. Assuming you are using an appropriate filtration, so that your processes are adapted, then they are predictable.

Comment by George Lowther — 4 February 11 @ 9:06 PM |

Thanks for your reply George.

One more thing that’s bothering me, in the context of finding an equivalent martingale measure. Correct if I am wrong below:

For discounted asset price process, Dt (assuming a constant discount rate), there are two reasons why the discounted asset price process could be predictable:

a) Dt will be a martingale under say Q-probability measure, and is hence predictable (under this probability measure only?).

b) If we assume say Geometric Brownian Motion Model for the asset price, the random component of the asset price will only be the Brownian Motion, which is predictable. Hence St (and Dt) is a predictable process.

I have read somewhere that Dt is predictable, which is not very intuitive since asset prices from the market cannot be predictable. Hence my two reasoning above: Predictable only under certain probability measure OR predictable for asset prices which we have assumed a certain model.

Hope you can understand where I am coming from and really appreciate if you could shed some light on this.

Comment by Su — 6 February 11 @ 1:30 AM |

Su,

The reason why processes such as Brownian motion are predictable is just because they are continuous and adapted. It does not have anything to do with the martingale property. In fact, it does not have much to do with the measure either. The property of being predictable is not affected by an equivalent measure change.

I think maybe you are a bit confused about the meaning of being predictable in the precise mathematical sense defined above, which is rather different from the (rather ill-defined) everyday notion of something being predictable. Here, predictability only means that you can (roughly speaking) predict the future value in an infinitesimal sense. Just because a Brownian motion is predictable does not mean that you can tell what value it will take at some future point in time. What you can say is that you can tell when the process is going to reach some value

ajust prior to it happening. This is simply because of continuity. If at some time then, by continuity, will be arbitrarily small for timessjust beforet.So, you can model asset prices by predictable processes. It simply means that they are continuous. Under Black-Scholes, the process is predictable in this sense. That is because it ignores the possibility that the price can suddenly jump.

And, apologies for the late response. Hope that helps clear things up a bit.

Comment by George Lowther — 25 February 11 @ 4:49 PM |

Hi George,

Appreciate the explanation. Definitely cleared things up!

Comment by Su — 12 March 11 @ 6:39 AM |

[…] is predictable, (see definition here) […]

Pingback by Levy Process-3: Ito’s formula « 01law's Blog — 27 April 11 @ 9:31 AM |

Nice proof of Lemma 2, although it took me a while to see why it does not get through when X is right continuous :). Adapt.

Comment by yaoliang — 9 June 11 @ 4:56 AM |

BTW, would that mean if the filtration is right continuous, then optional sigma field = predictable sigma field?

Comment by yaoliang — 9 June 11 @ 5:05 AM |

Hi Yaoliang. The proof of Lemma 2 doesn’t work if

Xis only right-continuous becauseXwould not converge to^{n}X. I’m not sure why you suggest that optional = predictable when the filtration is right-continuous though. This is not true. The filtration generated by a Poisson process is right-continuous (once you complete it), but Poisson processes are not predictable. However, it is often the case that optional = pedictable. For example, this happens if the filtration is generated by a Brownian motion — see my latest post Predictable Stopping Times.Comment by George Lowther — 9 June 11 @ 10:07 PM |

Thanks, George. Here is what I thought about right continuous processes: change the integral interval of X^n from [t-1/n,t] to [t, t+1/n]. Would this be enough to argue that X^n converges to X? But of course, X is only F_+ measurable, that’s why I thought option=predictable when the filtration is right continuous (then F_+=F, hence X is adapted). I believe my claim is false, but still couldn’t see why the proof of Lemma 2 cannot go through.

Comment by yaoliang — 9 June 11 @ 10:56 PM |

Lemma 2 relies on the fact that the processes

Xare both continuous and adapted and, then, the fact that they converge to^{n}Xmeans thatXis measurable with respect to the sigma algebra generated by the continuous and adapted processes.If

Xwas only assumed to be right-continuous then, as you say, you could defineXby integrating over the range [^{n}t,t+1/n] instead. However,Xwould not be adapted. So, you can say that^{n}Xis measurable with respect to the sigma-algebra generated by the continuous processes. But, it need not be measurable with respect to the sigma-algebra generated by the processes which are simultaneously continuous and adapted. It is true though, thatXis measurable with respect to the processes which are continuous and -adapted for any given . This is much weaker than being measurable with respect to the continous and -adapted processes.Comment by George Lowther — 9 June 11 @ 11:09 PM

Hi, George. You say “The filtration generated by a Poisson process is right-continuous (once you complete it)” – I know that was off-hand, but why is this the case? I’m trying to go through the mechanics of proving that statement, and it doesn’t seem to come to easily – the modulus for the right continuity of the process is stochastic.

Comment by Vlad — 21 January 17 @ 9:30 PM |

Hi. One method to prove what you want using results from these notes, is to apply the statements that, Poisson processes are Feller and Feller processes have right-continuous filtration.

A more direct proof would be to show that for bounded rvs U. For U a function of , with it follows from the independent increments property, and it is immediate for U measurable. It can be extended to general U using the monotone class lemma.

Comment by George Lowther — 28 February 17 @ 11:23 AM |

Thanks for the references and response, George. It seems like both approaches (Feller and the “direct proof”) only show almost surely, which is fine if the -algebras are known to be augmented already. However, in the natural filtration for a Poisson process is trivial, so it would seem that any approach showing conditional expectations are equivalent would suffer from the augmentation assumption. Is this assumption tight? I’d imagine so.

Something still seems off for the direct proof sketch, even with augmentation, though – why couldn’t we apply it to the augmented natural filtration for the Wiener process, which isn’t RC?

Comment by Vlad — 28 February 17 @ 3:54 PM

By the ‘augmented’ natural filtration, you just mean with the zero probability events added, right? In that case, the argument shows that the augmented natural filtration is right-continuous. The argument does indeed apply to the Wiener process, whose augmented natural filtration is right-continuous.

I was not making any claim about the non-augmented natural filtration although, now I think about it, it will still be right-continuous for the Poissoj process (but, not for the Wiener process).

Comment by George Lowther — 28 February 17 @ 5:07 PM

Hi George,

Thanks for this great post, it certainly clears things up. Nevertheless I have some trouble understanding the proof of lemma 4. I don’t completely understand how optional impies progressive. You proved that the generators of the optional sigma algebra, are progressively measurable. But I don’t see how this implies that for any optional process this also holds. Do you assume if the generators of a sigma algebra are B([o,t]) x Ft measurable on the subset [0,t] x omega then the same holds for the functions measurable with respect to this sigma algebra? I hope you understand my problem, and I would really appreciate it if you would explain this to me.

All the best,

GuidovM

Comment by GuidovM — 14 June 11 @ 11:15 PM |

GuidovM,

Yes I did assume that if the generators of a sigma algebra are B([0,t]) × F

_{t}measurable on the subset [0,t] × Ω then the same holds for the functions measurable with respect to this sigma algebra.Maybe I did not explain this bit well. You can show that the sets A ⊆ R

_{+}×Ω which are B([0,t])×F_{t}measurable on the subsets [0,t] × Ω forms a sigma-algebra (just check that it is closed under countable unions and taking complements). This is theprogressive sigma-algebra. Furthermore, the definition of progressive processes given in Definition 1 is equivalent to saying that the process is measurable with respect to the progressive sigma-algebra. So, the statement “optional ⇒ progressive” is the same as saying that optional sigma-algebra ⊆ progressive sigma-algebra. To prove this you only need to look at processes generating the optional sigma-algebra.I should add a statement to Definition 1 defining the progressive sigma-algebra and stating that progressive ⇔ measurable with respect to the progressive sigma-algebra. I’ll do this at some point – thanks for mentioning it.

Hope that clears things up.

Comment by George Lowther — 15 June 11 @ 1:35 AM |

Hi George,

First of all thanks for your great set of notes. They are among the best I have come across so far.

Quick question:

Is it true that if a function f is F1xF2 measurable and that its section f_y(x) is F1* measurable for all y in Omega2, then f is F1xF2* measurable?

Your proof of “optional => progressive” seems to rely on the latter fact.

Thanks in advance!

– Tigran

Comment by Tigran — 16 November 11 @ 2:51 PM |

Actually I just realized that your proof does not require what I mentioned.

Never mind!

Comment by Tigran — 16 November 11 @ 3:38 PM |

Hi.

No, I don’t rely on that fact. Actually, that’s not true. Consider for example F1, F2 to be the Borel sigma-algebra on the reals. Let f be the indicator function of a non-measurable subset of the diagonal in R

^{2}. Then, f_{y}(x) will be measurable for each y, but f itself is not measurable (I’m not sure what you mean by F1* and F1xF2*. Maybe the completion? That depends on having a measure though).Comment by George Lowther — 16 November 11 @ 10:46 PM |

Thanks for the reply. I realized after my initial comment that the proof doesn’t rely on this statement.

I also just realized how unclear my initial statement actually was (and how inconsistent the notation was as well).

What I meant to say is that suppose f(x,y) is measurable w.r.t. some joint sigma-algebra (F1xF2) and that its section in y (f_y(x)) is measurable w.r.t. a sigma-algebra smaller than F1 (this smaller sigma-algebra is denoted F1*) for each y.

Note: F1* does not depend on y.

The question is: Is f(x,y) measurable w.r.t. F1*xF2?

I have a feeling the answer in no.

Comment by Tigran — 21 November 11 @ 3:23 PM |

The answer is indeed no. Consider any probability space (Ω,F1,P) on which we have defined a uniformly distributed random variable

T: Ω → [0,1]. Let F1* consist of the sets in F1 with probability 0 or 1. Let F2 be the Borel sigma-algebra on [0,1].Now, look at the stochastic process Ω×[0,1] → R given by (ω,

t) ↦X(ω) = 1_{t}_{{T(ω)=t}}. This is F1×F2 measurable but not F1*×F2 measurable. However,Xis almost surely zero at each time_{t}t, so is F1* measurable.Comment by George Lowther — 21 November 11 @ 7:32 PM |

Thanks again for your answer George. It’s a nice counter-example.

The only bit that I have yet to understand completely is: why is I_{T(omega}=t} not F1*xF2-measurable.

It’s probably a trivial detail that I am missing and your input would be great to help me understand this.

P.S.: Once again my hat is off to the clarity of your notes. They are extremely helpful in getting a better intuition of the technicalities in stochastic analysis.

Comment by Tigran — 26 November 11 @ 2:55 PM |

It is true that this process is not F1*xF2 measurable. However, it is

nota trivial detail. Far from it. You can prove it by the measurable projection theorem. If it was measurable then this theorem would tell you that the set of ω ∈ Ω such that there exists at≤ 1/2 withX(ω) = 1 is measurable with respect to the completion of F1*, so has measure 0 or 1. However, this set is just {_{t}T≤ 1/2}, which has probability 1/2. SoXis not F1*xF2 measurable.Comment by George Lowther — 26 November 11 @ 3:11 PM

…actually the measurable projection theorem is very similar to the generalised form of the Debut theorem which says that the first time that a jointly measurable process hits a measurable set is itself measurable wrt the completion of the probability space (and is a stopping time if the process is progressive and the filtration is right continuous). In this case, the first time X hits 1 is T, which is not F1* measurable. So X is not F1*xF2 measurable.

Comment by George Lowther — 26 November 11 @ 3:32 PM

Dear George, thanks for very nice notes – the explanation is highly helpful. I didn’t get the definition of a predictable sigma-algebra: that is generated by left-continuous and adapted processes. W.r.t. which filtration processes generating predictable sigma-algebra should be adapted?

Thanks in advance,

Ilya.

Comment by Ilya — 17 February 12 @ 4:44 PM |

The definition of the predictable sigma-algebra does depend on what underlying filtration is used to define adapted processes. Usually, we are given one sigma-algebra (representing all observable events at time

t). Then, we work exclusively with respect to this sigma-algebra, which is used in the definition of adapted processes (and predictable processes, optional processes, stopping times, etc). If we have multiple sigma-algebras and it is not clear which one is being used from the context, then it should be stated. With multiple sigma-algebras, we have multiple predictable sigma-algebras (and multiple optional sigma algebras, etc). Where this happens in these notes I say something along the lines of “the -predictable sigma algebra” or “predictable with respect to â€ť).Comment by George Lowther — 22 February 12 @ 1:20 AM |

Hi!

I don’t quite understand why it is important to assume the usual conditions. What exactly do we miss if don’t assume them? Do stochastic integral stop being adapted? Are we unable to make Doob-Meyer decompositions? Martingales stop having a cadlag version? When we make a Girsanov transformation, the new Brownian motion generates a filtration which does not coincide with the filtration of the original Bm?

Thanks for these great notes!

Comment by Chilorio — 4 January 13 @ 6:22 AM |

Cadlag versions and stochastic integrals would both fail to be adapted. Check: compensated Poisson process X, pick its left-continuous version Y and the filtration generated by Y (= the predictable).

Comment by MathSwineT — 18 January 13 @ 7:43 AM |

George, I would like to ask a question about this very useful thread. In Lemma 2, when you define the process $X_t^n = \int_{t-1/n}^t etc… I can see that this definition works when $X_t $ is a step process or a simple process. But for general $X_t $ how do you handle them inside the integral?…. Are you saying that one should prove it using the functional monotone class theorem? In this case which version do you use as there is more than one fMonClassTheorem? Thank you in advance if you could throw a few more details in there.

Comment by Mario — 4 September 13 @ 7:09 AM |

Hi, I have a quite related question. Suppose X and Y are two continuous processes. For each t, Xt is adapted with respect to {Ys:s X(t,w) and (t, w) –> Y(t,w), is the joint sigma-field generated by X contained in that generated by Y? This kind of joint measurability question is interesting and difficult.

Comment by Chi Dong — 24 June 14 @ 12:29 AM |

Sorry, there seems to be some system error. My question is like this: Suppose X and Y are two continuous processes. For each t, Xt is adapted with respect to {Ys:s X(t,w) and (t, w) â€“> Y(t,w), is the joint sigma-field generated by X contained in that generated by Y?

Comment by Chi Dong — 24 June 14 @ 12:31 AM |

Sorry, still error. Suppose X and Y are two continuous processes. For each t, Xt is adapted with respect to {Ys:s< = t}.

Comment by Chi Dong — 24 June 14 @ 12:32 AM |

Considering X and Y as functions (t,w) –> X(t,w) and (t, w) â€“> Y(t,w), is the joint sigma-field generated by X contained in that generated by Y?

Comment by Chi Dong — 24 June 14 @ 12:32 AM |

Hi George Lowther

I could not understand what do you mean by the following? I would be grateful if you could explain it

“Given any set of real-valued functions on a set, which is closed under multiplication, the set of functions measurable with respect to the generated sigma-algebra can be identitified as follows. They form the smallest set of real-valued functions containing the generating set and which is closed under taking linear combinations and increasing limits. So, for example, the predictable processes form the smallest set containing the adapted left-continuous processes which is closed under linear combinations and such that the limit of an increasing sequence of predictable processes is predictable.”

Comment by vinayak chuni — 26 October 14 @ 11:39 PM |

[…] reading George Lowther’s blog and have a question about the proof of lemma 2. We want to verify that the predictable sigma […]

Pingback by Proof that the predictable sigma algebra is also generated by continuous and adapted processes - MathHub — 7 May 16 @ 6:20 PM |

In the proof of lemma 2, do you really need the functional monotone class theorem? Doesn’t continuity follow directly by choosing

Comment by Anonymous — 10 August 16 @ 12:37 AM |

I think you are correct — but why not just ?

Comment by George Lowther — 10 August 16 @ 10:02 PM |

Actually, I see why you used the min, but it is unnecessary. I updated the post. Thanks!

Comment by George Lowther — 10 August 16 @ 10:43 PM |

Hi George,

I have a question regarding adaptedness and filtration: is adaptedness a distributional property?

More precisely, Suppose is a probability space supporting a Brownian motion , where the filtration is the augmented brownian filtration and . Let be a jointly measurable process which is a martingale in its own natural filtration. I want to understand if there is condition in terms of the joint law of (on the canonical path space ) which is sufficient for to be -adapted.

My first guess is that if is a martingale w.r.t. , then is -adapted. Namely I want to show that enlarge the Brownian filtration by including future information will destroy the martingale property of the brownian motion. I am wondering if this is true or not. Thanks!

Comment by yan — 28 February 17 @ 9:57 PM |

Are you asking if any strict enlargement of the Brownian filtration (but keeping the same underlying probability space) will necessarily destroy the martingale property of the underlying Brownian motion?

If so, then yes that is true. By the martingale representation theorem, all martingale would have to retain the martingale property, in which case the result follows.

Comment by George Lowther — 1 March 17 @ 1:28 PM |

Thank you for your prompt response. As a follow-up question, I am wondering if this is still true in the discrete time case where is a random walk with jumps that are i.i.d normal random variables(say,with mean zero). I am not aware of any martingale representation in discrete time, except the binomial-tree model. Thanks again!

Comment by yan — 1 March 17 @ 3:05 PM |

In the discrete time case you mention, it is possible to enlarge the filtration while keeping the martingale property. For example, add the absolute sizes of all the jumps to .

Comment by George Lowther — 1 March 17 @ 3:44 PM

I agree with you that the martingale property can be preserved by enlarging the filtration. However, in this case it seems that we may lose the independent increment property(i.e., is no longer independent of ). Going back to my original question which involve the process . Is there a condition on the joint distribution of which will guarantee that is adapted to ? Namely, can we find two processes and , one being adapted and one not, such that has the same distribution as ? As you pointed out this is not possible in the continuous time case.

Thank you and looking forward to your reply!

Comment by yan — 1 March 17 @ 4:34 PM

On your question about the distribution of , I would consider the statement that S is W adapted to be a statement about this distribution, so I am not sure what further to say about it.

Comment by George Lowther — 2 March 17 @ 9:51 AM

what I am looking for is a condition on the joint distribution which can be expressed in terms of some equations. For example, that is martingale is equivalent to saying for all f. Unfortunately, this is not sufficient to give adaptedness as you pointed out. I am just curious if one can formulate condition of this kind which is sufficient. I hope this is a question that worth thinking. thank you!

Comment by yan — 2 March 17 @ 3:38 PM |

I think it should be enough to state that, for each , the distribution of conditional on Is preserved. In the case of independent increments, it is enough to say that is preserved.

Comment by George Lowther — 2 March 17 @ 7:18 PM

Hi!

You define a predictable sigma-algebra on {{\mathbb R}_+\times\Omega}, denoted by {\mathcal{P}} to be the sigma algebra generated by the left-continuous and adapted processes. You also write that a adapted process is one which in turn depends on the sigma algebra i.e that each r.v is measuable. This seems circular to me, do you mean that we take a left contious processes and then generate it’s natural filtration? Or have I gotten something wrong?

Comment by Anonymous — 17 July 17 @ 5:54 AM |

Traiteur Rabat Regal; Traiteur de ronome au MarocThis is my expert

Trackback by traiteur rabat — 29 August 17 @ 12:17 AM |

Assume that $L$ is a process of FV that is adapted to the filtration generated by a Wiener Process $W$ defined in a probability space $(\Omega, \mathcal{F}, P)$ and consider a, say, continuous function $f:[0,T]\to\mathbb{R}$ (I have f(t)=exp(-rt) with $r>0$ in mind). Let us define the stochastic process $X_t:=\int_0^t f(t)dL_t$. Is it the case that the expectation of $X_T$ with respect to $P$ is always well defined? Many thanks!

Comment by Santo — 12 January 18 @ 6:10 PM |

Hello. It occurs to me that the definition of natural filtration of a rv as the smallest filtration to which it is adapted is not obviously well-defined, because there is no total order on the set of filtrations. We can define a partial order by saying that one filtration is less than or equal to another iff $\forall t\geq 0:\ \mathscr F^1_t\subseteq \mathscr F^2_t$. It is imaginable that a rv would have multiple filtrations that are minimum elements of the poset of filtrations under that order. It seems to me that the simplest way to define a natural filtration is by the formula given in the sentence that follows the definition above, ie . We would then wish to prove that filtration is smaller than any filtration to which is adapted, under the above-defined order. That, together with the observation that is adapted to that filtration, would entail that filtration is a unique minimum element of the subset of all filtrations to which is adapted.

I hope this LaTeX works. It’s the first time I’ve tried using it in a WordPress environment đź™‚

Comment by sageandonions — 25 April 18 @ 6:45 AM |

Hi in my opinion what you do in the second part of your post is exactly what it means in the definition of this blog for a filtration to be minimal (and hence the natural filtration w.r.t. the process X), and being minimal has nothing to do with a total order on the set of all filtrations. So even if what you say in the second part makes sense it is redundant with the notions used here and there is no problem of definition. Best regards

Comment by TheBridge — 26 April 18 @ 1:28 PM |

It doesn’t require a total order but it does require a partial order and the existence of a minimal element of the set of adapted filtrations. Perhaps I missed it but I couldn’t see any definition of an order relation on filtrations above, which is why I suggested one (there is an implicit order relation used on sigma algebras, but not on filtrations). Having done that, it remains to be proved that the set of filtrations adapted to a process X has a minimal element under that order relation. I imagine one would do that by taking the filtration such that and proving that it is a minimal element under that order relation.

Comment by sageandonions — 27 April 18 @ 12:16 AM |

So far so good, you make explicit what is implicit in the post (but very clear from the context IMO). Nonetheless your precisions are exactly what the definition given here is about and I wouldn’t go as far as qualify the definition of natural filtration as not well defined (introducing pre-order is a bit overkill here). Best regards

Comment by TheBridge — 27 April 18 @ 9:53 AM

In the proof of Lemma 2, you reason as “In fact, as is bounded by , it has Lipschitz constant .” I doubt that “boundedness” implies “Lipschitz”, the counterexample being on .

Comment by xyz — 7 December 18 @ 6:49 PM |

What I meant is that as f is bounded, its

integralis Lipschitz.Comment by George Lowther — 7 December 18 @ 9:21 PM |