A single jump filtration (Ft)t∈R+ generated by a random variable γ with values in R‾+ on a probability space (Ω,F,P) is defined as follows: a set A∈F belongs to Ft if A∩{γ>t}t\}$]]> is either ∅ or {γ>t}t\}$]]>. A process M is proved to be a local martingale with respect to this filtration if and only if it has a representation Mt=F(t)1{t<γ}+L1{t⩾γ}, where F is a deterministic function and L is a random variable such that E|Mt|<∞ and E(Mt)=E(M0) for every t∈{t∈R+:P(γ⩾t)>0}0\}$]]>. This result seems to be new even in a special case that has been studied in the literature, namely, where F is the smallest σ-field with respect to which γ is measurable (and then the filtration is the smallest one with respect to which γ is a stopping time). As a consequence, a full description of all local martingales is given and they are classified according to their global behaviour.
Filtrationlocal martingaleprocesses with finite variationσ-martingalestopping time60G4460G07Introduction
Starting with Dellacherie [4], the following simple model has been studied and intensively used in applications. Given a random variable γ with positive values on a probability space (Ω,F,P), one considers the smallest filtration with respect to which γ is a stopping time (or, equivalently, the process 1{t⩾γ} is adapted). In particular, Dellacherie gives a formula for the compensator of this single jump process 1{t⩾γ}. Chou and Meyer [2] describe all local martingales with respect to this filtration and prove a martingale representation theorem. A significant contribution is done in a recent paper by Herdegen and Herrmann [13], where a classification, whether a local martingale in this model is a strict local martingale, or a uniformly integrable martingale, etc., is given. Let us also mention some related papers [1, 15, 16, 3, 8, 21, 12], where, in particular, local martingales with respect to the filtrations generated by jump processes or measures of certain kind are studied.
Let us clarify that in the above model every local martingale has the form
Mt=F(t)1{t<γ}+H(γ)1{t⩾γ},
or
Mt=F(t∧γ)−K(γ)1{t⩾γ},
where γ is a random variable with values in, say, (0,+∞), F, H, and K=F−H are deterministic functions. Denote by G the distribution function of γ, G‾(t)=1−G(t), tG=sup{t:G(t)<1} is the right endpoint of the distribution of γ. Assume that E|Mt|<∞, then
E(Mt)=F(t)G‾(t)+∫[0,t]H(s)dG(s),
where the corresponding Lebesgue–Stieltjes integral is finite. If (Mt) is a martingale, then E(Mt)=E(M0), and this equality can be written as
F(t)G‾(t)+∫[0,t]H(s)dG(s)=F(0)
and can be viewed as a functional equation concerning one of functions in (F,G,H) or (F,G,K), where other two functions are assumed to be given. In fact, this equation takes place for t<tG or t⩽tG, the latter in the case where tG<∞ and P(γ=tG)>00$]]>. Moreover, it turns out that this is not only the necessary condition but also the sufficient one for (Mt)t∈R+ given by (1) to be a local martingale. This consideration allows us to reduce problems to solving this functional equation. For example, to find the compensator F(t∧γ) of 1{t⩾γ} as in [4] one needs to find a solution F given G and K≡1. A possible way to explain the idea in [2] is the following: The terminal value M∞ of any local martingale M in this model is represented as H(γ), and to find a representation (1) for M it is enough to solve the equation for F given G and H; the linear dependence between H and F results in a representation theorem. Contrariwise, in [13] the authors suggest to find H from the equation for given F and G. This allows them to study global properties of M.
In this paper we consider a more general model, where all randomness appears “at time γ” but it may contain much more information than γ does. We start with a random variable γ on a probability space (Ω,F,P), and define a single jump filtration (Ft) in such way that nothing happens strictly before γ, γ is a stopping time with respect to it, and the σ-field Fγ of events that occur before or at time γ coincides with F (in fact, on the set {γ<∞}). We prove that every local martingale with respect to this filtration has the representation
Mt=F(t)1{t<γ}+L1{t⩾γ},
where now L is a random variable which is not necessarily a function of γ. However, denoting H(t)=E[L|γ=t], we come to the same functional equation of type (2).
Some results of the paper can be deduced from known results for marked point processes, at least if F is countably generated; this applies, for example, to Theorem 5 about the compensator of a single jump process. Another example is Corollary 1 which says that every local martingale is the sum of a local martingale of form (1) and an “orthogonal” local martingale, the latter being characterised, essentially, by the property F(t)≡0. The reader can recognize in this decomposition the representation of a local martingale as the sum of two stochastic integrals with respect to random measures, see [16] and [17]. However, our direct proofs are simpler due to the key feature of our paper. Namely, we obtain a simple necessary and sufficient condition for a process to be a local martingale and later exploit it. A description of all local martingales via a full description of all possible solutions to a functional equation of type (2) is a simple consequence of this necessary and sufficient condition. In particular, an absolute continuity type property of F with respect to G, considered as an assumption in [13], is proved to be a necessary condition. An elementary analysis of a functional equation of type (2) shows that, if γ has no atom at its right endpoint, there are different F satisfying the equation for given H and G. In particular, there is a local martingale M such that M0=1 and M∞=0; M is necessarily a closed supermartingale.
Another important feature of our model, in contrast to Dellacherie’s model, is that it admits σ-martingales which are not local martingales.
Let us also mention some other papers where processes of form (1) or (3) are considered. Processes of form (1) with tG=∞ are typical in the modelling of credit risk, see, e.g., [18] and [19, Chapter 7], where usually F is expressed via G and one needs to find H. Since tG=∞, such a process is a martingale. For example, in the simplest case F=1/G‾ and hence H=0. This process is the same that is mentioned in two paragraphs above. Single jump filtrations and processes of form (3) appear in [10] and [11]. It is interesting to note that, in [11], the random “time” γ is, in fact, the global maximum of a random process, say, a convergent continuous local martingale.
Section 2 contains our main results. In Theorem 1 we establish a necessary and sufficient condition for a process of type (3) to be a local martingale. This allows us to obtain a full description of all local martingales through a functional equation of type (2) in Theorem 2. A similar description is available for σ-martingales, see Theorem 3. Finally, Theorem 4 classifies local martingales in accordance with their global behaviour up to ∞. Section 3 contains the proofs of these results. In Section 4 we consider complementary questions. Namely, we find the compensator of a single jump process. We also consider submartingales of class (Σ), see [22], and show that their transformation via a change of time leads to processes of type (3). As a consequence, we reprove Theorem 4.1 of [22].
We use the following notation: R+=[0,+∞), R‾+=[0,+∞], a∧b=min{a,b}. The arrows ↑ and ↓ indicate monotone convergence, while lims⇈t stands for lims→t,s<t.
A real-valued function Z(t) defined at least for t∈[0,s) is called càdlàg on [0,s) if it is right-continuous at every t∈[0,s) and has a finite left-hand limit at every t∈(0,s); it is not assumed that it has a limit as t⇈s. If, additionally, a finite limit limt⇈sZ(t) exists, then Z(t) is called càdlàg on [0,s]. Functions Z of finite variation on compact intervals are understood as usually and are assumed to be càdlàg. The variation at 0 includes |Z(0)| as if Z is extended by 0 on negative axis. The total variation of Z over [0,t] is denoted by Var(Z)t. We say that Z has a finite variation over [0,s), s⩽∞, if limt⇈sVar(Z)t<∞. We denote Var(Z)∞:=limt→∞Var(Z)t.
A filtration on a probability space (Ω,F,P) is an increasing right-continuous family F=(Ft)t∈R+ of sub-σ-fields of F. No completeness assumption is made. As usual, we define F∞=σ(∪t∈R+Ft) and, for a stopping time τ the σ-field Fτ is defined by
Fτ={A∈F∞:A∩{τ⩽t}∈Ftfor everyt⩾0}.
A set B⊂Ω×R+ is evanescent if B⊆A×R+, where A∈F and P(A)=0. We say that two stochastic processes X and Y are indistinguishable if {X≠Y} is an evanescent set.
Since we do not suppose completeness of the filtration F, we cannot expect that processes that we consider have all paths càdlàg. Instead we consider processes whose almost all paths are càdlàg. Obviously, for any càdlàg process X adapted with respect to the completed filtration, there is an a.s. càdlàg F-adapted process indistinguishable from X. Furthermore, any F-adapted process X with a.s. càdlàg paths is indistinguishable from an F-optional process Y whose paths are right-continuous everywhere and have finite left-hand limits for t<ρ(ω) and t>ρ(ω)\rho (\omega )$]]>, where ρ is a F-stopping time with P(ρ<∞)=0; let us call such Y regular and ρ a moment of irregularity for Y. Dellacherie and Meyer [6, VI.5 (a), p. 70] prove that, if the filtration is not complete, every supermartingale X (with right-continuous expectation) has a modification Y with the above regularity property. If we are given just an adapted process X with almost all paths càdlàg, we define ρ and Y from values of X on a countable set exactly as is done in [6] in the case where X is a supermartingale. Using [5, Theorem IV.22, p. 94], we obtain that ρ(ω)=∞ and paths X·(ω) and Y·(ω) coincide for those ω for which X·(ω) is càdlàg everywhere. Moreover, if ρ(ω)<∞, then Yt(ω) is càdlàg for t<ρ(ω) and one may put Yt(ω)=0 for t⩾ρ(ω).
Processes with finite variation are adapted and not assumed to start from 0. A moment of irregularity for them has additionally the property that their paths have finite variation over [0,t] for all t<ρ(ω).
It is instructive to mention that, in our model, there is no need to use general results on the existence of (a.s.) càdlàg modifications for martingales since they can be proved directly. For example, if L is an integrable random variable with EL=0, then the process M given by (3) with F(t)=E[L|γ>t]1{t<tG}t]{\mathbb{1}_{\{t<{t_{G}}\}}}$]]> satisfies Mt=E[L|Ft] a.s. for an arbitrary t. It is trivial to check that this function F has finite variation over any [0,t] with P(γ>t)>0t)>0$]]> (and over [0,tG) if P(γ=tG<∞)>00$]]>). Thus M is regular. It may be that, if tG<∞ and P(γ=tG)=0, the function F has not a finite limit as t⇈tG, or, more generally, has unbounded variation over [0,tG). Then a moment of irregularity is given by
ρ(ω)=tG,ifγ⩾tG;+∞,otherwise.
It takes a finite value only on the set {γ⩾tG} of zero measure. In all other cases we may put ρ≡+∞. See Remark 2 in Section 2 for more details.
Main results
Let γ be a random variable with values in R‾+ on a probability space (Ω,F,P). We tacitly assume that P(γ>0)>00)>0$]]>. G(t)=P(γ⩽t), t∈R+, stands for the distribution function of γ and G‾(t)=1−G(t). Put also tG=sup{t∈R+:G(t)<1} and T={t∈R+:P(γ⩾t)>0}0\}$]]>. Note that P(γ∉T)=0. We will often distinguish between the following two cases:
P(γ=tG<∞)=0.
P(γ=tG<∞)>00$]]>.
It is clear that T=[0,tG) in Case A and T=[0,tG] in Case B.
We define Ft, t∈R+, as the collection of subsets A of Ω such that A∈F and A∩{t<γ} is either ∅ or coincides with {t<γ}.
It is shown in Proposition 1 that Ft is a σ-field for every t∈R+ and the family F=(Ft)t∈R+ is a filtration. We call this filtration a single jump filtration. It is determined by generating elements γ and F. In this paper we consider only single jump filtrations and, if necessary to indicate generating elements, we use the notation F(γ,F) for the single jump filtration generated by γ and F.
In this section a single jump filtration F=F(γ,F) is fixed. All notions depending on filtration (stopping times, martingales, local martingales, etc.) refer to this filtration F, unless otherwise specified.
(i) Ftis a σ-field and a random variable ξ isFt-measurable,t∈R+, if and only if ξ is constant on{t<γ}. ξ isF∞-measurable if and only if ξ is constant on{γ=∞}.
(ii) The family(Ft)t∈R+is increasing and right-continuous, i.e.F=(Ft)t∈R+is a filtration.
(iii) γ is a stopping time andFγ=F∞.
(iv) A random variable T with values inR‾+is a stopping time if and only if it satisfies the following property: if the set{T<γ}is not empty, then there is a number r such that{T<γ}={T=r<γ}={r<γ}.
(i) IfX=(Xt)t∈R+is an adapted process, then there is a deterministic functionF(t),0⩽t<tG, such thatXt=F(t)on{t<γ∧tG}. IfY=(Yt)t∈R+is an adapted process andP(Xt=Yt)=1for everyt∈R+, thenXt=Ytidentically on{t<γ∧tG}.
(ii) IfY=(Yt)t∈R+is a predictable process, then there is a measurable deterministic functionC(t),t∈T, such thatYt=C(t)on{t⩽γ},t∈T.
(iii) IfX=(Xt)t∈R+is a process with finite variation, thenF(t)in (i) has a finite variation over[0,t]for everyt<tGin Case A and over[0,tG)in Case B.
(iv) Every semimartingale is a process with finite variation.
(v) IfM=(Mt)t∈R+is a σ-martingale then there are a deterministic functionF(t),t∈R+, and a finite random variable L such that, up toP-indistinguishability,Mt=F(t)1{t<γ}+L1{t⩾γ}.
Statement (iv) is not surprising. If the σ-field F is countably generated, then our filtration is a special case of a filtration generated by a marked point process, and it is known, see [17], that then all martingales are of finite variation. In general, a single jump filtration is a special case of a jumping filtration, see [14], where again all martingales are of finite variation.
If M is a σ-martingale, then it is a process with finite variation due to (iv) and, hence, the function F(t) in (5) has a finite variation over [0,t] for every t<tG in Case A and over [0,tG) in Case B according to (iii).
According to (i), the function F(t) in (5) is uniquely determined for t<tG. Since P(γ>tG)=0{t_{G}})=0$]]>, the stochastic interval ⟦tG,γ⟦ is an evanescent set. Hence, F(t) can be defined arbitrarily for t⩾tG. For example, we can put it equal to 0 for t⩾tG. Then F(t) has a finite variation on compact intervals if tG=+∞ or in Case B. In Case A, if tG is finite, F(t) may have infinite variation over [0,tG) (and even not have a finite limit as t⇈tG), see Theorem 2 and Example 3 below. All other points are regular for F(t). Now put ρ(ω)=tG<+∞ if we are in Case A, tG<+∞, limt⇈tGVar(F)t=∞, and γ(ω)⩾tG, and let ρ(ω)=+∞ in all other cases. It follows that ρ is a moment of irregularity for the process in the right-hand side of (5).
In what follows, when we write that the process M has the representation (5), this means that M and the right-hand side of (5) are indistinguishable. Moreover, we tacitly assume that F(t) is right-continuous for t⩾tG to ensure that the right-hand side of (5) is right-continuous.
Propositions 1 and 2 explain why we call F a single jump filtration: all randomness appears at time γ. It is not so natural to describe local martingales with respect to F as single jump processes. As we will see, the function F in (5) need not be continuous, so local martingales may have several jumps.
Our main goal is to provide a complete description of all local martingales. According to Proposition 2 (v), a necessary condition is that it is represented in form (5). Thus, it is enough to study only processes of this form.
LetF(t),0⩽t<tG, be a deterministic càdlàg function, L be a random variable, and a processM=(Mt)t∈R+be given byMt=F(t)1{t<γ}+L1{t⩾γ}.The following statements are equivalent:
M=(Mt)t∈R+is a local martingale.
(Mt)t∈Tis a martingale.
E(|Mt|)<∞,t∈T,andE(Mt)=E(M0),t∈T.
In the case where F=σ{γ}, equivalence (i) and (ii) is proved in [2].
Concerning the last statement of the proposition, let us emphasize that if tG<∞ and P(γ=tG)=0, a local martingale M=(Mt)t∈R+ may not be a martingale on [0,tG]; obviously, if it is a martingale, then it is uniformly integrable, and necessary and sufficient conditions for this are given in Theorem 4.
If (6) and (7) hold, then
E(|L|1{γ⩽t})<∞,t∈T,
and one can define the conditional expectation H(t) of L given that γ=t for t∈T:
H(t)=E[L|γ=t].
More precisely, H(t) is a Borel function on T with finite values such that for any t∈TE(L1{γ⩽t})=∫[0,t]H(s)dG(s).
Note that the function H is dG-a.s. unique and is dG-integrable over any closed interval in T. It is convenient to introduce a notation for such functions.
Let Lloc1(dG) be the set of all Borel functions z on T such that
∫[0,t]|z(s)|dG(s)<∞for allt∈T.
Given a function Z:[0,tG)→R, let us write Z≪locG if there is z∈Lloc1(dG) such that Z(t)=Z(0)+∫(0,t]z(s)dG(s) for all t<tG; in this case we put dZdG(t):=z(t) for 0<t<tG. Let us emphasize that in Case B this definition implies that z is dG-integrable over [0,tG] and, hence, the function Z has a finite variation over [0,tG) and there is a finite limit limt⇈tGZ(t)=Z(0)+∫(0,tG)z(s)dG(s). Note also that in this definition the value z(0) can be chosen arbitrarily even if G(0)>00$]]>; the same refers to the value z(tG) in Case B. Correspondingly, dZ/dG is defined only for 0<t<tG.
Let G be a distribution function of a law on [0,+∞]. We will say that a pair (F,H) satisfies Condition M if F:[0,tG)→R,F≪locG,H:T→R,H∈Lloc1(dG),F(t)G‾(t)+∫(0,t]H(s)dG(s)=F(0)G‾(0),t<tG, and, additionally in Case B,
limt⇈tGF(t)=H(tG).
(a) Let H be any function satisfying (12). DefineF(t)=G‾(t)−1[F(0)G‾(0)−∫(0,t]H(s)dG(s)],0<t<tG,whereF(0)is an arbitrary real number in Case A andF(0)=G‾(0)−1∫(0,tG]H(s)dG(s)in Case B. Then the pair(F,H)satisfies Condition M. Conversely, if F is such that the pair(F,H)satisfies Condition M, then F satisfies (15) and, in Case B, (16) holds.
(b) Let F be any function satisfying (11). DefineH(0)arbitrarily,H(t)=F(t)−G‾(t−)dFdG(t),0<t<tG,H(tG)arbitrarily in Case A andH(tG)=limt⇈tGF(t)in Case B. Then the pair(F,H)satisfies Condition M. Conversely, if H is such that the pair(F,H)satisfies Condition M, then H satisfies (17) and, in Case B, (18) holds.
In order that a right-continuous processM=(Mt)t∈R+be a local martingale it is necessary and sufficient that there be a pair(F,H)satisfying Condition M and a random variableL′satisfyingE(|L′|1{γ⩽t})<∞,t∈T,andE[L′|γ]=0,such that, up toP-indistinguishability,Mt=F(t)1{t<γ}+(H(γ)+L′)1{t⩾γ},
The statement that the process M given by (20) with L′=0 is a local martingale if F≪locG and H is constructed as in part (b) of Proposition 3, is essentially due to Herdegen and Herrmann [13], though they formulate (17) in an equivalent form:
H(t)=F(t−)−G‾(t)dFdG(t),0<t<tG.
They also prove that, in Case B, if F has infinite variation on [0,tG) (and hence does not satisfy F≪locG), then M given by (6) is not a semimartingale, see [13, Lemma B.6]. (Note that this follows also from our Proposition 2 (iv).) We add that, also in Case B, if H is dG-integrable over (0,tG), F satisfies (15), but F(0) is greater or less than the right-hand side of (16), then M given by (20) with L′ satisfying (19), is a supermartingale or a submartingale, respectively, cf. Theorem 4.
The fact that H(0) can be chosen arbitrarily in Proposition 3 (b) says only that L can be an arbitrary integrable random variable on the set {γ=0}, which is evident ab initio. On the contrary, the fact that F(0) can be chosen arbitrarily in (a) in Case A is an interesting feature of this model. It says that, given the terminal value M∞ of M (on {γ<∞}), one can freely choose the initial value M0 of M (on {γ>0}0\}$]]>) to keep the property of being a local martingale for M.
Every local martingaleM=(Mt)t∈R+has a unique decomposition into the sumM=M′+M″of two local martingalesM′andM″, whereM′is adapted with respect to the smallest filtration making γ a stopping time, andM″which vanishes on{t<γ}and satisfiesEM0″=0.
If P(γ=0)=0, then it follows from the first property for M″ that M0″=0 a.s. and thus the second property holds automatically.
The smallest filtration making γ a stopping time is a single jump filtration F(γ,σ{γ}) generated by γ and the smallest σ-field σ{γ} with respect to which γ is measurable. Let M be a F-local martingale adapted to F(γ,σ{γ}). It follows from Theorem 1 that M is a F(γ,σ{γ})-local martingale.
As the next example shows, the product M′M″ of local martingales from the above decomposition may not be a local martingale because the first condition in (19) may fail. It will follow from Theorem 3 below that this product is always a σ-martingale.
Let γ have an exponential distribution, e.g., G‾(t)=e−t, F is given by (15) with H(t)=t−1/2 and an arbitrary F(0), Mt′=F(t)1{t<γ}+H(γ)1{t⩾γ}, Mt″=Yγ−1/21{t⩾γ}, where Y takes values ±1 with probabilities 1/2 and is independent of γ. It follows that M′ and M″ are local martingales but their product Mt′Mt″=Yγ−11{t⩾γ} does not satisfy the integrability condition (7) and, hence, is not a local martingale. This process is a classical example (due to Émery) of a σ-martingale which is not a local martingale, see, e.g., [9, Example 2.3, p. 86].
The previous example shows that our model admits σ-martingales which are not local martingales. In the next theorem we describe all σ-martingales in our model. In particular, it implies that if F=σ{γ}, then all σ-martingales that are integrable at 0 are local martingales.
In order that a right-continuous processM=(Mt)t∈R+be a σ-martingale it is necessary and sufficient that it have a representation (20), where a pair(F,H)satisfies Condition M and a random variableL′satisfiesE[|L′|1{γ>0}|γ]<∞andE[L′|γ]=0.0\}}}\big|\gamma \big]<\infty \hspace{2em}\textit{and}\hspace{2em}\mathsf{E}[{L^{\prime }}|\gamma ]=0.\]]]>
The next theorem complements the classification of the limit behaviour of local martingales that was considered in Herdegen and Herrmann [13] in the case where F=σ{γ}. Let us say that a local martingale M=(Mt)t∈R+ has
if the limit M∞=limt→∞Mt does not exist with positive probability or exists with probability one but is not integrable: E|M∞|=∞;
if M is a closed supermartingale (in particular, E|M∞|<∞) and E(M∞)<E(M0);
if M is a closed submartingale (in particular, E|M∞|<∞) and E(M∞)>E(M0)\mathsf{E}({M_{0}})$]]>;
if M is a uniformly integrable martingale (in particular, E|M∞|<∞ and E(M∞)=E(M0)) and E(supt|Mt|)=∞;
if M has an integrable variation: E(Var(M)∞)<∞.
LetM=(Mt)t∈R+be a local martingale with the representationMt=F(t)1{t<γ}+L1{t⩾γ},t∈R+,whereL=H(γ)+L′, a pair(F,H)satisfies Condition M and a random variableL′satisfies (19). Then in Case B the local martingale M has type 4. In Case A all types are possible. Namely,
M has type 1 if and only ifE(|L′|1{γ<∞})=∞or∫[0,tG)|H(s)|dG(s)=∞.
IfP(γ=∞)>00$]]>,E(|L′|1{γ<∞})<∞, and∫R+|H(s)|dG(s)<∞then M has type 4.
IfP(γ=∞)=0,E|L′|<∞, and∫[0,tG)|H(s)|dG(s)<∞then
M has type 2a (resp., 2b) if and only iflimt⇈tGF(t)G‾(t)>00$]]> (resp.,limt⇈tGF(t)G‾(t)<0);
M has type 3 if and only iflimt⇈tGF(t)G‾(t)=0and∫[0,tG)G‾(s)|dFdG(s)|dG(s)=∞;
M has type 4 if and only if∫[0,tG)G‾(s)|dFdG(s)|dG(s)<∞.
It follows from (13) that the limit limt→tGF(t)G‾(t) in (iii.i) and (iii.ii) exists. Also, ∫[0,tG)|H(s)|dG(s) in (i)–(iii) is finite if only if F(t)G‾(t) has a finite variation over [0,tG).
It follows from Theorem 4 that, in our model, every martingale M with E(supt|Mt|)<∞ has an integrable total variation. Of course, on general spaces, there exist martingales M having finite variation on compacts and such that E(supt|Mt|)<∞ and their total variation is not integrable, see, e.g., [9, Example 2.7, p. 103].
Assume that H:(0,1)→R is a monotone nondecreasing function and, for definiteness, that it is right-continuous. Then it is the upper quantile function of H(γ), where γ is uniformly distributed on (0,1). Assume also that H is integrable on (0,1) and ∫01H(s)ds=0, that is to say, that H(γ) has zero mean. Put
F(t)=−(1−t)−1∫0tH(s)ds=(1−t)−1∫t1H(s)ds.
We see that F satisfying (13) with F(0)=0 is the Hardy–Littlewood maximal function corresponding to H. If we define M by (23) with L=H(γ), then, by Theorem 4, M is a uniformly integrable martingale with M∞=H(γ) and suptMt=F(γ). This example is essentially the example of Dubins and Gilat [7] of a uniformly integrable martingale with a given distribution of its terminal value, having maximal (with respect to the stochastic partial order) maximum (in time).
([13, Example 3.14]).
Let Ω=(0,1] be equipped with the Borel σ-field F, and let P be the Lebesgue measure, γ(ω)=ω. Put H(t)≡0. Then F(t)=(1−t)−1 satisfies (13) with F(0)=1. By Theorem 4, M defined by (23) is a supermartingale and local martingale but not a martingale. This seems to be the simplest example of a local martingale with continuous time, which is not a martingale. Note that, for ω=1, the trajectory Mt(ω)=(1−t)−11{t<1} has not a finite left-hand limit at 1. Moreover, if N is a modification of M, for t<1, the values of Mt(ω) and Nt(ω) must coincide on the atom {t<γ}=(t,1] of Ft, having the positive measure. Hence, Nt(ω)=Mt(ω) for ω=1 for all t<1. This is an example of a right-continuous supermartingale which has not a modification with all paths càdlàg. Of course, the usual assumptions are not satisfied in this example.
Proofs
(i) and (iii) are evident from the definition of Ft, and (ii) follows easily from (i).
Let us prove (iv). To prove that T is a stopping time, we must check that {T⩽t<γ} is either ∅ or {t<γ} for all t∈R+. This is trivial if {T<γ}=∅. If there is a number r such that (4) holds, then {T⩽t<γ} is either ∅ if r>tt$]]> or {t<γ} if r⩽t.
Conversely, let T be a stopping time. If T⩾γ for all ω, then there is nothing to prove. Assume that the set {T<γ}≠∅. Then there are real numbers q such that {T⩽q<γ}≠∅. For such q, by the definition of Fq, {T⩽q<γ}={q<γ}, or, equivalently, {T⩽q}⊇{q<γ}. Let r be the greatest lower bound of such q. The sets {q<γ}↑{r<γ} and {T⩽q}↓{T⩽r} as q↓r. Thus,
{T<γ}=⋃q:{T⩽q<γ}≠∅{q<γ}={r<γ}⊆{T⩽r}.
Since {T⩽t<γ}=∅ for any t<r, we have (4). □
The first statement in (i) follows from Proposition 1 (i). Since P(t<γ∧tG)>00$]]> for every t<tG, we obtain that Xt and Yt take the same constant value on {t<γ∧tG}.
Since a random variable Yt is Ft−-measurable for a predictable process Y, Yt is constant on {t⩽γ}. Denote by C(t), t∈T, the value of Yt on {t⩽γ}. Since P(γ⩾t)>00$]]> for t∈T, there is an ω such that C(s)≡Ys(ω), s⩽t, and the measurability of C follows.
Let us prove (iii) in Case B. Then we obtain that Xt=F(t) for all t<tG on the set {γ=tG}, which has a positive probability. However, almost all paths of Xt have a finite variation over [0,tG), and the claim follows. The proof in Case A is similar.
Now let us prove (5) in the case where M=(Mt)t∈R+ is a uniformly integrable (a.s. càdlàg) martingale. We can find a random variable M∞ that is F∞-measurable and such that limn→∞Mn=M∞P-a.s. Since {t<γ} is an atom of Ft and has a positive probability for t<tG, we obtain from the martingale property that Mt(ω)=F(t) for all ω∈{t<γ}, where
F(t)=E(M∞1{t<γ})G‾(t),t<tG.
It is clear that the nominator and the denominator are right-continuous functions of bounded variation on [0,tG], hence F(t), 0⩽t<tG is a càdlàg function on T and has a finite variation on [0,tG) in Case B and on every [0,t], 0⩽t<tG, in Case A.
Now set L=M∞1{γ<∞}. Then L1{γ⩽t}=M∞1{γ⩽t} is Ft-measurable, and hence P-a.s.
Mt1{γ⩽t}=E(M∞1{γ⩽t}|Ft)=L1{γ⩽t}.
Thus we have obtained, that, for a given t∈R+, Mt is equal P-a.s. to the right-hand side of (5) with L and F(t) as above. Since both the left-hand side and the right-hand side of (5) are almost surely right-continuous, they are indistinguishable. Moreover, if we change F(t) for t⩾tG, the right-hand side of (5) will change on an evanescent set. Thus we can put, say, F(t)=0 for t⩾tG, and then the right-hand side of (5) is a regular right-continuous process with finite variation, and indistinguishable from M.
Now let M be a local martingale and {Tn} be a localizing sequence of stopping times, i.e. Tn↑∞ a.s. and MTn is a uniformly integrable martingale for each n. We have proved that almost all paths of MTn have finite variation. It follows that almost all paths of M have finite variation. This proves (iv).
Next, let M be a σ-martingale, i.e. M is a semimartingale and there is an increasing sequence of predictable sets Σn such that ∪nΣn=Ω×R+ and the integral process 1Σn·M is a uniformly integrable martingale for every n. It does not matter if we integrate over [0,t] or (0,t], so let us agree that the domain of integration does not include 0. Since the integrand is bounded and every semimartingale is a process with finite variation in our model, the integral can be considered in the Lebesgue–Stieltjes sense, as well as other integrals appearing in the proof. Since 1Σn·M is stopped at γ for every n with probability one, we have
∫1⟧γ,∞⟦∩Σn(t)dVar(M)t=∫1⟧γ,∞⟦(t)dVar(1Σn·M)t=0P-a.s.
for every n, therefore,
∫1⟧γ,∞⟦(t)dVar(M)t=0P-a.s.
Combining with (i), we prove representation (5). □
As it was already explained in the introduction, we can prove directly, without assuming that paths are a.s. càdlàg, that any uniformly integrable martingale has a regular modification. The proof is essentially the same as above where we proved that a.s. càdlàg uniformly integrable martingale has representation (5).
First, we prove that statements (ii) and (iii) are equivalent. The implication (ii)⇒(iii) follows trivially from the definition of a martingale. Conversely, let (iii) hold. The process (Mt)t∈T is right-continuous, adapted by Proposition 1 (i), and integrable, see (7). Moreover, due to (6),
Mt−Ms=0on{s⩾γ},
where 0⩽s<t∈T. Hence,
E[Mt−Ms|Fs]=0on{s⩾γ}.
But E[Mt−Ms|Fs] is Fs-measurable and, thus, equals a constant on {s<γ}. And this constant must be zero since E(Mt−Ms)=0 by (8).
The implication (ii)⇒(i) is trivial if tG=∞ or tG∈T. So we assume that tG<∞ and G‾(t)↓0 as t⇈tG. Let t1<⋯<tn<⋯<tG, tn→tG, be an increasing sequence, then G‾(tn)→0. Put
Tn=tn,ifγ>tn;+∞,otherwise.{t_{n}}\text{;}\\ {} +\infty ,& \text{otherwise.}\end{array}\right.\]]]>
Then Tn is a stopping time by Proposition 1 (iv), Tn↑∞ a.s., and Mt∧Tn=Mt∧tn. Hence, MTn is a martingale and M is a local martingale.
It remains to prove the implication (i)⇒(ii). Let M=(Mt)t∈R+ be a local martingale with a localizing sequence {Tn}, i.e. Tn↑∞ a.s. and MTn is a uniformly integrable martingale for each n. If P(Tn⩾γ)=1 for some n, then M=MTn is a uniformly integrable martingale, and there is nothing to prove. So assume that P(Tn<γ)>00$]]> for all n. By Proposition 1 (iv), there is a number rn such that {Tn<γ}={Tn=rn<γ}={rn<γ}. It follows from P(rn<γ)>00$]]> that rn<tG. In Case B we get P(Tn<γ)=P(rn<γ)⩾P(γ=tG)>00$]]> for every n, a contradiction with Tn→∞ a.s. In Case A, if P(γ=∞)>00$]]>, then it follows from Tn→∞ a.s. that rn→∞. In remaining cases where P(γ=tG)=0, we obtain from P(rn<γ)→0 that rn→tG, n→∞. The claim follows since Mt∧Tn=Mt∧rn, and hence (Mt)t⩽rn is a martingale. □
(a) It is obvious that (13) is equivalent to (15). It also follows from (13) that in Case B (14) is equivalent to (16). Thus it remains to prove that F defined in (a) satisfies F≪locG. Since G‾(s)⩾G‾(t)>00$]]> for any s<t<tG, we have
1G‾(t)=1G‾(0)+∫(0,t]1G‾(s)G‾(s−)dG(s),t<tG.
On the other hand, from (15)
F(t)G‾(t)=F(0)G‾(0)−∫(0,t]H(s)dG(s),t<tG.
Combining, we obtain from integration by parts that
F(t)=F(t)G‾(t)1G‾(t)=F(0)−∫(0,t]H(s)G‾(s−)dG(s)+∫(0,t]F(s)G‾(s−)dG(s),t<tG.
This shows that F≪locG in Case A. In Case B we must show additionally that the function |F(s)|+|H(s)|G‾(s−) is dG-integrable over (0,tG). But 1/G‾(s−)⩽1/P(γ=tG), s⩽tG, and F(s) is bounded on [0,tG) in view of (15). The claim follows.
(b) It is clear that the function H(t), t∈T, defined as in the statement, belongs to Lloc1(dG). Integrating by parts, we get, for t∈[0,tG),
F(t)G‾(t)=F(0)G‾(0)−∫(0,t]F(s)dG(s)+∫(0,t]G‾(s−)dF(s)=F(0)G‾(0)−∫(0,t]F(s)dG(s)+∫(0,t]G‾(s−)dFdG(s)dG(s)=F(0)G‾(0)−∫(0,t]H(s)dG(s),
i.e. (13) holds. Therefore, Condition M is satisfied. Conversely, let (13) hold. In the proof of part (a) we deduced from (15) (and, hence, from (13)) that
dFdG(t)=−H(t)G‾(t−)+F(t)G‾(t−),dG-a.s.,
and (17) follows. □
Let M=(Mt)t∈R+ be a local martingale. By Proposition 2 (v) and Theorem 1, M has representation (6) and, moreover, (7) and (8) hold. Define the function H(t), t∈T, by (10). Then, see (9),
E(|H(γ)|1{γ⩽t})⩽E(|L|1{γ⩽t})<∞,
which implies H∈Lloc1(dG). Putting L′=(L−H(γ))1{γ<∞}, we obtain (19) as well. Now it follows from (20) and the second relation in (19) that
E(Mt)=F(t)G‾(t)+∫[0,t]H(s)dG(s),t∈T,
so (13) and (14) follow from (8). Finally, (11) follows from Proposition 3 (a).
Conversely, let (20) hold true with a pair (F,H) satisfying Condition M and a random variable L′ satisfying (19). Then, putting L=H(γ)+L′, we obtain (6) and (7). It remains to note that (13) and (14) (in case B) imply (8). □
The required decomposition follows from (20) if we put
Mt′=F(t)1{t<γ}+H(γ)1{t⩾γ},Mt″=L′1{t⩾γ}.
Let a local martingale M with a representation (20) vanish on {t<γ} and EM0=0. Then F(t)≡0 for t<tG and 0=EM0=H(0)G‾(0)+E(L′1{γ=0})=H(0)G‾(0) in view of the second relation in (19). By Theorem 2, it follows from (13) and (14) that H(t)=0dG-a.s. Now, if M is also adapted with respect to the smallest filtration making γ a stopping time, then Mγ=(H(γ)+L′)1{γ<∞}=L′1{γ<∞} is σ{γ}-measurable. Using again the second relation in (19), we conclude that L′1{γ<∞}=0 a.s. This proves the unicity. □
To prove sufficiency it is enough to consider the case, where H≡0 and F≡0. In view of the first condition in (22), there exists a Borel function J:(0,∞]→R+ such that E[|L′||γ=t]=J(t). Put Σn=Ω×{t∈(0,+∞):J(t)⩽n} and consider the Lebesgue–Stieltjes integral process 1Σn·Mt=L′1{E[|L′||γ]⩽n}1{t⩾γ>0}0\}}}$]]>. By Theorem 2, cf. condition (19), it is a local martingale. Since Σn are predictable and ∪nΣn=Ω×R+, M is a σ-martingale.
Conversely, let M be a σ-martingale. It is easy to check that to prove necessity it is enough to consider the case M0=0. According to Proposition 2 (v) and Remark 1Mt=F(t)1{t<γ}+L1{t⩾γ},
where L is a random variable, F(t), 0⩽t<tG, is a deterministic function with finite variation over [0,t] for every t<tG in case A and over [0,tG) in Case B. By the definition of σ-martingales, there is an increasing sequence of predictable sets Σn such that ∪nΣn=Ω×R+ and the integral process 1Σn·M is a local martingale for every n. It was mentioned in the proof of Proposition 2 that the integral is understood as the Lebesgue–Stieltjes integral. By Proposition 2 (ii), there are Borel subsets Dn of R+ such that 1Σn(ω,t)1{γ(ω)⩾t}=1Dn(t)1{γ(ω)⩾t}, in particular,
∪nDn⊇T.
According to Theorem 2 and Proposition 3, 1Σn·M has a representation
1Σn·Mt=Fn(t)1{t<γ}+(Hn(γ)+Ln)1{t⩾γ}
where E(|Ln|1{γ⩽t})<∞, t∈T, E[Ln|γ]=0, Hn∈Lloc1(dG), Fn≪locG,
Hn(t)=Fn(t)−G‾(t−)dFndG(t)=Fn(t−)−G‾(t)dFndG(t),0<t<tG, and in Case B Hn(tG):=limt⇈tGFn(t).
Combining with (25), we get
Fn(t)=∫(0,t]1Dn(s)dF(s),0<t<tG,
and
Hn(γ)+Ln=∫(0,γ)1Dn(s)dF(s)+1Dn(γ)(L−F(γ−))a.s.
Since Fn≪locG, it follows from (28) and (26) that F≪locG. Substituting (28) in (29) and taking conditional expectation given γ, we get
Hn(γ)−Fn(γ−)=1Dn(γ)(H(γ)−F(γ−))a.s.,
i.e.
Hn(t)−Fn(t−)=1Dn(t)(H(t)−F(t−))dG(t)-a.s.
It follows from (28) and (27) that
−G‾(t)1Dn(t)dFdG(t)=−G‾(t)dFndG(t)=1Dn(t)(H(t)−F(t−))dG(t)-a.s.,
so, taking (26) into account, we obtain
H(t)=F(t)−G‾(t−)dFdG(t)dG(t)-a.s.
Additionally, in Case B, the left-hand side of (30) at t=tG vanishes, hence, H(tG):=limt⇈tGF(t). It remains to put L′=L−H(γ). □
In Case B
Var(M)∞⩽2Var(F)tG−+|L|,
and the first term is finite by Remark 1, while E|L|<∞ due to (7). Thus, we proceed to Case A.
(i) Note that
∫[0,tG)|H(s)|dG(s)=E(|H(γ)|1{γ<tG})=E(|H(γ)|1{γ<∞})
and E(|L|1{γ<∞})<∞ if and only if
E(|H(γ)|1{γ<∞})<∞andE(|L′|1{γ<∞})<∞.
Next, if M∞ is well defined, then
M∞=L1{γ<∞}+limt→∞F(t)1{γ=∞}.
Finally, if P(γ=∞)>00$]]>, then it follows from (13) that limt→∞F(t) exists and is finite if ∫[0,tG)|H(s)|dG(s)<∞. Now, combining all above, we arrive at (i).
(ii) If P(γ=∞)>00$]]>, then
Var(M)∞⩽2Var(F)∞+|L|1{γ<∞},
and the last term on the right has finite expectation by assumptions. Since G‾(t)⩾P(γ=∞)>00$]]> in the case under consideration, it follows from assumptions and (13) that F has a finite variation over R+.
From now on we assume that E(|L′|1{γ<∞})<∞, ∫[0,tG)|H(s)|dG(s)<∞, and P(γ=tG)=0. Then M is a martingale on [0,tG) by Theorem 1 and it coincides with M∞=L for t⩾tG. Hence, it is a (necessarily closed) submartingale (resp. supermartingale) if and only if E[L−Mt|Ft]⩾0 (resp. ⩽0), for t<tG. As in the proof of Theorem 1,
L−Mt=0on{t⩾γ},
hence,
E[L−Mt|Ft]=const1{t<γ}.
Taking expectations, we see that this constant has the same sign as E(L−Mt)=E(L−M0). However,
E(L−M0)=E(H(γ)−M0)=∫(0,tG)H(s)dG(s)−F(0)G‾(0)=−limt⇈tGF(t)G‾(t),
and (iii.i) follows.
The same proof shows that if ∫(0,tG)H(s)dG(s)=F(0)G‾(0), then M is a uniformly integrable martingale. Therefore, to prove (iii.ii) and (iii.iii) it is enough to show that E(supt|Mt|)<∞ implies (24), and that (24) implies E(Var(M)∞)<∞.
If M is a local martingale with E(supt|Mt|)<∞, then E(|ΔMγ|1{γ<∞})<∞. But |ΔMγ|1{γ<∞}=|L−F(γ−)|1{γ<∞}, hence, taking conditional expectation given γ, we get
∫[0,tG)|H(s)−F(s−)|dG(s)<∞.
In view of (21) which is equivalent to (17), we obtain (24).
Conversely, let (24) hold. Then
Var(M)∞=|L|1{γ=0}+|F(0)|1{γ>0}+∫(0,γ)|dFdG(s)|dG(s)+|L−F(γ−)|1{0<γ<∞}⩽2|L|1{γ<∞}+2|F(0)|1{γ>0}+2∫(0,γ)|dFdG(s)|dG(s)0\}}}+\underset{(0,\gamma )}{\int }\Big|\frac{dF}{dG}(s)\Big|\hspace{0.1667em}dG(s)+|L-F(\gamma -)|{\mathbb{1}_{\{0<\gamma <\infty \}}}\\ {} & \hspace{1em}\leqslant 2|L|{\mathbb{1}_{\{\gamma <\infty \}}}+2|F(0)|{\mathbb{1}_{\{\gamma >0\}}}+2\underset{(0,\gamma )}{\int }\Big|\frac{dF}{dG}(s)\Big|\hspace{0.1667em}dG(s)\end{aligned}\]]]>
and
E(∫(0,γ)|dFdG(s)|dG(s))=∫[0,tG)∫(0,u)|dFdG(s)|dG(s)dG(u)=∫[0,tG)|dFdG(s)|∫(s,tG)dG(u)dG(s)=∫[0,tG)G‾(s)|dFdG(s)|dG(s)<∞.
□
It follows from the last equalities in the proof that, due to (11) and (21) respectively, (24) implies that the following integrals are also finite:
∫[0,tG)|F(s−)|dG(s)<∞and∫[0,tG)|H(s)|dG(s)<∞.
However, it may happen that
∫[0,tG)|F(s)|dG(s)=∞,
see an example in [13, Remark 3.11].
ComplementsSingle jump processes and their compensators
Let us consider the same setting as in Section 2 and let V be a finite random variable. For simplicity, we assume that {γ=0}⊆{V=0}. Then
Xt=V1{t⩾γ}
is an adapted process of finite variation on compact intervals.
The processX=(Xt)R+is of locally integrable variation if and only ifE(|V|1{γ⩽t})<∞,t∈T.
Let (31) hold. If tG∈T, then E(|V|1{γ⩽tG})<∞ means that the process X itself has integrable variation. In Case A, put
Tn=tn,ifγ>tn;+∞,otherwise.{t_{n}}\text{;}\\ {} +\infty ,& \text{otherwise.}\end{array}\right.\]]]>
where tn⇈tG. Then Tn↑∞ a.s. and Var(XTn)∞=|V|1{γ⩽Tn}=|V|1{γ⩽tn}.
Conversely, let {Tn} be a localizing sequence of stopping times such that E(|V|1{γ⩽Tn})<∞. If P(γ⩽Tn)=1 for n large enough, then V is integrable. So assume that P(γ>Tn)>0{T_{n}})>0$]]> for every n. By Proposition 1 (iv), there are numbers rn such that {Tn<γ}={Tn=rn<γ}={rn<γ}. Thus, we have a sequence rn such that E(|V|1{γ⩽rn})<∞. Since Tn→∞ a.s. and the sequence {Tn} is increasing, in Case A it follows that rn↑tG, and in Case B we come to a contradiction by repeating the arguments in the concluding part of the proof of Theorem 1. □
From now on we will assume that X is a process of locally integrable variation, i.e. (31) holds. Our aim is to find its compensator. We can introduce a function K similarly as the function H is introduced in (10):
K(t)=E[V|γ=t],t∈T.
It is clear that K∈Lloc1(dG) and K(0)=0 if P(γ=0)>00$]]>. Now define
F(t)=∫(0,t]G‾(s−)−1K(s)dG(s),0⩽t<tG,
in particular, F(0)=0. It follows that, in Case B, the function F has a bounded variation on [0,tG) and has a finite limit as t⇈tG, so we put
F(tG)=limt⇈tGF(t).
The next theorem takes its origin in [4], where the case when V=1, γ is finite and tG=+∞ is considered.
Let V be a random variable satisfying (31), K and F defined in (32)–(34). Then the compensatorAtof the processXt=V1{t⩾γ}is given byAt=F(t∧γ)in Case AandAt=F(t∧γ)+K(tG)1{γ⩾tG}1{t⩾tG}in Case B.
The process t∧γ is adapted and continuous, hence, it is predictable. It follows that F(t∧γ) is predictable. Next, in Case B, the set {(ω,t):γ(ω)⩾tG,t⩾tG} coincides with the intersection of predictable sets
{γ⩾tG}×[tG,∞)=⋂n[{γ>tG−n−1}×(tG−n−1,∞)],{t_{G}}-{n^{-1}}\}\times ({t_{G}}-{n^{-1}},\infty )\Big],\]]]>
therefore, A is predictable. Hence it is enough to show that M=A−X is a local martingale.
We use Theorem 2 and Proposition 3 (b). M has the representation (6) with the same function F and L=F(γ)1{γ<∞}−V1{γ<∞}+K(tG)1{γ=tG<∞}. Define the function H as in Proposition 3 (b). Then it follows from (33) that H(t)=F(t)−K(t), 0<t<tG, and, in Case B, H(tG)=F(tG). On the other hand, we have E[L|γ=t]=F(t)−K(t)=H(t), 0<t<tG, and, in Case B, E[L|γ=tG]=F(tG)−K(tG)+K(tG)=F(tG)=H(tG). The claim follows. □
Example: submartingales of class (Σ)
Recall, see [22], that a nonnegative submartingale X=(Xt)t∈R+ is called a submartingale of class (Σ) if X0=0 and it can be decomposed as Xt=Nt+At, where N=(Nt)t∈R+, N0=0, is a local martingale, A=(At)t∈R+, A0=0, is a continuous increasing process, and the measure (dAt) is carried by the set {t:Xt=0}. A typical example is a process Xt=L‾t−Lt which is the difference between the running maximum L‾t of a continuous local martingale (Lt) and Lt itself.
Let X=(Xt)t∈R+ be a nonnegative submartingale with the Doob–Meyer decomposition Xt=Nt+At, where N0=A0=0, N is a local martingale, A is a predictable increasing process. Assume that A∞<∞ a.s. and put Ct=inf{s⩾0:As>t}t\}$]]>. Then, see [10, Lemma 3.1], X is of class (Σ) if and only if a.s.
ACt=A∞∧tandXCt=X∞1{t⩾A∞},
where a finite limit X∞:=limt→∞Xt exists a.s. by [10, Proposition 3.1]. Therefore, the process Mt=−NCt has the representation
Mt=t∧γ−V1{t⩾γ},whereγ=A∞andV=X∞.M may not be a local martingale. For example, take as L a Brownian motion stopped when it hits 1 and define X=L‾−L, then Mt=t∧1. However, if X is a submartingale of class (D) then N is a uniformly integrable martingale and M is also a uniformly integrable martingale (with respect to its own filtration and, by Theorem 1, with respect to the single jump filtration generated by γ on an original space). Now we can define a function K according to (32) and conclude that (33) is valid with F(t)=t. We may interpret (33) as the equation with known K and unknown G. This identity says that the Lebesgue measure on [0,tG) is absolutely continuous with respect to dG but not vice versa. However, if the function K(t) does not vanish (dG-a.s.) then we obtain from (33) that dG is equivalent to the Lebesgue measure on T, in particular, G is continuous, and
P(γ>t)=exp(−∫0tdtK(t)),t<tG.t)=\exp \Big(-{\underset{0}{\overset{t}{\int }}}\frac{dt}{K(t)}\Big),\hspace{1em}t<{t_{G}}.\]]]>
This statement coincides with Theorem 4.1 in [22]. If the function K(t) may vanish, analysis of equation (33) with F(t)=t, known K(t) and unknown G(t) is done in [23].
A kind of a converse statement is proved in [11]. If, say, a martingale M satisfies
Mt=t∧γ−V1{t⩾γ},
where γ<∞ and V⩾0, then, using Monroe’s theorem [20], we prove that there is a Brownian motion B and a finite stopping time T such that, for the stopped process L=BT, the joint law of its terminal value L∞ and its maximum L‾∞ coincides with that of M, that is, with the law of (γ−V,γ). In particular, this shows that a distribution function G is the law of the maximum of a uniformly integrable continuous martingale L with L0=0 if and only if, with F(t)=t, 0⩽t<tG, we have F≪locG, ∫[0,tG)|H(s)|dG(s)<∞, where H is defined by (17), and G(t)=o(t−1), see conditions for M to have type 3 or 4 in Theorem 4. This gives an alternative proof of the main result in [23].
Acknowledgments
We thank three anonymous referees for careful reading of the paper and constructive comments and suggestions for improving the presentation. A special thanks goes to the referee who suggested Theorem 3 on a characterisation of σ-martingales.
ReferencesBoel, R., Varaiya, P., Wong, E.: Martingales on jump processes. I. Representation results. 13(5), 999–1021 (1975). MR0400379. https://doi.org/10.1137/0313063Chou, C.-S., Meyer, P.-A.: Sur la représentation des martingales comme intégrales stochastiques dans les processus ponctuels. In: . Lecture Notes in Math., vol. 465, pp. 226–236. Springer (1975). MR0436310. https://doi.org/10.1007/BFb0102993Davis, M.H.A.: The representation of martingales of jump processes. 14(4), 623–638 (1976). MR0418221. https://doi.org/10.1137/0314041Dellacherie, C.: Un exemple de la théorie générale des processus. In: . Lecture Notes in Math., vol. 124, pp. 60–70. Springer (1970). MR0263157. https://doi.org/10.1007/BFb0059333Dellacherie, C., Meyer, P.-A.: . North-Holland Mathematics Studies, vol. 29, p. 189. North-Holland Publishing Co., Amsterdam-New York (1978). MR521810Dellacherie, C., Meyer, P.-A.: . North-Holland Mathematics Studies, vol. 72, p. 463. North-Holland Publishing Co., Amsterdam (1982). MR745449Dubins, L.E., Gilat, D.: On the distribution of maxima of martingales. 68(3), 337–338 (1978). MR0494473. https://doi.org/10.1090/S0002-9939-1978-0494473-4Elliott, R.J.: Stochastic integrals for martingales of a jump process with partially accessible jump times. 36(3), 213–226 (1976). MR0420846. https://doi.org/10.1007/BF00532546Gushchin, A.A.: , p. 185. ISTE Press, London; Elsevier Ltd, Oxford (2015). MR3410512Gushchin, A.A.: The joint law of terminal values of a nonnegative submartingale and its compensator. 62(2), 216–235 (2018). MR3649035. https://doi.org/10.1137/S0040585X97T988575Gushchin, A.A.: The joint law of a max-continuous local submartingale and its maximum. Theory Probab. Appl. 65 (2020)He, S.W.: Some remarks on single jump processes. In: . Lecture Notes in Math., vol. 986, pp. 346–348. Springer (1983). MR0770423. https://doi.org/10.1007/BFb0068327Herdegen, M., Herrmann, S.: Single jump processes and strict local martingales. 126(2), 337–359 (2016). MR3434986. https://doi.org/10.1016/j.spa.2015.09.003Jacod, J., Skorohod, A.V.: Jumping filtrations and martingales with finite variation. In: . Lecture Notes in Math., vol. 1583, pp. 21–35. Springer (1994). MR1329098. https://doi.org/10.1007/BFb0073831Jacod, J.: Multivariate point processes: predictable projection, Radon-Nikodým derivatives, representation of martingales. 31, 235–253 (1975). MR0380978. https://doi.org/10.1007/BF00536010Jacod, J.: Un théorème de représentation pour les martingales discontinues. 34(3), 225–244 (1976). MR0418222. https://doi.org/10.1007/BF00532705Jacod, J.: . Lecture Notes in Mathematics, vol. 714, p. 539. Springer (1979). MR542115Jeanblanc, M., Rutkowski, M.: Modelling of default risk: an overview. In: , pp. 171–269. Higher Education Press, Beijing (2000)Jeanblanc, M., Yor, M., Chesney, M.: . Springer Finance, p. 732. Springer (2009). MR2568861. https://doi.org/10.1007/978-1-84628-737-4Monroe, I.: On embedding right continuous martingales in Brownian motion. 43(4), 1293–1311 (1972). MR0343354. https://doi.org/10.1214/aoms/1177692480Neveu, J.: Processus ponctuels. In: , pp. 249–445. Springer (1977). MR0474493. https://doi.org/10.1007/BFb0097494Nikeghbali, A.: A class of remarkable submartingales. 116(6), 917–938 (2006). MR2254665. https://doi.org/10.1016/j.spa.2005.12.003Vallois, P.: Sur la loi du maximum et du temps local d’une martingale continue uniformement intégrable. 69(2), 399–427 (1994). MR1281971. https://doi.org/10.1112/plms/s3-69.2.399