This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Single jump filtrations and local martingales

A. A.Alexander A. Gushchinlabel=e1][email protected]\orcid0000-0002-0020-7496 [ \institutionSteklov Mathematical Institute, Gubkina 8, 119991 Moscow, \cnyRussia \institutionNational Research University Higher School of Economics,Pokrovsky Boulevard 11, 109028 Moscow, \cnyRussia
(2020; \sday24 9 2019; \sday30 4 2020; \sday1 5 2020)
Abstract

A single jump filtration (t)t+({\mathscr{F}}_{t})_{t\in\mathbb{R}_{+}} generated by a random variable γ\gamma with values in ¯+\overline{\mathbb{R}}_{+} on a probability space (Ω,,𝖯)(\Omega,{\mathscr{F}},\mathsf{P}) is defined as follows: a set AA\in{\mathscr{F}} belongs to t{\mathscr{F}}_{t} if A{γ>t}A\cap\{\gamma>t\} is either \varnothing\varnothing or {γ>t}\{\gamma>t\}. A process MM is proved to be a local martingale with respect to this filtration if and only if it has a representation Mt=F(t)\mathbh1{t<γ}+L\mathbh1{t\geqslantγ}M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+L{\mathbh{1}}_{\{t\geqslant\gamma\}}, where FF is a deterministic function and LL is a random variable such that 𝖤|Mt|<\mathsf{E}|M_{t}|<\infty and 𝖤(Mt)=𝖤(M0)\mathsf{E}(M_{t})=\mathsf{E}(M_{0}) for every t{t+:𝖯(γ\geqslantt)>0}t\in\{t\in\mathbb{R}_{+}\colon{\mathsf{P}}(\gamma\geqslant t)>0\}. This result seems to be new even in a special case that has been studied in the literature, namely, where {\mathscr{F}} is the smallest σ\sigma-field with respect to which γ\gamma is measurable (and then the filtration is the smallest one with respect to which γ\gamma is a stopping time). As a consequence, a full description of all local martingales is given and they are classified according to their global behaviour.

Filtration,
local martingale,
processes with finite variation,
σ\sigma-martingale,
stopping time,
60G44,
60G07,
doi:
10.15559/20-VMSTA153
keywords:
keywords:
[MSC2010]
volume: 7issue: 2articletype: research-article
\aid

VMSTA153 stmry"71 stmry"79 \DeclareMathOperator\VarVar \startlocaldefs \urlstylerm \allowdisplaybreaks \endlocaldefs

\pretitle

Research Article

\publishedonline\sday

25 5 2020

1 Introduction

Starting with Dellacherie [Dellacherie:70], the following simple model has been studied and intensively used in applications. Given a random variable γ\gamma with positive values on a probability space (Ω,,𝖯)(\Omega,{\mathscr{F}},\mathsf{P}), one considers the smallest filtration with respect to which γ\gamma is a stopping time (or, equivalently, the process \mathbh1{t\geqslantγ}{\mathbh{1}}_{\{t\geqslant\gamma\}} is adapted). In particular, Dellacherie gives a formula for the compensator of this single jump process \mathbh1{t\geqslantγ}{\mathbh{1}}_{\{t\geqslant\gamma\}}. Chou and Meyer [ChouMeyer:1975] describe all local martingales with respect to this filtration and prove a martingale representation theorem. A significant contribution is done in a recent paper by Herdegen and Herrmann [HerdegenHerrmann:16], where a classification, whether a local martingale in this model is a strict local martingale, or a uniformly integrable martingale, etc., is given. Let us also mention some related papers [BoelVaraiyaWong, jacod1975multivariate, Jacod1976, Davis:1976, Elliott:1976, Neveu:1977, He:1983], where, in particular, local martingales with respect to the filtrations generated by jump processes or measures of certain kind are studied.

Let us clarify that in the above model every local martingale has the form

Mt=F(t)\mathbh1{t<γ}+H(γ)\mathbh1{t\geqslantγ},M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+H(\gamma){\mathbh{1}}_{\{t\geqslant\gamma\}}, (1)

or

Mt=F(tγ)K(γ)\mathbh1{t\geqslantγ},M_{t}=F(t\wedge\gamma)-K(\gamma){\mathbh{1}}_{\{t\geqslant\gamma\}},

where γ\gamma is a random variable with values in, say, (0,+)(0,+\infty), FF, HH, and K=FHK=F-H are deterministic functions. Denote by GG the distribution function of γ\gamma, G¯(t)=1G(t){\overline{G}}(t)=1-G(t), tG=sup{t:G(t)<1}t_{G}=\sup\{t\colon G(t)<1\} is the right endpoint of the distribution of γ\gamma. Assume that 𝖤|Mt|<\mathsf{E}|M_{t}|<\infty, then

𝖤(Mt)=F(t)G¯(t)+[0,t]H(s)dG(s),{\mathsf{E}}(M_{t})=F(t){\overline{G}}(t)+\int_{[0,t]}H(s)\,dG(s),

where the corresponding Lebesgue–Stieltjes integral is finite. If (Mt)(M_{t}) is a martingale, then 𝖤(Mt)=𝖤(M0)\mathsf{E}(M_{t})=\mathsf{E}(M_{0}), and this equality can be written as

F(t)G¯(t)+[0,t]H(s)dG(s)=F(0)F(t){\overline{G}}(t)+\int_{[0,t]}H(s)\,dG(s)=F(0) (2)

and can be viewed as a functional equation concerning one of functions in (F,G,H)(F,G,H) or (F,G,K)(F,G,K), where other two functions are assumed to be given. In fact, this equation takes place for t<tGt<t_{G} or t\leqslanttGt\leqslant t_{G}, the latter in the case where tG<t_{G}<\infty and 𝖯(γ=tG)>0\mathsf{P}(\gamma=t_{G})>0. Moreover, it turns out that this is not only the necessary condition but also the sufficient one for (Mt)t+(M_{t})_{t\in\mathbb{R}_{+}} given by \eqrefrepr to be a local martingale. This consideration allows us to reduce problems to solving this functional equation. For example, to find the compensator F(tγ)F(t\wedge\gamma) of \mathbh1{t\geqslantγ}{\mathbh{1}}_{\{t\geqslant\gamma\}} as in [Dellacherie:70] one needs to find a solution FF given GG and K1K\equiv 1. A possible way to explain the idea in [ChouMeyer:1975] is the following: The terminal value MM_{\infty} of any local martingale MM in this model is represented as H(γ)H(\gamma), and to find a representation \eqrefrepr for MM it is enough to solve the equation for FF given GG and HH; the linear dependence between HH and FF results in a representation theorem. Contrariwise, in [HerdegenHerrmann:16] the authors suggest to find HH from the equation for given FF and GG. This allows them to study global properties of MM.

In this paper we consider a more general model, where all randomness appears “at time γ\gamma” but it may contain much more information than γ\gamma does. We start with a random variable γ\gamma on a probability space (Ω,,𝖯)(\Omega,{\mathscr{F}},\mathsf{P}), and define a single jump filtration (t)({\mathscr{F}}_{t}) in such way that nothing happens strictly before γ\gamma, γ\gamma is a stopping time with respect to it, and the σ\sigma-field γ{\mathscr{F}}_{\gamma} of events that occur before or at time γ\gamma coincides with {\mathscr{F}} (in fact, on the set {γ<}\{\gamma<\infty\}). We prove that every local martingale with respect to this filtration has the representation

Mt=F(t)\mathbh1{t<γ}+L\mathbh1{t\geqslantγ},M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+L{\mathbh{1}}_{\{t\geqslant\gamma\}}, (3)

where now LL is a random variable which is not necessarily a function of γ\gamma. However, denoting H(t)=𝖤[L|γ=t]H(t)=\mathsf{E}[L|\gamma=t], we come to the same functional equation of type \eqrefe:fe.

Some results of the paper can be deduced from known results for marked point processes, at least if {\mathscr{F}} is countably generated; this applies, for example, to Theorem LABEL:th:incr about the compensator of a single jump process. Another example is Corollary 1 which says that every local martingale is the sum of a local martingale of form \eqrefrepr and an “orthogonal” local martingale, the latter being characterised, essentially, by the property F(t)0F(t)\equiv 0. The reader can recognize in this decomposition the representation of a local martingale as the sum of two stochastic integrals with respect to random measures, see [Jacod1976] and [Jacod1979]. However, our direct proofs are simpler due to the key feature of our paper. Namely, we obtain a simple necessary and sufficient condition for a process to be a local martingale and later exploit it. A description of all local martingales via a full description of all possible solutions to a functional equation of type \eqrefe:fe is a simple consequence of this necessary and sufficient condition. In particular, an absolute continuity type property of FF with respect to GG, considered as an assumption in [HerdegenHerrmann:16], is proved to be a necessary condition. An elementary analysis of a functional equation of type \eqrefe:fe shows that, if γ\gamma has no atom at its right endpoint, there are different FF satisfying the equation for given HH and GG. In particular, there is a local martingale MM such that M0=1M_{0}=1 and M=0M_{\infty}=0; MM is necessarily a closed supermartingale.

Another important feature of our model, in contrast to Dellacherie’s model, is that it admits σ\sigma-martingales which are not local martingales.

Let us also mention some other papers where processes of form \eqrefrepr or \eqrefrepr2 are considered. Processes of form \eqrefrepr with tG=t_{G}=\infty are typical in the modelling of credit risk, see, e.g., [JeanblancRutkowski:2000] and [JeanblancYorChesney:2009, Chapter 7], where usually FF is expressed via GG and one needs to find HH. Since tG=t_{G}=\infty, such a process is a martingale. For example, in the simplest case F=1/G¯F=1/{\overline{G}} and hence H=0H=0. This process is the same that is mentioned in two paragraphs above. Single jump filtrations and processes of form \eqrefrepr2 appear in [Gushchin:18] and [Gushchin:20]. It is interesting to note that, in [Gushchin:20], the random “time” γ\gamma is, in fact, the global maximum of a random process, say, a convergent continuous local martingale.

Section 2 contains our main results. In Theorem 1 we establish a necessary and sufficient condition for a process of type \eqrefrepr2 to be a local martingale. This allows us to obtain a full description of all local martingales through a functional equation of type \eqrefe:fe in Theorem 2. A similar description is available for σ\sigma-martingales, see Theorem 3. Finally, Theorem 4 classifies local martingales in accordance with their global behaviour up to \infty. Section 3 contains the proofs of these results. In Section LABEL:sec:5 we consider complementary questions. Namely, we find the compensator of a single jump process. We also consider submartingales of class (Σ)(\Sigma), see [Nikeghbali:06], and show that their transformation via a change of time leads to processes of type \eqrefrepr2. As a consequence, we reprove Theorem 4.1 of [Nikeghbali:06].

We use the following notation: +=[0,+)\mathbb{R}_{+}=[0,+\infty), ¯+=[0,+]\overline{\mathbb{R}}_{+}=[0,+\infty], ab=min{a,b}a\wedge b=\min\,\{a,b\}. The arrows \uparrow and \downarrow indicate monotone convergence, while lims\upuparrowst\lim_{s\upuparrows t} stands for limst,s<t\lim_{s\to t,s<t}.

A real-valued function Z(t)Z(t) defined at least for t[0,s)t\in[0,s) is called càdlàg on [0,s)[0,s) if it is right-continuous at every t[0,s)t\in[0,s) and has a finite left-hand limit at every t(0,s)t\in(0,s); it is not assumed that it has a limit as t\upuparrowsst\upuparrows s. If, additionally, a finite limit limt\upuparrowssZ(t)\lim_{t\upuparrows s}Z(t) exists, then Z(t)Z(t) is called càdlàg on [0,s][0,s]. Functions ZZ of finite variation on compact intervals are understood as usually and are assumed to be càdlàg. The variation at 0 includes |Z(0)||Z(0)| as if ZZ is extended by 0 on negative axis. The total variation of ZZ over [0,t][0,t] is denoted by \Var(Z)t\Var(Z)_{t}. We say that ZZ has a finite variation over [0,s)[0,s), s\leqslants\leqslant\infty, if limt\upuparrowss\Var(Z)t<\lim_{t\upuparrows s}\Var(Z)_{t}<\infty. We denote \Var(Z):=limt\Var(Z)t\Var(Z)_{\infty}:=\lim_{t\to\infty}\Var(Z)_{t}.

A filtration on a probability space (Ω,,𝖯)(\Omega,{\mathscr{F}},\mathsf{P}) is an increasing right-continuous family 𝔽=(t)t+\mathbb{F}=({\mathscr{F}}_{t})_{t\in\mathbb{R}_{+}} of sub-σ\sigma-fields of {\mathscr{F}}. No completeness assumption is made. As usual, we define =σ(t+t){\mathscr{F}}_{\infty}=\sigma\bigl{(}\cup_{t\in\mathbb{R}_{+}}{\mathscr{F}}_{t}\bigr{)} and, for a stopping time τ\tau the σ\sigma-field τ{\mathscr{F}}_{\tau} is defined by

τ={A:A{τ\leqslantt}t\textforeveryt\geqslant0}.{\mathscr{F}}_{\tau}=\bigl{\{}A\in{\mathscr{F}}_{\infty}\colon A\cap\{\tau\leqslant t\}\in{\mathscr{F}}_{t}\text{forevery}t\geqslant 0\bigr{\}}.

A set BΩ×+B\subset\Omega\times\mathbb{R}_{+} is evanescent if BA×+B\subseteq A\times\mathbb{R}_{+}, where AA\in{\mathscr{F}} and 𝖯(A)=0\mathsf{P}(A)=0. We say that two stochastic processes XX and YY are indistinguishable if {XY}\{X\neq Y\} is an evanescent set.

Since we do not suppose completeness of the filtration 𝔽\mathbb{F}, we cannot expect that processes that we consider have all paths càdlàg. Instead we consider processes whose almost all paths are càdlàg. Obviously, for any càdlàg process XX adapted with respect to the completed filtration, there is an a.s. càdlàg 𝔽\mathbb{F}-adapted process indistinguishable from XX. Furthermore, any 𝔽\mathbb{F}-adapted process XX with a.s. càdlàg paths is indistinguishable from an 𝔽\mathbb{F}-optional process YY whose paths are right-continuous everywhere and have finite left-hand limits for t<ρ(ω)t<\rho(\omega) and t>ρ(ω)t>\rho(\omega), where ρ\rho is a 𝔽\mathbb{F}-stopping time with 𝖯(ρ<)=0\mathsf{P}(\rho<\infty)=0; let us call such YY regular and ρ\rho a moment of irregularity for YY. Dellacherie and Meyer [DellacherieMeyer:1982, VI.5 (a), p. 70] prove that, if the filtration is not complete, every supermartingale XX (with right-continuous expectation) has a modification YY with the above regularity property. If we are given just an adapted process XX with almost all paths càdlàg, we define ρ\rho and YY from values of XX on a countable set exactly as is done in [DellacherieMeyer:1982] in the case where XX is a supermartingale. Using [DellacherieMeyer:1978, Theorem IV.22, p. 94], we obtain that ρ(ω)=\rho(\omega)=\infty and paths X(ω)X_{\cdot}(\omega) and Y(ω)Y_{\cdot}(\omega) coincide for those ω\omega for which X(ω)X_{\cdot}(\omega) is càdlàg everywhere. Moreover, if ρ(ω)<\rho(\omega)<\infty, then Yt(ω)Y_{t}(\omega) is càdlàg for t<ρ(ω)t<\rho(\omega) and one may put Yt(ω)=0Y_{t}(\omega)=0 for t\geqslantρ(ω)t\geqslant\rho(\omega).

Processes with finite variation are adapted and not assumed to start from 0. A moment of irregularity for them has additionally the property that their paths have finite variation over [0,t][0,t] for all t<ρ(ω)t<\rho(\omega).

It is instructive to mention that, in our model, there is no need to use general results on the existence of (a.s.) càdlàg modifications for martingales since they can be proved directly. For example, if LL is an integrable random variable with 𝖤L=0\mathsf{E}L=0, then the process MM given by \eqrefrepr2 with F(t)=𝖤[L|γ>t]\mathbh1{t<tG}F(t)=\mathsf{E}[L|\gamma>t]{\mathbh{1}}_{\{t<t_{G}\}} satisfies Mt=𝖤[L|t]M_{t}=\mathsf{E}[L|{\mathscr{F}}_{t}] a.s. for an arbitrary tt. It is trivial to check that this function FF has finite variation over any [0,t][0,t] with 𝖯(γ>t)>0\mathsf{P}(\gamma>t)>0 (and over [0,tG)[0,t_{G}) if 𝖯(γ=tG<)>0\mathsf{P}(\gamma=t_{G}<\infty)>0). Thus MM is regular. It may be that, if tG<t_{G}<\infty and 𝖯(γ=tG)=0\mathsf{P}(\gamma=t_{G})=0, the function FF has not a finite limit as t\upuparrowstGt\upuparrows t_{G}, or, more generally, has unbounded variation over [0,tG)[0,t_{G}). Then a moment of irregularity is given by

ρ(ω)={tG,\textifγ\geqslantt_G;+,\textotherwise.\rho(\omega)=\left\{\begin{array}[]{ll}t_{G},&\text{if$\gamma\geqslant t_{G}$;}\\ +\infty,&\text{otherwise.}\end{array}\right.

It takes a finite value only on the set {γ\geqslanttG}\{\gamma\geqslant t_{G}\} of zero measure. In all other cases we may put ρ+\rho\equiv+\infty. See Remark 2 in Section 2 for more details.

2 Main results

Let γ\gamma be a random variable with values in ¯+\overline{\mathbb{R}}_{+} on a probability space (Ω,,𝖯)(\Omega,{\mathscr{F}},\mathsf{P}). We tacitly assume that 𝖯(γ>0)>0\mathsf{P}(\gamma>0)>0. G(t)=𝖯(γ\leqslantt)G(t)=\mathsf{P}(\gamma\leqslant t), t+t\in\mathbb{R}_{+}, stands for the distribution function of γ\gamma and G¯(t)=1G(t){\overline{G}}(t)=1-G(t). Put also tG=sup{t+:G(t)<1}t_{G}=\sup\,\{t\in\mathbb{R}_{+}\colon G(t)<1\} and 𝒯={t+:𝖯(γ\geqslantt)>0}{\mathcal{T}}=\{t\in\mathbb{R}_{+}\colon{\mathsf{P}}(\gamma\geqslant t)>0\}. Note that 𝖯(γ𝒯)=0\mathsf{P}(\gamma\notin{\mathcal{T}})=0. We will often distinguish between the following two cases:

Case A

𝖯(γ=tG<)=0\mathsf{P}(\gamma=t_{G}<\infty)=0.

Case B

𝖯(γ=tG<)>0\mathsf{P}(\gamma=t_{G}<\infty)>0.

It is clear that 𝒯=[0,tG){\mathcal{T}}=[0,t_{G}) in Case A and 𝒯=[0,tG]{\mathcal{T}}=[0,t_{G}] in Case B.

We define t{\mathscr{F}}_{t}, t+t\in\mathbb{R}_{+}, as the collection of subsets AA of Ω\Omega such that AA\in{\mathscr{F}} and A{t<γ}A\cap\{t<\gamma\} is either \varnothing\varnothing or coincides with {t<γ}\{t<\gamma\}.

It is shown in Proposition 1 that t{\mathscr{F}}_{t} is a σ\sigma-field for every t+t\in\mathbb{R}_{+} and the family 𝔽=(t)t+\mathbb{F}=({\mathscr{F}}_{t})_{t\in\mathbb{R}_{+}} is a filtration. We call this filtration a single jump filtration. It is determined by generating elements γ\gamma and {\mathscr{F}}. In this paper we consider only single jump filtrations and, if necessary to indicate generating elements, we use the notation 𝔽(γ,)\mathbb{F}(\gamma,{\mathscr{F}}) for the single jump filtration generated by γ\gamma and {\mathscr{F}}.

In this section a single jump filtration 𝔽=𝔽(γ,)\mathbb{F}=\mathbb{F}(\gamma,{\mathscr{F}}) is fixed. All notions depending on filtration (stopping times, martingales, local martingales, etc.) refer to this filtration 𝔽\mathbb{F}, unless otherwise specified.

Proposition 1

(i) t{\mathscr{F}}_{t} is a σ\sigma-field and a random variable ξ\xi is t{\mathscr{F}}_{t}-measurable, t+t\in\mathbb{R}_{+}, if and only if ξ\xi is constant on {t<γ}\{t<\gamma\}. ξ\xi is {\mathscr{F}}_{\infty}-measurable if and only if ξ\xi is constant on {γ=}\{\gamma=\infty\}.

(ii) The family (t)t+({\mathscr{F}}_{t})_{t\in\mathbb{R}_{+}} is increasing and right-continuous, i.e. 𝔽=(t)t+\mathbb{F}=({\mathscr{F}}_{t})_{t\in\mathbb{R}_{+}} is a filtration.

(iii) γ\gamma is a stopping time and γ={\mathscr{F}}_{\gamma}={\mathscr{F}}_{\infty}.

(iv) A random variable TT with values in ¯+\overline{\mathbb{R}}_{+} is a stopping time if and only if it satisfies the following property: if the set {T<γ}\{T<\gamma\} is not empty, then there is a number rr such that

{T<γ}={T=r<γ}={r<γ}.\{T<\gamma\}=\{T=r<\gamma\}=\{r<\gamma\}. (4)
Proposition 2

(i) If X=(Xt)t+X=(X_{t})_{t\in\mathbb{R}_{+}} is an adapted process, then there is a deterministic function F(t)F(t), 0\leqslantt<tG0\leqslant t<t_{G}, such that Xt=F(t)X_{t}=F(t) on {t<γtG}\{t<\gamma\wedge t_{G}\}. If Y=(Yt)t+Y=(Y_{t})_{t\in\mathbb{R}_{+}} is an adapted process and 𝖯(Xt=Yt)=1\mathsf{P}(X_{t}=Y_{t})=1 for every t+t\in\mathbb{R}_{+}, then Xt=YtX_{t}=Y_{t} identically on {t<γtG}\{t<\gamma\wedge t_{G}\}.

(ii) If Y=(Yt)t+Y=(Y_{t})_{t\in\mathbb{R}_{+}} is a predictable process, then there is a measurable deterministic function C(t)C(t), t𝒯t\in{\mathcal{T}}, such that Yt=C(t)Y_{t}=C(t) on {t\leqslantγ}\{t\leqslant\gamma\}, t𝒯t\in{\mathcal{T}}.

(iii) If X=(Xt)t+X=(X_{t})_{t\in\mathbb{R}_{+}} is a process with finite variation, then F(t)F(t) in (i) has a finite variation over [0,t][0,t] for every t<tGt<t_{G} in Case A and over [0,tG)[0,t_{G}) in Case B.

(iv) Every semimartingale is a process with finite variation.

(v) If M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} is a σ\sigma-martingale then there are a deterministic function F(t)F(t), t+t\in\mathbb{R}_{+}, and a finite random variable LL such that, up to 𝖯\mathsf{P}-indistinguishability,

Mt=F(t)\mathbh1{t<γ}+L\mathbh1{t\geqslantγ}.M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+L{\mathbh{1}}_{\{t\geqslant\gamma\}}. (5)

Statement (iv) is not surprising. If the σ\sigma-field {\mathscr{F}} is countably generated, then our filtration is a special case of a filtration generated by a marked point process, and it is known, see [Jacod1979], that then all martingales are of finite variation. In general, a single jump filtration is a special case of a jumping filtration, see [JacodSkorokhod1994], where again all martingales are of finite variation.

Remark 1.

If MM is a σ\sigma-martingale, then it is a process with finite variation due to (iv) and, hence, the function F(t)F(t) in \eqrefeq:mr has a finite variation over [0,t][0,t] for every t<tGt<t_{G} in Case A and over [0,tG)[0,t_{G}) in Case B according to (iii).

Remark 2.

According to (i), the function F(t)F(t) in \eqrefeq:mr is uniquely determined for t<tGt<t_{G}. Since 𝖯(γ>tG)=0\mathsf{P}(\gamma>t_{G})=0, the stochastic interval tG,γ\rlb\llbracket t_{G},\gamma\rlb is an evanescent set. Hence, F(t)F(t) can be defined arbitrarily for t\geqslanttGt\geqslant t_{G}. For example, we can put it equal to 0 for t\geqslanttGt\geqslant t_{G}. Then F(t)F(t) has a finite variation on compact intervals if tG=+t_{G}=+\infty or in Case B. In Case A, if tGt_{G} is finite, F(t)F(t) may have infinite variation over [0,tG)[0,t_{G}) (and even not have a finite limit as t\upuparrowstGt\upuparrows t_{G}), see Theorem 2 and Example 3 below. All other points are regular for F(t)F(t). Now put ρ(ω)=tG<+\rho(\omega)=t_{G}<+\infty if we are in Case A, tG<+t_{G}<+\infty, limt\upuparrowstG\Var(F)t=\lim_{t\upuparrows t_{G}}\Var(F)_{t}=\infty, and γ(ω)\geqslanttG\gamma(\omega)\geqslant t_{G}, and let ρ(ω)=+\rho(\omega)=+\infty in all other cases. It follows that ρ\rho is a moment of irregularity for the process in the right-hand side of \eqrefeq:mr.

In what follows, when we write that the process MM has the representation \eqrefeq:mr, this means that MM and the right-hand side of \eqrefeq:mr are indistinguishable. Moreover, we tacitly assume that F(t)F(t) is right-continuous for t\geqslanttGt\geqslant t_{G} to ensure that the right-hand side of \eqrefeq:mr is right-continuous.

Propositions 1 and 2 explain why we call 𝔽\mathbb{F} a single jump filtration: all randomness appears at time γ\gamma. It is not so natural to describe local martingales with respect to 𝔽\mathbb{F} as single jump processes. As we will see, the function FF in \eqrefeq:mr need not be continuous, so local martingales may have several jumps.

Our main goal is to provide a complete description of all local martingales. According to Proposition 2 (v), a necessary condition is that it is represented in form \eqrefeq:mr. Thus, it is enough to study only processes of this form.

Theorem 1

Let F(t)F(t), 0\leqslantt<tG0\leqslant t<t_{G}, be a deterministic càdlàg function, LL be a random variable, and a process M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} be given by

Mt=F(t)\mathbh1{t<γ}+L\mathbh1{t\geqslantγ}.M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+L{\mathbh{1}}_{\{t\geqslant\gamma\}}. (6)

The following statements are equivalent:

  1. (i)

    M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} is a local martingale.

  2. (ii)

    (Mt)t𝒯(M_{t})_{t\in{\mathcal{T}}} is a martingale.

  3. (iii)
    𝖤(|Mt|)<,t𝒯,{\mathsf{E}}\bigl{(}|M_{t}|\bigr{)}<\infty,\quad t\in{\mathcal{T}}, (7)

    and

    𝖤(Mt)=𝖤(M0),t𝒯.{\mathsf{E}}(M_{t})=\mathsf{E}(M_{0}),\quad t\in{\mathcal{T}}. (8)

In the case where =σ{γ}{\mathscr{F}}=\sigma\{\gamma\}, equivalence (i) and (ii) is proved in [ChouMeyer:1975].

Concerning the last statement of the proposition, let us emphasize that if tG<t_{G}<\infty and 𝖯(γ=tG)=0\mathsf{P}(\gamma=t_{G})=0, a local martingale M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} may not be a martingale on [0,tG][0,t_{G}]; obviously, if it is a martingale, then it is uniformly integrable, and necessary and sufficient conditions for this are given in Theorem 4.

If \eqrefeq:mr2 and \eqrefeq:integrab hold, then

𝖤(|L|\mathbh1{γ\leqslantt})<,t𝒯,{\mathsf{E}}\bigl{(}|L|{\mathbh{1}}_{\{\gamma\leqslant t\}}\bigr{)}<\infty,\quad t\in{\mathcal{T}}, (9)

and one can define the conditional expectation H(t)H(t) of LL given that γ=t\gamma=t for t𝒯t\in{\mathcal{T}}:

H(t)=𝖤[L|γ=t].H(t)=\mathsf{E}[L|\gamma=t]. (10)

More precisely, H(t)H(t) is a Borel function on 𝒯{\mathcal{T}} with finite values such that for any t𝒯t\in{\mathcal{T}}

𝖤(L\mathbh1{γ\leqslantt})=[0,t]H(s)dG(s).{\mathsf{E}}\bigl{(}L{\mathbh{1}}_{\{\gamma\leqslant t\}}\bigr{)}=\int_{[0,t]}H(s)\,dG(s).

Note that the function HH is dGdG-a.s. unique and is dGdG-integrable over any closed interval in 𝒯{\mathcal{T}}. It is convenient to introduce a notation for such functions.

Let L1loc(dG)L^{1}_{\mathrm{loc}}(dG) be the set of all Borel functions zz on 𝒯{\mathcal{T}} such that

[0,t]|z(s)|dG(s)<\textforallt𝒯.\int_{[0,t]}|z(s)|\,dG(s)<\infty\quad\text{forallt\in{\mathcal{T}}}.

Given a function Z:[0,tG)Z\colon[0,t_{G})\to\mathbb{R}, let us write Z\oversetlocGZ\overset{\mathrm{loc}}{\ll}G if there is zL1loc(dG)z\in L^{1}_{\mathrm{loc}}(dG) such that Z(t)=Z(0)+(0,t]z(s)dG(s)Z(t)=Z(0)+\int_{(0,t]}z(s)\,dG(s) for all t<tGt<t_{G}; in this case we put dZdG(t):=z(t)\frac{dZ}{dG}(t):=z(t) for 0<t<tG0<t<t_{G}. Let us emphasize that in Case B this definition implies that zz is dGdG-integrable over [0,tG][0,t_{G}] and, hence, the function ZZ has a finite variation over [0,tG)[0,t_{G}) and there is a finite limit limt\upuparrowstGZ(t)=Z(0)+(0,tG)z(s)dG(s)\lim_{t\upuparrows t_{G}}Z(t)=Z(0)+\int_{(0,t_{G})}z(s)\,dG(s). Note also that in this definition the value z(0)z(0) can be chosen arbitrarily even if G(0)>0G(0)>0; the same refers to the value z(tG)z(t_{G}) in Case B. Correspondingly, dZ/dGdZ/dG is defined only for 0<t<tG0<t<t_{G}.

Let GG be a distribution function of a law on [0,+][0,+\infty]. We will say that a pair (F,H)(F,H) satisfies Condition M if {gather} F:[0,t_G)→R,  F\oversetloc≪G,
H:TR,  H∈L^1_loc(dG),
F(t)¯G(t) + ∫_(0,t]H(s) dG(s) = F(0) ¯G(0),  t<t_G, and, additionally in Case B,

limt\upuparrowstGF(t)=H(tG).\lim_{t\upuparrows t_{G}}F(t)=H(t_{G}). (11)
Proposition 3

(a) Let HH be any function satisfying \eqrefeq:MH. Define

F(t)=G¯(t)1[F(0)G¯(0)(0,t]H(s)dG(s)],0<t<tG,F(t)={\overline{G}}(t)^{-1}\Bigl{[}F(0){\overline{G}}(0)-\int\limits_{(0,t]}H(s)\,dG(s)\Bigr{]},\quad 0<t<t_{G}, (12)

where F(0)F(0) is an arbitrary real number in Case A and

F(0)=G¯(0)1(0,tG]H(s)dG(s)F(0)={\overline{G}}(0)^{-1}\int\limits_{(0,t_{G}]}H(s)\,dG(s) (13)

in Case B. Then the pair (F,H)(F,H) satisfies Condition M. Conversely, if FF is such that the pair (F,H)(F,H) satisfies Condition M, then FF satisfies \eqrefeq:F and, in Case B, \eqrefeq:F0 holds.

(b) Let FF be any function satisfying \eqrefeq:MF. Define H(0)H(0) arbitrarily,

H(t)=F(t)G¯(t)dFdG(t),0<t<tG,H(t)=F(t)-{\overline{G}}(t-)\frac{dF}{dG}(t),\quad 0<t<t_{G}, (14)

H(tG)H(t_{G}) arbitrarily in Case A and

H(tG)=limt\upuparrowstGF(t)H(t_{G})=\lim_{t\upuparrows t_{G}}F(t) (15)

in Case B. Then the pair (F,H)(F,H) satisfies Condition M. Conversely, if HH is such that the pair (F,H)(F,H) satisfies Condition M, then HH satisfies \eqrefeq:H and, in Case B, \eqrefeq:Ht holds.

Theorem 2

In order that a right-continuous process M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} be a local martingale it is necessary and sufficient that there be a pair (F,H)(F,H) satisfying Condition M and a random variable LL^{\prime} satisfying

𝖤(|L|\mathbh1{γ\leqslantt})<,t𝒯,\textand𝖤[L|γ]=0,{\mathsf{E}}\bigl{(}|L^{\prime}|{\mathbh{1}}_{\{\gamma\leqslant t\}}\bigr{)}<\infty,\quad t\in{\mathcal{T}},\qquad\text{and}\qquad\mathsf{E}[L^{\prime}|\gamma]=0, (16)

such that, up to 𝖯\mathsf{P}-indistinguishability,

Mt=F(t)\mathbh1{t<γ}+(H(γ)+L)\mathbh1{t\geqslantγ},M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+\bigl{(}H(\gamma)+L^{\prime}\bigr{)}{\mathbh{1}}_{\{t\geqslant\gamma\}}, (17)

The statement that the process MM given by \eqrefeq:mr3 with L=0L^{\prime}=0 is a local martingale if F\oversetlocGF\overset{\mathrm{loc}}{\ll}G and HH is constructed as in part (b) of Proposition 3, is essentially due to Herdegen and Herrmann [HerdegenHerrmann:16], though they formulate \eqrefeq:H in an equivalent form:

H(t)=F(t)G¯(t)dFdG(t),0<t<tG.H(t)=F(t-)-{\overline{G}}(t)\frac{dF}{dG}(t),\quad 0<t<t_{G}. (18)

They also prove that, in Case B, if FF has infinite variation on [0,tG)[0,t_{G}) (and hence does not satisfy F\oversetlocGF\overset{\mathrm{loc}}{\ll}G), then MM given by \eqrefeq:mr2 is not a semimartingale, see [HerdegenHerrmann:16, Lemma B.6]. (Note that this follows also from our Proposition 2 (iv).) We add that, also in Case B, if HH is dGdG-integrable over (0,tG)(0,t_{G}), FF satisfies \eqrefeq:F, but F(0)F(0) is greater or less than the right-hand side of \eqrefeq:F0, then MM given by \eqrefeq:mr3 with LL^{\prime} satisfying \eqrefeq:integrab2, is a supermartingale or a submartingale, respectively, cf. Theorem 4.

The fact that H(0)H(0) can be chosen arbitrarily in Proposition 3 (b) says only that LL can be an arbitrary integrable random variable on the set {γ=0}\{\gamma=0\}, which is evident ab initio. On the contrary, the fact that F(0)F(0) can be chosen arbitrarily in (a) in Case A is an interesting feature of this model. It says that, given the terminal value MM_{\infty} of MM (on {γ<}\{\gamma<\infty\}), one can freely choose the initial value M0M_{0} of MM (on {γ>0}\{\gamma>0\}) to keep the property of being a local martingale for MM.

Corollary 1

Every local martingale M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} has a unique decomposition into the sum M=M+MM=M^{\prime}+M^{\prime\prime} of two local martingales MM^{\prime} and MM^{\prime\prime}, where MM^{\prime} is adapted with respect to the smallest filtration making γ\gamma a stopping time, and MM^{\prime\prime} which vanishes on {t<γ}\{t<\gamma\} and satisfies 𝖤M0=0\mathsf{E}M^{\prime\prime}_{0}=0.

Remark 3.

If 𝖯(γ=0)=0\mathsf{P}(\gamma=0)=0, then it follows from the first property for MM^{\prime\prime} that M0=0M^{\prime\prime}_{0}=0 a.s. and thus the second property holds automatically.

Remark 4.

The smallest filtration making γ\gamma a stopping time is a single jump filtration 𝔽(γ,σ{γ})\mathbb{F}(\gamma,\sigma\{\gamma\}) generated by γ\gamma and the smallest σ\sigma-field σ{γ}\sigma\{\gamma\} with respect to which γ\gamma is measurable. Let MM be a 𝔽\mathbb{F}-local martingale adapted to 𝔽(γ,σ{γ})\mathbb{F}(\gamma,\sigma\{\gamma\}). It follows from Theorem 1 that MM is a 𝔽(γ,σ{γ})\mathbb{F}(\gamma,\sigma\{\gamma\})-local martingale.

As the next example shows, the product MMM^{\prime}M^{\prime\prime} of local martingales from the above decomposition may not be a local martingale because the first condition in \eqrefeq:integrab2 may fail. It will follow from Theorem 3 below that this product is always a σ\sigma-martingale.

Example 1.

Let γ\gamma have an exponential distribution, e.g., G¯(t)=et{\overline{G}}(t)=e^{-t}, FF is given by \eqrefeq:F with H(t)=t1/2H(t)=t^{-1/2} and an arbitrary F(0)F(0), Mt=F(t)\mathbh1{t<γ}+H(γ)\mathbh1{t\geqslantγ}M^{\prime}_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+H(\gamma){\mathbh{1}}_{\{t\geqslant\gamma\}}, Mt=Yγ1/2\mathbh1{t\geqslantγ}M^{\prime\prime}_{t}=Y\gamma^{-1/2}{\mathbh{1}}_{\{t\geqslant\gamma\}}, where YY takes values ±1\pm 1 with probabilities 1/21/2 and is independent of γ\gamma. It follows that MM^{\prime} and MM^{\prime\prime} are local martingales but their product MtMt=Yγ1\mathbh1{t\geqslantγ}M^{\prime}_{t}M^{\prime\prime}_{t}=Y\gamma^{-1}{\mathbh{1}}_{\{t\geqslant\gamma\}} does not satisfy the integrability condition \eqrefeq:integrab and, hence, is not a local martingale. This process is a classical example (due to Émery) of a σ\sigma-martingale which is not a local martingale, see, e.g., [Gushchin:15, Example 2.3, p. 86].

The previous example shows that our model admits σ\sigma-martingales which are not local martingales. In the next theorem we describe all σ\sigma-martingales in our model. In particular, it implies that if =σ{γ}{\mathscr{F}}=\sigma\{\gamma\}, then all σ\sigma-martingales that are integrable at 0 are local martingales.

Theorem 3

In order that a right-continuous process M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} be a σ\sigma-martingale it is necessary and sufficient that it have a representation \eqrefeq:mr3, where a pair (F,H)(F,H) satisfies Condition M and a random variable LL^{\prime} satisfies

𝖤[|L|\mathbh1{γ>0}|γ]<\textand𝖤[L|γ]=0.{\mathsf{E}}\bigl{[}|L^{\prime}|{\mathbh{1}}_{\{\gamma>0\}}\big{|}\gamma\bigr{]}<\infty\qquad\text{and}\qquad\mathsf{E}[L^{\prime}|\gamma]=0. (19)

The next theorem complements the classification of the limit behaviour of local martingales that was considered in Herdegen and Herrmann [HerdegenHerrmann:16] in the case where =σ{γ}{\mathscr{F}}=\sigma\{\gamma\}. Let us say that a local martingale M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} has

type 1

if the limit M=limtMtM_{\infty}=\lim_{t\to\infty}M_{t} does not exist with positive probability or exists with probability one but is not integrable: 𝖤|M|=\mathsf{E}|M_{\infty}|=\infty;

type 2a

if MM is a closed supermartingale (in particular, 𝖤|M|<\mathsf{E}|M_{\infty}|<\infty) and 𝖤(M)<𝖤(M0)\mathsf{E}(M_{\infty})<\mathsf{E}(M_{0});

type 2b

if MM is a closed submartingale (in particular, 𝖤|M|<\mathsf{E}|M_{\infty}|<\infty) and 𝖤(M)>𝖤(M0)\mathsf{E}(M_{\infty})>\mathsf{E}(M_{0});

type 3

if MM is a uniformly integrable martingale (in particular, 𝖤|M|<\mathsf{E}|M_{\infty}|<\infty and 𝖤(M)=𝖤(M0)\mathsf{E}(M_{\infty})=\mathsf{E}(M_{0})) and 𝖤(supt|Mt|)=\mathsf{E}(\sup_{t}|M_{t}|)=\infty;

type 4

if MM has an integrable variation: 𝖤(\Var(M))<\mathsf{E}\bigl{(}\Var(M)_{\infty}\bigr{)}<\infty.

Theorem 4

Let M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} be a local martingale with the representation

Mt=F(t)\mathbh1{t<γ}+L\mathbh1{t\geqslantγ},t+,M_{t}=F(t){\mathbh{1}}_{\{t<\gamma\}}+L{\mathbh{1}}_{\{t\geqslant\gamma\}},\quad t\in\mathbb{R}_{+}, (20)

where L=H(γ)+LL=H(\gamma)+L^{\prime}, a pair (F,H)(F,H) satisfies Condition M and a random variable LL^{\prime} satisfies \eqrefeq:integrab2. Then in Case B the local martingale MM has type 4. In Case A all types are possible. Namely,

  1. (i)

    MM has type 1 if and only if 𝖤(|L|\mathbh1{γ<})=\mathsf{E}\bigl{(}|L^{\prime}|{\mathbh{1}}_{\{\gamma<\infty\}}\bigr{)}=\infty or [0,tG)|H(s)|dG(s)=\int_{[0,t_{G})}|H(s)|\,dG(s)=\infty.

  2. (ii)

    If 𝖯(γ=)>0\mathsf{P}(\gamma=\infty)>0, 𝖤(|L|\mathbh1{γ<})<\mathsf{E}\bigl{(}|L^{\prime}|{\mathbh{1}}_{\{\gamma<\infty\}}\bigr{)}<\infty, and +|H(s)|dG(s)<\int_{\mathbb{R}_{+}}|H(s)|\,dG(s)<\infty then MM has type 4.

  3. (iii)

    If 𝖯(γ=)=0\mathsf{P}(\gamma=\infty)=0, 𝖤|L|<\mathsf{E}|L^{\prime}|<\infty, and [0,tG)|H(s)|dG(s)<\int_{[0,t_{G})}|H(s)|\,dG(s)<\infty then

    1. (iii.i)

      MM has type 2a (resp., 2b) if and only if limt\upuparrowstGF(t)G¯(t)>0\lim_{t\upuparrows t_{G}}F(t){\overline{G}}(t)>0 (resp., limt\upuparrowstGF(t)G¯(t)<0)\lim_{t\upuparrows t_{G}}F(t){\overline{G}}(t)<0);

    2. (iii.ii)

      MM has type 3 if and only if

      limt\upuparrowstGF(t)G¯(t)=0\textand[0,tG)G¯(s)|dFdG(s)|dG(s)=;\lim_{t\upuparrows t_{G}}F(t){\overline{G}}(t)=0\qquad\text{and}\qquad\int_{[0,t_{G})}{\overline{G}}(s)\Bigl{|}\frac{dF}{dG}(s)\Bigr{|}\,dG(s)=\infty;
    3. (iii.iii)

      MM has type 4 if and only if

      [0,tG)G¯(s)|dFdG(s)|dG(s)<.\int\limits_{[0,t_{G})}{\overline{G}}(s)\Bigl{|}\frac{dF}{dG}(s)\Bigr{|}\,dG(s)<\infty. (21)
Remark 5.

It follows from \eqrefeq:main that the limit limttGF(t)G¯(t)\lim_{t\to t_{G}}F(t){\overline{G}}(t) in (iii.i) and (iii.ii) exists. Also, [0,tG)|H(s)|dG(s)\int_{[0,t_{G})}|H(s)|\,dG(s) in (i)–(iii) is finite if only if F(t)G¯(t)F(t){\overline{G}}(t) has a finite variation over [0,tG)[0,t_{G}).

Remark 6.

It follows from Theorem 4 that, in our model, every martingale MM with 𝖤(supt|Mt|)<\mathsf{E}(\sup_{t}|M_{t}|)<\infty has an integrable total variation. Of course, on general spaces, there exist martingales MM having finite variation on compacts and such that𝖤(supt|Mt|)<\mathsf{E}(\sup_{t}|M_{t}|)<\infty and their total variation is not integrable, see, e.g., [Gushchin:15, Example 2.7, p. 103].

Example 2.

Assume that H:(0,1)H\colon(0,1)\to\mathbb{R} is a monotone nondecreasing function and, for definiteness, that it is right-continuous. Then it is the upper quantile function of H(γ)H(\gamma), where γ\gamma is uniformly distributed on (0,1)(0,1). Assume also that HH is integrable on (0,1)(0,1) and 01H(s)ds=0\int_{0}^{1}H(s)\,ds=0, that is to say, that H(γ)H(\gamma) has zero mean. Put

F(t)=(1t)10tH(s)ds=(1t)1t1H(s)ds.F(t)=-(1-t)^{-1}\int\limits_{0}^{t}H(s)\,ds=(1-t)^{-1}\int\limits_{t}^{1}H(s)\,ds.

We see that FF satisfying \eqrefeq:main with F(0)=0F(0)=0 is the Hardy–Littlewood maximal function corresponding to HH. If we define MM by \eqrefeq:mrU with L=H(γ)L=H(\gamma), then, by Theorem 4, MM is a uniformly integrable martingale with M=H(γ)M_{\infty}=H(\gamma) and suptMt=F(γ)\sup_{t}M_{t}=F(\gamma). This example is essentially the example of Dubins and Gilat [DubinsGilat] of a uniformly integrable martingale with a given distribution of its terminal value, having maximal (with respect to the stochastic partial order) maximum (in time).

Example 3 ([HerdegenHerrmann:16, Example 3.14]).

Let Ω=(0,1]\Omega=(0,1] be equipped with the Borel σ\sigma-field {\mathscr{F}}, and let 𝖯\mathsf{P} be the Lebesgue measure, γ(ω)=ω\gamma(\omega)=\omega. Put H(t)0H(t)\equiv 0. Then F(t)=(1t)1F(t)=(1-t)^{-1} satisfies \eqrefeq:main with F(0)=1F(0)=1. By Theorem 4, MM defined by \eqrefeq:mrU is a supermartingale and local martingale but not a martingale. This seems to be the simplest example of a local martingale with continuous time, which is not a martingale. Note that, for ω=1\omega=1, the trajectory Mt(ω)=(1t)1\mathbh1{t<1}M_{t}(\omega)=(1-t)^{-1}{\mathbh{1}}_{\{t<1\}} has not a finite left-hand limit at 11. Moreover, if NN is a modification of MM, for t<1t<1, the values of Mt(ω)M_{t}(\omega) and Nt(ω)N_{t}(\omega) must coincide on the atom {t<γ}=(t,1]\{t<\gamma\}=(t,1] of t{\mathscr{F}}_{t}, having the positive measure. Hence, Nt(ω)=Mt(ω)N_{t}(\omega)=M_{t}(\omega) for ω=1\omega=1 for all t<1t<1. This is an example of a right-continuous supermartingale which has not a modification with all paths càdlàg. Of course, the usual assumptions are not satisfied in this example.

3 Proofs

Proof of Proposition 1.

(i) and (iii) are evident from the definition of t{\mathscr{F}}_{t}, and (ii) follows easily from (i).

Let us prove (iv). To prove that TT is a stopping time, we must check that {T\leqslantt<γ}\{T\leqslant t<\gamma\} is either \varnothing\varnothing or {t<γ}\{t<\gamma\} for all t+t\in\mathbb{R}_{+}. This is trivial if {T<γ}=\varnothing\{T<\gamma\}=\varnothing. If there is a number rr such that \eqrefeq:NS-ST holds, then {T\leqslantt<γ}\{T\leqslant t<\gamma\} is either \varnothing\varnothing if r>tr>t or {t<γ}\{t<\gamma\} if r\leqslanttr\leqslant t.

Conversely, let TT be a stopping time. If T\geqslantγT\geqslant\gamma for all ω\omega, then there is nothing to prove. Assume that the set {T<γ}\varnothing\{T<\gamma\}\neq\varnothing. Then there are real numbers qq such that {T\leqslantq<γ}\varnothing\{T\leqslant q<\gamma\}\neq\varnothing. For such qq, by the definition of q{\mathscr{F}}_{q}, {T\leqslantq<γ}={q<γ}\{T\leqslant q<\gamma\}=\{q<\gamma\}, or, equivalently, {T\leqslantq}{q<γ}\{T\leqslant q\}\supseteq\{q<\gamma\}. Let rr be the greatest lower bound of such qq. The sets {q<γ}{r<γ}\{q<\gamma\}\uparrow\{r<\gamma\} and {T\leqslantq}{T\leqslantr}\{T\leqslant q\}\downarrow\{T\leqslant r\} as qrq\downarrow r. Thus,

{T<γ}=q:{T\leqslantq<γ}\varnothing{q<γ}={r<γ}{T\leqslantr}.\{T<\gamma\}=\bigcup_{q\colon\{T\leqslant q<\gamma\}\neq\varnothing}\{q<\gamma\}=\{r<\gamma\}\subseteq\{T\leqslant r\}.

Since {T\leqslantt<γ}=\varnothing\{T\leqslant t<\gamma\}=\varnothing for any t<rt<r, we have \eqrefeq:NS-ST. ∎

Proof of Proposition 2.

The first statement in (i) follows from Proposition 1 (i). Since 𝖯(t<γtG)>0\mathsf{P}(t<\gamma\wedge t_{G})>0 for every t<tGt<t_{G}, we obtain that XtX_{t} and YtY_{t} take the same constant value on {t<γtG}\{t<\gamma\wedge t_{G}\}.

Since a random variable YtY_{t} is t{\mathscr{F}}_{t-}-measurable for a predictable process YY, YtY_{t} is constant on {t\leqslantγ}\{t\leqslant\gamma\}. Denote by C(t)C(t), t𝒯t\in{\mathcal{T}}, the value of YtY_{t} on {t\leqslantγ}\{t\leqslant\gamma\}. Since 𝖯(γ\geqslantt)>0\mathsf{P}(\gamma\geqslant t)>0 for t𝒯t\in{\mathcal{T}}, there is an ω\omega such that C(s)Ys(ω)C(s)\equiv Y_{s}(\omega), s\leqslantts\leqslant t, and the measurability of CC follows.

Let us prove (iii) in Case B. Then we obtain that Xt=F(t)X_{t}=F(t) for all t<tGt<t_{G} on the set {γ=tG}\{\gamma=t_{G}\}, which has a positive probability. However, almost all paths of XtX_{t} have a finite variation over [0,tG)[0,t_{G}), and the claim follows. The proof in Case A is similar.

Now let us prove \eqrefeq:mr in the case where M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} is a uniformly integrable (a.s. càdlàg) martingale. We can find a random variable MM_{\infty} that is {\mathscr{F}}_{\infty}-measurable and such that limnMn=M\lim_{n\to\infty}M_{n}=M_{\infty} 𝖯\mathsf{P}-a.s. Since {t<γ}\{t<\gamma\} is an atom of t{\mathscr{F}}_{t} and has a positive probability for t<tGt<t_{G}, we obtain from the martingale property that Mt(ω)=F(t)M_{t}(\omega)=F(t) for all ω{t<γ}\omega\in\{t<\gamma\}, where

F(t)=𝖤(M\mathbh1{t<γ})G¯(t),t<tG.F(t)=\frac{\mathsf{E}\bigl{(}M_{\infty}{\mathbh{1}}_{\{t<\gamma\}}\bigr{)}}{{\overline{G}}(t)},\quad t<t_{G}.

It is clear that the nominator and the denominator are right-continuous functions of bounded variation on [0,tG][0,t_{G}], hence F(t)F(t), 0\leqslantt<tG0\leqslant t<t_{G} is a càdlàg function on 𝒯{\mathcal{T}} and has a finite variation on [0,tG)[0,t_{G}) in Case B and on every [0,t][0,t], 0\leqslantt<tG0\leqslant t<t_{G}, in Case A.

Now set L=M\mathbh1{γ<}L=M_{\infty}{\mathbh{1}}_{\{\gamma<\infty\}}. Then L\mathbh1{γ\leqslantt}=M\mathbh1{γ\leqslantt}L{\mathbh{1}}_{\{\gamma\leqslant t\}}=M_{\infty}{\mathbh{1}}_{\{\gamma\leqslant t\}} is t{\mathscr{F}}_{t}-measurable, and hence 𝖯\mathsf{P}-a.s.

Mt\mathbh1{γ\leqslantt}=𝖤(M\mathbh1{γ\leqslantt}|t)=L\mathbh1{γ\leqslantt}.M_{t}{\mathbh{1}}_{\{\gamma\leqslant t\}}=\mathsf{E}(M_{\infty}{\mathbh{1}}_{\{\gamma\leqslant t\}}|{\mathscr{F}}_{t})=L{\mathbh{1}}_{\{\gamma\leqslant t\}}.

Thus we have obtained, that, for a given t+t\in\mathbb{R}_{+}, MtM_{t} is equal 𝖯\mathsf{P}-a.s. to the right-hand side of \eqrefeq:mr with LL and F(t)F(t) as above. Since both the left-hand side and the right-hand side of \eqrefeq:mr are almost surely right-continuous, they are indistinguishable. Moreover, if we change F(t)F(t) for t\geqslanttGt\geqslant t_{G}, the right-hand side of \eqrefeq:mr will change on an evanescent set. Thus we can put, say, F(t)=0F(t)=0 for t\geqslanttGt\geqslant t_{G}, and then the right-hand side of \eqrefeq:mr is a regular right-continuous process with finite variation, and indistinguishable from MM.

Now let MM be a local martingale and {Tn}\{T_{n}\} be a localizing sequence of stopping times, i.e. TnT_{n}\uparrow\infty a.s. and MTnM^{T_{n}} is a uniformly integrable martingale for each nn. We have proved that almost all paths of MTnM^{T_{n}} have finite variation. It follows that almost all paths of MM have finite variation. This proves (iv).

Next, let MM be a σ\sigma-martingale, i.e. MM is a semimartingale and there is an increasing sequence of predictable sets Σn\Sigma_{n} such that nΣn=Ω×+\cup_{n}\Sigma_{n}=\Omega\times\mathbb{R}_{+} and the integral process \mathbh1ΣnM{\mathbh{1}}_{\Sigma_{n}}\cdot M is a uniformly integrable martingale for every nn. It does not matter if we integrate over [0,t][0,t] or (0,t](0,t], so let us agree that the domain of integration does not include 0. Since the integrand is bounded and every semimartingale is a process with finite variation in our model, the integral can be considered in the Lebesgue–Stieltjes sense, as well as other integrals appearing in the proof. Since \mathbh1ΣnM{\mathbh{1}}_{\Sigma_{n}}\cdot M is stopped at γ\gamma for every nn with probability one, we have

\mathbh1\lrbγ,\rlbΣn(t)d\Var(M)t=\mathbh1\lrbγ,\rlb(t)d\Var(\mathbh1ΣnM)t=0𝖯\texta.s.\int{\mathbh{1}}_{\lrb\gamma,\infty\rlb\cap\Sigma_{n}}(t)\,d\Var(M)_{t}=\int{\mathbh{1}}_{\lrb\gamma,\infty\rlb}(t)\,d\Var({\mathbh{1}}_{\Sigma_{n}}\cdot M)_{t}=0\quad\mathsf{P}\text{-a.s.}

for every nn, therefore,

\mathbh1\lrbγ,\rlb(t)d\Var(M)t=0𝖯\texta.s.\int{\mathbh{1}}_{\lrb\gamma,\infty\rlb}(t)\,d\Var(M)_{t}=0\quad\mathsf{P}\text{-a.s.}

Combining with (i), we prove representation \eqrefeq:mr. ∎

Remark 7.

As it was already explained in the introduction, we can prove directly, without assuming that paths are a.s. càdlàg, that any uniformly integrable martingale has a regular modification. The proof is essentially the same as above where we proved that a.s. càdlàg uniformly integrable martingale has representation \eqrefeq:mr.

Proof of Theorem 1.

First, we prove that statements (ii) and (iii) are equivalent. The implication (ii)\Rightarrow(iii) follows trivially from the definition of a martingale. Conversely, let (iii) hold. The process (Mt)t𝒯(M_{t})_{t\in{\mathcal{T}}} is right-continuous, adapted by Proposition 1 (i), and integrable, see \eqrefeq:integrab. Moreover, due to \eqrefeq:mr2,

MtMs=0\texton{s\geqslantγ},M_{t}-M_{s}=0\quad\text{on}\ \{s\geqslant\gamma\},

where 0\leqslants<t𝒯0\leqslant s<t\in{\mathcal{T}}. Hence,

𝖤[MtMs|s]=0\texton{s\geqslantγ}.{\mathsf{E}}[M_{t}-M_{s}|{\mathscr{F}}_{s}]=0\quad\text{on}\ \{s\geqslant\gamma\}.

But 𝖤[MtMs|s]\mathsf{E}[M_{t}-M_{s}|{\mathscr{F}}_{s}] is s{\mathscr{F}}_{s}-measurable and, thus, equals a constant on {s<γ}\{s<\gamma\}. And this constant must be zero since 𝖤(MtMs)=0\mathsf{E}(M_{t}-M_{s})=0 by \eqrefeq:main+.

The implication (ii)\Rightarrow(i) is trivial if tG=t_{G}=\infty or tG𝒯t_{G}\in{\mathcal{T}}. So we assume that tG<t_{G}<\infty and G¯(t)0{\overline{G}}(t)\downarrow 0 as t\upuparrowstGt\upuparrows t_{G}. Let t1<<tn<<tGt_{1}<\cdots<t_{n}<\cdots<t_{G}, tntGt_{n}\to t_{G}, be an increasing sequence, then G¯(tn)0{\overline{G}}(t_{n})\to 0. Put

Tn={tn,\textifγ>t_n;+,\textotherwise.T_{n}=\left\{\begin{array}[]{ll}t_{n},&\text{if$\gamma>t_{n}$;}\\ +\infty,&\text{otherwise.}\end{array}\right.

Then TnT_{n} is a stopping time by Proposition 1 (iv), TnT_{n}\uparrow\infty a.s., and MtTn=MttnM_{t\wedge T_{n}}=M_{t\wedge t_{n}}. Hence, MTnM^{T_{n}} is a martingale and MM is a local martingale.

It remains to prove the implication (i)\Rightarrow(ii). Let M=(Mt)t+M=(M_{t})_{t\in\mathbb{R}_{+}} be a local martingale with a localizing sequence {Tn}\{T_{n}\}, i.e. TnT_{n}\uparrow\infty a.s. and MTnM^{T_{n}} is a uniformly integrable martingale for each nn. If 𝖯(Tn\geqslantγ)=1\mathsf{P}(T_{n}\geqslant\gamma)=1 for some nn, then M=MTnM=M^{T_{n}} is a uniformly integrable martingale, and there is nothing to prove. So assume that 𝖯(Tn<γ)>0\mathsf{P}(T_{n}<\gamma)>0 for all nn. By Proposition 1 (iv), there is a number rnr_{n} such that {Tn<γ}={Tn=rn<γ}={rn<γ}\{T_{n}<\gamma\}=\{T_{n}=r_{n}<\gamma\}=\{r_{n}<\gamma\}. It follows from 𝖯(rn<γ)>0\mathsf{P}(r_{n}<\gamma)>0 that rn<tGr_{n}<t_{G}. In Case B we get 𝖯(Tn<γ)=𝖯(rn<γ)\geqslant𝖯(γ=tG)>0\mathsf{P}(T_{n}<\gamma)=\mathsf{P}(r_{n}<\gamma)\geqslant{\mathsf{P}}(\gamma=t_{G})>0 for every nn, a contradiction with TnT_{n}\to\infty a.s. In Case A, if 𝖯(γ=)>0\mathsf{P}(\gamma=\infty)>0, then it follows from TnT_{n}\to\infty a.s. that rnr_{n}\to\infty. In remaining cases where 𝖯(γ=tG)=0\mathsf{P}(\gamma=t_{G})=0, we obtain from 𝖯(rn<γ)0\mathsf{P}(r_{n}<\gamma)\to 0 that rntGr_{n}\to t_{G}, nn\to\infty. The claim follows since MtTn=MtrnM_{t\wedge T_{n}}=M_{t\wedge r_{n}}, and hence (Mt)t\leqslantrn(M_{t})_{t\leqslant r_{n}} is a martingale. ∎

Proof of Proposition 3.

(a) It is obvious that \eqrefeq:main is equivalent to \eqrefeq:F. It also follows from \eqrefeq:main that in Case B \eqrefeq:mainB is equivalent to \eqrefeq:F0. Thus it remains to prove that FF defined in (a) satisfies F\oversetlocGF\overset{\mathrm{loc}}{\ll}G. Since G¯(s)\geqslantG¯(t)>0{\overline{G}}(s)\geqslant{\overline{G}}(t)>0 for any s<t<tGs<t<t_{G}, we have

1G¯(t)=1G¯(0)+(0,t]1G¯(s)G¯(s)dG(s),t<tG.\frac{1}{{\overline{G}}(t)}=\frac{1}{{\overline{G}}(0)}+\int\limits_{(0,t]}\frac{1}{{\overline{G}}(s){\overline{G}}(s-)}\,dG(s),\quad t<t_{G}.

On the other hand, from \eqrefeq:F

F(t)G¯(t)=F(0)G¯(0)(0,t]H(s)dG(s),t<tG.F(t){\overline{G}}(t)=F(0){\overline{G}}(0)-\int\limits_{(0,t]}H(s)\,dG(s),\quad t<t_{G}.

Combining, we obtain from integration by parts that

F(t)=F(t)G¯(t)1G¯(t)=F(0)(0,t]H(s)G¯(s)dG(s)+(0,t]F(s)G¯(s)dG(s),t<tG.F(t)=F(t){\overline{G}}(t)\frac{1}{{\overline{G}}(t)}=F(0)-\int\limits_{(0,t]}\frac{H(s)}{{\overline{G}}(s-)}\,dG(s)+\int\limits_{(0,t]}\frac{F(s)}{{\overline{G}}(s-)}\,dG(s),\quad t<t_{G}.

This shows that F\oversetlocGF\overset{\mathrm{loc}}{\ll}G in Case A. In Case B we must show additionally that the function |F(s)|+|H(s)|G¯(s)\frac{|F(s)|+|H(s)|}{{\overline{G}}(s-)} is dGdG-integrable over (0,tG)(0,t_{G}). But 1/G¯(s)\leqslant1/𝖯(γ=tG)1/{\overline{G}}(s-)\leqslant 1/{\mathsf{P}}(\gamma=t_{G}), s\leqslanttGs\leqslant t_{G}, and F(s)F(s) is bounded on [0,tG)[0,t_{G}) in view of \eqrefeq:F. The claim follows.

(b) It is clear that the function H(t)H(t), t𝒯t\in{\mathcal{T}}, defined as in the statement, belongs to L1loc(dG)L^{1}_{\mathrm{loc}}(dG). Integrating by parts, we get, for t[0,tG)t\in[0,t_{G}), {align*} F(t)¯G(t) &= F(0)¯G(0)-∫_(0,t] F(s)  dG(s) + ∫_(0,t] ¯G(s-) dF(s)
= F(0)¯G(0)-∫_(0,t] F(s) dG(s) + ∫_(0,t] ¯G(s-)dFdG(s) dG(s)
= F(0)