This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Stopping Times Occurring Simultaneously

Philip Protter
Department of Statistics
Columbia University
New York, NY, 10027
ORCID 0000-0003-1344-0403
[email protected]
Supported in part by NSF grant DMS-2106433
   Alejandra Quintos
Department of Statistics
University of Wisconsin-Madison
Madison, WI, 53706
ORCID 0000-0003-3447-3255
[email protected]
Supported in part by the Office of the Vice Chancellor for Research and Graduate Education at the University of Wisconsin-Madison with funding from the Wisconsin Alumni Research Foundation and by the Fulbright-García Robles ProgramCorresponding author
(April 15, 2024)

ABSTRACT

Stopping times are used in applications to model random arrivals. A standard assumption in many models is that they are conditionally independent, given an underlying filtration. This is a widely useful assumption, but there are circumstances where it seems to be unnecessarily strong. We use a modified Cox construction along with the bivariate exponential introduced by Marshall and Olkin (1967) to create a family of stopping times, which are not necessarily conditionally independent, allowing for a positive probability for them to be equal. We show that our initial construction only allows for positive dependence between stopping times, but we also propose a joint distribution that allows for negative dependence while preserving the property of non-zero probability of equality. We indicate applications to modeling COVID-19 contagion (and epidemics in general), civil engineering, and to credit risk.

1 INTRODUCTION

Probability models are ubiquitous in modern society. The timing of a random event is often crucial to the analysis of the reliability of a system or to the danger of a default within a credit risk context. Such random times are, of course, referred to as stopping times. Some examples of random times of intrinsic interest range from the quotidian banal such as bus arrivals, customers arriving at a restaurant, to the less banal, such as the time a cancer metastasizes, or an individual contracts a contagious disease such as COVID-19. Stopping times appear in Civil Engineering as they model metal fatigue in aircraft, the times of collapse of a bridge, or extreme events such as the recent collapse of the condominium towers in Surfside Florida. A particularly common use of stopping times is in the theory of Credit Risk, within the discipline of Mathematical Finance.

Indeed, we take the approach pioneered by researchers in Credit Risk, the seminal event being the publication of the book by David Lando in 2004 (see [25]). The “Cox Construction” that Lando uses begins with a filtration of observable events, satisfying the usual hypotheses (see Protter [30]) for the formal definition of the usual hypotheses). To fix notation, let (Ω,,,(t)t0)\left(\Omega,\mathcal{F},\mathbb{P},(\mathcal{F}_{t})_{t\geq 0}\right) be the probability space on which (t)t0\left(\mathcal{F}_{t}\right)_{t\geq 0} is the filtration of observable events. We will add a random time τ\tau to our model by creating an exponential random variable ZZ (of parameter 1) that is independent of \mathcal{F} and all of 𝔽=(t)t0\mathbb{F}=(\mathcal{F}_{t})_{t\geq 0}. We choose an t\mathcal{F}_{t} predictable increasing process (At)t0\left(A_{t}\right)_{t\geq 0}, and we create the desired random time τ\tau by writing:

τ=inft0{AtZ}.\tau=\inf_{t\geq 0}\{A_{t}\geq Z\}. (1)

τ\tau is then a totally inaccessible stopping time for an enlarged (by ZZ) filtration 𝔾=(𝒢t)t0\mathbb{G}=(\mathcal{G}_{t})_{t\geq 0}, this is

𝒢t=tσ(Z).\mathcal{G}_{t}=\mathcal{F}_{t}\vee\sigma(Z).

With τ\tau being a stopping time, the process 1{τt}1_{\{\tau\geq t\}} is adapted and non-decreasing, hence it is a submartingale, whence there exists, via the Doob-Meyer decomposition theorem, a unique, 𝔾\mathbb{G}-predictable, increasing process (Ct)t0(C_{t})_{t\geq 0} such that 1{τt}Ct1_{\{\tau\geq t\}}-C_{t} is a martingale. It is easy to show that the process (Ct)t0\left(C_{t}\right)_{t\geq 0} is in fact the process (At)t0\left(A_{t}\right)_{t\geq 0} of (1). The process (At)t0\left(A_{t}\right)_{t\geq 0} is called the compensator of the stopping time τ\tau.

If we repeat this procedure to construct two stopping time τ1\tau_{1} and τ2\tau_{2}, using two independent exponentials of parameter 1, Z1Z_{1} and Z2Z_{2}, which are independent of each other as well as the underlying filtration 𝔽\mathbb{F}, then we have two totally inaccessible stopping times which are also conditionally independent of each other, and we have (τ1=τ2)=0.\mathbb{P}(\tau_{1}=\tau_{2})=0. This is the approach of Lando and many others, with some notable exceptions such as Bielecki et al. [3] and Jiao and Li [22].

In applications, often the process (At)t0\left(A_{t}\right)_{t\geq 0} is assumed to be of the form

At=0tαs𝑑s, for an adapted process αs0 all s0.A_{t}=\int_{0}^{t}\alpha_{s}ds,\text{ for an adapted process }\alpha_{s}\geq 0\text{ all }s\geq 0. (2)

Sufficient conditions for AtA_{t} to be of the form (2) are known (see, e.g., Ethier and Kurtz [14], Guo and Zeng [18], Zeng [36], and for a general result, Janson et al. [20]).

In this paper we are concerned with models where one can have (τ1=τ2)>0\mathbb{P}(\tau_{1}=\tau_{2})>0, and some of the ramifications of such a model. This model arises in applications when η1\eta_{1} and η2\eta_{2} are constructed as functionals of Cox processes with independent exponentials Z1Z_{1} and Z2Z_{2}, but with an added complication: there is a third stopping time η3\eta_{3}, and τ1=η1η3\tau_{1}=\eta_{1}\wedge\eta_{3}, while τ2=η2η3\tau_{2}=\eta_{2}\wedge\eta_{3}. This is a natural situation in Credit Risk for example, as we indicate in Section 3, but also in other domains, such as Civil Engineering, and disease contagion which we treat in Section 4. The resulting stopping times of interest, τ1\tau_{1} and τ2\tau_{2}, are no longer conditionally independent, and in simple cases, the bivariate exponential distribution of Marshall and Olkin [28] comes into play. These models do not have densities in +×+\mathbb{R}_{+}\times\mathbb{R}_{+}, leading to a two dimensional cumulative distribution function with a singular component. We explore the consequences of such a phenomenon in some detail, and we explain its utility for various kinds of applications caused by the confluence of stopping times that arise naturally in the modeling of random events.

One could argue that it is not of vital importance to have two stopping times with a positive probability of being equal, but rather just to have them be close to each other, even arbitrarily close. We study this situation in Section 2.5.

In many easy to imagine examples, the times τ1\tau_{1} and τ2\tau_{2} are positively correlated. It is possible to imagine, however, situations where they would naturally be negatively correlated. To cover that situation, we slightly modify our constructions, as exhibited in Section 3.2. As an example, consider the recent scandal with the Boeing 737 Max airplanes. The two horrifically deadly crashes occurred with the confluence of the airplane’s software malfunctioning and the panic of inexperienced and poorly trained pilots. Since airplanes are so carefully constructed, software failures are rare, which made the Boeing 737 Max failures unanticipated; however, unfortunately, poorly trained pilots have recently become rather common, especially in airlines based in poorly regulated countries. (The two examples were Lion Air in Malaysia and Ethiopian Airlines, in Ethiopia). That duality between those “rare” and “common” events is due to their negative correlation.

Our work in this paper falls into a popular thread of prior research. Cox and Lewis [11] studied the case of multiple event types occurring in a continuum, but differently from us, they do not generalize the model proposed by Marshall and Olkin [28]. Moreover, they mostly consider “regular” processes, i.e., where events cannot occur at the same time, which is an important contribution of our work. Diggle and Milne [13] also proposed a multivariate version of the Cox Process but their model does not allow for (τ1=τ2)>0\mathbb{P}(\tau_{1}=\tau_{2})>0 either. Another example of a multivariate point process is given in Brown et al. [7].

There is existing work on modeling simultaneous defaults, but different from us, under Merton’s structural risk model (see Li [26] and Bielecki et al. [4]). Kay Giesecke [15], in a seminal paper concerning Credit Risk published in 2003, was the first (to our knowledge) to consider the Marshall and Olkin model of the bivariate exponential. Later, in 2013, Bielecki et al. [3] also worked in a similar model, which is also discussed in Chapter 8 of Crépey et al. [12]. Sun et al. [31] in 2017 developed a similar model to the one we present here, but restricted their attention to multidimensional Lévy subordinator processes. In this paper, we develop the ideas present in Giesecke [15], in Bielecki et al. [3], and in Crépey et al. [12], and go beyond them.

Brigo et al. [5] argued that a reasonable trade-off between applicability and model flexibility might consist in modeling a multivariate survival indicator process such that each subvector of it is a continuous-time Markov chain, which, as they show, is a property of the Marshall–Olkin distribution. Hence, they claimed that the Marshall–Olkin distribution is the natural and unique choice for modeling default times. This strengthens our choice of joint distribution of the stopping times.

Another closely related work to ours is Lindskog and McNeil [27] who consider a Poisson shock model of arbitrary dimension with both fatal and not-necessarily-fatal shocks. Jiao and Li [22] also allow for simultaneous defaults without using a pure Cox process construction. Inspired by the Jacod Criterion, they use what is known as a density approach, and they can include cases where a stopping time meets another stopping time, specified in advance.

There are also other types of generalizations of Cox Processes (see Gueye and Jeanblanc [16]), where they generalize, not the number of stopping times, but the form of the process AtA_{t}. They assume that AtA_{t} is not necessarily of the form given in equation (2).

The organization of this paper is as follows. Section 2, presents the survival function of two conditionally dependent (as opposed to conditionally independent) stopping times, an interpretation of it, a decomposition of it into its singular and absolutely continuous parts, and a series of results exploring the special properties such a modeling approach has. For example, not only do we treat the case where (τ1=τ2)>0\mathbb{P}(\tau_{1}=\tau_{2})>0, but more generally we study when the two stopping times are ‘close’ to each other, in various metrics. Section 3 provides two generalizations of our model. In the first one, we extend to an arbitrary (but finite) number of such stopping times and in the second one, we allow for a slightly different dependency between the stopping times. Section 4 shows the applicability of our results by providing examples in Epidemiology (such as the case of COVID-19 and its variants) and in Civil Engineering (e.g., the recent condo collapse of Champlain Towers in Florida).

2 THE SURVIVAL FUNCTION

As a starting point, we will consider the case of having two stopping times, which we will generalize to nn stopping times in Section 3.1. Let us fix a filtered probability space (Ω,,,𝔽)(\Omega,\mathcal{F},\mathbb{P},\mathbb{F}) large enough to support an d\mathbb{R}^{d}-valued càdlàg stochastic process X={Xt,t0}X=\left\{X_{t},t\geq 0\right\} and three i.i.d exponential random variables with parameter 1 (Zi,i=1,2,3Z_{i},\ i=1,2,3) and independent of XX. Then, define:

τ1:=min(η1,η3)andτ2:=min(η2,η3)\tau_{1}:=\min\left(\eta_{1},\eta_{3}\right)\quad\textnormal{and}\quad\tau_{2}:=\min\left(\eta_{2},\eta_{3}\right) (3)

where

ηi:=inf{s:AsiZi},Asi:=0sαi(Xr)𝑑r,ZiiidExp(1)\eta_{i}:=\inf\left\{s:A^{i}_{s}\geq Z_{i}\right\},\qquad A^{i}_{s}:=\int_{0}^{s}\alpha^{i}(X_{r})dr,\qquad Z_{i}\overset{\text{iid}}{\sim}\text{Exp}(1) (4)

and

αi:d[0,) for i=1,2,3 are non-random positive continuous functions.\alpha^{i}:\mathbb{R}^{d}\mapsto[0,\infty)\text{ for }i=1,2,3\text{ are non-random positive continuous functions.}

Note that this characterization of αi\alpha^{i} implies that AsiA^{i}_{s} are continuous and strictly increasing.

Theorem 1 (Survival function).

Using the definitions above and assuming that limsAs=\lim_{s\rightarrow\infty}A_{s}=\infty a.s. and that Asi<s0A^{i}_{s}<\infty\ \forall s\geq 0. Then,

F¯(τ1,τ2)(s,t):=(τ1>s,τ2>t)=𝔼[exp[As1At2Ast3]].\overline{F}_{(\tau_{1},\tau_{2})}(s,t):=\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t\right)=\mathbb{E}\left[\exp\left[-A^{1}_{s}-A^{2}_{t}-A^{3}_{s\vee t}\right]\right].
Proof.

Let ~(s,t)():=(|(Xu)0u(st))\tilde{\mathbb{P}}_{(s,t)}(\cdot):=\mathbb{P}\left(\cdot|(X_{u})_{0\leq u\leq(s\vee t)}\right). By definition of τ1\tau_{1} and τ2\tau_{2}, we have,

~(s,t)(τ1>s,τ2>t)\displaystyle\tilde{\mathbb{P}}_{(s,t)}\left(\tau_{1}>s,\tau_{2}>t\right) =~(s,t)(min(η1,η3)>s,min(η2,η3)>t)\displaystyle=\tilde{\mathbb{P}}_{(s,t)}\left(\min(\eta_{1},\eta_{3})>s,\min(\eta_{2},\eta_{3})>t\right)
=~(s,t)(η1>s)~(s,t)(η2>t)~(s,t)(η3>s,η3>t)\displaystyle=\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{1}>s\right)\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{2}>t\right)\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{3}>s,\eta_{3}>t\right)
=~(s,t)(η1>s)~(s,t)(η2>t)~(s,t)(η3>max(s,t))\displaystyle=\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{1}>s\right)\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{2}>t\right)\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{3}>\max(s,t)\right)
=exp[As1At2Ast3]\displaystyle=\exp\left[-A^{1}_{s}-A^{2}_{t}-A^{3}_{s\vee t}\right]
={exp[As1At2As3]s>texp[As1At2At3]s<t.\displaystyle=\begin{cases}\exp\left[-A^{1}_{s}-A^{2}_{t}-A^{3}_{s}\right]&s>t\\ \exp\left[-A^{1}_{s}-A^{2}_{t}-A^{3}_{t}\right]&s<t.\end{cases}

The result follows by taking the expectation. ∎

Remark 2.1.

From Theorem 1, it is clear that τ1\tau_{1} is not independent of τ2\tau_{2} and, as we will show in Subsection (2.2), their joint distribution has an absolutely continuous and a non-trivial singular part, which means that:

(τ1=τ2)>0.\mathbb{P}\left(\tau_{1}=\tau_{2}\right)>0. (5)
Remark 2.2.

Note that when Xt=t+X_{t}=t\in\mathbb{R}^{+} (i.e., there is no randomness coming from XtX_{t}), Theorem 1 is a generalization of the bivariate exponential (BVE) distribution introduced by Marshall and Olkin [28]. In the specific case where αi(Xs)=αi\alpha^{i}(X_{s})=\alpha^{i}\in\mathbb{R}. i.e., the intensity is constant, we recover the Marshall and Olkin BVE with parameter (α1,α2,α3)(\alpha^{1},\alpha^{2},\alpha^{3}). This is (τ1,τ2)BVE(α1,α2,α3)(\tau_{1},\tau_{2})\sim\text{BVE}(\alpha^{1},\alpha^{2},\alpha^{3}).

Remark 2.3.

The marginal distribution of τi\tau_{i} is given by:

\displaystyle\mathbb{P} (τi>s)=(min(ηi,η3)>s)=(ηi>s,η3>s)=𝔼[(ηi>s,η3>s(Xu)us)]\displaystyle(\tau_{i}>s)=\mathbb{P}(\min(\eta_{i},\eta_{3})>s)=\mathbb{P}\left(\eta_{i}>s,\eta_{3}>s\right)=\mathbb{E}\left[\mathbb{P}\left(\eta_{i}>s,\eta_{3}>s\mid\left(X_{u}\right)_{u\leq s}\right)\right]
=𝔼[(ηi>s(Xu)us)(η3>s(Xu)us)]=𝔼[exp(0s(αi+α3)(r)𝑑r)]\displaystyle=\mathbb{E}\left[\mathbb{P}\left(\eta_{i}>s\mid\left(X_{u}\right)_{u\leq s}\right)\mathbb{P}\left(\eta_{3}>s\mid\left(X_{u}\right)_{u\leq s}\right)\right]=\mathbb{E}\left[\exp\left(-\int_{0}^{s}(\alpha^{i}+\alpha^{3})(r)dr\right)\right] (6)

which coincides with the marginal distribution in the Cox construction, see Lando [24] [25]. That is, each of the stopping times is functional of a Cox process. However, by construction, they are not independent of each other.

Remark 2.4.

αi()\alpha^{i}() is usually known as the intensity of ηi\eta_{i} in the progressive enlargement of 𝔽\mathbb{F} with ηi\eta_{i}. (See Corollary 2.26 in Aksamit and Jeanblanc [1]).

Remark 2.5.

The intensity of τi\tau_{i} for i=1,2i=1,2 in the progressive enlargement of 𝔽\mathbb{F} with τi\tau_{i} is (αi+α3)()\left(\alpha^{i}+\alpha^{3}\right)\left(\cdot\right).

Caveat. In the rest of the paper, for easiness of notation, sometimes we write αsi\alpha^{i}_{s} instead of αi(Xs)\alpha^{i}\left(X_{s}\right) for i=1,2,3i=1,2,3.

2.1 Interpretation of Joint Distribution

A way to interpret the distribution given in Theorem 1 is the following: Suppose we have a two-component system where each component’s life is represented by τ1\tau_{1} and τ2\tau_{2} respectively. Any component dies after receiving a shock. Shocks are governed by three independent Cox processes Λ1(t,α1(Xt))\Lambda_{1}\left(t,\alpha^{1}(X_{t})\right), Λ2(t,α2(Xt))\Lambda_{2}\left(t,\alpha^{2}(X_{t})\right) and Λ3(t,α3(Xt))\Lambda_{3}\left(t,\alpha^{3}(X_{t})\right), in other words, Λi(t,αi(Xt))\Lambda_{i}\left(t,\alpha^{i}\left(X_{t}\right)\right) is an inhomogeneous Poisson process with stochastic intensity αi(Xt)\alpha^{i}\left(X_{t}\right). For j=1,2j=1,2, Λj(t,αj(Xt))\Lambda_{j}\left(t,\alpha^{j}(X_{t})\right) represents the number of shocks through time that affect only component jj, while Λ3(t,α3(Xt))\Lambda_{3}\left(t,\alpha^{3}(X_{t})\right) represents the number of shocks through time that affect both components.

Events in the process Λ1(t,α1(Xt))\Lambda_{1}\left(t,\alpha^{1}(X_{t})\right) are shocks to component 1, events in the process Λ2(t,α2(Xt))\Lambda_{2}\left(t,\alpha^{2}(X_{t})\right) are shocks to component 2 and events in the process Λ3(t,α3(Xt))\Lambda_{3}\left(t,\alpha^{3}(X_{t})\right) are shocks to both components.

Hence, we get,

(τ1>s,τ2>t)=𝔼[(τ1>s,τ2>t|(Xu)0u(st))]=𝔼[~(s,t)[Λ1(s,α1(Xs))=0]~(s,t)[Λ2(t,α2(Xt))=0]~(s,t)[Λ3(st,α3(Xst))=0]]=𝔼[exp(As1At2Ast3)]\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t\right)=\mathbb{E}\left[\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|(X_{u})_{0\leq u\leq(s\vee t)}\right)\right]\\ =\mathbb{E}\left[\tilde{\mathbb{P}}_{(s,t)}\left[\Lambda_{1}\left(s,\alpha^{1}(X_{s})\right)=0\right]\tilde{\mathbb{P}}_{(s,t)}\left[\Lambda_{2}\left(t,\alpha^{2}(X_{t})\right)=0\right]\tilde{\mathbb{P}}_{(s,t)}\left[\Lambda_{3}\left(s\vee t,\alpha^{3}(X_{s\vee t})\right)=0\right]\right]\\ =\mathbb{E}\left[\exp\left(-A^{1}_{s}-A^{2}_{t}-A^{3}_{s\vee t}\right)\right] (7)

Which coincides with Theorem 1. In the second line of the previous expression ~(s,t)()\tilde{\mathbb{P}}_{(s,t)}(\cdot) means (|(Xu)0u(st))\mathbb{P}\left(\cdot|(X_{u})_{0\leq u\leq(s\vee t)}\right).

2.2 Decomposition of the Joint Distribution

As found in Theorem 1 and assuming that Xt=t+X_{t}=t\in\mathbb{R}^{+}, we have that the joint survival function is:

F¯(τ1,τ2)(s,t)=exp[As1At2Ast3]\overline{F}_{(\tau_{1},\tau_{2})}(s,t)=\exp\left[-A^{1}_{s}-A^{2}_{t}-A^{3}_{s\vee t}\right]
Theorem 2.

Under the conditions of Theorem 1 and assuming that Xt=tX_{t}=t\in\mathbb{R}, the absolutely continuous and singular parts of F¯(τ1,τ2)(s,t)\overline{F}_{(\tau_{1},\tau_{2})}(s,t) are given by:

F¯(τ1,τ2)(s,t)=βF¯a(s,t)+(1β)F¯sing(s,t)\overline{F}_{(\tau_{1},\tau_{2})}(s,t)=\beta\overline{F}_{\text{a}}(s,t)+(1-\beta)\overline{F}_{\text{sing}}(s,t) (8)

where

β:\displaystyle\beta: =0(αs1+αs2)exp[As1As2As3]𝑑s\displaystyle=\int\limits_{0}^{\infty}\left(\alpha^{1}_{s}+\alpha^{2}_{s}\right)\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds
F¯a(s,t)\displaystyle\overline{F}_{\text{a}}(s,t) =1β[eAs1At2Ast3stαx3exp(i=13Axi)𝑑x]\displaystyle=\frac{1}{\beta}\left[e^{-A^{1}_{s}-A^{2}_{t}-A^{3}_{s\vee t}}-\int_{s\vee t}^{\infty}\alpha^{3}_{x}\exp\left(-\sum_{i=1}^{3}A^{i}_{x}\right)dx\right]
F¯sing(s,t)\displaystyle\overline{F}_{\text{sing}}(s,t) =11β[stαx3exp[Ax1Ax2Ax3]𝑑x].\displaystyle=\frac{1}{1-\beta}\left[\int_{s\vee t}^{\infty}\alpha^{3}_{x}\exp\left[-A^{1}_{x}-A^{2}_{x}-A^{3}_{x}\right]dx\right].
Proof.

Let Λi\Lambda_{i} be the waiting time to the first shock in the process Λi(t,αti)\Lambda_{i}(t,\alpha^{i}_{t}) defined in Section 2.1. (Recall that we are assuming that Xt=tX_{t}=t). By the assumptions made in Section 2.1, Λi\Lambda_{i} are independent and the survival function of each one of them is given by:

(Λi>t)=exp[Ati] for all t0.\mathbb{P}\left(\Lambda_{i}>t\right)=\exp\left[-A^{i}_{t}\right]\text{ for all }t\geq 0. (9)

Hence, the density is equal to:

fΛi(s)=αsiexp(Asi) for all s0.f_{\Lambda_{i}}(s)=\alpha^{i}_{s}\exp\left(-A^{i}_{s}\right)\text{ for all }s\geq 0. (10)

F¯(τ1,τ2)(s,t)\overline{F}_{(\tau_{1},\tau_{2})}(s,t) can be written as,

F¯(τ1,τ2)(s,t)=(τ1>s,τ2>t|B)(B)+(τ1>s,τ2>t|Bc)(Bc)\overline{F}_{(\tau_{1},\tau_{2})}(s,t)=\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|B\right)\mathbb{P}\left(B\right)+\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|B^{c}\right)\mathbb{P}\left(B^{c}\right) (11)

where B:={Λ3>min(Λ1,Λ2)}B:=\left\{\Lambda_{3}>\min(\Lambda_{1},\Lambda_{2})\right\} and Bc={Λ3min(Λ1,Λ2)}B^{c}=\left\{\Lambda_{3}\leq\min(\Lambda_{1},\Lambda_{2})\right\}

Now, using the density given in (10):

(Bc)=\displaystyle\mathbb{P}\left(B^{c}\right)= (Λ3Λ1,Λ3Λ2)\displaystyle\mathbb{P}\left(\Lambda_{3}\leq\Lambda_{1},\ \Lambda_{3}\leq\Lambda_{2}\right)
=\displaystyle= (Λ3Λ1Λ2)+(Λ3Λ2Λ1)\displaystyle\mathbb{P}\left(\Lambda_{3}\leq\Lambda_{1}\leq\Lambda_{2}\right)+\mathbb{P}\left(\Lambda_{3}\leq\Lambda_{2}\leq\Lambda_{1}\right)
=\displaystyle= 0zxαx1αy2αz3exp[Ax1Ay2Az3]𝑑y𝑑x𝑑z\displaystyle\int\limits_{0}^{\infty}\int\limits_{z}^{\infty}\int\limits_{x}^{\infty}\alpha^{1}_{x}\alpha^{2}_{y}\alpha^{3}_{z}\exp\left[-A^{1}_{x}-A^{2}_{y}-A^{3}_{z}\right]dydxdz
+0zyαx1αy2αz3exp[Ax1Ay2Az3]𝑑x𝑑y𝑑z\displaystyle+\int\limits_{0}^{\infty}\int\limits_{z}^{\infty}\int\limits_{y}^{\infty}\alpha^{1}_{x}\alpha^{2}_{y}\alpha^{3}_{z}\exp\left[-A^{1}_{x}-A^{2}_{y}-A^{3}_{z}\right]dxdydz
=\displaystyle= 0αs3exp[As1As2As3]𝑑s\displaystyle\int_{0}^{\infty}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds (12)

From this, (B)=0(αs1+αs2)exp[As1As2As3]𝑑s\mathbb{P}\left(B\right)=\int_{0}^{\infty}\left(\alpha^{1}_{s}+\alpha^{2}_{s}\right)\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds

Since (min(Λ1,Λ2,Λ3)>t)=i=13(Λi>t)=exp[i=13Ati]\mathbb{P}\left(\min(\Lambda_{1},\Lambda_{2},\Lambda_{3})>t\right)=\prod_{i=1}^{3}\mathbb{P}\left(\Lambda_{i}>t\right)=\exp\left[-\sum_{i=1}^{3}A^{i}_{t}\right], we get

(τ1>s,τ2>t|Bc)\displaystyle\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|B^{c}\right) =[Λ1>s,Λ2>t,Λ3>(st)|Λ3min(Λ1,Λ2)]\displaystyle=\mathbb{P}\left[\Lambda_{1}>s,\Lambda_{2}>t,\Lambda_{3}>(s\vee t)|\Lambda_{3}\leq\min(\Lambda_{1},\Lambda_{2})\right]
=[Λ3>(st)|Λ3min(Λ1,Λ2)]\displaystyle=\mathbb{P}\left[\Lambda_{3}>(s\vee t)|\Lambda_{3}\leq\min(\Lambda_{1},\Lambda_{2})\right]
=[(st)<Λ3min(Λ1,Λ2)][Λ3min(Λ1,Λ2)]\displaystyle=\frac{\mathbb{P}\left[\left(s\vee t\right)<\Lambda_{3}\leq\min(\Lambda_{1},\Lambda_{2})\right]}{\mathbb{P}\left[\Lambda_{3}\leq\min(\Lambda_{1},\Lambda_{2})\right]}
=stαx3exp[Ax1Ax2Ax3]𝑑x0αx3exp[Ax1Ax2Ax3]𝑑x\displaystyle=\frac{\int_{s\vee t}^{\infty}\alpha^{3}_{x}\exp\left[-A^{1}_{x}-A^{2}_{x}-A^{3}_{x}\right]dx}{\int_{0}^{\infty}\alpha^{3}_{x}\exp\left[-A^{1}_{x}-A^{2}_{x}-A^{3}_{x}\right]dx} (13)

With all these elements, (τ1>s,τ2>t|B)\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|B\right) is obtained by subtraction.

We can verify that (τ1>s,τ2>t|Bc)\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|B^{c}\right) is a singular distribution since its mixed second partial derivative is zero when sts\neq t. Conversely, (τ1>s,τ2>t|B)\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t|B\right) is absolutely continuous since its mixed second partial derivative is a density. ∎

Remark 2.6.

The value of (Bc)\mathbb{P}\left(B^{c}\right) corresponds to (τ1=τ2)\mathbb{P}(\tau_{1}=\tau_{2}). This is,

(τ1=τ2)=0αs3exp[As1As2As3]𝑑s.\mathbb{P}(\tau_{1}=\tau_{2})=\int_{0}^{\infty}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds. (14)
Corollary 2.1.

Let αui:=αi(Xu)\alpha^{i}_{u}:=\alpha^{i}(X_{u}), where αi()\alpha^{i}(\cdot) is a positive continuous function and XX is an d\mathbb{R}^{d}-valued stochastic process adapted to the filtration 𝔽\mathbb{F} and independent of Z1,Z2 and Z3Z_{1},Z_{2}\text{ and }Z_{3}. Then,

(τ1=τ2)=𝔼(0αs3exp[As1As2As3]𝑑s).\mathbb{P}\left(\tau_{1}=\tau_{2}\right)=\mathbb{E}\left(\int_{0}^{\infty}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds\right).
Proof.

By a similar calculation to get (Bc)\mathbb{P}(B^{c}) in the proof of Theorem 2, we get

(τ1=τ2|(Xu)u0)=0αs3exp[As1As2As3]𝑑s.\mathbb{P}\left(\tau_{1}=\tau_{2}|(X_{u})_{u\geq 0}\right)=\int_{0}^{\infty}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds.

The result follows by taking the expectation. ∎

2.3 Estimating the Probability of Equality in Two Stopping Times

Now, we are interested in finding estimates for (τ1=τ2)\mathbb{P}(\tau_{1}=\tau_{2}) (see equation (14) in Section 2.2) under different assumptions for αi()\alpha^{i}(\cdot) (the intensity of ηi\eta_{i}) or AsiA^{i}_{s} (the compensator of ηi\eta_{i}). As in Theorem 1, we assume limsAsi=\lim_{s\rightarrow\infty}A^{i}_{s}=\infty a.s. for every i=1,2,3i=1,2,3.

Example 2.1 (Constant intensity).

If αi(Xs)=αi+\alpha^{i}(X_{s})=\alpha^{i}\in\mathbb{R}^{+} for all s0s\geq 0, it follows that:

(τ1=τ2)\displaystyle\mathbb{P}(\tau_{1}=\tau_{2}) =0α3exp[(α1+α2+α3)s]𝑑s\displaystyle=\int_{0}^{\infty}\alpha^{3}\exp\left[-(\alpha^{1}+\alpha^{2}+\alpha^{3})s\right]ds
=α3α1+α2+α3\displaystyle=\frac{\alpha^{3}}{\alpha^{1}+\alpha^{2}+\alpha^{3}}
Example 2.2 (Same intensity).

If α1(Xs)=α2(Xs)=α3(Xs)=:α(Xs)\alpha^{1}(X_{s})=\alpha^{2}(X_{s})=\alpha^{3}(X_{s})=:\alpha(X_{s}) for all s0s\geq 0, we have that As1=As2=As3=:AsA^{1}_{s}=A^{2}_{s}=A^{3}_{s}=:A_{s} for all s0s\geq 0. Moreover, it is straighforward to get,

(τ1=τ2)=13\mathbb{P}\left(\tau_{1}=\tau_{2}\right)=\frac{1}{3}
Example 2.3 (Bounded intensity).

Assume Liβ(Xs)αi(Xs)Uiβ(Xs)L_{i}\beta(X_{s})\leq\alpha^{i}(X_{s})\leq U_{i}\beta(X_{s}) a.s. for all s0s\geq 0, for β()\beta(\cdot) positive, non-random, and integrable with 0β(Xs)𝑑s=\int_{0}^{\infty}\beta(X_{s})ds=\infty a.s., where Li,UiL_{i},U_{i} are positive real random variables, independent of everything else, and such that 0<Li<Ui<0<L_{i}<U_{i}<\infty. Then, we get

𝔼[3u1+u2+u3](τ1=τ2)𝔼[u31+2+3]\mathbb{E}\left[\frac{\ell_{3}}{u_{1}+u_{2}+u_{3}}\right]\leq\mathbb{P}(\tau_{1}=\tau_{2})\leq\mathbb{E}\left[\frac{u_{3}}{\ell_{1}+\ell_{2}+\ell_{3}}\right]
Example 2.4 (Intensity bounded by sum of intensities).

If L[α1(Xs)+α2(Xs)]α3(Xs)U[α1(Xs)+α2(Xs)]L\left[\alpha^{1}(X_{s})+\alpha^{2}(X_{s})\right]\leq\alpha^{3}(X_{s})\leq U\left[\alpha^{1}(X_{s})+\alpha^{2}(X_{s})\right] for a.s. all s0s\geq 0 where L,UL,U are positive real random variables, independent of everything else, and such that 0<L<U<L+1<0<L<U<L+1<\infty. Then:

𝔼[LU+1](τ1=τ2)𝔼[UL+1]\mathbb{E}\left[\frac{L}{U+1}\right]\leq\mathbb{P}(\tau_{1}=\tau_{2})\leq\mathbb{E}\left[\frac{U}{L+1}\right]

The proofs of these last two examples are straightforward conditioning arguments and, hence, left to the reader.

2.4 Conditional Probabilities

Through this Section, let ~():=(|(Xu)u0)\tilde{\mathbb{P}}(\cdot):=\mathbb{P}\left(\cdot|\left(X_{u}\right)_{u\geq 0}\right).

Proposition 2.1.
(τ1=τ2,τ1t)=𝔼[0tαs3exp[As1As2As3]𝑑s]\mathbb{P}\left(\tau_{1}=\tau_{2},\tau_{1}\leq t\right)=\mathbb{E}\left[\int_{0}^{t}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds\right]
Proof.

By the definition of τi\tau_{i}, we have

~(τ1=τ2,τ1>t)=\displaystyle\tilde{\mathbb{P}}(\tau_{1}=\tau_{2},\tau_{1}>t)= ~(min(η1,η3)=min(η2,η3),min(η1,η3)>t)\displaystyle\tilde{\mathbb{P}}\left(\min(\eta_{1},\eta_{3})=\min(\eta_{2},\eta_{3}),\min(\eta_{1},\eta_{3})>t\right)
=\displaystyle= ~(min(η1,η3)=η3=min(η2,η3),min(η1,η3)>t)\displaystyle\tilde{\mathbb{P}}\left(\min(\eta_{1},\eta_{3})=\eta_{3}=\min(\eta_{2},\eta_{3}),\min(\eta_{1},\eta_{3})>t\right)
=\displaystyle= ~(η3<η1,η3<η2,η1>t,η3>t)\displaystyle\tilde{\mathbb{P}}\left(\eta_{3}<\eta_{1},\eta_{3}<\eta_{2},\eta_{1}>t,\eta_{3}>t\right)
=\displaystyle= ~(t<η3<η1<η2)+~(t<η3<η2<η1)\displaystyle\tilde{\mathbb{P}}\left(t<\eta_{3}<\eta_{1}<\eta_{2}\right)+\tilde{\mathbb{P}}\left(t<\eta_{3}<\eta_{2}<\eta_{1}\right)

Recall that, by definition, ηi\eta_{i} are independent with density under ~\tilde{\mathbb{P}} equal to αsiexp(Asi)\alpha^{i}_{s}\exp\left(-A^{i}_{s}\right)

~(τ1=τ2,τ1>t)=\displaystyle\tilde{\mathbb{P}}(\tau_{1}=\tau_{2},\tau_{1}>t)= tzxαx1αy2αz3exp[Ax1Ay2Az3]𝑑y𝑑x𝑑z\displaystyle\int\limits_{t}^{\infty}\int\limits_{z}^{\infty}\int\limits_{x}^{\infty}\alpha^{1}_{x}\alpha^{2}_{y}\alpha^{3}_{z}\exp\left[-A^{1}_{x}-A^{2}_{y}-A^{3}_{z}\right]dydxdz
+tzyαx1αy2αz3exp[Ax1Ay2Az3]𝑑x𝑑y𝑑z\displaystyle+\int\limits_{t}^{\infty}\int\limits_{z}^{\infty}\int\limits_{y}^{\infty}\alpha^{1}_{x}\alpha^{2}_{y}\alpha^{3}_{z}\exp\left[-A^{1}_{x}-A^{2}_{y}-A^{3}_{z}\right]dxdydz
=\displaystyle= tαs3exp[As1As2As3]𝑑s\displaystyle\int_{t}^{\infty}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds

This implies,

~(τ1=τ2,τ1t)=0tαs3exp[As1As2As3]𝑑s\tilde{\mathbb{P}}(\tau_{1}=\tau_{2},\tau_{1}\leq t)=\int_{0}^{t}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds

Take the expectation to get the result. ∎

Remark 2.7.

When taking the limit tt\rightarrow\infty in the previous proposition, we recover, as expected, the result of Corollary 2.1.

The following two propositions are straightforward applications of Proposition 2.1. Hence, the proofs are left to the reader.

Proposition 2.2.
(τ1=τ2|τ1t)=𝔼[0tαs3exp[As1As2As3]𝑑s1exp[At1At3]]\mathbb{P}\left(\tau_{1}=\tau_{2}|\tau_{1}\leq t\right)=\mathbb{E}\left[\frac{\int_{0}^{t}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds}{1-\exp[-A^{1}_{t}-A^{3}_{t}]}\right]
Proposition 2.3.
(τ1=τ2|τ1t,τ2t)=𝔼[0tαs3exp[As1As2As3]𝑑s1eAt3(eAt2+eAt1eAt1At2)]\mathbb{P}\left(\tau_{1}=\tau_{2}|\tau_{1}\leq t,\tau_{2}\leq t\right)=\mathbb{E}\left[\frac{\int_{0}^{t}\alpha^{3}_{s}\exp\left[-A^{1}_{s}-A^{2}_{s}-A^{3}_{s}\right]ds}{1-e^{-A^{3}_{t}}\left(e^{-A^{2}_{t}}+e^{-A^{1}_{t}}-e^{-A^{1}_{t}-A^{2}_{t}}\right)}\right]

2.5 Distance Between Stopping Times

Proposition 2.4.

For s<ts<t,

(s<τ1t,sτ2t)=𝔼[e(As1+As2+As3)+e(At1+At2+At3)eAs1At2At3eAt1As2At3]\mathbb{P}\left(s<\tau_{1}\leq t,s\leq\tau_{2}\leq t\right)=\mathbb{E}\left[e^{-\left(A^{1}_{s}+A^{2}_{s}+A^{3}_{s}\right)}+e^{-\left(A^{1}_{t}+A^{2}_{t}+A^{3}_{t}\right)}\right.\\ \left.-e^{-A^{1}_{s}-A^{2}_{t}-A^{3}_{t}}-e^{-A^{1}_{t}-A^{2}_{s}-A^{3}_{t}}\right]

In particular, if s=rεs=r-\varepsilon and t=r+εt=r+\varepsilon, we get

(rε<τ1r+ε,rετ2r+ε)=𝔼[e(Arε1+Arε2+Arε3)+e(Ar+ε1+Ar+ε2+Ar+ε3)eArε1Ar+ε2Ar+ε3eAr+ε1Arε2Ar+ε3]\mathbb{P}\left(r-\varepsilon<\tau_{1}\leq r+\varepsilon,r-\varepsilon\leq\tau_{2}\leq r+\varepsilon\right)=\mathbb{E}\left[e^{-\left(A^{1}_{r-\varepsilon}+A^{2}_{r-\varepsilon}+A^{3}_{r-\varepsilon}\right)}\right.\\ \left.+e^{-\left(A^{1}_{r+\varepsilon}+A^{2}_{r+\varepsilon}+A^{3}_{r+\varepsilon}\right)}-e^{-A^{1}_{r-\varepsilon}-A^{2}_{r+\varepsilon}-A^{3}_{r+\varepsilon}}-e^{-A^{1}_{r+\varepsilon}-A^{2}_{r-\varepsilon}-A^{3}_{r+\varepsilon}}\right]
Proof.

It follows by noticing that:

(s<τ1t,sτ2t)=(τ1>s,τ2>s)(τ1>s,τ2>t)(τ1>t,τ2>s)+(τ1>t,τ2>t)\mathbb{P}\left(s<\tau_{1}\leq t,s\leq\tau_{2}\leq t\right)=\mathbb{P}\left(\tau_{1}>s,\tau_{2}>s\right)-\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t\right)\\ -\mathbb{P}\left(\tau_{1}>t,\tau_{2}>s\right)+\mathbb{P}\left(\tau_{1}>t,\tau_{2}>t\right)

Then, conditioning on (Xt)t0\left(X_{t}\right)_{t\geq 0}, use Theorem 1, and take the expectation. ∎

The next proposition shows that the probability of the two stopping times happening over the next time interval (t,t+ε)(t,t+\varepsilon) with ε0\varepsilon\approx 0 is equal to the expectation of α3(Xt)\alpha^{3}\left(X_{t}\right), which is the common part of the intensities of τ1\tau_{1} and τ2\tau_{2} (recall Remark 2.5)

An anonymous referee has pointed out a relation of the following proposition to Aven’s Lemma. This is, of course, correct, see Aven [2], Ethier and Kurtz [14], and Zeng [36].

Proposition 2.5.
limε0(τ1(t,t+ε],τ2(t,t+ε]|τ1>t,τ2>t)ε=𝔼(α3(Xt))\lim_{\varepsilon\rightarrow 0}\frac{\mathbb{P}\left(\tau_{1}\in\left(t,t+\varepsilon\right],\tau_{2}\in\left(t,t+\varepsilon\right]|\tau_{1}>t,\tau_{2}>t\right)}{\varepsilon}=\mathbb{E}\left(\alpha^{3}(X_{t})\right)
Proof.

Let ~():=(|(Xu)u0)\tilde{\mathbb{P}}(\cdot):=\mathbb{P}\left(\cdot|\left(X_{u}\right)_{u\geq 0}\right) and note that:

~(τ1(t,t+ε],τ2(t,t+ε]|τ1>t,τ2>t)=~(τ1(t,t+ε],τ2(t,t+ε])~(τ1>t,τ2>t)\tilde{\mathbb{P}}\left(\tau_{1}\in\left(t,t+\varepsilon\right],\tau_{2}\in\left(t,t+\varepsilon\right]|\tau_{1}>t,\tau_{2}>t\right)=\frac{\tilde{\mathbb{P}}\left(\tau_{1}\in\left(t,t+\varepsilon\right],\tau_{2}\in\left(t,t+\varepsilon\right]\right)}{\tilde{\mathbb{P}}\left(\tau_{1}>t,\tau_{2}>t\right)}

Using Proposition 2.4 with tt and t+εt+\varepsilon, dividing by ε\varepsilon, and using L’Hôpital’s rule to get the limit:

limε0~(τ1(t,t+ε],τ2(t,t+ε]|τ1>t,τ2>t)ε=α3(Xt).\lim_{\varepsilon\rightarrow 0}\frac{\tilde{\mathbb{P}}\left(\tau_{1}\in\left(t,t+\varepsilon\right],\tau_{2}\in\left(t,t+\varepsilon\right]|\tau_{1}>t,\tau_{2}>t\right)}{\varepsilon}=\alpha^{3}(X_{t}).

Finally, given ~(τ1(t,t+ε],τ2(t,t+ε]|τ1>t,τ2>t)\tilde{\mathbb{P}}\left(\tau_{1}\in\left(t,t+\varepsilon\right],\tau_{2}\in\left(t,t+\varepsilon\right]|\tau_{1}>t,\tau_{2}>t\right) is bounded by 1 and using a conditioning argument, we can conclude that:

limε0(τ1(t,t+ε],τ2(t,t+ε]|τ1>t,τ2>t)ε=𝔼(α3(Xt)).\lim_{\varepsilon\rightarrow 0}\frac{\mathbb{P}\left(\tau_{1}\in\left(t,t+\varepsilon\right],\tau_{2}\in\left(t,t+\varepsilon\right]|\tau_{1}>t,\tau_{2}>t\right)}{\varepsilon}=\mathbb{E}\left(\alpha^{3}(X_{t})\right).

Proposition 2.5 motivates the following one:

Proposition 2.6.

If s<ts<t

limε0(τ1s+ε,τ2t+ε|τ1>s,τ2>t)ε2=𝔼[α1(Xs)(α2(Xt)+α3(Xt))]\lim_{\varepsilon\rightarrow 0}\frac{\mathbb{P}\left(\tau_{1}\leq s+\varepsilon,\tau_{2}\leq t+\varepsilon|\tau_{1}>s,\tau_{2}>t\right)}{\varepsilon^{2}}=\mathbb{E}\left[\alpha^{1}(X_{s})\left(\alpha^{2}(X_{t})+\alpha^{3}(X_{t})\right)\right]
Proof.

As s<ts<t, we can always find a sufficiently small ε\varepsilon such that s+ε<ts+\varepsilon<t. Hence, without loss of generality, we assume s<s+ε<t<t+εs<s+\varepsilon<t<t+\varepsilon and we proceed as in the proof of Proposition 2.5, but differently from there, we use L’Hôpital’s rule twice to get the desired limit. ∎

The next couple of propositions provide a measure, in two different metrics, of how close the two stopping are from each other.

Proposition 2.7.

(Distance in probability)

(|τ1τ2|ε)=1𝔼[0αx1eAx1(Ax+ε2+Ax+ε3)𝑑x+0αx2eAx2(Ax+ε1+Ax+ε3)𝑑x]\mathbb{P}(|\tau_{1}-\tau_{2}|\leq\varepsilon)=1-\mathbb{E}\left[\int\limits_{0}^{\infty}\alpha^{1}_{x}e^{-A^{1}_{x}-\left(A^{2}_{x+\varepsilon}+A^{3}_{x+\varepsilon}\right)}\,dx+\int\limits_{0}^{\infty}\alpha^{2}_{x}e^{-A^{2}_{x}-\left(A^{1}_{x+\varepsilon}+A^{3}_{x+\varepsilon}\right)}\,dx\right]
Proof.

Let τ(1):=min(τ1,τ2)\tau_{(1)}:=\min\left(\tau_{1},\tau_{2}\right) and τ(2):=max(τ1,τ2)\tau_{(2)}:=\max\left(\tau_{1},\tau_{2}\right). Similarly, η(1):=min(η1,η2,η3)\eta_{(1)}:=\min\left(\eta_{1},\eta_{2},\eta_{3}\right), η(3):=max(η1,η2,η3)\eta_{(3)}:=\max\left(\eta_{1},\eta_{2},\eta_{3}\right), and η(2)\eta_{(2)} be the second largest from (η1,η2,η3)\left(\eta_{1},\eta_{2},\eta_{3}\right). Also, let ~():=(|(Xu)u0)\tilde{\mathbb{P}}(\cdot):=\mathbb{P}\left(\cdot|\left(X_{u}\right)_{u\geq 0}\right). Then,

~(τ(2)>τ(1)+ε)=\displaystyle\tilde{\mathbb{P}}\left(\tau_{(2)}>\tau_{(1)}+\varepsilon\right)= ~(η(3)η(2)>η(1)+ε,η3η(2))\displaystyle\tilde{\mathbb{P}}\left(\eta_{(3)}\geq\eta_{(2)}>\eta_{(1)}+\varepsilon,\eta_{3}\geq\eta_{(2)}\right)
=\displaystyle= 0f1(x1)x1+εf2(x2)x2f3(x3)𝑑x3𝑑x2𝑑x1\displaystyle\int\limits_{0}^{\infty}f_{1}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{2}(x_{2})\int\limits_{x_{2}}^{\infty}f_{3}(x_{3})\,dx_{3}\,dx_{2}\,dx_{1}
+0f2(x1)x1+εf1(x2)x2f3(x3)𝑑x3𝑑x2𝑑x1\displaystyle+\int\limits_{0}^{\infty}f_{2}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{1}(x_{2})\int\limits_{x_{2}}^{\infty}f_{3}(x_{3})\,dx_{3}\,dx_{2}\,dx_{1}
+0f1(x1)x1+εf3(x2)x2f2(x3)𝑑x3𝑑x2𝑑x1\displaystyle+\int\limits_{0}^{\infty}f_{1}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{3}(x_{2})\int\limits_{x_{2}}^{\infty}f_{2}(x_{3})\,dx_{3}\,dx_{2}\,dx_{1}
+0f2(x1)x1+εf3(x2)x2f1(x3)𝑑x3𝑑x2𝑑x1\displaystyle+\int\limits_{0}^{\infty}f_{2}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{3}(x_{2})\int\limits_{x_{2}}^{\infty}f_{1}(x_{3})\,dx_{3}\,dx_{2}\,dx_{1}
where fjf_{j} stands for the density of ηj\eta_{j} under ~\tilde{\mathbb{P}}, i.e., fj(y)=αyjeAyj:=αj(Xy)eAyjf_{j}(y)=\alpha^{j}_{y}e^{-A^{j}_{y}}:=\alpha^{j}(X_{y})e^{-A^{j}_{y}}
=\displaystyle= 0f1(x1)x1+εf2(x2)eAx23𝑑x2𝑑x1+0f2(x1)x1+εf1(x2)eAx23𝑑x2𝑑x1\displaystyle\int\limits_{0}^{\infty}f_{1}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{2}(x_{2})e^{-A^{3}_{x_{2}}}\,dx_{2}\,dx_{1}+\int\limits_{0}^{\infty}f_{2}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{1}(x_{2})e^{-A^{3}_{x_{2}}}\,dx_{2}\,dx_{1}
+0f1(x1)x1+εf3(x2)eAx22𝑑x2𝑑x1+0f2(x1)x1+εf3(x2)eAx21𝑑x2𝑑x1\displaystyle+\int\limits_{0}^{\infty}f_{1}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{3}(x_{2})e^{-A^{2}_{x_{2}}}\,dx_{2}\,dx_{1}+\int\limits_{0}^{\infty}f_{2}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}f_{3}(x_{2})e^{-A^{1}_{x_{2}}}\,dx_{2}\,dx_{1}
=\displaystyle= 0f1(x1)x1+ε[αx22+αx23]eAx22Ax23𝑑x2𝑑x1\displaystyle\int\limits_{0}^{\infty}f_{1}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}\left[\alpha^{2}_{x_{2}}+\alpha^{3}_{x_{2}}\right]e^{-A^{2}_{x_{2}}-A^{3}_{x_{2}}}\,dx_{2}\,dx_{1}
+0f2(x1)x1+ε[αx21+αx23]eAx21Ax23𝑑x2𝑑x1\displaystyle+\int\limits_{0}^{\infty}f_{2}(x_{1})\int\limits_{x_{1}+\varepsilon}^{\infty}\left[\alpha^{1}_{x_{2}}+\alpha^{3}_{x_{2}}\right]e^{-A^{1}_{x_{2}}-A^{3}_{x_{2}}}\,dx_{2}\,dx_{1}
=\displaystyle= 0αx11eAx11(Ax1+ε2+Ax1+ε3)𝑑x1+0αx12eAx12(Ax1+ε1+Ax1+ε3)𝑑x1\displaystyle\int\limits_{0}^{\infty}\alpha^{1}_{x_{1}}e^{-A^{1}_{x_{1}}-\left(A^{2}_{x_{1}+\varepsilon}+A^{3}_{x_{1}+\varepsilon}\right)}\,dx_{1}+\int\limits_{0}^{\infty}\alpha^{2}_{x_{1}}e^{-A^{2}_{x_{1}}-\left(A^{1}_{x_{1}+\varepsilon}+A^{3}_{x_{1}+\varepsilon}\right)}\,dx_{1}

Use this expression, along with the following equality, and then take the expectation to get the desired result

~(|τ1τ2|ε)=1~(τ(2)τ(1)>ε)\displaystyle\tilde{\mathbb{P}}\left(|\tau_{1}-\tau_{2}|\leq\varepsilon\right)=1-\tilde{\mathbb{P}}\left(\tau_{(2)}-\tau_{(1)}>\varepsilon\right)

Proposition 2.8 (L2L^{2} distance).

If 𝔼(τi2)<\mathbb{E}\left(\tau_{i}^{2}\right)<\infty and limxx2eAxi=0\lim_{x\rightarrow\infty}x^{2}e^{-A^{i}_{x}}=0 a.s. for i=1,2i=1,2 (in other words, eAxie^{-A^{i}_{x}} goes faster to 0 than x2x^{2} goes to infinity when xx\rightarrow\infty), we have

𝔼[(τ1τ2)2]=2𝔼[0xeAx1Ax3dx+0xeAx2Ax3dx00yeAx1Ay2Ay3dxdy0yeAx1Ay2Ax3dxdy]\mathbb{E}\left[\left(\tau_{1}-\tau_{2}\right)^{2}\right]=2\mathbb{E}\left[\int_{0}^{\infty}xe^{-A^{1}_{x}-A^{3}_{x}}dx+\int_{0}^{\infty}xe^{-A^{2}_{x}-A^{3}_{x}}dx\right.\\ \left.-\int_{0}^{\infty}\int_{0}^{y}e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{y}}dxdy-\int_{0}^{\infty}\int_{y}^{\infty}e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{x}}dxdy\right]
Proof.

Let 𝔼~():=𝔼(|(Xu)u0)\tilde{\mathbb{E}}(\cdot):=\mathbb{E}(\cdot|(X_{u})_{u\geq 0}). FτiF_{\tau_{i}} and F¯τi\overline{F}_{\tau_{i}} stand for the cumulative and the survival distribution functions of τi\tau_{i} given (Xu)u0(X_{u})_{u\geq 0}. Similarly, F(x,y)F(x,y) and F¯(x,y)\overline{F}(x,y) stand for the cumulative and the survival joint distribution functions of τ1,τ2\tau_{1},\tau_{2} given (Xu)u0(X_{u})_{u\geq 0}.

We expand the square and handle each term separately. For 𝔼~(τi2)<\tilde{\mathbb{E}}\left(\tau_{i}^{2}\right)<\infty, we use integration by parts in the following way:

𝔼~(τi2)\displaystyle\tilde{\mathbb{E}}\left(\tau_{i}^{2}\right) =0x2𝑑Fτi(x)\displaystyle=\int_{0}^{\infty}x^{2}dF_{\tau_{i}}(x)
=20xF¯τi(x)𝑑x\displaystyle=2\int_{0}^{\infty}x\overline{F}_{\tau_{i}}(x)dx
=20xeAxiAx3𝑑x\displaystyle=2\int_{0}^{\infty}xe^{-A^{i}_{x}-A^{3}_{x}}dx

To find 𝔼~(τ1τ2)\tilde{\mathbb{E}}(\tau_{1}\tau_{2}), we exploit the result of Young [35] on integration by parts in two or more dimensions. If G(0,y)0G(y,0)G(0,y)\equiv 0\equiv G(y,0) and GG is of bounded variation on finite intervals, then:

00G(x,y)F(dx,dy)=00F¯(x,y)G(dx,dy)\int_{0}^{\infty}\int_{0}^{\infty}G(x,y)F(dx,dy)=\int_{0}^{\infty}\int_{0}^{\infty}\overline{F}(x,y)G(dx,dy) (15)

This equality implies that, for i,j>0i,j>0:

00xiyjF(dx,dy)=00ijxi1yj1F¯(x,y)𝑑x𝑑y\int_{0}^{\infty}\int_{0}^{\infty}x^{i}y^{j}F(dx,dy)=\int_{0}^{\infty}\int_{0}^{\infty}ijx^{i-1}y^{j-1}\overline{F}(x,y)dxdy (16)

Hence,

𝔼~(τ1τ2)\displaystyle\tilde{\mathbb{E}}\left(\tau_{1}\tau_{2}\right) =00xyF(τ1,τ2)(dx,dy)\displaystyle=\int_{0}^{\infty}\int_{0}^{\infty}xyF_{\left(\tau_{1},\tau_{2}\right)}(dx,dy)
=00eAx1Ay2Axy3𝑑x𝑑y\displaystyle=\int_{0}^{\infty}\int_{0}^{\infty}e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{x\vee y}}dxdy
=00yeAx1Ay2Ay3𝑑x𝑑y+0yeAx1Ay2Ax3𝑑x𝑑y\displaystyle=\int_{0}^{\infty}\int_{0}^{y}e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{y}}dxdy+\int_{0}^{\infty}\int_{y}^{\infty}e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{x}}dxdy (17)

The result follows by taking the expectation. ∎

Remark 2.8.

In the case of independence of τ1\tau_{1} and τ2\tau_{2}, as As3=0A^{3}_{s}=0 for all s0s\geq 0, we get that:

𝔼[(τ1τ2)2]=2𝔼[0xeAx1𝑑x+0xeAx2𝑑x0eAx1𝑑x0eAx2𝑑x]\mathbb{E}\left[\left(\tau_{1}-\tau_{2}\right)^{2}\right]=2\mathbb{E}\left[\int_{0}^{\infty}xe^{-A^{1}_{x}}dx+\int_{0}^{\infty}xe^{-A^{2}_{x}}dx-\int_{0}^{\infty}e^{-A^{1}_{x}}dx\int_{0}^{\infty}e^{-A^{2}_{x}}dx\right] (18)

3 GENERALIZATIONS

3.1 Generalization to K Stopping Times.

Given the interpretation introduced in Section (2.1), there is a natural way to extend our model to more than two stopping times. We explicitly motivate and present the case of 3 stopping times. Suppose we have a three-component system where each component’s life is represented by τ1\tau_{1}, τ2\tau_{2}, and τ3\tau_{3} respectively. Any component dies after receiving a shock, which are governed by 6 Cox processes, which are independent given (Xt)t0(X_{t})_{t\geq 0}:

Λ1(t,α1(Xt)),Λ2(t,α2(Xt)),Λ3(t,α3(Xt)),Λ(1,2)(t,α(1,2)(Xt)),Λ(2,3)(t,α(2,3)(Xt)),Λ(1,3)(t,α(1,3)(Xt)) and Λ(1,2,3)(t,α(1,2,3)(Xt))\Lambda_{1}\left(t,\alpha^{1}(X_{t})\right),\Lambda_{2}\left(t,\alpha^{2}(X_{t})\right),\Lambda_{3}\left(t,\alpha^{3}(X_{t})\right),\Lambda_{(1,2)}\left(t,\alpha^{(1,2)}(X_{t})\right),\Lambda_{(2,3)}\left(t,\alpha^{(2,3)}(X_{t})\right),\\ \Lambda_{(1,3)}\left(t,\alpha^{(1,3)}(X_{t})\right)\text{ and }\Lambda_{(1,2,3)}\left(t,\alpha^{(1,2,3)}(X_{t})\right)

Events in the process Λi(t,αi(Xt))\Lambda_{i}\left(t,\alpha^{i}(X_{t})\right) are shocks to only component ii (for i=1,2,3i=1,2,3), events in the process Λ(i,j)(t,α(i,j)(Xt))\Lambda_{(i,j)}\left(t,\alpha^{(i,j)}(X_{t})\right) are shocks to component ii and jj (for i,j=1,2,3i,j=1,2,3 and iji\neq j), and events in the process Λ(1,2,3)(t,α(1,2,3)(Xt))\Lambda_{(1,2,3)}\left(t,\alpha^{(1,2,3)}(X_{t})\right) are shocks to the 3 components.

In this way, considering ~():=(|(Xu)u0)\tilde{\mathbb{P}}(\cdot):=\mathbb{P}\left(\cdot|(X_{u})_{u\geq 0}\right) and Asj=0sαj(Xu)𝑑uA^{j}_{s}=\int_{0}^{s}\alpha^{j}(X_{u})du we have:

(τ1>s1,τ2>s2,τ3>s3)=𝔼(~[Λ1(s1,α1(Xs1))=0]~[Λ2(s2,α2(Xs2))=0]~[Λ3(s3,α3(Xs3))=0]~[Λ(1,2)(s1s2,α(1,2)(Xs1s2))=0]~[Λ(1,3)(s1s3,α(1,3)(Xs1s3))=0]~[Λ(2,3)(s2s3,α(2,3)(Xs2s3))=0]~[Λ(1,2,3)(s1s2s3,α(1,2,3)(Xs1s2s3))=0])=𝔼(exp[As11As22As33As1s2(1,2)As1s3(1,3)As2s3(2,3)As1s2s3(1,2,3)])\mathbb{P}(\tau_{1}>s_{1},\tau_{2}>s_{2},\tau_{3}>s_{3})=\mathbb{E}\Bigl{(}\tilde{\mathbb{P}}\left[\Lambda_{1}\left(s_{1},\alpha^{1}(X_{s_{1}})\right)=0\right]\tilde{\mathbb{P}}\left[\Lambda_{2}\left(s_{2},\alpha^{2}(X_{s_{2}})\right)=0\right]\\ \tilde{\mathbb{P}}\left[\Lambda_{3}\left(s_{3},\alpha^{3}(X_{s_{3}})\right)=0\right]\tilde{\mathbb{P}}\left[\Lambda_{(1,2)}\left(s_{1}\vee s_{2},\alpha^{(1,2)}(X_{s_{1}\vee s_{2}})\right)=0\right]\tilde{\mathbb{P}}\left[\Lambda_{(1,3)}\left(s_{1}\vee s_{3},\alpha^{(1,3)}(X_{s_{1}\vee s_{3}})\right)=0\right]\\ \tilde{\mathbb{P}}\left[\Lambda_{(2,3)}\left(s_{2}\vee s_{3},\alpha^{(2,3)}(X_{s_{2}\vee s_{3}})\right)=0\right]\tilde{\mathbb{P}}\left[\Lambda_{(1,2,3)}\left(s_{1}\vee s_{2}\vee s_{3},\alpha^{(1,2,3)}(X_{s_{1}\vee s_{2}\vee s_{3}})\right)=0\right]\Bigr{)}\\ =\mathbb{E}\left(\exp\left[-A^{1}_{s_{1}}-A^{2}_{s_{2}}-A^{3}_{s_{3}}-A^{(1,2)}_{s_{1}\vee s_{2}}-A^{(1,3)}_{s_{1}\vee s_{3}}-A^{(2,3)}_{s_{2}\vee s_{3}}-A^{(1,2,3)}_{s_{1}\vee s_{2}\vee s_{3}}\right]\right) (19)

By using a similar technique we can generalize to any number KK of stopping times and in Section 4.1, we will present an application of this natural extension of our model. However, as the number of stopping times KK increases, handling the expression presented in equation (19) becomes cumbersome. In Jarrow et al. [21], we propose and study in detail a slightly different approach that makes easier to generalize to more than two stopping times.

3.2 A More General Distribution

In this Section, we generalize the result from Section 2 to get a distribution that allows τ1\tau_{1} and τ2\tau_{2} to have a negative covariance 111We thank Philip Ernst and Guodong Pang for suggesting this idea in an earlier version of this paper.. Recalling Equation (2.5) and that 𝔼~(τi)=0eAsiAs3𝑑s\tilde{\mathbb{E}}\left(\tau_{i}\right)=\int_{0}^{\infty}e^{-A^{i}_{s}-A^{3}_{s}}ds, one can see that the covariance of τ1\tau_{1} and τ2\tau_{2} in the previous model is:

Cov(τ1,τ2)\displaystyle\text{Cov}\left(\tau_{1},\tau_{2}\right) =𝔼[00(eAx1Ay2Axy3eAx1Ay2Ax3Ay3)𝑑x𝑑y]\displaystyle=\mathbb{E}\left[\int_{0}^{\infty}\int_{0}^{\infty}\left(e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{x\vee y}}-e^{-A^{1}_{x}-A^{2}_{y}-A^{3}_{x}-A^{3}_{y}}\right)dxdy\right]
=𝔼[00eAx1Ay2(eAxy3eAx3Ay3)𝑑x𝑑y]\displaystyle=\mathbb{E}\left[\int_{0}^{\infty}\int_{0}^{\infty}e^{-A^{1}_{x}-A^{2}_{y}}\left(e^{-A^{3}_{x\vee y}}-e^{-A^{3}_{x}-A^{3}_{y}}\right)dxdy\right] (20)

As Axy3<Ax3+Ay3A^{3}_{x\vee y}<A^{3}_{x}+A^{3}_{y} for all x,y0x,y\geq 0, it is clear that the covariance is always nonnegative.

To get the property of a negative covariance, we stop assuming that the underlying exponential random variables (i.e., Z1,Z2,Z3Z_{1},Z_{2},Z_{3}) are independent (recall equation (4) where we use these random variables to define ηi\eta_{i} and AsiA^{i}_{s}). Now, we assume that (Z1,Z2)(Z_{1},Z_{2}) follow the Gumbel Bivariate Distribution (see Gumbel [17]). This is, the joint survival function of (Z1,Z2)(Z_{1},Z_{2}) is:

(Z1>s,Z2>t)=estδst for 0δ1\mathbb{P}(Z_{1}>s,Z_{2}>t)=e^{-s-t-\delta st}\text{ for }0\leq\delta\leq 1 (21)

While Z3Exp(1)Z_{3}\sim\text{Exp}(1) and it is independent of (Z1,Z2)(Z_{1},Z_{2}). The rest of the definitions given in (4) remain the same. Then, it is easy to check that:

(η1>s,η2>t)=𝔼[exp(As1At2δAs1At2)]\mathbb{P}\left(\eta_{1}>s,\eta_{2}>t\right)=\mathbb{E}\left[\exp\left(-A^{1}_{s}-A^{2}_{t}-\delta A^{1}_{s}A^{2}_{t}\right)\right] (22)
(η3>s)=𝔼[exp(As3)]\mathbb{P}\left(\eta_{3}>s\right)=\mathbb{E}\left[\exp\left(-A^{3}_{s}\right)\right] (23)

Define τ1\tau_{1} and τ2\tau_{2} as in equation (3). Then, we can get the join distribution of (τ1,τ2)(\tau_{1},\tau_{2}):

Theorem 3 (Survival function).

Suppose αi:d[0,)\alpha^{i}:\mathbb{R}^{d}\mapsto[0,\infty) for i=1,2,3i=1,2,3 are non-random positive continuous functions, which implies that AsiA^{i}_{s} are continuous and strictly increasing. Assume further that limsAsi=\lim_{s\rightarrow\infty}A^{i}_{s}=\infty a.s. Then,

F¯(τ1,τ2)(s,t):=(τ1>s,τ2>t)=𝔼[exp[As1At2δAs1At2Ast3]].\overline{F}_{(\tau_{1},\tau_{2})}(s,t):=\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t\right)=\mathbb{E}\left[\exp\left[-A^{1}_{s}-A^{2}_{t}-\delta A^{1}_{s}A^{2}_{t}-A^{3}_{s\vee t}\right]\right].
Proof.

Let ~(s,t)():=(|(Xu)0u(st))\tilde{\mathbb{P}}_{(s,t)}(\cdot):=\mathbb{P}\left(\cdot|(X_{u})_{0\leq u\leq(s\vee t)}\right). By definition of τ1\tau_{1} and τ2\tau_{2}, we have,

~(s,t)(τ1>s,τ2>t)\displaystyle\tilde{\mathbb{P}}_{(s,t)}\left(\tau_{1}>s,\tau_{2}>t\right) =~(s,t)(min(η1,η3)>s,min(η2,η3)>t)\displaystyle=\tilde{\mathbb{P}}_{(s,t)}\left(\min(\eta_{1},\eta_{3})>s,\min(\eta_{2},\eta_{3})>t\right)
=~(s,t)(η1>s,η2>t)~(s,t)(η3>s,η3>t)\displaystyle=\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{1}>s,\eta_{2}>t\right)\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{3}>s,\eta_{3}>t\right)
=~(s,t)(η1>s,η2>t)~(s,t)(η3>max(s,t))\displaystyle=\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{1}>s,\eta_{2}>t\right)\tilde{\mathbb{P}}_{(s,t)}\left(\eta_{3}>\max(s,t)\right)
=exp[As1At2δAs1At2Ast3]\displaystyle=\exp\left[-A^{1}_{s}-A^{2}_{t}-\delta A^{1}_{s}A^{2}_{t}-A^{3}_{s\vee t}\right]
={exp[As1At2δAs1At2As3]s>texp[As1At2δAs1At2At3]st.\displaystyle=\begin{cases}\exp\left[-A^{1}_{s}-A^{2}_{t}-\delta A^{1}_{s}A^{2}_{t}-A^{3}_{s}\right]&s>t\\ \exp\left[-A^{1}_{s}-A^{2}_{t}-\delta A^{1}_{s}A^{2}_{t}-A^{3}_{t}\right]&s\leq t.\end{cases}

The result follows by taking the expectation. ∎

Remark 3.1.

From Theorem 3, it is clear that τ1\tau_{1} is not independent of τ2\tau_{2}.

Remark 3.2.

The marginal distribution of τi\tau_{i} is given by:

(τi>s)=(min(ηi,η3)>s)=𝔼[exp(0s(αri+αr3)𝑑r)]\mathbb{P}(\tau_{i}>s)=\mathbb{P}(\min(\eta_{i},\eta_{3})>s)=\mathbb{E}\left[\exp\left(-\int_{0}^{s}(\alpha^{i}_{r}+\alpha^{3}_{r})dr\right)\right] (24)

which, as in Section 2, coincides with the marginal distribution in the Cox construction, see Lando [24] [25].

Proposition 3.1 (Probability of the two stopping times being equal).
(τ1=τ2)=𝔼[0αt3exp[(At1+At2+At3)δAt1At2]𝑑t]\mathbb{P}\left(\tau_{1}=\tau_{2}\right)=\mathbb{E}\left[\int_{0}^{\infty}\alpha^{3}_{t}\exp\left[-(A^{1}_{t}+A^{2}_{t}+A^{3}_{t})-\delta A^{1}_{t}A^{2}_{t}\right]dt\right] (25)

which is greater than 0 as long as αt3\alpha^{3}_{t} is not identically equal to 0

The proof of this Proposition follows by a similar argument as the one used in the proof of Theorem 2 to get (BC)\mathbb{P}\left(B^{C}\right), so we leave it to the reader.

This is the general set-up. For the rest of this Section, we assume that for i=1,2,3i=1,2,3, we have αi(Xt)λi+\alpha^{i}(X_{t})\equiv\lambda_{i}\in\mathbb{R}^{+} for all t0t\geq 0 to get tractable computations. Under this assumption, note the following:

  1. 1.

    (η1>s,η2>t)=exp[λ1sλ2tδλ1λ2st]\mathbb{P}\left(\eta_{1}>s,\eta_{2}>t\right)=\exp\left[-\lambda_{1}s-\lambda_{2}t-\delta\lambda_{1}\lambda_{2}st\right]

  2. 2.

    (τ1>s,τ2>t)=exp[λ1sλ2tδλ1λ2stλ3(st)]\mathbb{P}\left(\tau_{1}>s,\tau_{2}>t\right)=\exp\left[-\lambda_{1}s-\lambda_{2}t-\delta\lambda_{1}\lambda_{2}st-\lambda_{3}(s\vee t)\right]. That is (τ1,τ2)(\tau_{1},\tau_{2}) follow a bivariate exponential (BVE) that is a combination of the Marshall-Olkin BVE (see Marshall and Olkin [28]) and the Gumbel BVE (see [17]). Actually, if δ=0\delta=0, we recover the BVE of Marshall and Olkin and if λ3=0\lambda_{3}=0 we recover the BVE of Gumbel.

  3. 3.

    (τ1>t)=exp((λ1+λ3)t)\mathbb{P}(\tau_{1}>t)=\exp\left(-(\lambda_{1}+\lambda_{3})t\right) and (τ2>t)=exp((λ2+λ3)t)\mathbb{P}(\tau_{2}>t)=\exp\left(-(\lambda_{2}+\lambda_{3})t\right), i.e., marginally τiExp(λi+λ3)\tau_{i}\sim\text{Exp}(\lambda_{i}+\lambda_{3}).

Before showing that one can get a negative covariance between the stopping times, we show in the next proposition that, by allowing the possibility of a negative covariance, the probability of the two stopping times being equal is smaller.

Proposition 3.2.

Fix some values of λ1,λ2,λ3\lambda_{1},\lambda_{2},\lambda_{3}. If δ0\delta\neq 0, the probability of τ1\tau_{1} being equal to τ2\tau_{2} is smaller than the probability of τ1\tau_{1} being equal to τ2\tau_{2} for the case of δ=0\delta=0

Proof.

Let δ0(A)\mathbb{P}_{\delta\neq 0}(A) and δ=0(A)\mathbb{P}_{\delta=0}(A) stand for the probability of an event AA under the law of (τ1,τ2)(\tau_{1},\tau_{2}) when δ0\delta\neq 0 and when δ=0\delta=0 respectively

Recall from Example 2.1 that:

δ=0(τ1=τ2)=λ3λ1+λ2+λ3\mathbb{P}_{\delta=0}\left(\tau_{1}=\tau_{2}\right)=\frac{\lambda_{3}}{\lambda_{1}+\lambda_{2}+\lambda_{3}} (26)

Also, setting αti=λi\alpha^{i}_{t}=\lambda_{i} in (25), we get:

δ0(τ1=τ2)\displaystyle\mathbb{P}_{\delta\neq 0}\left(\tau_{1}=\tau_{2}\right) =0λ3exp[(λ1+λ2+λ3)tδλ1λ2t2]𝑑t\displaystyle=\int_{0}^{\infty}\lambda_{3}\exp\left[-(\lambda_{1}+\lambda_{2}+\lambda_{3})t-\delta\lambda_{1}\lambda_{2}t^{2}\right]dt
=λ3exp[(λ1+λ2+λ3)24δλ1λ2]0exp[(δλ1λ2t+λ1+λ2+λ32δλ1λ2)2]𝑑t\displaystyle=\lambda_{3}\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\delta\lambda_{1}\lambda_{2}}\right]\int_{0}^{\infty}\exp\left[-\left(\sqrt{\delta\lambda_{1}\lambda_{2}}t+\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{2\sqrt{\delta\lambda_{1}\lambda_{2}}}\right)^{2}\right]dt
=λ3δλ1λ2exp[(λ1+λ2+λ3)24δλ1λ2]λ1+λ2+λ32δλ1λ2eu2𝑑u\displaystyle=\frac{\lambda_{3}}{\sqrt{\delta\lambda_{1}\lambda_{2}}}\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\delta\lambda_{1}\lambda_{2}}\right]\int_{\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{2\sqrt{\delta\lambda_{1}\lambda_{2}}}}^{\infty}e^{-u^{2}}du (27)

where the second equality follows by completing the square and it is only valid if δ0\delta\neq 0. The third equality is just a change of variable.

Then, using the well-known bound xeu2𝑑uex22x for all x>0\int_{x}^{\infty}e^{-u^{2}}du\leq\frac{e^{-x^{2}}}{2x}\text{ for all }x>0, we have that:

λ3δλ1λ2\displaystyle\frac{\lambda_{3}}{\sqrt{\delta\lambda_{1}\lambda_{2}}} exp[(λ1+λ2+λ3)24δλ1λ2]λ1+λ2+λ32δλ1λ2eu2𝑑u\displaystyle\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\delta\lambda_{1}\lambda_{2}}\right]\int_{\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{2\sqrt{\delta\lambda_{1}\lambda_{2}}}}^{\infty}e^{-u^{2}}du
λ3δλ1λ2exp[(λ1+λ2+λ3)24δλ1λ2]exp[(λ1+λ2+λ3)24δλ1λ2]2δλ1λ22(λ1+λ2+λ3)\displaystyle\leq\frac{\lambda_{3}}{\sqrt{\delta\lambda_{1}\lambda_{2}}}\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\delta\lambda_{1}\lambda_{2}}\right]\exp\left[-\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\delta\lambda_{1}\lambda_{2}}\right]\frac{2\sqrt{\delta\lambda_{1}\lambda_{2}}}{2\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)}
=λ3λ1+λ2+λ3\displaystyle=\frac{\lambda_{3}}{\lambda_{1}+\lambda_{2}+\lambda_{3}}

which shows, as desired:

δ0(τ1=τ2)δ=0(τ1=τ2)\mathbb{P}_{\delta\neq 0}\left(\tau_{1}=\tau_{2}\right)\leq\mathbb{P}_{\delta=0}\left(\tau_{1}=\tau_{2}\right) (28)

Proposition 3.3 (Negative Covariance).

Suppose λ1=c1λ3\lambda_{1}=c_{1}\lambda_{3}, λ2=c2λ3\lambda_{2}=c_{2}\lambda_{3} and δ=1\delta=1. If 5c1c24(c1+c2+1)5c_{1}c_{2}\geq 4\left(c_{1}+c_{2}+1\right), then Cov(τ1,τ2)<0\text{Cov}\left(\tau_{1},\tau_{2}\right)<0.

Proof.

Using the result of Young [35], as in the proof of Proposition 2.8 as well as Theorem 3, we have that:

𝔼(τ1τ2)=\displaystyle\mathbb{E}\left(\tau_{1}\tau_{2}\right)= 00exp[λ1xλ2yλ3(xy)λ1λ2xy]𝑑x𝑑y\displaystyle\int_{0}^{\infty}\int_{0}^{\infty}\exp\left[-\lambda_{1}x-\lambda_{2}y-\lambda_{3}(x\vee y)-\lambda_{1}\lambda_{2}xy\right]dxdy
=\displaystyle= 00yexp[λ1xλ2yλ1λ2xyλ3y]𝑑x𝑑y\displaystyle\int_{0}^{\infty}\int_{0}^{y}\exp\left[-\lambda_{1}x-\lambda_{2}y-\lambda_{1}\lambda_{2}xy-\lambda_{3}y\right]dxdy
+0yexp[λ1xλ2yλ1λ2xyλ3x]𝑑x𝑑y\displaystyle+\int_{0}^{\infty}\int_{y}^{\infty}\exp\left[-\lambda_{1}x-\lambda_{2}y-\lambda_{1}\lambda_{2}xy-\lambda_{3}x\right]dxdy
=\displaystyle= 0xexp[λ1xλ2yλ1λ2xyλ3y]𝑑y𝑑x\displaystyle\int_{0}^{\infty}\int_{x}^{\infty}\exp\left[-\lambda_{1}x-\lambda_{2}y-\lambda_{1}\lambda_{2}xy-\lambda_{3}y\right]dydx
+0yexp[λ1xλ2yλ1λ2xyλ3x]𝑑x𝑑y\displaystyle+\int_{0}^{\infty}\int_{y}^{\infty}\exp\left[-\lambda_{1}x-\lambda_{2}y-\lambda_{1}\lambda_{2}xy-\lambda_{3}x\right]dxdy
=\displaystyle= 01λ2+λ3+λ1λ2xexp[(λ1+λ2+λ3)xλ1λ2x2)]dx\displaystyle\int_{0}^{\infty}\frac{1}{\lambda_{2}+\lambda_{3}+\lambda_{1}\lambda_{2}x}\exp\left[-(\lambda_{1}+\lambda_{2}+\lambda_{3})x-\lambda_{1}\lambda_{2}x^{2}\right)]dx
+01λ1+λ3+λ1λ2yexp[(λ1+λ2+λ3)yλ1λ2y2)]dy\displaystyle+\int_{0}^{\infty}\frac{1}{\lambda_{1}+\lambda_{3}+\lambda_{1}\lambda_{2}y}\exp\left[-(\lambda_{1}+\lambda_{2}+\lambda_{3})y-\lambda_{1}\lambda_{2}y^{2}\right)]dy (29)

We reduce the previous equation even further by completing the square, for instance, take the first summand:

01λ2+λ3+λ1λ2xexp[(λ1+λ2+λ3)xλ1λ2x2)]dx=exp[(λ1+λ2+λ3)24λ1λ2]01λ2+λ3+λ1λ2xexp[(λ1λ2x+12(λ1+λ2+λ3λ1λ2))2]𝑑x=2λ1λ2exp[(λ1+λ2+λ3)24λ1λ2]12(λ1+λ2+λ3λ1λ2)12λ1λ2u+λ2+λ3λ1eu2𝑑u\int_{0}^{\infty}\frac{1}{\lambda_{2}+\lambda_{3}+\lambda_{1}\lambda_{2}x}\exp\left[-(\lambda_{1}+\lambda_{2}+\lambda_{3})x-\lambda_{1}\lambda_{2}x^{2}\right)]dx\\ =\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\lambda_{1}\lambda_{2}}\right]\int_{0}^{\infty}\frac{1}{\lambda_{2}+\lambda_{3}+\lambda_{1}\lambda_{2}x}\exp\left[-\left(\sqrt{\lambda_{1}\lambda_{2}}x+\frac{1}{2}\left(\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{\sqrt{\lambda_{1}\lambda_{2}}}\right)\right)^{2}\right]dx\\ =\frac{2}{\sqrt{\lambda_{1}\lambda_{2}}}\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\lambda_{1}\lambda_{2}}\right]\int\limits_{\frac{1}{2}\left(\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{\sqrt{\lambda_{1}\lambda_{2}}}\right)}^{\infty}\frac{1}{2\sqrt{\lambda_{1}\lambda_{2}}u+\lambda_{2}+\lambda_{3}-\lambda_{1}}e^{-u^{2}}du (30)

Similarly, the second summand in (3.2) reduces to:

01λ1+λ3+λ1λ2yexp[(λ1+λ2+λ3)yλ1λ2y2)]dy=2λ1λ2exp[(λ1+λ2+λ3)24λ1λ2]12(λ1+λ2+λ3λ1λ2)12λ1λ2u+λ1+λ3λ2eu2𝑑u\int_{0}^{\infty}\frac{1}{\lambda_{1}+\lambda_{3}+\lambda_{1}\lambda_{2}y}\exp\left[-(\lambda_{1}+\lambda_{2}+\lambda_{3})y-\lambda_{1}\lambda_{2}y^{2}\right)]dy\\ =\frac{2}{\sqrt{\lambda_{1}\lambda_{2}}}\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\lambda_{1}\lambda_{2}}\right]\int\limits_{\frac{1}{2}\left(\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{\sqrt{\lambda_{1}\lambda_{2}}}\right)}^{\infty}\frac{1}{2\sqrt{\lambda_{1}\lambda_{2}}u+\lambda_{1}+\lambda_{3}-\lambda_{2}}e^{-u^{2}}du (31)

Using the fact that 𝔼(τi)=1λi+λ3\mathbb{E}\left(\tau_{i}\right)=\frac{1}{\lambda_{i}+\lambda_{3}} and equations (30) and (31), we get that:

Cov(τ1,τ2)=2λ1λ2exp[(λ1+λ2+λ3)24λ1λ2]12(λ1+λ2+λ3λ1λ2)[12λ1λ2u+λ2+λ3λ1+12λ1λ2u+λ1+λ3λ2]eu2du(1λ1+λ3)(1λ2+λ3)\text{Cov}\left(\tau_{1},\tau_{2}\right)=\frac{2}{\sqrt{\lambda_{1}\lambda_{2}}}\exp\left[\frac{\left(\lambda_{1}+\lambda_{2}+\lambda_{3}\right)^{2}}{4\lambda_{1}\lambda_{2}}\right]\int\limits_{\frac{1}{2}\left(\frac{\lambda_{1}+\lambda_{2}+\lambda_{3}}{\sqrt{\lambda_{1}\lambda_{2}}}\right)}^{\infty}\left[\frac{1}{2\sqrt{\lambda_{1}\lambda_{2}}u+\lambda_{2}+\lambda_{3}-\lambda_{1}}\right.\\ \left.+\frac{1}{2\sqrt{\lambda_{1}\lambda_{2}}u+\lambda_{1}+\lambda_{3}-\lambda_{2}}\right]e^{-u^{2}}du-\left(\frac{1}{\lambda_{1}+\lambda_{3}}\right)\left(\frac{1}{\lambda_{2}+\lambda_{3}}\right) (32)

Using the assumption λ1=c1λ3\lambda_{1}=c_{1}\lambda_{3} and λ2=c2λ3\lambda_{2}=c_{2}\lambda_{3}, we rewrite the previous expression to get:

Cov(τ1,τ2)=2λ32c1c2exp[(c1+c2+1)24c1c2]c1+c2+12c1c2[12c1c2u+c2c1+1+12c1c2u+c1c2+1]eu2du1λ32(1c1+1)(1c2+1)\text{Cov}\left(\tau_{1},\tau_{2}\right)=\frac{2}{\lambda_{3}^{2}\sqrt{c_{1}c_{2}}}\exp\left[\frac{\left(c_{1}+c_{2}+1\right)^{2}}{4c_{1}c_{2}}\right]\int\limits_{\frac{c_{1}+c_{2}+1}{2\sqrt{c_{1}c_{2}}}}^{\infty}\left[\frac{1}{2\sqrt{c_{1}c_{2}}u+c_{2}-c_{1}+1}\right.\\ \left.+\frac{1}{2\sqrt{c_{1}c_{2}}u+c_{1}-c_{2}+1}\right]e^{-u^{2}}du-\frac{1}{\lambda_{3}^{2}}\left(\frac{1}{c_{1}+1}\right)\left(\frac{1}{c_{2}+1}\right) (33)

As uc1+c2+12c1c2u\geq\frac{c_{1}+c_{2}+1}{2\sqrt{c_{1}c_{2}}}, we have that 12c1c2u+c2c1+112(c2+1)\frac{1}{2\sqrt{c_{1}c_{2}}u+c_{2}-c_{1}+1}\leq\frac{1}{2\left(c_{2}+1\right)} and 12c1c2u+c1c2+112(c1+1)\frac{1}{2\sqrt{c_{1}c_{2}}u+c_{1}-c_{2}+1}\leq\frac{1}{2\left(c_{1}+1\right)}. Then:

Cov(τ1,τ2)1λ32c1c2(1c2+1+1c1+1)exp[(c1+c2+1)24c1c2]c1+c2+12c1c2eu2𝑑u1λ32(1c1+1)(1c2+1)=1λ32(c1+1)(c2+1)[1c1c2(c1+c2+2)exp[(c1+c2+1)24c1c2]c1+c2+12c1c2eu2𝑑u1]\text{Cov}\left(\tau_{1},\tau_{2}\right)\leq\frac{1}{\lambda_{3}^{2}\sqrt{c_{1}c_{2}}}\left(\frac{1}{c_{2}+1}+\frac{1}{c_{1}+1}\right)\exp\left[\frac{\left(c_{1}+c_{2}+1\right)^{2}}{4c_{1}c_{2}}\right]\int\limits_{\frac{c_{1}+c_{2}+1}{2\sqrt{c_{1}c_{2}}}}^{\infty}e^{-u^{2}}du\\ -\frac{1}{\lambda_{3}^{2}}\left(\frac{1}{c_{1}+1}\right)\left(\frac{1}{c_{2}+1}\right)\\ =\frac{1}{\lambda_{3}^{2}\left(c_{1}+1\right)\left(c_{2}+1\right)}\left[\frac{1}{\sqrt{c_{1}c_{2}}}\left(c_{1}+c_{2}+2\right)\exp\left[\frac{\left(c_{1}+c_{2}+1\right)^{2}}{4c_{1}c_{2}}\right]\int\limits_{\frac{c_{1}+c_{2}+1}{2\sqrt{c_{1}c_{2}}}}^{\infty}e^{-u^{2}}du-1\right] (34)

Hence, to get a negative covariance, it suffices to show:

1c1c2(c1+c2+2)exp[(c1+c2+1)24c1c2]c1+c2+12c1c2eu2𝑑u<1\frac{1}{\sqrt{c_{1}c_{2}}}\left(c_{1}+c_{2}+2\right)\exp\left[\frac{\left(c_{1}+c_{2}+1\right)^{2}}{4c_{1}c_{2}}\right]\int\limits_{\frac{c_{1}+c_{2}+1}{2\sqrt{c_{1}c_{2}}}}^{\infty}e^{-u^{2}}du<1 (35)

Set k:=c1+c2+1c1c2k:=\frac{c_{1}+c_{2}+1}{\sqrt{c_{1}c_{2}}} and note that our initial assumption implies that 1c1c254(c1c2c1+c2+1)=54k\frac{1}{\sqrt{c_{1}c_{2}}}\leq\frac{5}{4}\left(\frac{\sqrt{c_{1}c_{2}}}{c_{1}+c_{2}+1}\right)=\frac{5}{4k}. Hence:

1c1c2(c1+c2+2)exp[(c1+c2+1)24c1c2]c1+c2+12c1c2eu2𝑑u(k+54k)e14k2k2eu2𝑑u\frac{1}{\sqrt{c_{1}c_{2}}}\left(c_{1}+c_{2}+2\right)\exp\left[\frac{\left(c_{1}+c_{2}+1\right)^{2}}{4c_{1}c_{2}}\right]\int\limits_{\frac{c_{1}+c_{2}+1}{2\sqrt{c_{1}c_{2}}}}^{\infty}e^{-u^{2}}du\leq\left(k+\frac{5}{4k}\right)e^{\frac{1}{4}k^{2}}\int\limits_{\frac{k}{2}}^{\infty}e^{-u^{2}}du (36)

By the proof in the Appendix, one can see that the previous integral is always smaller than one when k>2k>2, which is the case here because:

(c1c2)2+1>0c12c1c2+c2+1>0c1+c2+1>2c1c2\left(\sqrt{c_{1}}-\sqrt{c_{2}}\right)^{2}+1>0\implies c_{1}-2\sqrt{c_{1}c_{2}}+c_{2}+1>0\implies c_{1}+c_{2}+1>2\sqrt{c_{1}c_{2}}

Hence:

k=c1+c2+1c1c2>2c1c2c1c2>2k=\frac{c_{1}+c_{2}+1}{\sqrt{c_{1}c_{2}}}>\frac{2\sqrt{c_{1}c_{2}}}{\sqrt{c_{1}c_{2}}}>2 (37)

This shows that equation (35) holds and consequently, we have a negative covariance. ∎

Remark 3.3.

If c:=c1=c2c:=c_{1}=c_{2}, the condition of Proposition 3.3 reduces to 5c28c405c^{2}-8c-4\geq 0 which is equivalent to c2c\geq 2. The interpretation of this is that to get a negative covariance when λ1=λ2\lambda_{1}=\lambda_{2}, it suffices to have λ1,λ2\lambda_{1},\lambda_{2} sufficiently large compared to λ3\lambda_{3}.

4 APPLICATIONS

4.1 Application to Epidemiology

Suppose we are interested in knowing what is the probability of nn people getting infected with COVID-19 at the exact same time. The time to infection of each person can be modeled as a stopping time. Some current models (see Britton and Pardoux [6]) assume independence of these stopping times and thus the probability of them being equal is 0. However, using our model we can weaken the independence assumption and conclude that:

Theorem 4.

If (τ1,τ2,,τn)\left(\tau_{1},\tau_{2},\dots,\tau_{n}\right) follows the joint distribution described in Section 3.1, then:

(τ1=τ2==τn)=\displaystyle\mathbb{P}\left(\tau_{1}=\tau_{2}=\dots=\tau_{n}\right)= 𝔼[0α(1,2,,n)(u)ek=1nj𝒞knAuj𝑑u]\displaystyle\mathbb{E}\left[\int_{0}^{\infty}\alpha^{(1,2,\dots,n)}(u)e^{-\sum\limits_{k=1}^{n}\sum\limits_{j\in\mathcal{C}^{n}_{k}}A^{j}_{u}}du\right] (38)

The innermost sum in the exponent is taken over all the possible combinations 𝒞kn\mathcal{C}^{n}_{k} of (nk)\binom{n}{k}. For example, if k=3k=3, jj could be (1,3,5)(1,3,5), (2,3,n)(2,3,n), etc; if k=nk=n, jj can only be (1,2,,n)(1,2,\dots,n)

Remark 4.1.

To be more specific if n=3n=3, then we get,

(τ1=τ2=τ3)=\displaystyle\mathbb{P}\left(\tau_{1}=\tau_{2}=\tau_{3}\right)= 𝔼[0α(1,2,3)(u)exp[Au1Au2Au3\displaystyle\mathbb{E}\Biggl{[}\int_{0}^{\infty}\alpha^{(1,2,3)}(u)\exp\left[-A^{1}_{u}-A^{2}_{u}-A^{3}_{u}\right.
Au(1,2)Au(1,3)Au(2,3)Au(1,2,3)]]du\displaystyle\left.-A^{(1,2)}_{u}-A^{(1,3)}_{u}-A^{(2,3)}_{u}-A^{(1,2,3)}_{u}\right]\Biggr{]}du (39)
Proof.

Using the notation from Section 3, events in the process Λ(1,2,,n)\Lambda_{(1,2,\dots,n)} are shocks to the nn components of the system. Hence

{τ1=τ2==τn1=τn}={Λ(1,2,n)min(Λ1,,Λn,Λ(1,2),,Λ(n1,n),Λ(1,2,3),,Λ(1,2,,n1))}\left\{\tau_{1}=\tau_{2}=\dots=\tau_{n-1}=\tau_{n}\right\}=\\ \left\{\Lambda_{(1,2,\dots\ n)}\leq\min\left(\Lambda_{1},\dots,\Lambda_{n},\Lambda_{(1,2)},\dots,\Lambda_{(n-1,n)},\Lambda_{(1,2,3)},\dots,\Lambda_{(1,2,\dots,n-1)}\right)\right\}

Then, we find the distribution of:

M:=min(Λ1,,Λn,Λ(1,2),,Λ(n1,n),Λ(1,2,3),,Λ(1,2,,n1))M:=\min\left(\Lambda_{1},\dots,\Lambda_{n},\Lambda_{(1,2)},\dots,\Lambda_{(n-1,n)},\Lambda_{(1,2,3)},\dots,\Lambda_{(1,2,\dots,n-1)}\right)

under the measure ~()\tilde{\mathbb{P}}(\cdot) which stands for (|(Xs)s0)\mathbb{P}\left(\cdot|(X_{s})_{s\geq 0}\right)

~[M>t]=~[Λ1>t,,Λn>t,,Λ(n,n)>t,Λ(1,2,3)>t,,Λ(1,2,,n1)>t]=exp[At1AtnAt(1,2)At(n1,n)At(1,2,3)At(1,2,,n1)]\tilde{\mathbb{P}}\left[M>t\right]=\tilde{\mathbb{P}}\left[\Lambda_{1}>t,\dots,\Lambda_{n}>t,\dots,\Lambda_{(n,n)}>t,\Lambda_{(1,2,3)}>t,\dots,\Lambda_{(1,2,\dots,n-1)}>t\right]\\ =\exp\left[-A^{1}_{t}-\dots-A^{n}_{t}-A^{(1,2)}_{t}-\dots-A^{(n-1,n)}_{t}-A^{(1,2,3)}_{t}-\dots-A^{(1,2,\dots,n-1)}_{t}\right] (40)

Hence, MM, given (Xs)s0\left(X_{s}\right)_{s\geq 0}, has a continuous distribution with density equal to:

fM(t)=[k=1n1j𝒞knαtj]ek=1n1j𝒞knαtjAtjf_{M}(t)=\left[\sum\limits_{k=1}^{n-1}\sum\limits_{j\in\mathcal{C}^{n}_{k}}\alpha^{j}_{t}\right]e^{-\sum\limits_{k=1}^{n-1}\sum\limits_{j\in\mathcal{C}^{n}_{k}}\alpha^{j}_{t}A^{j}_{t}} (41)

The innermost sums are taken over all the possible combinations 𝒞kn\mathcal{C}^{n}_{k} of (nk)\binom{n}{k}. Then,

~[Λ1,2,,n<M]\displaystyle\tilde{\mathbb{P}}\left[\Lambda_{1,2,\dots,n}<M\right] =0xαx(1,2,,n)eAx(1,2,,n)fM(y)𝑑y𝑑x\displaystyle=\int_{0}^{\infty}\int_{x}^{\infty}\alpha^{(1,2,\dots,n)}_{x}e^{-A^{(1,2,\dots,n)}_{x}}f_{M}(y)\,dy\,dx
=0αu(1,2,,n)ek=1nj𝒞knAuj𝑑u\displaystyle=\int_{0}^{\infty}\alpha^{(1,2,\dots,n)}_{u}e^{-\sum\limits_{k=1}^{n}\sum\limits_{j\in\mathcal{C}^{n}_{k}}A^{j}_{u}}\,du (42)

The result follows by taking the expectation. ∎

Corollary 4.1.

If αi(Xt)=α(Xt)\alpha^{i}(X_{t})=\alpha(X_{t}) for all i=1,2,,ni=1,2,\dots,n; α(i,j)(Xt)=12α(Xt)\alpha^{(i,j)}(X_{t})=\frac{1}{2}\alpha(X_{t}) for all size 2 combinations in (n2)\binom{n}{2}; α(i,j,k)(Xt)=13α(Xt)\alpha^{(i,j,k)}(X_{t})=\frac{1}{3}\alpha(X_{t}) for all size 3 combinations in (n3)\binom{n}{3}; \dots; α(1,2,,n)(Xt)=1nα(Xt)\alpha^{(1,2,\dots,n)}(X_{t})=\frac{1}{n}\alpha(X_{t}). Then, we have:

(τ1=τ2==τn)=1n[1(n1)+12(n2)++1n(nn)]\displaystyle\mathbb{P}\left(\tau_{1}=\tau_{2}=\dots=\tau_{n}\right)=\frac{1}{n}\left[\frac{1}{\binom{n}{1}+\frac{1}{2}\binom{n}{2}+\dots+\frac{1}{n}\binom{n}{n}}\right] (43)
Corollary 4.2.

If αi(Xt)=α(Xt)\alpha^{i}(X_{t})=\alpha(X_{t}) for all i=1,2,,ni=1,2,\dots,n; α(i,j)(Xt)=2α(Xt)\alpha^{(i,j)}(X_{t})=2\alpha(X_{t}) for all size 2 combinations in (n2)\binom{n}{2}; α(i,j,k)(Xt)=3α(Xt)\alpha^{(i,j,k)}(X_{t})=3\alpha(X_{t}) for all size 3 combinations in (n3)\binom{n}{3}; \dots; α(1,2,,n)(Xt)=nα(Xt)\alpha^{(1,2,\dots,n)}(X_{t})=n\alpha(X_{t}). Then, we have:

(τ1=τ2==τn)\displaystyle\mathbb{P}\left(\tau_{1}=\tau_{2}=\dots=\tau_{n}\right) =n[1(n1)+2(n2)++n(nn)]\displaystyle=n\left[\frac{1}{\binom{n}{1}+2\binom{n}{2}+\dots+n\binom{n}{n}}\right]
=12n1\displaystyle=\frac{1}{2^{n-1}} (44)

4.2 Application to Engineering

A classic problem in Operations Research, basically studied in queuing theory, is that of a complicated machine. The machine fails if one of its key parts fails. Knowing this, designers create a certain redundancy by doubling key components, so that if one fails, there is a back-up ready to assume its duties. To save money, however, one can have one back-up for two components. Suppose η1\eta_{1} is the (random) failure time of one component, and η2\eta_{2} is the (again, random) failure time of the second component. If they fail at the same time, then the machine itself will fail, since the solitary back-up cannot replace both components simultaneously. One usually considers such a situation to be unlikely, even very unlikely. If it were to happen, however, we would be interested in P(η1=η2)P(\eta_{1}=\eta_{2}). In the conventional models, P(η1=η2)P(\eta_{1}=\eta_{2}) is zero, since they are each exponential, and conditionally independent. However if we consider a third time, η3\eta_{3}, and if this third stopping time is the (once again, random) time of an external shock (such as the failure of the air conditioning unit, or a power failure with a surge when the power resumes, etc.), and if we let

τ1=η1η3 and τ2=η2η3\tau_{1}=\eta_{1}\wedge\eta_{3}\text{ and }\tau_{2}=\eta_{2}\wedge\eta_{3} (45)

then we are in the case where P(τ1=τ2)=α>0P(\tau_{1}=\tau_{2})=\alpha>0, and in special cases we can calculate the probability α\alpha with precision, and in other, more complicated situations, we can give upper and lower bounds, both, for α\alpha.

A different kind of example, exemplified by the recent, and quite dramatic example is that of the collapse of the Champlain Towers South, in Surfside, Florida (just north of Miami Beach). The twelve story towers fell at night (1:30AM), and killed 98 people who were in their apartments at the time, and presumably even in their beds. The towers had a pool deck above a covered garage. The towers holding up the pool deck were too thin, and not strong enough to withstand the stresses imposed on them over four decades. Water was seen pouring into the parking garage only minutes before the collapse.

In the main building that collapsed, structural columns were too narrow to accommodate enough rebar, meaning that contractors had to choose between cramming extra steel into a too-small column (which can create air pockets that accelerate corrosion) or inadequately attaching floor slabs to their supports. Our model would have several stopping times, each one representing the failure of a different component of the structure. Choosing two important ones, such as: (1) The corrosion of the rebar supports within the concrete, due to the salt air and massive strains due to violent weather, which plagues the Florida coast during hurricane season; (2) The use of a low quality grade of concrete, violating regulations of the local government in the construction of the towers, leading to concrete integrity decay due to 40 years of seaside weather.

One way to model this is to take a vector of two Cox constructions, using two independent exponentials Z1Z_{1} and Z2Z_{2} to construct our failure times τ1\tau_{1} and τ2\tau_{2}. This then gives us that τ1\tau_{1} and τ2\tau_{2} are conditionally independent, given the underlying filtration 𝔽=(t)t0\mathbb{F}=(\mathcal{F}_{t})_{t\geq 0}. This leads to P(τ1=τ2)=0P(\tau_{1}=\tau_{2})=0, even if they have the exact same compensators.

Other models have been proposed, such as the general framework presented in the book of Anna Aksamit and Monique Jeanblanc [1], where the random variables Z1Z_{1} and Z2Z_{2} are multivariate exponentials, but with a joint density describing how they relate to each other. However, even in this more general setting, we have P(τ1=τ2)=0P(\tau_{1}=\tau_{2})=0.

Assuming it is the stopping times occurring simultaneously that causes the collapse, we want a model that allows us to have P(τ1=τ2)>0P(\tau_{1}=\tau_{2})>0. Let’s assume we have three standard Cox Constructions, with independent exponentials (Z1;Z2;Z3)(Z_{1};Z_{2};Z_{3}), with different compensators (0tαsi𝑑s\int_{0}^{t}\alpha^{i}_{s}ds for i=1,2,3i=1,2,3).

Call the three stopping times η1,η2,η3\eta_{1},\eta_{2},\eta_{3}. The time η3\eta_{3} could be anything, such as a hurricane putting heavy stress on the building, or an earthquake, or some other external factor. (Current forensic analysis, which is ongoing as we write this, suggests that the collapse of the pool-deck created a seismic shock sufficient to precipitate the collapse of the south tower, and weakened the structural integrity of the north tower to such an extent that it was condemned and then deliberately destroyed.) However in this case we can take η3\eta_{3} to be the time of the flooding of the parking structure under the swimming pool and pool deck in general.

Simulations commissioned by the newspaper The Washington Post and done by a team led by Khalid M. Mosalam of the University of California, Berkeley, show how that might have happened and indicate that it is a plausible scenario (Swaine et al. [32]).

As in (45), we define

τ1=η1η3 and τ2=η2η3,\tau_{1}=\eta_{1}\wedge\eta_{3}\text{ and }\tau_{2}=\eta_{2}\wedge\eta_{3}, (46)

where we take the stopping time η3\eta_{3} to be the time of the collapse of the pool deck. As the noted engineer H. Petroski has pointed out [29] it is often the case that multiple things happen at once in order to precipitate a disaster such as the fall of the Champlain Tower South. Indeed, the forensic engineer R. Leon of Virginia Tech is quoted in American Society of Civil Engineers, 2021 [34]: “I think it is way too early to tell,”said Roberto Leon, P.E., F.SEI, Dist.M.ASCE, the D.H. Burrows Professor of Construction Engineering in the Charles Edward Via Jr. Department of Civil and Environmental Engineering at Virginia Tech.“It’s going to require a very careful forensic approach here, because I don’t think the building collapsed just because of one reason. What we tend to find in forensic investigations is that three or four things have to happen for a collapse to occur that is so catastrophic.”

Professor Leon is a widely respected authority in forensic civil engineering, and the key insight for us is his last statement that three or four things have to happen simultaneously for a catastrophic collapse.

Less dramatic examples, but quite pertinent to the recent attention being paid to the decay of infrastructure around the US, provide more examples of the utility of this approach. A first example is the Interstate 10 Twin Span Bridge over Lake Pontchartrain north of New Orleans, LA. It was rendered completely unusable by Hurricane Katrina, but the naive explanation was demonstrated false by the fact that several other bridges that had the same structural design remained intact. Upon investigation, it was determined that air trapped beneath the deck of the Interstate 10 bridges was a major contributing factor to the bridge’s collapse. While major, it was not the only contributing factor (Chen et al. [8]).

A final example is the derailment of an Amtrak train near Joplin, Montana in September, 2021. 154 people were on board the train, and 44 passengers and crew were taken to area hospitals with injuries. The train was traveling at between 75 and 78 mph, just below the speed limit of 79 mph on that section of track when its emergency brakes were activated. The two locomotives and two railcars remained on the rails and eight cars derailed. Investigations of these types of events take years, but preliminary speculation is that the accident could have been caused by problems with the railroad or track, such as a rail that buckled under high heat, or the track itself giving way when the train passed over. Both might also be possible, leading to the two stopping times τ1\tau_{1} and τ2\tau_{2}, and a situation where P(τ1=τ2)>0P(\tau_{1}=\tau_{2})>0. See Hanson and Brown [19].

5 APPENDIX

Showing that equation (36) is smaller than 1 for any k>2k>2 is equivalent to showing the following inequality:

xeu2𝑑u8x16x2+5ex2 for all x1\int_{x}^{\infty}e^{-u^{2}}du\leq\frac{8x}{16x^{2}+5}e^{-x^{2}}\text{ for all }x\geq 1 (47)

This is equivalent to the problem of finding an upper bound for the complementary error (erfc) function (recall that ercf(x)=2πxeu2𝑑u\text{ercf}(x)=\frac{2}{\pi}\int_{x}^{\infty}e^{-u^{2}}du), which has been widely studied in the literature. A classical result is the Chernoff-Rubin bound (see Chernoff [9]). More recent work can be found in Chiani et al [10], Karagiannidis and Lioumpas [23], and in Tanash and Riihonen [33]. However, to our knowledge, there is no bound of the type we need here.

We will show a slightly tighter bound than the one in (36), namely, we would like to optimize the following bound in \ell

xeu2𝑑u2x4x2+ex2 for all x1\int_{x}^{\infty}e^{-u^{2}}du\leq\frac{2x}{4x^{2}+\ell}e^{-x^{2}}\text{ for all }x\geq 1 (48)

When we say optimize, we mean that 2x4x2+ex2\frac{2x}{4x^{2}+\ell}e^{-x^{2}} is as close as possible (yet bigger) to xeu2𝑑u\int_{x}^{\infty}e^{-u^{2}}du for all x1x\geq 1.

In other words, define

h(x,):=2x4x2+ex2xeu2𝑑uh(x,\ell):=\frac{2x}{4x^{2}+\ell}e^{-x^{2}}-\int_{x}^{\infty}e^{-u^{2}}du (49)

We want to find \ell^{*} such that:

  1. 1.

    h(x,)h(x,)h(x,\ell^{*})\leq h(x,\ell) for all x1x\geq 1 and \ell\neq\ell^{*}.

  2. 2.

    h(x,)0h(x,\ell^{*})\geq 0 for all x1x\geq 1

To get the first item in the previous list, we need \ell to be as large as possible. This is because h(x,)h(x,\ell) is decreasing in \ell. However, if we wish to have h(x,)0h(x,\ell)\geq 0, \ell cannot be indiscriminately large.

To find \ell^{*}, let us first look at xh(x,)\frac{\partial}{\partial x}h(x,\ell).

xh(x,)=ex2[2(4x2)(4x2+)24x24x2+]+1\frac{\partial}{\partial x}h(x,\ell)=e^{-x^{2}}\left[-\frac{2\left(4x^{2}-\ell\right)}{\left(4x^{2}+\ell\right)^{2}}-\frac{4x^{2}}{4x^{2}+\ell}\right]+1 (50)

We can find that:

xh(x,)<0 when x>2+284\frac{\partial}{\partial x}h(x,\ell)<0\text{ when }x>\sqrt{\frac{\ell^{2}+2\ell}{8-4\ell}} (51)

Hence h(x,)h(x,\ell) is increasing when x[1,2+284)x\in\left[1,\sqrt{\frac{\ell^{2}+2\ell}{8-4\ell}}\right), h(x,)h(x,\ell) is maximized when x=2+284x=\sqrt{\frac{\ell^{2}+2\ell}{8-4\ell}} and it is decreasing when x(2+284,)x\in\left(\sqrt{\frac{\ell^{2}+2\ell}{8-4\ell}},\infty\right).

Also note that

limxh(x,)=0 for any  positive\lim_{x\rightarrow\infty}h(x,\ell)=0\text{ for any }\ell\text{ positive} (52)

Hence, to satisfy the second item in the list above (i.e., h(x,)0h(x,\ell^{*})\geq 0 for all x1x\geq 1), it suffices to pick \ell^{*} such that h(1,)=0h(1,\ell^{*})=0, i.e.,

h(1,)=24+e11eu2𝑑u=0h(1,\ell^{*})=\frac{2}{4+\ell^{*}}e^{-1}-\int_{1}^{\infty}e^{-u^{2}}du=0 (53)

Solving for \ell^{*}, we find that:

=2e1eu2𝑑u41.27935\ell^{*}=\frac{2}{e\int_{1}^{\infty}e^{-u^{2}}du}-4\approx 1.27935 (54)

We can then conclude that the maximum of h(x,)h(x,\ell^{*}) occurs when

x=()2+2841.2043Recall that 1.27935x=\sqrt{\frac{(\ell^{*})^{2}+2\ell^{*}}{8-4\ell^{*}}}\approx 1.2043\quad\text{Recall that }\ell^{*}\approx 1.27935 (55)

We can then numerically calculate that

h(1.2043,)0.00131266h(1.2043,\ell^{*})\approx 0.00131266 (56)

This last equation means that the maximum difference between 2x4x2+ex2\frac{2x}{4x^{2}+\ell}e^{-x^{2}} and xeu2𝑑u\int_{x}^{\infty}e^{-u^{2}}du is 0.001312660.00131266

As h(x,)h(x,\ell) is decreasing as \ell increases, h(x,)0h(x,\ell)\geq 0 for all x1x\geq 1 as long as <1.27935\ell<\ell^{*}\approx 1.27935. In (47), we set =54<\ell=\frac{5}{4}<\ell^{*} and hence the work above shows that equation (36) is indeed smaller than 1.

References

  • Aksamit and Jeanblanc [2017] A. Aksamit and M. Jeanblanc. Enlargement of Filtration with Finance in View. SpringerBriefs in Quantitative Finance. Springer, Cham, 1 edition, 2017. doi: https://doi.org/10.1007/978-3-319-41255-9.
  • Aven [1985] T. Aven. A Theorem for Determining the Compensator of a Counting Process. Scandinavian Journal of Statistics, 12(1):69–72, 1985. ISSN 03036898, 14679469. URL http://www.jstor.org/stable/4615974.
  • Bielecki et al. [2013] T. Bielecki, A. Cousin, S. Crépey, and A. Herbertsson. In search of a grand unifying theory. Creditflux Newsletter Analysis, pages 20–21, July 2013.
  • Bielecki et al. [2018] T. R. Bielecki, M. Jeanblanc, and A. D. Sezer. Joint densities of hitting times for finite state Markov processes. Turkish Journal of Mathematics, 42(2):586–608, 2018. doi: https://doi.org/10.3906/mat-1608-29. URL https://journals.tubitak.gov.tr/math/vol42/iss2/14.
  • Brigo et al. [2016] D. Brigo, J.-F. Mai, and M. Scherer. Markov multi-variate survival indicators for default simulation as a new characterization of the Marshall–Olkin law. Statistics & Probability Letters, 114:60–66, 2016. ISSN 0167-7152. doi: https://doi.org/10.1016/j.spl.2016.03.013. URL https://www.sciencedirect.com/science/article/pii/S016771521530167X.
  • Britton and Pardoux [2019] T. Britton and E. Pardoux. Chapter 1 Stochastic Epidemic Models, pages 5–19. Springer International Publishing, Cham, 2019. ISBN 978-3-030-30900-8.
  • Brown et al. [1981] T. C. Brown, B. W. Silverman, and R. K. Milne. A class of two-type point processes. Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete, 58:299–308, 1981. URL https://doi.org/10.1007/BF00542637.
  • Chen et al. [2007] G. Chen, E. C. W. III, D. Hoffman, R. Luna, and A. Sevi. Analysis of the Interstate 10 Twin Bridges Collapse During Hurricane Katrina. Science and the storms-the USGS response to the hurricanes of 2005, pages 35–42, 2007. doi: 10.3133/cir13063D.
  • Chernoff [1952] H. Chernoff. A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations. The Annals of Mathematical Statistics, 23(4):493 – 507, 1952. doi: 10.1214/aoms/1177729330. URL https://doi.org/10.1214/aoms/1177729330.
  • Chiani et al. [2003] M. Chiani, D. Dardari, and M. K. Simon. New exponential bounds and approximations for the computation of error probability in fading channels. IEEE Transactions on Wireless Communications, 2(4):840–845, 2003. doi: 10.1109/TWC.2003.814350.
  • Cox and Lewis [1972] D. R. Cox and P. A. W. Lewis. Multivariate Point Processes. Berkeley Symposium on Mathematical Statistics and Probability, pages 401–448, 1972. doi: doi:10.1525/9780520375918-024. URL https://doi.org/10.1525/9780520375918-024.
  • Crépey et al. [2014] S. Crépey, T. R. Bielecki, and D. Brigo. Counterparty Risk and Funding: A Tale of Two Puzzles. Chapman & Hall, 2014.
  • Diggle and Milne [1983] P. J. Diggle and R. K. Milne. Bivariate Cox Processes: Some Models for Bivariate Spatial Point Patterns. Journal of the Royal Statistical Society. Series B (Methodological), 45(1):11–21, 1983. ISSN 00359246. URL http://www.jstor.org/stable/2345617.
  • Ethier and Kurtz [1986] S. N. Ethier and T. G. Kurtz. Markov Processes: Characterization and Convergence. Wiley, 1986. ISBN 9780471081869.
  • Giesecke [2003] K. Giesecke. A Simple Exponential Model for Dependent Defaults. The Journal of Fixed Income, 13(3):74–83, 2003. ISSN 1059-8596. doi: 10.3905/jfi.2003.319362. URL https://jfi.pm-research.com/content/13/3/74.
  • Gueye and Jeanblanc [2021] D. Gueye and M. Jeanblanc. Generalized Cox Model for Default Times. Working paper, June 2021. URL https://hal.archives-ouvertes.fr/hal-03264864.
  • Gumbel [1960] E. J. Gumbel. Bivariate Exponential Distributions. Journal of the American Statistical Association, 55(292):698–707, 1960. ISSN 01621459. URL http://www.jstor.org/stable/2281591.
  • Guo and Zeng [2008] X. Guo and Y. Zeng. Intensity Process and Compensator: A New Filtration Expansion Approach and the Jeulin-Yor Theorem. The Annals of Applied Probability, 18(1):120–142, 2008. ISSN 10505164. URL http://www.jstor.org/stable/25442749.
  • Hanson and Brown [2021] A. B. Hanson and M. Brown. Cause of Montana Amtrak derailment that killed 3, injured dozens still under investigation. Great Falls Tribune, October 27, 2021 2021.
  • Janson et al. [2011] S. Janson, S. M’Baye, and P. Protter. Absolutely Continuous Compensators. International Journal of Theoretical and Applied Finance, 14(03):335–351, 2011. doi: 10.1142/S0219024911006565. URL https://doi.org/10.1142/S0219024911006565.
  • Jarrow et al. [2022] R. Jarrow, P. Protter, and A. Quintos. Computing the Probability of a Financial Market Failure: A New Measure of Systemic Risk. Annals of Operations Research, 2022. doi: 10.1007/s10479-022-05146-9. URL https://doi.org/10.1007/s10479-022-05146-9.
  • Jiao and Li [2015] Y. Jiao and S. Li. Generalized density approach in progressive enlargement of filtrations. Electronic Journal of Probability, 20:1 – 21, 2015. doi: 10.1214/EJP.v20-3296. URL https://doi.org/10.1214/EJP.v20-3296.
  • Karagiannidis and Lioumpas [2007] G. K. Karagiannidis and A. S. Lioumpas. An Improved Approximation for the Gaussian Q-Function. IEEE Communications Letters, 11(8):644–646, 2007. doi: 10.1109/LCOMM.2007.070470.
  • Lando [1998] D. Lando. On Cox processes and credit risky securities. Review of Derivatives Research, 2:99–120, 1998. doi: https://doi.org/10.1007/BF01531332.
  • Lando [2004] D. Lando. Credit risk modeling : theory and applications. Princeton University Press, 2004. ISBN 0691089299.
  • Li [2016] W. Li. Probability of Default and Default Correlations. Journal of Risk and Financial Management, 9(3), 2016. ISSN 1911-8074. doi: 10.3390/jrfm9030007. URL https://www.mdpi.com/1911-8074/9/3/7.
  • Lindskog and McNeil [2003] F. Lindskog and A. J. McNeil. Common Poisson Shock Models: Applications to Insurance and Credit Risk Modelling. ASTIN Bulletin, 33(2):209–238, 2003. doi: 10.1017/S0515036100013441.
  • Marshall and Olkin [1967] A. W. Marshall and I. Olkin. A Multivariate Exponential Distribution. Journal of the American Statistical Association, 62(317):30–44, 1967. ISSN 01621459. URL http://www.jstor.org/stable/2282907.
  • Petroski [2021] H. Petroski. What lessons will be learned from the Florida Condo Collapse? The deadly catastrophic failure has put a lens on building maintenance. American Scientist, 109:278–281, September–October 2021 2021.
  • Protter [2005] P. Protter. Stochastic integration and differential equations, volume 21. Springer-Verlag Berlin Heidelberg, 2 edition, 2005. doi: 10.1007/978-3-662-10061-5.
  • Sun et al. [2017] Y. Sun, R. Mendoza-Arriaga, and V. Linetsky. Marshall–Olkin distributions, subordinators, efficient simulation, and applications to credit risk. Advances in Applied Probability, 49(2):481–514, 2017. doi: 10.1017/apr.2017.10.
  • Swaine et al. [2021] J. Swaine, E. Brown, J. S. Lee, A. Mirza, and M. Kelly. How a collapsed pool deck could have caused a Florida condo building to fall. The Washington Post, August 12, 2021 2021.
  • Tanash and Riihonen [2021] I. M. Tanash and T. Riihonen. Improved Coefficients for the Karagiannidis–Lioumpas Approximations and Bounds to the Gaussian Q-Function. IEEE Communications Letters, 25(5):1468–1471, 2021. doi: 10.1109/LCOMM.2021.3052257.
  • Walpole [2021] B. Walpole. Quest for answers begins following Florida building collapse. American Society of Civil Engineers, Civil Engineering Source, June 25, 2021.
  • Young [1916] W. Young. On multiple integration by parts and the second theorem of the mean. Proceedings of the London Mathematical Society, 2(16):273–293, 1916. URL https://londmathsoc.onlinelibrary.wiley.com/doi/pdf/10.1112/plms/s2-16.1.273.
  • Zeng [2006] Y. Zeng. Compensators of stopping times (Ph.D. thesis). Cornell University Mathematics Department, 2006.