This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Modeling and Quickest Detection of a Rapidly Approaching Object

\nameTim Brucksa, Taposh Banerjeeb, and Rahul Mishrac CONTACT Taposh Banerjee. Email: [email protected] aUniversity of Texas at San Antonio; bUniversity of Pittsburgh; cU. R. Rao Satellite Center, Bangalore, India.
Abstract

The problem of detecting the presence of a signal that can lead to a disaster is studied. A decision-maker collects data sequentially over time. At some point in time, called the change point, the distribution of data changes. This change in distribution could be due to an event or a sudden arrival of an enemy object. If not detected quickly, this change has the potential to cause a major disaster. In space and military applications, the values of the measurements can stochastically grow with time as the enemy object moves closer to the target. A new class of stochastic processes, called exploding processes, is introduced to model stochastically growing data. An algorithm is proposed and shown to be asymptotically optimal as the mean time to a false alarm goes to infinity.

keywords:
Quickest change detection, nonstationary processes, monotone likelihood ratio, stochastic dominance, minimax optimality.

1 Introduction

In the problem of quickest change detection (QCD), a decision maker collects a sequence of measurements. The values in the sequence are seen as a realization of a stochastic process. It is assumed that the law of this stochastic process initially follows a distribution that is believed to be normal. At some point in time, called the change point, the statistical properties of the measurements or the law of the process changes. The goal of the QCD problem is to detect this change in distribution as quickly as possible while avoiding many false alarms [19, 12, 15]. The most common change point model is the abrupt and persistent change model [9, 14, 5, 16, 17, 10]. In this model,the law of the process abruptly changes at the change point and persists with that new law forever. For example, in Fig. 1 (left), we have plotted the mean values of a sequence of Gaussian random variables. At the change point (8080 in the figure), the mean abruptly changes from 0 to 44 and stays at 44 forever.

Refer to caption
Refer to caption
Figure 1: Left figure: Classical abrupt change point model. Figure shows the mean of a sequence of Gaussian random variables. This mean is 0 before change and is 44 after change. Right figure: In some space or military applications, the post-change measurements can have an exploding effect (linear, super-linear, or sublinear). The values of measurements can stochastically increase as the enemy object comes closer.

In some military and space applications, the post-change law can have an exploding nature.

  1. 1.

    Space scientists are concerned with the detection of debris or other hazardous objects approaching a satellite and destroying it (see Fig. 2). As the target object rapidly approaches the satellite, the mean of the measurements will have an exploding nature as shown in Fig. 1 (right). In Section 1.1, we provide a detailed discussion on this application.

  2. 2.

    Another classical example is enemy object detection in military applications. As a missile approaches a target or a torpedo approaches a submarine or ship, the values of the measurements are expected to increase with time.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 2: Space scientists are concerned with debris or other hazardous objects approaching a satellite and destroying it. A missile (or torpedo) can quickly approach a target (or submarine) and destroy it. Source: https://images.google.com/.

In this paper, we propose a new class of stochastic processes to capture the stochastically growing nature of a process. We then obtain an optimal algorithm for detecting a change in distribution in this new class of processes.

1.1 Satellite Application

In this section, we provide a brief discussion on satellite safety, the main motivation behind the quickest change detection problem studied in this paper.

Refer to caption
Refer to caption
Refer to caption
Figure 3: Left figure: Debris density in earth orbits; Right figure: Debris hit in ISS; Bottom figure: Debris size distribution.

The problem of detection of the approaching object in space and tracking them finds many concrete applications as follows:

  1. 1.

    In Fig. 3 (left and bottom), we provide the size and density of objects present in Earth orbits. These objects are potentially hazardous for Astronauts doing Extra-Vehicular Activities (EVA) and can be detrimental to the space resources such as operational satellites, ISS. In Fig. 3 (right), we can see the impact of debris hit in ISS. In [8, 20], the authors have mapped the debris and detect them using on-board systems.

  2. 2.

    Some of the major space accidents due to in space collisions and space debris are as follows: a) A piece of space junk damaged the robotic arm of Lab within the ISS; A puncture in the thermal covering of robotic arm was observed. b) The 1996 collision between the French Cerise with a catalogued debris from an Ariane rocket which tore the stabilization boom of the satellite. c) Collision of Iridium-33 (a part of commercial communication network) with the Kosmos 2251, a first major satellite collision in space of 2 satellites in Earth Orbit on account of congestion and not deorbiting the defunct satellite. This resulted in more than thousand debris particles which could have potentially lead to the escalation of debris collision. d) The recent collision in March 2021 between Yunhai-1 02 and debris from the Zenit-2 rocket created a debris of over 20 pieces.

  3. 3.

    Debris: Artificial space debris along with the meteoroids are referred as MMOD (Micrometeoroids and Orbital Debris). These objects can cause sandblasting, especially to spacecraft appendages and optics that are difficult to be protected by any shield [1]. Some of the types of space debris can be as follows: a) spare space parts, b) deorbitted satellites, and c) launch vehicle parts.

  4. 4.

    The orbital bodies such as functional and defunct satellites and their parts, lv parts, and the states of most of the catalogued objects are expected to evolve deterministically with the tolerances that are well modeled using the statistical approach illustrated in this paper. These reasonable assumptions are exploited to arrive at the useful results to enable autonomy in the future satellites. Thus the results in this paper can prove useful in Autonomous Meteors’ Avoidance Mechanism.

  5. 5.

    This work envisages to firm-up the more realistic assumptions pertaining to situations such as debris characterised by Kessler syndrome and intelligent chaser systems that can be in space or on-ground and can be a potential threat to cause disaster to space resources.

  6. 6.

    The immediate application of this work is to the autonomous spacecraft operations, which imparts substantial operational or commercial advantages to space industry. As the complexity of space operations is one of the major overheads associated with the satellite services, controlling this component can be critical in gaining an edge in the prevailing competition within the present space sector.

  7. 7.

    One recent example can be observed when the ESA satellite required a manoeuver to avoid a likely collision with the SPACEX’s Starlink satellite.

2 Model and Problem Formulation

2.1 Data Model

We introduce a new class of stochastic processes defined to model exploding nature of the post-change process:

Definition 2.1.

We say that a process {Xn}n1\{X_{n}\}_{n\geq 1} with densities {fn}n0\{f_{n}\}_{n\geq 0} is an exploding process if

  1. 1.

    {Xn}\{X_{n}\} are jointly independent,

  2. 2.

    Xnfn1X_{n}\sim f_{n-1}, n\forall n,

  3. 3.

    fn+1(x)fn(x)\frac{f_{n+1}(x)}{f_{n}(x)} is increasing in xx, n\forall n. We use fnfn+1f_{n}\prec f_{n+1} to denote this monotone likelihood ratio (MLR) order.

If fn+1=fnf_{n+1}=f_{n}, for all nn, then we get an independent and identically distributed (i.i.d.) process. Note that MLR dominance implies stochastic dominance [4]:

xfn(x)𝑑xxfn+1(x)𝑑x,x,n.\int_{x}^{\infty}f_{n}(x)dx\leq\int_{x}^{\infty}f_{n+1}(x)dx,\quad\forall x,n.

2.2 Change point model

We assume that before change, data is i.i.d. with density gf0g\prec f_{0}, and after change, an exploding process with sequence {fn}n0\{f_{n}\}_{n\geq 0} of densities. Mathematically, there is a discrete-time ν\nu such that

Xn{g,n<ν,fnνnν.X_{n}\sim\begin{cases}g,&\quad\forall n<\nu,\\ f_{n-\nu}&\quad\forall n\geq\nu.\end{cases}

Here we have used the notation XfX\sim f to denote that the random variable XX has law ff. Note that the post-change density of an observation depends on the location of the change point ν\nu. Our goal is to detect this change as quickly as possible, subject to a constraint on the rate of false alarms.

2.3 Problem Formulation

To solve the change detection problem, we are interested in two popular minimax formulations for the quickest change detection. To state the formulations, we define, for 1ν1\leq\nu\leq\infty, 𝖤ν\mathsf{E}_{\nu} as the expectation when the change occurs at time ν\nu. We consider the problem formulation of Lorden [7]:

minτsupν1ess sup𝖤ν[(τν+1)+|X1,,Xν1],subj. to 𝖤[τ]γ,\begin{split}\min_{\tau}&\quad\sup_{\nu\geq 1}\;\text{ess sup}\;\mathsf{E}_{\nu}[(\tau-\nu+1)^{+}|X_{1},\dots,X_{\nu-1}],\\ \text{subj. to }&\quad\mathsf{E}_{\infty}[\tau]\geq\gamma,\end{split} (1)

where γ\gamma is a constraint on the mean time to a false alarm. We will also consider the formulation of Pollak [11]:

minτsupν1𝖤ν[τν|τν],subj. to 𝖤[τ]γ,\begin{split}\min_{\tau}&\quad\sup_{\nu\geq 1}\;\mathsf{E}_{\nu}[\tau-\nu|\tau\geq\nu],\\ \text{subj. to }&\quad\mathsf{E}_{\infty}[\tau]\geq\gamma,\end{split} (2)

where again γ\gamma is a constraint on the mean time to a false alarm.

3 Candidate Algorithm: Exploding Cumulative Sum Algorithm

We propose to use the following exploding Cumulative Sum (EX-CUSUM) statistic for change detection:

Wn=max1kni=knlogfik(Xi)g(Xi).W_{n}=\max_{1\leq k\leq n}\;\sum_{i=k}^{n}\;\log\frac{f_{i-k}(X_{i})}{g(X_{i})}. (3)

The term i=knlogfik(Xi)g(Xi)\sum_{i=k}^{n}\;\log\frac{f_{i-k}(X_{i})}{g(X_{i})} is the log-likelihood ratio of the observations between post-change and pre-change distributions, conditioned that the change occurs at time kk. Since we do not know the change point, we take the maximum of all possible values at time nn, i.e., 1kn1\leq k\leq n.

In the above statistic, note that the likelihood ratio of an observation XiX_{i} at time ii,

fik(Xi)g(Xi),\frac{f_{i-k}(X_{i})}{g(X_{i})},

depends on the relative distance iki-k between time ii and the hypothesis kk of the change point. It is because of this reason the EX-CUSUM statistic is not a special case of the generalized CUSUM statistic of Lai [5].

To detect the change in law from i.i.d. with law gg to an exploding process {fn}\{f_{n}\}, we stop the first time the EX-CUSUM statistic is above a threshold AA:

τec=inf{n1:Wn>A}.\tau_{ec}=\inf\{n\geq 1:W_{n}>A\}. (4)

We select the threshold AA to control the rate of false alarms: the higher the threshold, the smaller the mean time to a false alarm 𝖤[τec]\mathsf{E}_{\infty}[\tau_{ec}] (this fact will be formally proved below).

Our goal in this paper is to characterize conditions under which the EX-CUSUM algorithm is asymptotically optimal for Lorden’s and Pollak’s problems in (1) and (2).

4 Asymptotic Lower Bound on the Performance

4.1 Lai’s Asymptotic Lower Bound

In [5], Lai reported a general minimax theory for the quickest change detection. We first review it and discuss its limitations.

It is assumed in [5] that the pre-change densities are f0(Xi|X1,,Xi1)f_{0}(X_{i}|X_{1},\dots,X_{i-1}) at time ii and the post-change densities are f1(Xi|X1,,Xi1)f_{1}(X_{i}|X_{1},\dots,X_{i-1}) giving us the log-likelihood ratio

Zi=logf1(Xi|X1,,Xi1)f0(Xi|X1,,Xi1).Z_{i}=\log\frac{f_{1}(X_{i}|X_{1},\dots,X_{i-1})}{f_{0}(X_{i}|X_{1},\dots,X_{i-1})}.

Note that the densities are general conditional densities allowing for data dependence but are not a function of the hypothesis on the change point (as defined in the paper). Lai showed that if ZiZ_{i}s are such that there exists an information number I>0I>0 satisfying

limnsupν1ess sup 𝖯ν(maxtni=νν+tZiI(1+δ)n|X1,,Xν1)=0,\begin{split}\lim_{n\to\infty}\;\sup_{\nu\geq 1}\;\text{ess sup }\mathsf{P}_{\nu}&\left(\max_{t\leq n}\sum_{i=\nu}^{\nu+t}Z_{i}\geq I(1+\delta)n\;\;\bigg{|}\;X_{1},\dots,X_{\nu-1}\right)=0,\end{split} (5)

then we have the universal lower bound as γ\gamma\to\infty,

minτsupν1ess sup𝖤ν[(τν+1)+|X1,,Xν1]minτsupν1𝖤ν[τν|τν]logγI(1+o(1)).\begin{split}\min_{\tau}\;\sup_{\nu\geq 1}\;&\text{ess sup}\;\mathsf{E}_{\nu}[(\tau-\nu+1)^{+}|X_{1},\dots,X_{\nu-1}]\\ &\geq\quad\min_{\tau}\;\sup_{\nu\geq 1}\;\mathsf{E}_{\nu}[\tau-\nu|\tau\geq\nu]\\ &\quad\geq\quad\frac{\log\gamma}{I}(1+o(1)).\end{split} (6)

Here the minimum over τ\tau is over those stopping times satisfying 𝖤[τ]γ\mathsf{E}_{\infty}[\tau]\geq\gamma. Lai further showed that under certain additional conditions on the ZiZ_{i}s, the generalized CUSUM algorithm,

τc=min{n1:max1kni=knZilog(γ)},\tau_{c}=\min\left\{n\geq 1:\max_{1\leq k\leq n}\sum_{i=k}^{n}Z_{i}\geq\log(\gamma)\right\},

is asymptotically optimal for both Lorden and Pollak’s problems. Specifically,

𝖤[τc]γ.\mathsf{E}_{\infty}[\tau_{c}]\geq\gamma.

Further, if ZiZ_{i}s satisfy

limnsupkν1ess sup 𝖯ν(1ni=kk+nZiIδ|X1,,Xk1)=0,\begin{split}\lim_{n\to\infty}\;\sup_{k\geq\nu\geq 1}\;\text{ess sup }\mathsf{P}_{\nu}&\left(\frac{1}{n}\sum_{i=k}^{k+n}Z_{i}\leq I-\delta\;\bigg{|}\;X_{1},\dots,X_{k-1}\right)=0,\end{split} (7)

then as γ\gamma\to\infty, τc\tau_{c} achieves the lower bound:

supν1ess sup𝖤ν[(τcν+1)+|X1,,Xν1]logγI(1+o(1)),γ.\begin{split}\sup_{\nu\geq 1}\;\text{ess sup}\;\mathsf{E}_{\nu}&[(\tau_{c}-\nu+1)^{+}|X_{1},\dots,X_{\nu-1}]\leq\frac{\log\gamma}{I}(1+o(1)),\quad\gamma\to\infty.\end{split} (8)

It is not clear if these results are valid for the exploding process setting because in the latter setting likelihood ratios do depend on where the change point is. We discuss this next.

4.2 Sufficient Conditions for Change Point Dependent Likelihoods

In this section, we extend Lai’s results to the case of change point-dependent likelihoods. We continue to allow data dependence across time to state the more general result. Define the log-likelihood ratio at time nn when change occurs at ν\nu as

Zn,ν=logfn,ν(Xn|X1,,Xn1)f0(Xn|X1,,Xn1).Z_{n,\nu}=\log\frac{f_{n,\nu}(X_{n}|X_{1},\dots,X_{n-1})}{f_{0}(X_{n}|X_{1},\dots,X_{n-1})}.
Theorem 4.1.
  1. 1.

    Let there exist a positive number II such that the bivariate log likelihood ratios {Zn,ν}\{Z_{n,\nu}\} satisfy the following condition:

    limnsupν1ess sup 𝖯ν(maxtni=νν+tZi,νI(1+δ)n|X1,,Xν1)=0.\begin{split}\lim_{n\to\infty}\;\sup_{\nu\geq 1}\;\text{ess sup }\mathsf{P}_{\nu}&\left(\max_{t\leq n}\sum_{i=\nu}^{\nu+t}Z_{i,\nu}\geq I(1+\delta)n\;\bigg{|}\;X_{1},\dots,X_{\nu-1}\right)=0.\end{split} (9)

    Then, we have the universal lower bound as γ\gamma\to\infty,

    minτsupν1ess sup𝖤ν[(τν+1)+|X1,,Xν1]minτsupν1𝖤ν[τν|τν]logγI(1+o(1)).\begin{split}\min_{\tau}\;\sup_{\nu\geq 1}\;&\text{ess sup}\;\mathsf{E}_{\nu}[(\tau-\nu+1)^{+}|X_{1},\dots,X_{\nu-1}]\\ &\geq\quad\min_{\tau}\;\sup_{\nu\geq 1}\;\mathsf{E}_{\nu}[\tau-\nu|\tau\geq\nu]\\ &\quad\geq\quad\frac{\log\gamma}{I}(1+o(1)).\end{split} (10)

    Here the minimum over τ\tau is over those stopping times satisfying 𝖤[τ]γ\mathsf{E}_{\infty}[\tau]\geq\gamma.

  2. 2.

    The following modified generalized CUSUM algorithm,

    τmc=min{n1:max1kni=knZi,klog(γ)},\tau_{mc}=\min\left\{n\geq 1:\max_{1\leq k\leq n}\sum_{i=k}^{n}Z_{i,k}\geq\log(\gamma)\right\},

    satisfies

    𝖤[τmc]γ.\mathsf{E}_{\infty}[\tau_{mc}]\geq\gamma.
  3. 3.

    Let the bivariate log-likelihood ratios {Zn,ν}\{Z_{n,\nu}\} also satisfy

    limnsupkν1ess sup 𝖯ν(1ni=kk+nZi,kIδ|X1,,Xk1)=0.\begin{split}\lim_{n\to\infty}\;\sup_{k\geq\nu\geq 1}\;\text{ess sup }\mathsf{P}_{\nu}&\left(\frac{1}{n}\sum_{i=k}^{k+n}Z_{i,k}\leq I-\delta\;\bigg{|}\;X_{1},\dots,X_{k-1}\right)=0.\end{split} (11)

    Then as γ\gamma\to\infty, τmc\tau_{mc} achieves the lower bound:

    supν1ess sup𝖤ν[(τmcν+1)+|X1,,Xν1]logγI(1+o(1)),γ.\begin{split}\sup_{\nu\geq 1}\;\text{ess sup}\;\mathsf{E}_{\nu}&[(\tau_{mc}-\nu+1)^{+}|X_{1},\dots,X_{\nu-1}]\leq\frac{\log\gamma}{I}(1+o(1)),\quad\gamma\to\infty.\end{split} (12)
Proof.

The lower bound result goes through by replacing ZiZ_{i}s by Zn,νZ_{n,\nu}s in [5]. This is because the proof relies on a change of measure argument to get the lower bound. Since the likelihood ratios here are a function of Zn,νZ_{n,\nu}, the same change of measure argument works. The proof of detection delay also goes through provided the above conditions are satisfied and everywhere ZiZ_{i} are replaced by Zi,kZ_{i,k}. It is not clear to the authors whether Lai’s proof for the mean time to a false alarm result can be extended to the time-dependent setting. But, one can use another technique based on the Shiryaev-Roberts statistic and optional sampling theorem. Since the latter technique is classical [15], we skip the details. ∎

While this paper was being written, the authors were made aware of another paper [6] in which more general sufficient conditions for optimality (more general than those in [5]) have been established. In comparison with [6], we provide a novel way (using MLR order) to model stochastically growing nonstationary post-change process. In addition, our proof techniques for verifying these optimality conditions are slightly different and should be of independent interest.

5 Optimality of EX-CUSUM Algorithm

In this section, we first simplify the conditions in Theorem 4.1 for exploding processes as defined in Definition 2.1. We then provide additional comments on the simplified conditions to guarantee the optimality of the EX-CUSUM algorithm.

5.1 Simplifying Lower Bound Condition for Exploding Processes

In the theorem below, we show that the lower bound condition (9) can be simplified in the case of exploding processes.

Theorem 5.1.

To satisfy

limnsupν1ess sup 𝖯ν(maxtni=νν+tZi,νI(1+δ)n|X1,,Xν1)=0\begin{split}\lim_{n\to\infty}\;\sup_{\nu\geq 1}\;\text{ess sup }\mathsf{P}_{\nu}&\left(\max_{t\leq n}\sum_{i=\nu}^{\nu+t}Z_{i,\nu}\geq I(1+\delta)n\;\bigg{|}\;X_{1},\dots,X_{\nu-1}\right)=0\end{split} (13)

for some 0<I<0<I<\infty, it is sufficient that

1nk=1nZk,1=1nk=1nlogfk1(Xk)g(Xk)I,a.s. under 𝖯1.\frac{1}{n}\sum_{k=1}^{n}Z_{k,1}=\frac{1}{n}\sum_{k=1}^{n}\log\frac{f_{k-1}(X_{k})}{g(X_{k})}\;\to\;I,\quad\text{a.s. under }\mathsf{P}_{1}.
Proof.

Because of independence, we have

supν1ess sup 𝖯ν(maxtni=νν+tZi,νI(1+δ)n|X1,,Xν1)=supν1𝖯ν(maxtni=νν+tlogfiν(Xi)g(Xi)I(1+δ)n).\begin{split}\sup_{\nu\geq 1}\;&\text{ess sup }\mathsf{P}_{\nu}\left(\max_{t\leq n}\sum_{i=\nu}^{\nu+t}Z_{i,\nu}\geq I(1+\delta)n\;\bigg{|}\;X_{1},\dots,X_{\nu-1}\right)\\ &=\sup_{\nu\geq 1}\mathsf{P}_{\nu}\left(\max_{t\leq n}\sum_{i=\nu}^{\nu+t}\log\frac{f_{i-\nu}(X_{i})}{g(X_{i})}\geq I(1+\delta)n\right).\end{split}

Since likelihood ratios are computed relative to the change point ν\nu, the probability on the right is not a function of ν\nu. This implies

supν1𝖯ν(maxtni=νν+tlogfiν(Xi)g(Xi)I(1+δ)n)=𝖯1(maxtni=11+tlogfi1(Xi)g(Xi)I(1+δ)n)=𝖯1(1nmaxtni=11+tlogfi1(Xi)g(Xi)I(1+δ)).\begin{split}\sup_{\nu\geq 1}\;\mathsf{P}_{\nu}\left(\max_{t\leq n}\sum_{i=\nu}^{\nu+t}\log\frac{f_{i-\nu}(X_{i})}{g(X_{i})}\geq I(1+\delta)n\right)&=\mathsf{P}_{1}\left(\max_{t\leq n}\sum_{i=1}^{1+t}\log\frac{f_{i-1}(X_{i})}{g(X_{i})}\geq I(1+\delta)n\right)\\ &=\mathsf{P}_{1}\left(\frac{1}{n}\max_{t\leq n}\sum_{i=1}^{1+t}\log\frac{f_{i-1}(X_{i})}{g(X_{i})}\geq I(1+\delta)\right).\end{split}

Note that if the post-change process evolves differently for different change points, then the above simplification may not be true. Now, if

1ni=1nlogfi1(Xi)g(Xi)I,a.s. under𝖯1,\frac{1}{n}\sum_{i=1}^{n}\log\frac{f_{i-1}(X_{i})}{g(X_{i})}\;\to\;I,\quad\text{a.s. under}\;\mathsf{P}_{1},

then

1nmaxtni=11+tlogfi,1(Xi)g(Xi)I,a.s. under𝖯1.\frac{1}{n}\max_{t\leq n}\sum_{i=1}^{1+t}\log\frac{f_{i,1}(X_{i})}{g(X_{i})}\to I,\quad\text{a.s. under}\;\mathsf{P}_{1}.

For proof of the above fact, see the Proof of Theorem 5.1 in [2]. This proves the theorem because convergence almost surely implies convergence in probability. ∎

We still need to characterize conditions under which

1ni=1nlogfi1(Xi)g(Xi)I,a.s. under𝖯1,\frac{1}{n}\sum_{i=1}^{n}\log\frac{f_{i-1}(X_{i})}{g(X_{i})}\;\to\;I,\quad\text{a.s. under}\;\mathsf{P}_{1},

for an exploding process. This will be done in Section 5.4.

5.2 Controlling the False Alarm

In this section, we show that by setting the threshold A=logγA=\log\gamma in the EX-CUSUM algorithm, the constraints on the mean time to a false alarm can be satisfied. This is the content of the next theorem.

Recall that

Wn=max1kni=knlogfik(Xi)g(Xi),W_{n}=\max_{1\leq k\leq n}\;\sum_{i=k}^{n}\;\log\frac{f_{i-k}(X_{i})}{g(X_{i})},
τec=inf{n1:Wn>A}.\tau_{ec}=\inf\{n\geq 1:W_{n}>A\}.
Theorem 5.2.

Setting A=log(γ)A=\log(\gamma) ensures that

𝖤[τec]γ.\mathsf{E}_{\infty}[\tau_{ec}]\geq\gamma.
Proof.

This proof technique is standard and has been used, for example, in [15]. Because logarithm is monotonic and the maximum of positive quantities is always less than their sum, we have

τec=inf{n1:max1kni=knlogfik(Xi)g(Xi)>A}=inf{n1:max1kni=knfik(Xi)g(Xi)>eA}inf{n1:1kni=knfik(Xi)g(Xi)>eA}:=τesr.\begin{split}\tau_{ec}&=\inf\left\{n\geq 1:\max_{1\leq k\leq n}\;\sum_{i=k}^{n}\;\log\frac{f_{i-k}(X_{i})}{g(X_{i})}>A\right\}\\ &=\inf\left\{n\geq 1:\max_{1\leq k\leq n}\;\prod_{i=k}^{n}\;\frac{f_{i-k}(X_{i})}{g(X_{i})}>e^{A}\right\}\\ &\geq\inf\left\{n\geq 1:\sum_{1\leq k\leq n}\;\prod_{i=k}^{n}\;\frac{f_{i-k}(X_{i})}{g(X_{i})}>e^{A}\right\}:=\tau_{esr}.\end{split}

The process

Rnn:=1kni=knfik(Xi)g(Xi)nR_{n}-n:=\sum_{1\leq k\leq n}\;\prod_{i=k}^{n}\;\frac{f_{i-k}(X_{i})}{g(X_{i})}-n

is a 𝖯\mathsf{P}_{\infty}-martingale. Assuming 𝖤[τesr]<\mathsf{E}[\tau_{esr}]<\infty (otherwise the false alarm constraint is trivially satisfied), we have

𝖤[|Rnn|;{τesr>n}]𝖤[eA+τesr;{τesr>n}]0,n.\begin{split}\mathsf{E}\left[|R_{n}-n|;\{\tau_{esr}>n\}\right]&\leq\mathsf{E}\left[e^{A}+\tau_{esr};\{\tau_{esr}>n\}\right]\\ &\to 0,\quad n\to\infty.\end{split}

Thus, by Doob’s optional sampling theorem [3],

𝖤[Rτesrτesr]=0,\mathsf{E}\left[R_{\tau_{esr}}-\tau_{esr}\right]=0,

and

𝖤[τesr]=𝖤[Rτesr]eA.\mathsf{E}\left[\tau_{esr}\right]=\mathsf{E}\left[R_{\tau_{esr}}\right]\geq e^{A}.

Now, set A=logγA=\log\gamma to complete the proof. ∎

5.3 Simplifying Upper Bound Condition for Delay Analysis of an Exploding Process

In this section, we simplify the condition (11) for exploding processes. To complete this step, we need an intermediate result.

Lemma 5.3.

Let f(x1,x2,,xn)f(x_{1},x_{2},\dots,x_{n}) be a continuous function increasing in each of its arguments, with other arguments fixed. If {Xn}\{X_{n}\} is a stochastic process generated according to an exploding process, then for all n,m,tn,m,t,

𝖯(f(Xn,Xn+1,,Xn+m)t)𝖯(f(Xn+1,Xn+2,,Xn+m+1)t).\begin{split}\mathsf{P}&\left(f(X_{n},X_{n+1},\dots,X_{n+m})\geq t\right)\leq\mathsf{P}\left(f(X_{n+1},X_{n+2},\dots,X_{n+m+1})\geq t\right).\end{split} (14)
Proof.

The proof is similar to the one given in [18]. Our proof requires an additional randomization step. ∎

As in Theorem 5.1, we show below that an almost sure condition is sufficient to satisfy the condition in (11).

Theorem 5.4.

To satisfy, for all δ>0\delta>0,

limnsupkν1ess sup 𝖯ν(1ni=kk+nZi,kIδ|X1,,Xk1)=0\begin{split}\lim_{n\to\infty}\;\sup_{k\geq\nu\geq 1}\;\text{ess sup }\mathsf{P}_{\nu}&\left(\frac{1}{n}\sum_{i=k}^{k+n}Z_{i,k}\leq I-\delta\;\bigg{|}\;X_{1},\dots,X_{k-1}\right)=0\end{split} (15)

for some 0<I<0<I<\infty, it is sufficient that the log-likelihoods are continuous and

1nk=1nZk,1=1nk=1nlogfk1(Xk)g(Xk)I,a.s. under 𝖯1.\frac{1}{n}\sum_{k=1}^{n}Z_{k,1}=\frac{1}{n}\sum_{k=1}^{n}\log\frac{f_{k-1}(X_{k})}{g(X_{k})}\;\to\;I,\quad\text{a.s. under }\mathsf{P}_{1}.
Proof.

Due to independence and the nature of exploding processes, we have

supkν1ess sup 𝖯ν(1ni=kk+nZi,kIδ|X1,,Xk1)=supkν1𝖯ν(1ni=kk+nlogfik(Xi)g(Xi)Iδ)=supk1𝖯1(1ni=kk+nlogfik(Xi)g(Xi)Iδ).\begin{split}\sup_{k\geq\nu\geq 1}\;&\text{ess sup }\mathsf{P}_{\nu}\left(\frac{1}{n}\sum_{i=k}^{k+n}Z_{i,k}\leq I-\delta\;\bigg{|}\;X_{1},\dots,X_{k-1}\right)\\ &=\sup_{k\geq\nu\geq 1}\mathsf{P}_{\nu}\left(\frac{1}{n}\sum_{i=k}^{k+n}\log\frac{f_{i-k}(X_{i})}{g(X_{i})}\leq I-\delta\right)\\ &=\sup_{k\geq 1}\mathsf{P}_{1}\left(\frac{1}{n}\sum_{i=k}^{k+n}\log\frac{f_{i-k}(X_{i})}{g(X_{i})}\leq I-\delta\right).\end{split}

Because of Lemma 5.3, the random variables

1ni=kk+nlogfik(Xi)g(Xi)\frac{1}{n}\sum_{i=k}^{k+n}\log\frac{f_{i-k}(X_{i})}{g(X_{i})}

becomes stochastically bigger as kk increases. Thus, the maximum probability over kk is achieved at k=1k=1. This gives

supk1𝖯1(1ni=kk+nlogfik(Xi)g(Xi)Iδ)=𝖯1(1ni=11+nlogfi1(Xi)g(Xi)Iδ).\begin{split}\sup_{k\geq 1}\;&\mathsf{P}_{1}\left(\frac{1}{n}\sum_{i=k}^{k+n}\log\frac{f_{i-k}(X_{i})}{g(X_{i})}\leq I-\delta\right)=\mathsf{P}_{1}\left(\frac{1}{n}\sum_{i=1}^{1+n}\log\frac{f_{i-1}(X_{i})}{g(X_{i})}\leq I-\delta\right).\end{split}

The last term goes to zero if

1nk=1nlogfk1(Xk)g(Xk)I,a.s. under 𝖯1.\frac{1}{n}\sum_{k=1}^{n}\log\frac{f_{k-1}(X_{k})}{g(X_{k})}\;\to\;I,\quad\text{a.s. under }\mathsf{P}_{1}.

This completes the proof.

Thus, for the optimality of the exploding CUSUM algorithms, it is enough to find conditions under which the almost sure convergence stated in the previous theorem is satisfied.

5.4 A Law of Large Numbers for Independent and Non-Identically Distributed Random Variables

In this section, we give conditions on exploding processes to guarantee

1nk=1nlogfk1(Xk)g(Xk)I,a.s. under 𝖯1.\frac{1}{n}\sum_{k=1}^{n}\log\frac{f_{k-1}(X_{k})}{g(X_{k})}\;\to\;I,\quad\text{a.s. under }\mathsf{P}_{1}.

We first recall Cantelli’s strong law of large numbers. The proof can be found in [13].

Lemma 5.5.

Let Y1,Y2,Y_{1},Y_{2},\dots be independent random variables with finite fourth moments, and let

𝖤[|Yn𝖤[Yn]|4]C,n1,\mathsf{E}[|Y_{n}-\mathsf{E}[Y_{n}]|^{4}]\leq C,\quad n\geq 1,

for some constant CC. Then, as nn\to\infty,

Sn𝖤[Sn]n0,almost surely.\frac{S_{n}-\mathsf{E}[S_{n}]}{n}\to 0,\quad\text{almost surely}.

Here Sn=Y1+Y2+YnS_{n}=Y_{1}+Y_{2}+\dots Y_{n}.

Cantelli’s strong law of large numbers provides us the needed tool to state our conditions.

Theorem 5.6.

We assume the following conditions hold for every k1k\geq 1:

𝖤[(logfk1(Xk)g(Xk))4]<,Xkfk1,𝖤[(logfk1(Xk)g(Xk)D(fk1g))4]C,Xkfk1,\begin{split}\mathsf{E}&\left[\left(\log\frac{f_{k-1}(X_{k})}{g(X_{k})}\right)^{4}\right]<\infty,\quad X_{k}\sim f_{k-1},\\ \mathsf{E}&\left[\left(\log\frac{f_{k-1}(X_{k})}{g(X_{k})}-D(f_{k-1}\;\|\;g)\right)^{4}\right]\leq C,\quad X_{k}\sim f_{k-1},\end{split} (16)

where CC is a constant. Further, there exists an I>0I>0 such that

1nk=1nD(fk1g)I,n.\begin{split}\frac{1}{n}&\sum_{k=1}^{n}D(f_{k-1}\;\|\;g)\quad\to\quad I,\quad n\to\infty.\end{split} (17)

Then, by Cantelli’s strong law of large numbers (Lemma 5.5),

1nk=1nlogfk1(Xk)g(Xk)I,a.s. under 𝖯1.\frac{1}{n}\sum_{k=1}^{n}\log\frac{f_{k-1}(X_{k})}{g(X_{k})}\;\to\;I,\quad\text{a.s. under }\mathsf{P}_{1}.

5.5 Asymptotic Optimality of EX-CUSUM Algorithm

The previous theorem provides further simplification on the conditions needed for the optimality of the EX-CUSUM algorithm. We state this as a theorem:

Theorem 5.7.

Under moment conditions stated in Theorem 5.6, if there exists an I>0I>0, such that

1nk=1nD(fk1g)I,n,\frac{1}{n}\sum_{k=1}^{n}D(f_{k-1}\;\|\;g)\quad\to\quad I,\quad n\to\infty,

then the EX-CUSUM algorithm is asymptotically optimal for both Lorden’s and Pollak’s minimax formulations.

5.6 Gaussian Example

We now give an example of an exploding process model for which all the conditions stated in this paper are satisfied. We assume that

g=𝒩(0,1).g=\mathcal{N}(0,1).

and

fn=𝒩(μn,1)f_{n}=\mathcal{N}(\mu_{n},1)

with

0μnμ.0\leq\mu_{n}\uparrow\mu.

All the likelihood ratios in this example are monotone because 0μnμ0\leq\mu_{n}\to\mu. Also, log-likelihood ratios are continuous because they are linear. The fourth-moment condition on log-likelihood ratios is satisfied because of Gaussianity.

Further, the log-likelihood ratio between fnf_{n} and gg is given by

logfn(Xn+1)g(Xn+1)=μn(Xn+1μn2).\log\frac{f_{n}(X_{n+1})}{g(X_{n+1})}=\mu_{n}(X_{n+1}-\frac{\mu_{n}}{2}).

Thus,

D(fng)=𝖤Xn+1fn[logfn(Xn+1)g(Xn+1)]=μn[𝖤Xn+1fn[Xn+1]μn2]=μn(μnμn2)=μn22.\begin{split}D(f_{n}\;\|\;g)&=\mathsf{E}_{X_{n+1}\sim f_{n}}\left[\log\frac{f_{n}(X_{n+1})}{g(X_{n+1})}\right]=\mu_{n}\left[\mathsf{E}_{X_{n+1}\sim f_{n}}[X_{n+1}]-\frac{\mu_{n}}{2}\right]\\ &=\mu_{n}\left(\mu_{n}-\frac{\mu_{n}}{2}\right)=\frac{\mu_{n}^{2}}{2}.\end{split} (18)

This implies

D(fng)=μn22μ22.D(f_{n}\;\|\;g)=\frac{\mu_{n}^{2}}{2}\to\frac{\mu^{2}}{2}.

This further implies

1nk=1nD(fk1g)μ22\frac{1}{n}\sum_{k=1}^{n}D(f_{k-1}\;\|\;g)\quad\to\quad\frac{\mu^{2}}{2}

by Cesaro sum limit.

Finally, we need to check if the fourth central moments of the log-likelihoods are uniformly bounded. Let

ξn:=logfn(Xn+1)g(Xn+1)=μn(Xn+1μn2).\xi_{n}:=\log\frac{f_{n}(X_{n+1})}{g(X_{n+1})}=\mu_{n}(X_{n+1}-\frac{\mu_{n}}{2}).

Also, let

Ξn:=𝖤Xn+1fn[ξn]=μn22.\Xi_{n}:=\mathsf{E}_{X_{n+1}\sim f_{n}}[\xi_{n}]=\frac{\mu_{n}^{2}}{2}.

Then,

𝖤Xn+1fn[(ξnΞn)4]=𝖤Xn+1fn[(μn(Xn+1μn2)μn22)4]=𝖤Xn+1fn[(μnXn+1μn2)4]=𝖤Xn+1fn[(μn(Xn+1μn))4]=μn4𝖤Xn+1fn[((Xn+1μn))4]3μ4.\begin{split}\mathsf{E}_{X_{n+1}\sim f_{n}}\left[(\xi_{n}-\Xi_{n})^{4}\right]&=\mathsf{E}_{X_{n+1}\sim f_{n}}\left[\left(\mu_{n}(X_{n+1}-\frac{\mu_{n}}{2})-\frac{\mu_{n}^{2}}{2}\right)^{4}\right]\\ &\quad=\mathsf{E}_{X_{n+1}\sim f_{n}}\left[\left(\mu_{n}X_{n+1}-\mu_{n}^{2}\right)^{4}\right]\\ &\quad=\mathsf{E}_{X_{n+1}\sim f_{n}}\left[\left(\mu_{n}(X_{n+1}-\mu_{n})\right)^{4}\right]\\ &\quad=\mu_{n}^{4}\mathsf{E}_{X_{n+1}\sim f_{n}}\left[\left((X_{n+1}-\mu_{n})\right)^{4}\right]\\ &\quad\leq 3\mu^{4}.\end{split} (19)

The last inequality is true because μnμ\mu_{n}\uparrow\mu and because under fnf_{n}, Xn+1μn𝒩(0,1)X_{n+1}-\mu_{n}\sim\mathcal{N}(0,1), and the latter’s fourth moment is exactly equal to 33. Thus, the fourth central moments are uniformly bounded by C=3μ4C=3\mu^{4}. All the above arguments show that the EX-CUSUM algorithm is asymptotically optimal for this example.

6 Numerical Results

In this section, we apply the EX-CUSUM algorithm to the Gaussian data example discussed in Section 5.6. To generate data in Fig. 4, we used

μn=arctan(n)μ=π2.\mu_{n}=\arctan(n)\to\mu=\frac{\pi}{2}.

As seen in the figure, the EX-CUSUM statistic stays close to zero before the change point of 8080 and grows toward \infty after the change. This growth can be detected using a well-designed threshold.

Refer to caption
Refer to caption
Figure 4: EX-CUSUM algorithm applied to Gaussian data.

7 Conclusions

We introduced a new class of stochastic processes to model exploding nature of post-change observations in some change-point problems. Such observations are common in satellite and military applications where an approaching enemy object can cause the observation to grow stochastically over time. We proved that under mild conditions on the growth of Kullback-Leibler divergence, our proposed EX-CUSUM algorithm is asymptotically optimal, as the mean time to false alarm grows to infinity. In our future work, we will investigate the exact optimality of the algorithm and also obtain computationally efficient methods for detecting changes.

Acknowledgment

A part of this work was presented at the 58th Annual Allerton Conference on Communication, Control, and Computing, Monticello, IL, September 2022. The work of Taposh Banerjee was partially supported by the National Science Foundation under Grant 1917164 through a subcontract from Vanderbilt University. Tim Brucks passed away while we were working on this paper. This paper is dedicated to his memory.

References

  • [1] The threat of orbital debris and protecting nasa space assets from satellite collisions, http://images.spaceref.com/news/2009/ODMediaBriefing28Apr09-1.pdf. Accessed: 2023-02-03.
  • [2] T. Banerjee, P. Gurram, and G.T. Whipps, A Bayesian theory of change detection in statistically periodic random processes, IEEE Transactions on Information Theory 67 (2021), pp. 2562–2580.
  • [3] Y.S. Chow, H. Robbins, and D. Siegmund, Great expectations: the theory of optimal stopping, Houghton Mifflin, 1971.
  • [4] V. Krishnamurthy, Partially Observed Markov Decision Processes, Cambridge University Press, 2016.
  • [5] T.L. Lai, Information bounds and quick detection of parameter changes in stochastic systems, IEEE Trans. Inf. Theory 44 (1998), pp. 2917 –2929.
  • [6] Y. Liang, A.G. Tartakovsky, and V.V. Veeravalli, Quickest change detection with non-stationary post-change observations, To appear in IEEE Transactions on Information Theory (2022).
  • [7] G. Lorden, Procedures for reacting to a change in distribution, Ann. Math. Statist. 42 (1971), pp. 1897–1908.
  • [8] S. Mohammed, M.J.S. Teja, P. Kora, D.E. Vunnava, and M. Lokesh, Space Debris Detection Unit for Spacecrafts, in 2022 International Conference on Computer Communication and Informatics (ICCCI). IEEE, 2022, pp. 1–2.
  • [9] G.V. Moustakides, Optimal stopping times for detecting changes in distributions, Ann. Statist. 14 (1986), pp. 1379–1387.
  • [10] S. Pergamenchtchikov and A.G. Tartakovsky, Asymptotically optimal pointwise and minimax quickest change-point detection for dependent data, Statistical Inference for Stochastic Processes 21 (2018), pp. 217–259.
  • [11] M. Pollak, Optimal detection of a change in distribution, Ann. Statist. 13 (1985), pp. 206–227.
  • [12] H.V. Poor and O. Hadjiliadis, Quickest detection, Cambridge University Press, 2009.
  • [13] A.N. Shiryaev, Probability-2, Vol. 95, Springer, 2019.
  • [14] A.N. Shiryayev, Optimal Stopping Rules, Springer-Verlag, New York, 1978.
  • [15] A.G. Tartakovsky, I.V. Nikiforov, and M. Basseville, Sequential Analysis: Hypothesis Testing and Change-Point Detection, Statistics, CRC Press, 2014.
  • [16] A.G. Tartakovsky and V.V. Veeravalli, General asymptotic Bayesian theory of quickest change detection, SIAM Theory of Prob. and App. 49 (2005), pp. 458–497.
  • [17] A.G. Tartakovsky, On asymptotic optimality in sequential changepoint detection: Non-iid case, IEEE Transactions on Information Theory 63 (2017), pp. 3433–3450.
  • [18] J. Unnikrishnan, V.V. Veeravalli, and S.P. Meyn, Minimax robust quickest change detection, IEEE Trans. Inf. Theory 57 (2011), pp. 1604 –1614.
  • [19] V.V. Veeravalli and T. Banerjee, Quickest Change Detection, Academic Press Library in Signal Processing: Volume 3 – Array and Statistical Signal Processing, 2014.
  • [20] Y. Xiang, J. Xi, M. Cong, Y. Yang, C. Ren, and L. Han, Space debris detection with fast grid-based learning, in 2020 IEEE 3rd International Conference of Safe Production and Informatization (IICSPI). IEEE, 2020, pp. 205–209.