This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Behavioral event detection and rate estimation for autonomous vehicle evaluation

Maria A. Terres, Aiyou Chen, Rachel Zhou, Claire M. McLeod

Waymo Data Science
(Date: January 1, 2025)
Abstract.

Autonomous vehicles are continually increasing their presence on public roads. However, before any new autonomous driving software can be approved, it must first undergo a rigorous assessment of driving quality. These quality evaluations typically focus on estimating the frequency of (undesirable) behavioral events. While rate estimation would be straight-forward with complete data, in the autonomous driving setting this estimation is greatly complicated by the fact that detecting these events within large driving logs is a non-trivial task that often involves human reviewers. In this paper we outline a streaming partial tiered event review configuration that ensures both high recall and high precision on the events of interest. In addition, the framework allows for valid streaming estimates at any phase of the data collection process, even when labels are incomplete, for which we develop the maximum likelihood estimate and show it is unbiased. Constructing honest and effective confidence intervals (CI) for these rate estimates, particularly for rare safety-critical events, is a novel and challenging statistical problem due to the complexity of the data likelihood. We develop and compare several CI approximations, including a novel Gamma CI method that approximates the exact but intractable distribution with a weighted sum of independent Poisson random variables. There is a clear trade-off between statistical coverage and interval width across the different CI methods, and the extent of this trade-off varies depending on the specific application settings (e.g., rare vs. common events). In particular, we argue that our proposed CI method is the best-suited when estimating the rate of safety-critical events where guaranteed coverage of the true parameter value is a prerequisite to safely launching a new ADS on public roads.

1. Introduction

1.1. Autonomous Driving Systems

An “autonomous driving system" (ADS) refers to a level 3, 4, or 5 system [12] that is capable of performing a dynamic driving task on a sustained basis. It is comprised of both software and hardware components, and any vehicle equipped with this system may be referred to as an "Autonomous Vehicle" (AV).

Development of safe and high quality ADSs is a very active area of research within both academia and industry [15, 22]. Waymo is a prominent player in this space, with its origins in 2009 as the Google Self-Driving Car Project. Today Waymo produces Level 4 ADSs which currently provide fully autonomous rides, i.e., with no human in the driver seat, to members of the public in Arizona and in California. Waymo also has parallel work in the autonomous trucking and delivery space.

At Waymo, rigorous validation and evaluation of the ADS performance is a core tenet of the deployment philosophy. Prior to releasing AVs equipped with a new ADS onto public roads, we must have confidence that the safety and other performance metrics meet our stringent quality bar. Previous Waymo publications have already outlined the safety and readiness determinations [20, 13]. Complementing those publications, here we articulate some of the statistical and data science challenges that arise during the evaluation process.

1.2. ADS Performance Evaluation

ADS performance evaluation typically involves three primary components:

  1. (1)

    Definition of specific undesirable111There are many desirable behavioral events that are also of interest to track. For simplicity, in this paper we assume that these events can all be negated to produce a corresponding undesirable event. E.g., a desirable event may be “sufficiently pullover to pick up passenger out of the travel lane,” and the associated negative metrics would be things like “pullover had X percent overlap with travel lane” or “failed to pullover.” behavioral events, along with associated severity levels or other relevant annotations.

  2. (2)

    Rate estimation for the behavioral events’ relative frequencies on-road, typically in terms of the number of incidents per driving mile (IPM).

  3. (3)

    Accommodation of product requirements from the stakeholders who use these data for decision-making.

More details on each of these components are provided below.

The behavioral events of interest referenced in component (1) span a spectrum from critical safety situations such as collisions (very rare) to behavioral quirks such as unnecessary slowing (more common). While on its surface this aspect of performance evaluation appears straightforward, in practice rigorous definitions and delineation of edge cases can be a challenge. These challenges are left as out of scope for the current paper. Rather, we focus on data and statistical questions associated with the components (2) and (3), and how we are addressing these challenges at Waymo.

There are several data and statistical questions that arise when producing the rate estimates referenced in component (2):

  1. a)

    What data is best suited to produce these rate estimates? Waymo has autonomously driven tens of millions of miles on public roads as well as tens of billions of miles in simulation. The former have minimal observation error but may have sparse coverage of rare events; the latter may be subject to simulation artifacts but can more easily ensure coverage of rare events.

  2. b)

    Given the driving data, how can we reliably count these behavioral events? The ease of measurement for an individual class depends not only on the event’s frequency (denser events are easier to measure empirically on smaller datasets), but also our ability to algorithmically identify relevant events in the driving data. Identification of behavioral events in a purely algorithmic fashion is difficult and usually requires human review, often of varying levels of expertise ("tiered event review").

  3. c)

    How can we quantify statistical uncertainty via reliable confidence intervals for the rate estimates? IPM is not a typical metric of interest within the technological industry outside of AVs, nor is it a quantity that receives broad attention from the academic literature. Moreover, the uncertainty introduced from imperfect event detection in (b) further complicates the statistical theory. Finally, these confidence intervals must accommodate a broad range of events, including rare event scenarios, for which traditional asymptotic methods can fail easily.

These statistical challenges are further constrained by the product requirements referred to in component (3):

  • In order to drive efficient decision-making at Waymo, these rate estimates must be available as soon as possible after mileage collection. As such, we need the ability to compute rate estimates even when not all candidate events have been reviewed. Early rate estimates may have greater uncertainty than the final estimates, but they should still be unbiased and provide robust confidence intervals.

  • The most practically important event classes are very sparse (< 5-10 events), which means the most challenging statistical scenarios are also the most important.

The rest of this paper outlines how we address 2(b) and 2(c) while accommodating these product constraints. In Section 2 we introduce a streaming partial tiered event review framework. Then, Sections 3, 4, 5 outline the data generation model, estimation, and confidence interval construction implied by this framework. Section 6 provides numerical studies to assess the performance of the proposed methods for both rare and common event cases. Finally, Section 7 concludes the paper with some discussion. Technical details are provided in the appendix, including a relation between Poisson, multivariate Hypergeometric and multinomial distributions, which may be of independent interest.

2. The Role of Human Review in Event Detection

2.1. Behavioral Event Detection Overview

As described above, computing the rate at which a behavioral event occurs will require us to first identify and count those events within our driving data. To facilitate this detection at scale, we rely on a set of sophisticated algorithms that can automatically identify events of interest. These classification problems are conceptually similar to many other data science applications in industry. However, unlike most other application settings, edge cases are incredibly important within the autonomous vehicles industry and cannot be swept aside or averaged out. These edge cases, for us, are frequently the most interesting and most important events.

Take, for example, safety-relevant behavioral events which occur very rarely. Errors in their classification will show up as a very small percentage of the algorithm’s aggregate performance metrics, but lack of detection for these events can result in a very serious lack of information when deciding whether to put an ADS on public roads. This risk is compounded by the fact that estimating the recall loss by these algorithms can be very challenging and time-consuming since it requires finding sparse False Negatives in a very large collection of True Negatives.

To ensure confidence in our safety estimates at Waymo, we supplement these algorithmic solutions with a final stage of human review. When tuning any algorithmic classifier, one can adjust the operation point to favor either precision or recall, with very few algorithms being able to achieve high performance in both simultaneously. Our solution allows us to tune the algorithms for very high recall, and then lean on a final layer of human reviewers to fix any gaps in precision. Combined, the algorithm and human reviewers represent a multi-stage strategy that is both scalable and provides trustworthy rate estimates.

2.2. Review of Candidate Events

Let 0\mathcal{E}_{0} refer to the corpus of candidate events identified by the algorithms. For simplicity, we assume the algorithm feeding events to the human reviewers has perfect recall, so the human reviewers only need to address precision. Then, for any candidate event in 0\mathcal{E}_{0}, we ultimately want to confirm whether the event was a True Positive or False Positive detection of the behavioral event of interest.222Note that detailed instructions and rigorous training are often needed to ensure that human reviewers provide high quality labels, particularly for the most nuanced behavioral events. For the purposes of this paper, we assume the human labels are perfect and, as such, recall from this system is perfect.

A well-optimized event review configuration should maintain both high precision and high recall for True Positive events. If the precision is low, then the rate estimates will be too high; if the recall is low, then the rate estimates will be too low. With this as our motivation, we have developed an event review configuration at Waymo referred to as streaming partial tiered event review. The core concepts motivating this configuration are outlined below.

Under a complete event review configuration, all candidate events would be reviewed. However, for many applications this can be inefficient and labor intensive. Instead, reviewing a subset of the candidate events, i.e., partial event review, is often sufficient to estimate event rates with acceptable levels of uncertainty. Beyond efficiency, supporting estimates based on partial event review also ensures that we can fulfill the product requirements by providing trustworthy estimates at any stage of the event review process (e.g., checking results after only 10% of events have received review), and that the uncertainty in those estimates will continually be reduced as more events are reviewed.333In addition to ongoing event review, the on-road ADS logs and simulated autonomous driving provide a continuous stream of candidate events that are being fed into the event review pipeline. As a result, the corpus of candidate events is constantly changing. For simplicity, this is not addressed here, and we instead assume the full corpus of candidate events is available and unchanging prior to the event review initiation.

In its simplest form, partial event review utilizes uniform random sampling to select a subset of candidate events for labeling, but this is unlikely to be the most efficient strategy for most use-cases. To reduce sampling error, stratified random sampling can instead be used; this involves partitioning candidate events into mutually exclusive, cumulatively exhaustive strata based on pertinent event metadata (e.g. road characteristics, algorithmic model scores). The sampling probability may be identical across strata (i.e. proportional allocation), or may differ across strata (e.g. Neyman allocation or alternative allocation schemes). In some sections of this paper we simplify to assume a single stratum, without loss of generality, while in other sections we outline the statistical implications of this stratification in the data.

For operational efficiency, an event review configuration may utilize several tiers of human reviewers with increasing levels of expertise in the higher tiers. This tiered configuration is arranged such that each tier improves the precision of the candidate event corpus while maintaining recall. For example, in a three tiered process, the first tier would exclude clear false positive candidate events while escalating all other events to the second tier. The second tier would exclude additional false positive candidate events, while escalating the remainder to the third tier. The third tier would then make the final determination of false positive vs. true positive status. Each tier may itself consist of evaluation by a single reviewer or multiple reviewers who reach consensus.

Operationally, partial tiered event review can be implemented as either a batched or streaming process. Under a batched process, any candidate event that is reviewed in Tier 1 will complete its journey through the system and be confirmed as either a False Positive or make it to the final tier and be confirmed as a True Positive. In this setting, the rate estimates do not need to account for the details of how candidate events are escalated between tiers; the rate estimates are ultimately identical between a partial tiered setting and a partial untiered setting. However, at Waymo it is often the case that we need to produce unbiased rate estimates while many events are still only partway through the event review pipeline. E.g., a candidate event may have already been reviewed by Tier 1 and escalated, but hasn’t received further review yet. This is considered to be a streaming process where rate estimates are made based on the data available at any snapshot in time.

3. Statistical Formulation for Partial Tiered Event Review

Given a collection of driving data representing mm autonomous driving miles, we are interested in providing an estimate and confidence interval for the rate of occurrence (θ\theta) for the behavioral event of interest. Let TT be the number of tiers in the event review configuration and xTx_{T} be the total number of True Positive behavioral events in the driving data. We assume that xTx_{T} follows a Poisson distribution with mean θm\theta m. Note that xTx_{T} is not directly observed; instead, a corpus of candidate events, 0\mathcal{E}_{0} are sent through the streaming tiered partial event review process described in Section 2.2, resulting in a collection of partial event review results for each tier.444For simplicity, assume that the algorithmic approach to identify the corpus of candidate events has perfect recall. With this assumption, we don’t need to consider its role in the following sections.

We first discuss the statistical framing for a single stratum, and then extend this to accommodate multiple strata in a straightforward manner.

3.1. One stratum

Let e0=0e_{0}=||\mathcal{E}_{0}|| be the number of candidate events in the full corpus.

3.1.1. Partial Event Review

Due to the partial event review configuration, only a subset of the e0e_{0} candidate events will be randomly sampled and receive review in Tier 1; denoted by n1e0n_{1}\leq e_{0}. The reviewers will improve the precision of the event pool by clearly labeling the most obvious False Positives (removing them from further review) and escalating any ambiguous events to Tier 2. Let e1n1e_{1}\leq n_{1} be the number of events that Tier 1 escalates for review by Tier 2. This process proceeds at each tier with ntet1n_{t}\leq e_{t-1} denoting the number of events reviewed by Tier tt, and etnte_{t}\leq n_{t} denoting the number of events escalated by Tier tt up to Tier t+1t+1. Finally, eTe_{T} will be the number of candidate events formally confirmed as True Positives by Tier TT.

Refer to caption
Figure 3.1. Illustration of the sampling and escalation of events in the partial event review process from Tier 1 to Tier 3 (assuming T=3T=3 for a single stratum), where x3x_{3} is a latent variable representing the number of True Positive events and xtx_{t} for t{0,1,2}t\in\{0,1,2\} represents the number of candidate events which if reviewed would be rejected as False Positive by Tier-(t+1)(t+1) (formally defined in Algorithm 1).

Figure 3.1 provides an illustration of the sampling and escalation process in a single stratum under a partial event review configuration with 3 Tiers, and the detailed data generation process is described next.

3.1.2. Data Generation Model

Recall that our goal is to estimate the True Positive rate, θ\theta, for the behavioral event of interest. Under complete event review, the total confirmed True Positive count eTe_{T} would be equal to the total True Positive count in 0\mathcal{E}_{0}, and the estimate of θ\theta would be a simple function of eTe_{T}. However, because our event review is partial, we know that eTxTe_{T}\leq x_{T}; i.e., there are additional True Positives within the event review pipeline that have not yet been reviewed by the top tier. We have developed an estimation scheme that includes the contribution from these as-of-yet unreviewed True Positives. Towards this end, we propose a data generation model for the full set of candidate events in the event review pipeline, including both True Positives and False Positives.

To introduce the data generation model, let’s consider the potential outcome of each candidate event in 0\mathcal{E}_{0} under a complete event review configuration. Under the complete event review configuration, a candidate event would first be reviewed by Tier-1, then it would be either rejected as False Positive by Tier-1, or escalated by Tier-1; were the candidate escalated by Tier-1, it would be reviewed by Tier-2, and then it would be either rejected by Tier-2 as False Positive, or escalated by Tier-2, and so on. That is, the event review process continues until the candidate event would be either rejected as False Positive by Tier-tt (labeled as FP-tt) for some t{1,,T}t\in\{1,\cdots,T\} or would not be rejected by any Tier (labeled as TP). Whenever a candidate event is rejected as False Positive, it is removed from the event review pipeline for further consideration.

Let xTx_{T} be the number of candidate events in 0\mathcal{E}_{0} which would be labeled as TP, and let xt1x_{t-1} be the number of candidate events in 0\mathcal{E}_{0} which would be labeled as FP-tt for t{1,,T}t\in\{1,\cdots,T\}, under the complete event review configuration. It is natural to assume that the xtx_{t} are independent counts and that they each follow some Poisson process, i.e. for mm autonomous miles,

(3.1) xtPoisson(λtm) for t=0,T.x_{t}\sim Poisson(\lambda_{t}m)\text{ for }t=0,\cdots T.

Note that s=tTxs\sum_{s=t}^{T}x_{s} has a simple interpretation: it is the total number of candidate events in 0\mathcal{E}_{0} which if fully reviewed would not be rejected by Tier-tt. Again, in the complete event review case these xsx_{s} would be observed, but in the partial event review configuration these are instead treated as latent variables; only their total is observed, i.e.

e0=t=0Txt.e_{0}=\sum_{t=0}^{T}x_{t}.

Now given the set of candidate events 0\mathcal{E}_{0}, we draw a random sample of size n1n_{1} without replacement for Tier-1 review, among which the subset of candidate events with labels other than FP-11 — that is, not rejected by Tier-1 — will be denoted as 1\mathcal{E}_{1} and will be escalated for Tier-2 review. The process continues as follows: for t=2,,Tt=2,\cdots,T, we draw a random sample of size ntn_{t} from t1\mathcal{E}_{t-1} without replacement for Tier-tt review, among which the subset of candidate events with labels other than FP-tt — that is, not rejected by Tier-tt — will be denoted as t\mathcal{E}_{t} and will be escalated for Tier-(t+1)(t+1) review. Let ete_{t} be the number of candidate events escalated by Tier-tt:

et=t.e_{t}=||\mathcal{E}_{t}||.

Obviously, ntetn_{t}-e_{t} is the number of candidate events which are reviewed but rejected as False Positives by Tier-tt. It is also important to note that any candidate event in t\mathcal{E}_{t} will be either TP or FP-ss for some s{t+1,,T}s\in\{t+1,\cdots,T\}.

Algorithm 1 Data generation model for partial event review with single stratum

Configuration: mm miles and TT tiers.

Parameters: {λt:0tT}\{\lambda_{t}:0\leq t\leq T\} and {πt:1tT}\{\pi_{t}:1\leq t\leq T\}.

  1. (1)

    Draw

    xt\displaystyle x_{t} Poisson(mλt),t{0,,T},\displaystyle\sim Poisson(m\lambda_{t}),\forall t\in\{0,\cdots,T\},

    where xTx_{T} is the number of (latent) True Positive events (labeled as TP), and for t=1,,Tt=1,\cdots,T,

    • xt1x_{t-1} is the number of (latent) candidate events which are labeled as FP-tt — that is, they would be escalated by Tier-(t1)(t-1) but rejected as False Positive by Tier-tt.

    Then the total number of candidate events in the event review pipeline is

    e0\displaystyle e_{0} =txt.\displaystyle=\sum_{t}x_{t}.

    Let 0\mathcal{E}_{0} consist of all e0e_{0} candidate events.

  2. (2)

    For tier t=1,,Tt=1,\cdots,T,

    • If et1>0e_{t-1}>0:

      • Let ntn_{t} be the number of candidate events which are drawn randomly from t1\mathcal{E}_{t-1} without replacement and are reviewed by Tier-tt:

        nt\displaystyle n_{t} max(1,Binomial(et1,πt)).\displaystyle\sim\max(1,Binomial(e_{t-1},\pi_{t})).
      • Let t\mathcal{E}_{t} be the subset of the ntn_{t} candidate events with labels other than FP-tt — that is, they are escalated by Tier-tt.

      • Set et=te_{t}=||\mathcal{E}_{t}||.

    • Else:

      • Set ns=0n_{s}=0 and es=0e_{s}=0 for s=t,,Ts=t,\cdots,T.

      • Break (i.e. early termination).

  3. (3)

    Output: {et:0tT}\{e_{t}:0\leq t\leq T\} and {nt:1tT}\{n_{t}:1\leq t\leq T\} as the observed data.

Of course, if t\mathcal{E}_{t} is empty for some t<Tt<T, i.e. et=0e_{t}=0 (no candidate events are escalated by Tier-tt), then the partial event review process terminates at Tier-tt.

It is convenient to model the sample size ntn_{t} by a Binomial process. Note that if et1>0e_{t-1}>0 but nt=0n_{t}=0, i.e. there are et1e_{t-1} candidate events escalated for Tier-tt event review, but none is reviewed by Tier-tt, then θ\theta is not identifiable. To make the estimation problem well defined, whenever et1>0e_{t-1}>0, we need nt1n_{t}\geq 1, and to be precise, we use

(3.2) ntmax(1,Binomial(et1,πt)),\displaystyle n_{t}\sim\max(1,Binomial(e_{t-1},\pi_{t})),

where Binomial(et1,πt)Binomial(e_{t-1},\pi_{t}) stands for the random variable from a Binomial distribution with parameters et1e_{t-1} and πt(0,1]\pi_{t}\in(0,1].

This data generation model is summarized as Algorithm 1, which takes {λt:0tT}\{\lambda_{t}:0\leq t\leq T\} (associated with the latent variables xtx_{t}) and {πt:1tT}\{\pi_{t}:1\leq t\leq T\} (associated with tier level sampling) as the input parameters, and outputs the tier level sample sizes (ntn_{t}) and the numbers of tier level escalations (ete_{t}) as the observed data.

3.2. Multi-strata

Now suppose we sample from the candidate events via a stratified sampling scheme with HH comprehensive and mutually exclusive strata. We can generalize all the notations in Section 3.1 in a straightforward way by an additional subscript h{1H}h\in\{1\cdots H\} indicating the stratum. The detailed data generation model is provided as Algorithm 2, which calls Algorithm 1 for each stratum hh by taking {λht:0tT}\{\lambda_{ht}:0\leq t\leq T\} (associated with latent variables) and {πht:1tT}\{\pi_{ht}:1\leq t\leq T\} (associated with tier level sampling) as the input parameters, and outputing the tier level sample sizes (nhtn_{ht}) and the numbers of tier level escalations (ehte_{ht}) as the observed data. The parameter of interest can be written as

(3.3) θ=h=1TλhT.\displaystyle\theta=\sum_{h=1}^{T}\lambda_{hT}.
Algorithm 2 Data generation model for partial event review with multiple strata

Configuration: mm miles, HH strata, and TT tiers.

Parameters: {λht:0tT}\{\lambda_{ht}:0\leq t\leq T\} and {πht:1tT}\{\pi_{ht}:1\leq t\leq T\} for h=1,,Hh=1,\ldots,H.

  • For each h{1,,H}h\in\{1,\cdots,H\},

    • use {λht:0tT}\{\lambda_{ht}:0\leq t\leq T\} and {πht:1tT}\{\pi_{ht}:1\leq t\leq T\} as the input parameters and apply Algorithm 1 to generate the output for stratum-hh: {eht:0tT}\{e_{ht}:0\leq t\leq T\} and {nht:1tT}\{n_{ht}:1\leq t\leq T\}.

  • Output: {eht}\{e_{ht}\} and {nht}\{n_{ht}\} as the observed data for all strata.

4. Point estimation

To estimate θ\theta, it is desirable to derive the maximum likelihood estimate (MLE) due to its various optimality properties ([1, 21, 9]) with the hope that it is also unbiased. Instead of deriving the MLE directly, we first derive the point estimate based on a heuristic argument, then show that it is indeed the MLE. This is due to the complexity of the likelihood function under the partial event review configuration, with more details provided in the Appendix. The estimate is also unbiased.

Since the data across strata are independent, the rates associated with different strata can be estimated independently. Without loss of generality, we may consider a single stratum, using the same notation as in Section 3.1.

A natural estimate of mλTm\lambda_{T} is the total number of true positives within e0e_{0} (all candidate events in the stratum). This quantity is unfortunately not observable under partial event review and must be estimated instead. We start the argument by assuming eT>0e_{T}>0.

First, consider the total number of True Positive events present in the final tier, Tier TT. Recall that e(T1)e_{(T-1)} candidate events will be escalated to tier TT, a random sample of size nTn_{T} will be reviewed, and eTe_{T} of those reviewed events will be confirmed as True Positives. Then, one can estimate the total expected number of true positives that were present in Tier T1T-1 as below by up-weighting the observed true positives by the inverse of the sampling weight:

eT(nTe(T1))1.\displaystyle e_{T}\left(\frac{n_{T}}{e_{(T-1)}}\right)^{-1}.

By the same logic, one can estimate the expected number of true positives that were present in Tier T2T-2 to be

eT(nTe(T1))1(n(T1)e(T2))1\displaystyle e_{T}\left(\frac{n_{T}}{e_{(T-1)}}\right)^{-1}\left(\frac{n_{(T-1)}}{e_{(T-2)}}\right)^{-1} .

Iterating through each of the tiers leads to the estimate for the total number of true positives in all e0e_{0} candidate events in the stratum:

eT(nTe(T1))1(n(T1)e(T2))1(n1e0)1\displaystyle e_{T}\left(\frac{n_{T}}{e_{(T-1)}}\right)^{-1}\left(\frac{n_{(T-1)}}{e_{(T-2)}}\right)^{-1}\cdots\left(\frac{n_{1}}{e_{0}}\right)^{-1} .

Normalizing the above by mm after simplification leads to the estimate of λT\lambda_{T}:

(4.1) λ^T\displaystyle\hat{\lambda}_{T} =1mt=0Tett=1Tnt.\displaystyle=\frac{1}{m}\frac{\prod_{t=0}^{T}e_{t}}{\prod_{t=1}^{T}n_{t}}.

Note, an edge case occurs when eT=0e_{T}=0, or even et=0e_{t}=0 for some t<Tt<T; that is, when no events are escalated to Tier (t+1)(t+1) for some tt. In these cases the partial event review process would terminate early, and the natural estimate is λ^T=0\hat{\lambda}_{T}=0.

For t{0,,T}t\in\{0,\cdots,T\}, let

Λt=s=tTλs.\Lambda_{t}=\sum_{s=t}^{T}\lambda_{s}.

That is, Λt\Lambda_{t} is the overall Poisson rate for candidate events in 0\mathcal{E}_{0} which if fully reviewed would not be rejected by Tier-tt for t{1,,T}t\in\{1,\cdots,T\}. Note Λ0ΛT\Lambda_{0}\geq\cdots\geq\Lambda_{T}.

Similar to the heuristic derivation of (4.1), one can estimate Λt\Lambda_{t} as follows:

Λ^0=e0m\hat{\Lambda}_{0}=\frac{e_{0}}{m}

and for t1t\geq 1,

(4.2) Λ^t=1ms=0tess=1tns.\displaystyle\hat{\Lambda}_{t}=\frac{1}{m}\frac{\prod_{s=0}^{t}e_{s}}{\prod_{s=1}^{t}n_{s}}.

Note that Λ^t\hat{\Lambda}_{t} is well defined whenever et11e_{t-1}\geq 1, which is guaranteed by the sampling model (3.2). If Λ^t=0\hat{\Lambda}_{t}=0, which may occur due to et=0e_{t}=0 (i.e. early termination if t<Tt<T, see Algorithm 1), then Λ^s=0\hat{\Lambda}_{s}=0 for sts\geq t due to monotonicity. Therefore, the estimates are always well-defined. Note that Λ^Tλ^T\hat{\Lambda}_{T}\equiv\hat{\lambda}_{T}. In fact, the estimates are not only MLE but unbiased.

Theorem 1.

Under the Poisson model (3.1) and the sampling model (3.2),

  1. (1)

    {Λ^t:0tT}\{\hat{\Lambda}_{t}:0\leq t\leq T\} as defined above is the MLE of {Λt:0tT}\{\Lambda_{t}:0\leq t\leq T\}.

  2. (2)

    The point estimates are unbiased, i.e. E(Λ^t)=ΛtE(\hat{\Lambda}_{t})=\Lambda_{t}, t{0,,T}\forall t\in\{0,\cdots,T\}.

  3. (3)

    The MLE for {λt}\{\lambda_{t}\} for t=0,,T1t=0,\cdots,T-1 is given by

    (4.3) λ^t\displaystyle\hat{\lambda}_{t} =Λ^tΛ^(t+1).\displaystyle=\hat{\Lambda}_{t}-\hat{\Lambda}_{(t+1)}.

The derivation of the MLE for {Λt}\{\Lambda_{t}\} (and thus {λt}\{\lambda_{t}\}) relies on a novel application of the Expectation-Maximization algorithm [5]. The complete proof is provided in the Appendix.

For multi-strata, since the data across strata are assumed to follow independent random processes, we can apply the MLE estimator described in Theorem 1 to each stratum hh and obtain the estimate λ^ht\hat{\lambda}_{ht} for t=0,,Tt=0,\cdots,T. Then the MLE of θ\theta (3.3) is given by

(4.4) θ^\displaystyle\hat{\theta} =hλ^hT.\displaystyle=\sum_{h}\hat{\lambda}_{hT}.

which is unbiased due to additivity and Theorem 1.

5. Confidence Interval

As described in Algorithm 2, our model involves a single parameter of interest θ\theta but a large set of nuisance parameters λht\lambda_{ht} and πht\pi_{ht}. Due to the complexity of the model, it is hard to derive an “exact” confidence interval for θ\theta. As such, we consider several possible strategies for computing confidence intervals (CIs), then assess the relative metris of each strategy.

We first describe two standard approaches based on a parametric bootstrap and a normal approximation (Wald CI) respectively. These methods are expected to work well asymptotically as the expected True Positive event count increases, i.e., as the mileage mm goes to infinity. However, the CIs may be less reliable when the true positive events are rare. Due to the product importance of these rare event settings, we propose a novel approach by approximating the exact model with a simpler model – weighted sum of independent Poissons (WSIP) – and then applying the Gamma method ([7]), one of the most established methods for WSIP. The Gamma method is expected to work well not only for scenarios with high event counts (common events), but also for scenarios with relatively few events (rare events).

5.1. Parametric Bootstrap CI

With the mileage mm fixed, conceptually, the model for the data generation can be written as p(𝒟|{m,H,T,{λht,πht}})p(\mathcal{D}|\{m,H,T,\{\lambda_{ht},\pi_{ht}\}\}), where 𝒟\mathcal{D} consists of {eht:1hH,0tT}\{e_{ht}:1\leq h\leq H,0\leq t\leq T\} and {nht:1hH,1tT}\{n_{ht}:1\leq h\leq H,1\leq t\leq T\}. The detailed data generation procedure is provided by Algorithm 2. This is a parametric model, and the parametric bootstrap CI [6] can be constructed as follows:

Let {λ^ht,π^ht}\{\hat{\lambda}_{ht},\hat{\pi}_{ht}\} be the point estimates of {λht,πht}\{\lambda_{ht},\pi_{ht}\} respectively, where λ^ht\hat{\lambda}_{ht}s are obtained by applying the MLE estimator described in Theorem 1 to each stratum hh separately, and the following empirical estimate is used for πht\pi_{ht} according to (3.2):

(5.1) π^ht\displaystyle\hat{\pi}_{ht} =nhteh(t1)\displaystyle=\frac{n_{ht}}{e_{h(t-1)}}

which is also MLE when nht>1n_{ht}>1. It may be worth noting that when nht=1n_{ht}=1 and eh(t1)>1e_{h(t-1)}>1, the MLE for πht\pi_{ht} by the sampling model (3.2) is simply 0, which is close to (5.1) for large eh(t1)e_{h(t-1)} but would be less desirable in general.

Then draw BB i.i.d. samples of 𝒟\mathcal{D} according to the generative process outlined in Algorithm 2 by plugging in the parameter estimates from the observed data. For b=1,,Bb=1,\cdots,B,

𝒟(b)\displaystyle\mathcal{D}^{(b)} p(|{m,H,T,{λ^ht,π^ht}}).\displaystyle\sim p(\cdot|\{m,H,T,\{\hat{\lambda}_{ht},\hat{\pi}_{ht}\}\}).

For each draw, calculate the parameter estimates {λ^ht(b),π^ht(b)}\{\hat{\lambda}_{ht}^{(b)},\hat{\pi}_{ht}^{(b)}\} by replacing {eht,nht}\{e_{ht},n_{ht}\} in (4.3) and (5.1) with corresponding quantities in 𝒟(b)\mathcal{D}^{(b)}, and let

θ^(b)\displaystyle\hat{\theta}^{(b)} =h=1Hλ^hT(b).\displaystyle=\sum_{h=1}^{H}\hat{\lambda}_{hT}^{(b)}.

The 1α1-\alpha confidence interval for θ\theta is given by the α/2\alpha/2 and 1α/21-\alpha/2 quantiles of {θ^(b):b=1,,B}\{\hat{\theta}^{(b)}:b=1,\cdots,B\}.

5.2. Wald CI

In order to construct the Wald CI, we need to estimate var(θ^)var(\hat{\theta}). Unfortunately, the exact formula for var(θ^)var(\hat{\theta}) is extremely complicated. Instead, we derive a first order approximation by assuming mm is large.

It may be interesting to note that λ^hT\hat{\lambda}_{hT} as defined by (4.1) with index hh added to indicate the stratum can be rewritten as

λ^hT\displaystyle\hat{\lambda}_{hT} =ehTmπ^h\displaystyle=\frac{e_{hT}}{m\hat{\pi}_{h}}

where

π^h\displaystyle\hat{\pi}_{h} =t=1Tπ^ht\displaystyle=\prod_{t=1}^{T}\hat{\pi}_{ht}

and π^ht\hat{\pi}_{ht}s are defined by (5.1). Thus θ^\hat{\theta} in (4.4) can be rewritten as

(5.2) θ^\displaystyle\hat{\theta} =1mh=1HehTπ^h.\displaystyle=\frac{1}{m}\sum_{h=1}^{H}\frac{e_{hT}}{\hat{\pi}_{h}}.
Lemma 2.

If a random variable XPoisson(λ)X\sim Poisson(\lambda), and YXBinomial(X,π)Y\mid X\sim Binomial(X,\pi), then YPoisson(λπ)Y\sim Poisson(\lambda\pi).

For large mm, we may approximate the model (3.2) with subscript hh added to indicate stratum as

nht\displaystyle n_{ht} Binomial(eh(t1),πht).\displaystyle\sim Binomial(e_{h(t-1)},\pi_{ht}).

Then by iteratively applying Lemma 2, we have

(5.3) ehT\displaystyle e_{hT} Poisson(mλhTπh),\displaystyle\sim Poisson(m\lambda_{hT}\pi_{h}),

where πh=t=1Tπht\pi_{h}=\prod_{t=1}^{T}\pi_{ht}. By combining this with (5.2), we can obtain the following result.

Theorem 3.

Assume πht>0\pi_{ht}>0 and λht>0\lambda_{ht}>0 for all h,th,t. As mm\rightarrow\infty, we have

m(θ^θ)\displaystyle\sqrt{m}(\hat{\theta}-\theta) 𝒩(0,h=1HλhTπh1),\displaystyle\rightarrow\mathcal{N}(0,\sum_{h=1}^{H}\lambda_{hT}\pi_{h}^{-1}),

where 𝒩(μ,σ2)\mathcal{N}(\mu,\sigma^{2}) denotes the normal distribution with mean μ\mu and variance σ2\sigma^{2}.

As m,m\rightarrow\infty, π^htπht\hat{\pi}_{ht}\rightarrow\pi_{ht} and thus π^hπh\hat{\pi}_{h}\rightarrow\pi_{h}, the proof follows from Slutsky’s theorem.

From Theorem 3, the Wald CI can then be constructed by θ^±zα/21mh=1Hλ^hTπ^h1\hat{\theta}\pm z_{\alpha/2}\sqrt{\frac{1}{m}\sum_{h=1}^{H}\hat{\lambda}_{hT}\hat{\pi}_{h}^{-1}} based on the plug-in principle, where zα/2z_{\alpha/2} is the 1α/21-\alpha/2 quantile of 𝒩(0,1)\mathcal{N}(0,1).

As is well known, proper coverage from the Wald CI requires the asymptotic normality to hold well. As such, it may not be desirable when the expected event counts are low (i.e., for rare events or for low mileage); for example, its lower bound may become negative.

5.3. Gamma CI based on an approximate weighted sum of independent Poissons

Given the limitations of the normal asymptotic approximation outlined above, we propose an alternative approach to estimating the CIs for θ^\hat{\theta}. Consider the estimate in (5.2), which can be approximated by the quantity below:

θ^\displaystyle\hat{\theta} 1mh=1HehTπh,\displaystyle\approx\frac{1}{m}\sum_{h=1}^{H}\frac{e_{hT}}{\pi_{h}},

where by (5.3) the right hand side is a weighted sum of independent Poissons (WSIP). The approximate WSIP becomes clear by rewriting θ^\hat{\theta} as,

θ^\displaystyle\hat{\theta} =h=1HwhehT\displaystyle=\sum_{h=1}^{H}w_{h}e_{hT}
wh\displaystyle w_{h} =1mπ^h for h=1,,H.\displaystyle=\frac{1}{m\hat{\pi}_{h}}\text{ for }h=1,\cdots,H.

The distribution for WSIP random variables has been studied extensively in the statistics literature [10, 16, 17, 8]. A WSIP is close to normally distributed when there are high event counts, but can be far from normal for rare events. This again motivates the consideration of a CI methodology that does not rely on normality. In particular, the original Gamma method proposed by [7] has been conjectured and also shown numerically to provide good coverage properties for the distribution mean in all simulated scenarios [10, 8].

Our third CI methodology leverages these WSIP properties from the literature and can be described as follows:

  • The lower bound is the α/2\alpha/2 quantile of the Gamma distribution with mean θ^\hat{\theta} and variance h=1Hwh2ehT\sum_{h=1}^{H}w_{h}^{2}e_{hT}.

  • The upper bound is the 1α/21-\alpha/2 quantile of the Gamma distribution with mean θ^+wM\hat{\theta}+w_{M} and variance h=1Hwh2ehT+wM2\sum_{h=1}^{H}w_{h}^{2}e_{hT}+w_{M}^{2}, where wM=max1hHwhw_{M}=\max_{1\leq h\leq H}w_{h}.

Note that the above is based on the original Gamma method, but one may also apply other Gamma methods for WSIP such as the mid-p Gamma interval [8].

6. Numerical Studies

6.1. Simulation Setup

We provide several simulation studies to assess and compare the performance of the three confidence interval methods defined in Section 5. The first two simulation studies compare the performance when the behavioral event is rare vs. when the behavioral event is common, using fixed values for most parameters. The third study illustrates that these coverage properties hold for a wide range of parameter values, beyond the specific values chosen for the first two examples. Throughout the simulation section we set m=1m=1 for convenience (the unit could be 1 mile, 10K miles, or one million miles).

In each study, we simulate the data in a consistent manner. Following Algorithm 2, we generate the xhsPoisson(λhs),h=1,,Hx_{hs}\sim Poisson(\lambda_{hs}),h=1,\ldots,H and s=0,,Ts=0,\dots,T. We set H=5H=5 strata and assume T=3T=3 tiers. Then eh0=s=0Txhse_{h0}=\sum_{s=0}^{T}x_{hs} will be the total number of candidate events entering the tiered event review pipeline in stratum hh.

For the first two simulation studies in Section 6.2 and 6.3, we set the sampling rates

{πht}=(π10.50.95π10.60.96π10.70.97π10.80.98π10.90.99).\{\pi_{ht}\}=\begin{pmatrix}\pi_{1}&0.5&0.95\\ \pi_{1}&0.6&0.96\\ \pi_{1}&0.7&0.97\\ \pi_{1}&0.8&0.98\\ \pi_{1}&0.9&0.99\end{pmatrix}.

Note that when this event review methodology is put into practice one can imagine a fixed set of events to review, and the Tier 1 sampling rate will steadily increase over time as reviewers complete more events. We simulate this event review progression by varying the sampling rate at Tier 1, π1[0.1,1.0]\pi_{1}\in[0.1,1.0]. In this way, we can verify whether the confidence intervals have good coverage properties throughout the entire event review process.

For the third simulation study in Section 6.4, we consider a wide range of scenarios. For each scenario, we generate each Tier 1 sampling rate πh1Uniform (0,1)\pi_{h1}\sim\text{Uniform }(0,1), and generate πhtUniform (πh(t1),1),t=2,3,\pi_{ht}\sim\text{Uniform }(\pi_{h(t-1)},1),t=2,3, so that the sampling rate for each strata will not decrease as Tier goes up. {λht}\{\lambda_{ht}\} is specified later.

Given each set of parameter values ({λht}\{\lambda_{ht}\} and {πht}\{\pi_{ht}\}), we simulate the tiered partial event review data {eht}\{e_{ht}\} and {nht}\{n_{ht}\} following Algorithm 2 and replicate 1000 times. This allows us to obtain the empirical coverage and average CI width under each set of parameters.

Refer to caption
Figure 6.1. Common Events. Average one-sided coverage error of 90%90\% confidence intervals for lower bound (top left panel), upper bound (top right panel), total coverage error (bottom left panel) and average lengths of 90%90\% confidence intervals (bottom right panel). These results are based on estimating the total positive event rate θ=58\theta=58 under model (3.1) over 1000 replications. The horizontal lines in the first three panels represent the test level α=0.05\alpha=0.05.

6.2. Simulation for Common Events

Some behavioral events are relatively common, and the observed event counts will be relatively high. As described earlier, this can result in better behaved quantities that more closely follow a normal distribution. Here we assess the performance of our proposed CI methodologies in this setting.

To represent this common event scenario, the values below are used for {λht}\{\lambda_{ht}\}:

{λht}=(1052.518201525102030855625103012415).\{\lambda_{ht}\}=\begin{pmatrix}10&5&2.5&18\\ 20&15&25&10\\ 20&30&8&5\\ 5&6&25&10\\ 30&12&4&15\end{pmatrix}.

With this parameter setting, θ=58\theta=58.

Figure 6.1 illustrates the coverage properties for the three CI methods when the behavioral event is common. The x-axis represents the sampling rate, allowing us to assess how well these CIs fulfill the product requirement of adequate coverage regardless of the sampling rate. For all three methods, the coverage rates are more biased (over- or under-covering) when sampling rates are low, with improvements in coverage as event review progresses. As sampling rates increase, the parametric bootstrap rapidly converges close to expected coverage rates, and also produces the narrowest intervals. However, the under-coverage at low sampling rates is non-negligible (82% coverage vs. the desired 90%). In contrast, the Gamma CI consistently over-covers (98% coverage for low sampling rates) and produces the widest intervals. However, the Gamma method also provides the most symmetric errors, with similar error rates for the upper and lower bounds; this guarantees that the intervals from this approach will not systematically under- or over-estimate the rates. Finally, the Wald CI falls between the other two methods, with intervals that somewhat over-cover.

Refer to caption
Figure 6.2. Rare Events. Average one-sided coverage error of 90%90\% confidence intervals for lower bound (top left panel), upper bound (top right panel), total coverage error (bottom left panel) and average lengths of 90%90\% confidence intervals (bottom right panel). These results are based on estimating the total positive event rate θ=11\theta=11 under model (3.1) over 1000 replications. The horizontal lines in the first three panels represent the test level α=0.05\alpha=0.05.

6.3. Simulation for Rare Events

Next, we consider more rare events, where the observed event counts will be quite low. In this setting, the estimates are not going to cleanly follow a normal distribution, making the CI calculation potentially more challenging. Here we assess the performance of our proposed CI methodologies in this setting, again tracking how the coverage properties evolve as more events receive review.

To represent this scenario, we use the same λht\lambda_{ht} as before for t=0,1,2t=0,1,2 but reduce λh3\lambda_{h3} for all hh:

{λht}=(1052.54201525220308156252301242).\{\lambda_{ht}\}=\begin{pmatrix}10&5&2.5&4\\ 20&15&25&2\\ 20&30&8&1\\ 5&6&25&2\\ 30&12&4&2\end{pmatrix}.

With this parameter setting, θ=11\theta=11.

Figure 6.2 reports the coverage properties for the three confidence interval methods when the behavioral event is rare. The patterns in this figure are largely similar to the patterns in Figure 6.1. The primary difference between the rare event case and the common event case is the performance at low sampling rates. The under-coverage by the Wald and Parametric Bootstrap is severe at the lowest sampling rates, off by as much as 30 percentage points (60% coverage vs. the desired 90%). In this case, the over-coverage produced by the Gamma CI (97% vs. the desired 90%) is clearly preferable. That said, the Gamma CI is much wider than the Parametric Bootstrap, as much as 3x wider for low sampling rates.

Refer to caption
Figure 6.3. Average error rates and CI widths of 90%\% confidence intervals by different methods versus the the expected number of observed True Positives from m=1m=1 mileage over 10510^{5} scenarios. The lines are the minimum (black), 25%\% percentile (gray), median (black), 75%\% percentile (gray), and maximum (black) calculated from moving subsets of the data, where the subset for the value at expected number of observed True Positives includes all observations with var(weights) \in [expected number - 1, expected number + 1]. Reference lines for the error rate panels are drawn at the nominal error rates for a central 90%\% confidence interval.

6.4. Comprehensive simulation

In each of the simulations above, the majority of parameters were fixed, while the first tier sampling rate was allowed to vary. In our final simulation we continue to vary the sampling rate, as described earlier, but consider 10510^{5} different scenarios covering a wide range of parameter values λht\lambda_{ht}. In each scenario, we simulate the underlying Poisson rates as follows:

λhtExp(μht),h=1,,H and t=0,T,\lambda_{ht}\sim\text{Exp}(\mu_{ht}),h=1,\cdots,H\text{ and }t=0,\cdots T,

where μhtUniform(1,4)\mu_{ht}\sim\text{Uniform}(1,4) independently. This simulation study is inspired by the simulation setup in [8].

Figure 6.3 reports the performance comparison. Unlike Figures 6.1 and 6.2 where the x-axis represented a range of sampling rates, the x-axis here corresponds to the expected observed number of true positive events per mile, i.e. 1mhE(ehT)\frac{1}{m}\sum_{h}E(e_{hT}) which equals hλhTt=1Tπht\sum_{h}\lambda_{hT}\prod_{t=1}^{T}\pi_{ht}. In the earlier examples, the underlying event rates were fixed and only the sampling rates varied; in this study, both the sampling rates and the event rates are varying simultaneously. In this setup, this x-axis allows for consideration of how the coverage properties evolve in the asymptotic sense, regardless of whether it is increased review or increased event occurrence.

The pattern is overall similar to the previous examples, for example, for the lower bound, both Gamma method and Wald method maintain the right coverage. However, the Wald method can be badly under-covered for its upper bound, while parametric bootstrap can be badly under-covered for both lower bound and upper bound. On the other hand, the Gamma method is the only method that always produces reliable confidence intervals for all scenarios. Not surprisingly, this method also produces the widest intervals on average.

7. Summary and Discussion

Before launching any new autonomous driving software onto public roads, it is critical to first evaluate the driving quality of the software. To this end, estimation of behavioral event rates has become a statistical focus area within the autonomous vehicles industry. This estimation process often requires more sophisticated statistical consideration than one might expect at first glance.

As described earlier, accurate estimation of these event rates relies on accurate detection of the behavioral events within the driving logs. For many such events, we rely on algorithmic detectors combined with a partial tiered event review process where human reviewers reject or escalate the candidate events at each tier. Importantly, this event review process is a streaming process that rarely provides complete review of all events in the system. As a result, there are many candidate events with missing outcome labels, complicating the total event count estimation. In addition, each of these events may have already progressed to a different stage within the event review pipeline, and this partial information can be used to better estimate the candidate event’s likelihood of being a True Positive.

While the point estimate for the event rate is fairly intuitive, how to construct a reliable confidence interval for the rate is less straightforward. As defined, the data generation model has a very large number of latent variables and nuisance parameters. In addition, the practical application of this work within the autonomous vehicles industry necessitates strong performance for a range of behavioral event types ranging from benign common events to more safety-relevant rare events. As shown through the simulation studies, the different CI methods have different strengths and weaknesses, and these vary depending on the sampling rate and/or event frequency. Thus, the choice of confidence interval may depend on the characteristics of the behavioral event of interest. For a blanket approach that ensures consistent conservative coverage, we propose the Gamma method based on the weighted sum of independent Poissons.

The Parametric Bootstrap CI is consistently the narrowest CI, and it also frequently under-covers. For common behavioral events with a reasonable rate of review, this method’s under-coverage is typically not severe. However, when the sampling rate is low, e.g. under 25%, this method can drastically undercover.

The Wald CI is largely similar to the Parametric Bootstrap, except that the Wald CIs are somewhat wider and provide greater coverage. For applications where it is practical to require minimum rates of review, this method may be attractive to some users. With sufficient guardrails (i.e., only for common events and required minimum rates of review before reporting), this method provides close to the intended coverage and provides narrower intervals that can inform downstream product decision-making by stakeholders. Of course, we do not recommend use of this method unless the practitioner can guarantee the application is a setting where the method achieves desired coverage.

In many cases the above methods are not appropriate, for example when the user does not have a priori information about the relative frequency of the behavioral event, the behavioral event is known to be rare, or there is a product requirement to report rates at all stages of the event review process. In each of these settings, we recommend using the Gamma method based on the weighted sum of independent Poissons, which was shown to maintain consistent coverage across all simulations.

The Gamma method is a conservative strategy, which consistently over-covers relative to the target coverage rate and has interval widths that are consistently wider than the other methods. While this over-coverage is not strictly desirable (we’d prefer to achieve the target coverage rate precisely!), we believe that a conservative approach is the best strategy when estimating the rate of high priority safety-relevant behavioral events. In addition, the upper and lower error rates are generally quite similar, giving confidence that these intervals will not systematically over- or under-estimate the rate of occurrence for undesirable behavioral events. These properties are particularly important given the very-real implications of launching ADS on public roads.

In summary, we have outlined the benefits of the streaming partial tiered event review, but we have also illustrated that estimating confidence intervals for behavioral event rates under this scheme is quite challenging. While the Gamma method had the most consistent coverage, the utility of the resulting confidence intervals is limited by their excessive width and over-coverage in many settings. These findings are similar to the properties observed in the original Gamma method [7]. Future work should further refine and improve these strategies while maintaining the strong properties of the existing approach.

A vast literature exists on the subject of sampling [4], but application of the complex mixed sampling technique as presented is new for the area of autonomous vehicles. How to construct a valid confidence interval in the presence of many nuisance parameters is still very much an open problem for rare events, and the standard approaches such as bootstrap and Wald CI may fail easily for small event counts (e.g. [19]). The most relevant work to ours include the profile likelihood-based method [18], the Buehler method [2], and the hybrid sampling method [3], which however cannot be easily applied for our problem, see [14, 11] for some recent reviews on the subject.

8. Appendix: proof of Theorem 1

We first present a theorem about the relationship between Poisson, multivariate Hypergeometric and multinomial distributions, which may be of independent interest and is used to prove Theorem 1. We then proceed to derive the MLE by making use of the EM algorithm. Results 1) and 3) of Theorem 1 follows directly from Proposition 9, and result 2) follows from Lemma 5. In the end, we also provide an illustrative example for T=2T=2 on how to solve the EM equations.

8.1. Relation between Poisson, multivariate Hypergeometric and multinomial distributions

Consider an urn of balls, where each ball has a color kk with k{0,1,,K}k\in\{0,1,\cdots,K\}. Let zkz_{k} be the number of balls with color kk, and let sz=k=0Kzks_{z}=\sum_{k=0}^{K}z_{k} and z\vec{z} be the column vector of (z0,,zK)(z_{0},\cdots,z_{K}).

Let nszn\leq s_{z} be a random number whose distribution only depends on szs_{z}.

Randomly draw n1n\geq 1 balls from the urn without replacement, and let yky_{k} be the number of balls with color kk for k{1,,K}k\in\{1,\cdots,K\}. Let sy=k=1Kyks_{y}=\sum_{k=1}^{K}y_{k} and y\vec{y} be the column vector of (y1,,yK)(y_{1},\cdots,y_{K}).

Then, conditional on (z,n)(\vec{z},n), y\vec{y} follows a multivariate Hypergeometric distribution:

p(y|z,n)\displaystyle p(\vec{y}|\vec{z},n) =(z0nsy)k=1K(zkyk)(szn).\displaystyle=\frac{{z_{0}\choose n-s_{y}}\prod_{k=1}^{K}{z_{k}\choose y_{k}}}{{s_{z}\choose n}}.
Theorem 4.

If zkPoisson(λk)z_{k}\sim Poisson(\lambda_{k}), k=0,1,,Kk=0,1,\cdots,K, then the following results hold.

1) Conditional on (y,sz,n)(\vec{y},s_{z},n), the distribution of z\vec{z} is Multinomial with shifted mean:

z|(y,sz,n)\displaystyle\vec{z}|(\vec{y},s_{z},n) [nsyy]+Multinomial(szn,[Λ1λ0Λ1λK])\displaystyle\sim\left[\begin{array}[]{c}n-s_{y}\\ \vec{y}\end{array}\right]+Multinomial(s_{z}-n,\left[\begin{array}[]{c}\Lambda^{-1}\lambda_{0}\\ \cdots\\ \Lambda^{-1}\lambda_{K}\end{array}\right])

where Λ=k=0Kλk\Lambda=\sum_{k=0}^{K}\lambda_{k}.

2) Conditional on sys_{y}, y\vec{y} is independent of (sz,n)(s_{z},n), and follows a conditional Poisson distribution:

p(y|sy)\displaystyle p(\vec{y}|s_{y}) =C(sy)×k=1Keλkλkykyk!\displaystyle=C(s_{y})\times\prod_{k=1}^{K}\frac{e^{-\lambda_{k}}\lambda_{k}^{y_{k}}}{y_{k}!}

where C(s)=(y1++yk=sk=1Keλkλkykyk!)1C(s)=\left(\sum_{y_{1}+\cdots+y_{k}=s}\prod_{k=1}^{K}\frac{e^{-\lambda_{k}}\lambda_{k}^{y_{k}}}{y_{k}!}\right){}^{-1} is the normalization factor and only depends on ss. In other words,

y|(sz,n,sy)\displaystyle\vec{y}|(s_{z},n,s_{y}) u|k=1Kuk=sy\displaystyle\sim\vec{u}|\sum_{k=1}^{K}u_{k}=s_{y}

where u=(u1,,uK)\vec{u}=(u_{1},\cdots,u_{K}) and ukPoisson(λk)u_{k}\sim Poisson(\lambda_{k}) are independent.

Proof.

Note that the joint probability can be written as

p(z,n,y)\displaystyle p(\vec{z},n,\vec{y}) =p(z)p(n|sz)p(y|z,n).\displaystyle=p(\vec{z})p(n|s_{z})p(\vec{y}|\vec{z},n).

So given (y,sz,n)(\vec{y},s_{z},n), we have

p(z|y,sz,n)\displaystyle p(\vec{z}|\vec{y},s_{z},n) p(z)p(y|z,n)\displaystyle\propto p(\vec{z})p(\vec{y}|\vec{z},n)
=k=0Keλkλkzkzk!×(z0nsy)k=1K(zkyk)(szn)\displaystyle=\prod_{k=0}^{K}\frac{e^{-\lambda_{k}}\lambda_{k}^{z_{k}}}{z_{k}!}\times\frac{{z_{0}\choose n-s_{y}}\prod_{k=1}^{K}{z_{k}\choose y_{k}}}{{s_{z}\choose n}}
λ0z0k=1Kλkzkyk(z0n+sy)!k=1K(zkyk)!\displaystyle\propto\frac{\lambda_{0}^{z_{0}}\prod_{k=1}^{K}\lambda_{k}^{z_{k}-y_{k}}}{(z_{0}-n+s_{y})!\prod_{k=1}^{K}(z_{k}-y_{k})!}
(szn(z0(nsy))(z1y1)(zKyk))p0z0(nsy)k=1Kpkzkyk\displaystyle\propto{s_{z}-n\choose\begin{array}[]{c}\begin{array}[]{cccc}(z_{0}-(n-s_{y}))&(z_{1}-y_{1})&\cdots&(z_{K}-y_{k})\end{array}\end{array}}p_{0}^{z_{0}-(n-s_{y})}\prod_{k=1}^{K}p_{k}^{z_{k}-y_{k}}

where pk=Λ1λkp_{k}=\Lambda^{-1}\lambda_{k}. This completes the proof of 1).

The proof of 2) follows from the Bayes theorem, i.e.

p(y|sz,n,sy)\displaystyle p(\vec{y}|s_{z},n,s_{y}) =p(z,y|sz,n,sy)p(z|y,sz,n)\displaystyle=\frac{p(\vec{z},\vec{y}|s_{z},n,s_{y})}{p(\vec{z}|\vec{y},s_{z},n)}
p(z)p(y|z,n)p(z|y,sz,n).\displaystyle\propto\frac{p(\vec{z})p(\vec{y}|\vec{z},n)}{p(\vec{z}|\vec{y},s_{z},n)}.

8.2. Likelihood inference for a single stratum

The data for a single stratum can be described as (e0,n1,e1,,nT,eT)(e_{0},n_{1},e_{1},\cdots,n_{T},e_{T}). Note that in case the partial event review process is terminated earlier due to no escalation, say et=0e_{t}=0 for some tier t<Tt<T, then there will be no data from higher tiers (i.e., ns=0n_{s}=0 and es=0e_{s}=0 for s>ts>t according to Algorithm 1), so we may simply reset TT to tt for the data likelihood calculation. Let e=(e0,,eT)\vec{e}=(e_{0},\cdots,e_{T}) and n=(n1,,nT)\vec{n}=(n_{1},\cdots,n_{T}). The observed data consists of e\vec{e} and n\vec{n}. Without loss of generality, we will take the mileage m=1m=1.

To follow Algorithm 1, let 0\mathcal{E}_{0} consist of all e0e_{0} candidate events. Let t\mathcal{E}_{t} be the set of ete_{t} events which are escalated by Tier-tt for t=1,,Tt=1,\cdots,T, so T\mathcal{E}_{T} consists of the eTe_{T} events which are True Positive. Furthermore, each candidate event is associated with a label, which is either FP-tt (if the event would be escalated by Tier-(t1)(t-1) but rejected as False Positive by Tier-tt) for some t{1,,T}t\in\{1,\cdots,T\}, or TP (if the event would not be rejected by any Tier), under the complete event review configuration.

For s{0,,T1}s\in\{0,\cdots,T-1\}, let xtsx_{ts} be the number of FP-tt events in s\mathcal{E}_{s}, st<T\forall s\leq t<T, and let xTsx_{Ts} be the number of TP events in s\mathcal{E}_{s}, then

es=t=sTxts.e_{s}=\sum_{t=s}^{T}x_{ts}.

For convenience, we also introduce xTTeTx_{TT}\equiv e_{T}. With some abuse of notation, (x00,,xT0)(x_{00},\cdots,x_{T0}) is identical to (x0,,xT)(x_{0},\cdots,x_{T}) as defined by (3.1).

Let xs=(xss,,xTs)\vec{x}_{s}=(x_{ss},\cdots,x_{Ts}), for s=0,,Ts=0,\cdots,T. Table 1 summarizes the notation of {xts:0stT}\{x_{ts}:0\leq s\leq t\leq T\} and relation with e\vec{e}.

Table 1. Notation for {xts:0stT}\{x_{ts}:0\leq s\leq t\leq T\}, where xtsx_{ts} is the number of FP-tt events escalated by Tier-ss for 1st<T1\leq s\leq t<T, xTsx_{Ts} is the number of TP events escalated by Tier-ss, and {xt0}\{x_{t0}\} is identical to {xt}\{x_{t}\} as defined in Algorithm 1.
Tier-0 Tier-1 \cdots Tier-(T1)(T-1) Tier-TT
FP-1 x00x_{00} 0 0 0 0
FP-2 x10x_{10} x11x_{11} 0 0 0
\cdots \cdots \cdots \cdots 0 0
FP-TT x(T1)0x_{(T-1)0} x(T1)1x_{(T-1)1} \cdots x(T1)(T1)x_{(T-1)(T-1)} 0
TP xT0x_{T0} xT1x_{T1} \cdots xT(T1)x_{T(T-1)} xTTx_{TT}
Total e0e_{0} e1e_{1} \cdots eT1e_{T-1} eTe_{T}

According to the sampling procedure described in Algorithm 1, the following conditional independence holds:

(8.1) P(xs|x0,,xs1,n1,,ns)\displaystyle P(\vec{x}_{s}|\vec{x}_{0},\cdots,\vec{x}_{s-1},n_{1},\cdots,n_{s}) =P(xs|xs1,ns).\displaystyle=P(\vec{x}_{s}|\vec{x}_{s-1},n_{s}).

Also note that nsn_{s} only depends on es1e_{s-1}. Therefore the full likelihood function of {xs:0sT}\{\vec{x}_{s}:0\leq s\leq T\} and {ns:1sT}\{n_{s}:1\leq s\leq T\} can be written as

(8.2) L\displaystyle L =P(x0)×s=1T1P(xs|xs1,ns)×P(eT|xT1,nT)×s=1TP(ns|es1).\displaystyle=P(\vec{x}_{0})\times\prod_{s=1}^{T-1}P(\vec{x}_{s}|\vec{x}_{s-1},n_{s})\times P(e_{T}|\vec{x}_{T-1},n_{T})\times\prod_{s=1}^{T}P(n_{s}|e_{s-1}).

According to the Poisson assumption described in Algorithm 1,

P(x0)\displaystyle P(\vec{x}_{0}) =t=0Teλtλtxt0xt0!\displaystyle=\prod_{t=0}^{T}\frac{e^{-\lambda_{t}}\lambda_{t}^{x_{t0}}}{x_{t0}!}

According to the sampling procedure,

(8.3) P(xs|xs1,ns)\displaystyle P(\vec{x}_{s}|\vec{x}_{s-1},n_{s}) =(x(s1)(s1)nses)t=sT(xt(s1)xts)(es1ns).\displaystyle=\frac{{x_{(s-1)(s-1)}\choose n_{s}-e_{s}}\prod_{t=s}^{T}{x_{t(s-1)}\choose x_{ts}}}{{e_{s-1}\choose n_{s}}}.
Lemma 5.

Let Λ^t\hat{\Lambda}_{t} be defined by (4.2), then t{0,,T}\forall t\in\{0,\cdots,T\}

E(Λ^t)=Λt.\displaystyle E(\hat{\Lambda}_{t})=\Lambda_{t}.
Proof.

It is sufficient to prove the result for t=Tt=T, since the proof is the same for other tt. By (8.3), conditional on xT1\vec{x}_{T-1} and nTn_{T}, xTTx_{TT} (which is eTe_{T}) follows a hypergeometric distribution. Using (8.1), for eT1>0e_{T-1}>0, we have

E(eT|x0,,xT1,n1,,nT)=\displaystyle E(e_{T}|\vec{x}_{0},\cdots,\vec{x}_{T-1},n_{1},\cdots,n_{T})= E(eT|xT1,nT)\displaystyle E(e_{T}|\vec{x}_{T-1},n_{T})
=\displaystyle= xT(T1)nTeT1.\displaystyle\frac{x_{T(T-1)}n_{T}}{e_{T-1}}.

Note that

E(Λ^T)\displaystyle E(\hat{\Lambda}_{T}) =E(Λ^TI(e0:(T1)>0))\displaystyle=E(\hat{\Lambda}_{T}I(e_{0:(T-1)}>0))
=E(t=0Tett=1TntI(e0:(T1)>0)),\displaystyle=E\left(\frac{\prod_{t=0}^{T}e_{t}}{\prod_{t=1}^{T}n_{t}}I(e_{0:(T-1)}>0)\right),

where I(e0:t>0)I(e_{0:t}>0) indicates whether es>0e_{s}>0 for all s{0,,t}s\in\{0,\cdots,t\}. Then

E(Λ^T)\displaystyle E(\hat{\Lambda}_{T}) =E(t=0T1ett=1TntI(e0:(T1)>0)×E(eT|x0,,xT1,n0,,nT))\displaystyle=E\left(\frac{\prod_{t=0}^{T-1}e_{t}}{\prod_{t=1}^{T}n_{t}}I(e_{0:(T-1)}>0)\times E(e_{T}|\vec{x}_{0},\cdots,\vec{x}_{T-1},n_{0},\cdots,n_{T})\right)
=E(t=0T2ett=1T1ntI(e0:(T2)>0)×xT(T1)).\displaystyle=E\left(\frac{\prod_{t=0}^{T-2}e_{t}}{\prod_{t=1}^{T-1}n_{t}}I(e_{0:(T-2)}>0)\times x_{T(T-1)}\right).

By iteratively applying (8.1) for es1>0e_{s-1}>0,

E(xTs|x0,,xs1,n1,,ns)\displaystyle E(x_{Ts}|\vec{x}_{0},\cdots,\vec{x}_{s-1},n_{1},\cdots,n_{s}) =E(xTs|xs1,ns)=xT(s1)nses1\displaystyle=E(x_{Ts}|\vec{x}_{s-1},n_{s})=\frac{x_{T(s-1)}n_{s}}{e_{s-1}}

for s=T1,,1s=T-1,\cdots,1, we get

E(Λ^T)\displaystyle E(\hat{\Lambda}_{T}) =E(xT0I(e0>0))=E(xT0)\displaystyle=E(x_{T0}I(e_{0}>0))=E(x_{T0})

which is λT\lambda_{T} by (3.1) since xT0xTx_{T0}\equiv x_{T}. ∎

In order to drive the MLE, we apply the EM algorithm.

8.3. The Expectation-Maximization algorithm

Lemma 6.

The M-step simplifies to

(8.4) λt\displaystyle\lambda_{t} =E(xt0|e,n) for t=1,,T,\displaystyle=E(x_{t0}|\vec{e},\vec{n})\text{ for }t=1,\cdots,T,

which implies t=0Tλt=e0\sum_{t=0}^{T}\lambda_{t}=e_{0}, since t=0Txt0e0\sum_{t=0}^{T}x_{t0}\equiv e_{0}.

Proof.

Note that in the full likelihood function (8.2), only the term P(x0)P(\vec{x}_{0}) involves the parameters of interest. So the M-step, which maximizes E(logL|e,n)E(\log L|\vec{e},\vec{n}) w.r.t. (λ0,,λT)(\lambda_{0},\cdots,\lambda_{T}), is equivalent to maximization of

(λ0,,λT)\displaystyle\ell(\lambda_{0},\cdots,\lambda_{T}) E(logP(x0)|e,n)\displaystyle\equiv E(\log P(\vec{x}_{0})|\vec{e},\vec{n})
=t=0TE(xt0|e,n)×logλtλt+const\displaystyle=\sum_{t=0}^{T}E(x_{t0}|\vec{e},\vec{n})\times\log\lambda_{t}-\lambda_{t}+const

where constconst does not depend on (λ0,,λT)(\lambda_{0},\cdots,\lambda_{T}).

The conclusion follows from

λt\displaystyle\frac{\partial\ell}{\partial\lambda_{t}} =E(xt0|e,n)λt1=0.\displaystyle=\frac{E(x_{t0}|\vec{e},\vec{n})}{\lambda_{t}}-1=0.

For s=0,,Ts=0,\cdots,T let

Λs\displaystyle\Lambda_{s} t=sTλt.\displaystyle\equiv\sum_{t=s}^{T}\lambda_{t}.

In order to derive the E-step, we need the result below which follows from Theorem 4. The detailed proof is omitted for conciseness.

Lemma 7.

Conditional on (e0,x1,n1)(e_{0},\vec{x}_{1},n_{1}), x0\vec{x}_{0} is independent of (x2,,xT,n2,,nT)(\vec{x}_{2},\cdots,\vec{x}_{T},n_{2},\cdots,n_{T}), and follows a multinomial distribution with mean shift as follows:

x0|(e0,x1,n1)\displaystyle\vec{x}_{0}\Big{|}(e_{0},\vec{x}_{1},n_{1}) [n1e1x1]+Multinomial(e0n1,(Λ01λ0Λ01λT)).\displaystyle\sim\left[\begin{array}[]{c}n_{1}-e_{1}\\ \vec{x}_{1}\end{array}\right]+Multinomial\left(e_{0}-n_{1},\left(\begin{array}[]{c}\Lambda_{0}^{-1}\lambda_{0}\\ \cdots\\ \Lambda_{0}^{-1}\lambda_{T}\end{array}\right)\right).

Furthermore, for each s{1,,T1}s\in\{1,\cdots,T-1\}, conditional (es,xs+1,ns+1)(e_{s},\vec{x}_{s+1},n_{s+1}), xs\vec{x}_{s} is independent of (e0,,es1,xs+2,,xT,n1,,ns,ns+2,,nT)(e_{0},\cdots,e_{s-1},\vec{x}_{s+2},\cdots,\vec{x}_{T},n_{1},\cdots,n_{s},n_{s+2},\cdots,n_{T}), and follows multinomial with mean shift:

xs|(es,xs+1,ns+1)\displaystyle\vec{x}_{s}\Big{|}(e_{s},\vec{x}_{s+1},n_{s+1}) [ns+1es+1xs+1]+Multinomial(esns+1,(Λs1λsΛs1λT)).\displaystyle\sim\left[\begin{array}[]{c}n_{s+1}-e_{s+1}\\ \vec{x}_{s+1}\end{array}\right]+Multinomial\left(e_{s}-n_{s+1},\left(\begin{array}[]{c}\Lambda_{s}^{-1}\lambda_{s}\\ \cdots\\ \Lambda_{s}^{-1}\lambda_{T}\end{array}\right)\right).

The result below follows trivially from Lemma 7.

Lemma 8.

The E-step can be calculated as below: for s=0,,T1s=0,\cdots,T-1,

(8.5) E(xss|e,n)\displaystyle E(x_{ss}|\vec{e},\vec{n}) =(ns+1es+1)+(esns+1)×λsΛs\displaystyle=(n_{s+1}-e_{s+1})+(e_{s}-n_{s+1})\times\frac{\lambda_{s}}{\Lambda_{s}}
(8.6) E(xts|e,n)\displaystyle E(x_{ts}|\vec{e},\vec{n}) =E(xt(s+1)|e,n)+(esns+1)×λtΛs,for t=s+1,,T.\displaystyle=E(x_{t(s+1)}|\vec{e},\vec{n})+(e_{s}-n_{s+1})\times\frac{\lambda_{t}}{\Lambda_{s}},\text{for }t=s+1,\cdots,T.

The convergence point satisfies the equations from both the M-step and E-step, i.e. (8.4), (8.5) and (8.6), which is summarized below.

Proposition 9.

The EM algorithm converges to a unique solution as below:

Λ^t\displaystyle\hat{\Lambda}_{t} =s=0tess=1tns\displaystyle=\frac{\prod_{s=0}^{t}e_{s}}{\prod_{s=1}^{t}n_{s}}

for t=0,,Tt=0,\cdots,T. Thus the MLE of (λ0,,λT)(\lambda_{0},\cdots,\lambda_{T}) is given by λ^t=Λ^tΛ^t+1\hat{\lambda}_{t}=\hat{\Lambda}_{t}-\hat{\Lambda}_{t+1} for t=0,,T1t=0,\cdots,T-1 and λ^T=Λ^T\hat{\lambda}_{T}=\hat{\Lambda}_{T}.

The detail for solving the equations is omitted for conciseness. Instead, we provide the detailed proof for T=2T=2 to illustrate the idea.

8.4. Illustration with T=2T=2

For illustration, below shows the complete E-step and M-step for T=2T=2, for which, the notation of xstx_{st} is listed in Table 2 for convenience.

Table 2. Notation for {xts:0st2}\{x_{ts}:0\leq s\leq t\leq 2\}.
Tier-0 Tier-1 Tier-2
FP-1 x00x_{00} 0 0
FP-2 x10x_{10} x11x_{11} 0
TP x20x_{20} x21x_{21} x22x_{22}
Total e0e_{0} e1e_{1} e2e_{2}

According to the description of the EM algorithm in the previous section, we have, with obs(e0,n1,e1,n2,e2)obs\equiv(e_{0},n_{1},e_{1},n_{2},e_{2}) ,

  • M-step:

    λt\displaystyle\lambda_{t} =E(xt0|obs), for t=0,1,2.\displaystyle=E(x_{t0}|obs),\text{ for }t=0,1,2.
  • E-step:

    E(x00|obs)\displaystyle E(x_{00}|obs) =(e0n1)λ0t=02λt+(n1e1)\displaystyle=(e_{0}-n_{1})\frac{\lambda_{0}}{\sum_{t=0}^{2}\lambda_{t}}+(n_{1}-e_{1})
    E(x10|obs)\displaystyle E(x_{10}|obs) =(e0n1)λ1t=02λt+(e1n2)λ1t=12λt+(n2e2)\displaystyle=(e_{0}-n_{1})\frac{\lambda_{1}}{\sum_{t=0}^{2}\lambda_{t}}+(e_{1}-n_{2})\frac{\lambda_{1}}{\sum_{t=1}^{2}\lambda_{t}}+(n_{2}-e_{2})
    E(x20|obs)\displaystyle E(x_{20}|obs) =(e0n1)λ2t=02λt+(e1n2)λ2t=12λt+e2.\displaystyle=(e_{0}-n_{1})\frac{\lambda_{2}}{\sum_{t=0}^{2}\lambda_{t}}+(e_{1}-n_{2})\frac{\lambda_{2}}{\sum_{t=1}^{2}\lambda_{t}}+e_{2}.

The convergence point of the EM algorithm, i.e. (λ0,λ1,λ2)(\lambda_{0},\lambda_{1},\lambda_{2}) satisfies the equations in both E-step and M-step, which simplifies to:

(8.7) λ0\displaystyle\lambda_{0} =(e0n1)λ0e0+(n1e1)\displaystyle=(e_{0}-n_{1})\frac{\lambda_{0}}{e_{0}}+(n_{1}-e_{1})
(8.8) λ1\displaystyle\lambda_{1} =(e0n1)λ1e0+(e1n2)λ1λ1+λ2+(n2e2)\displaystyle=(e_{0}-n_{1})\frac{\lambda_{1}}{e_{0}}+(e_{1}-n_{2})\frac{\lambda_{1}}{\lambda_{1}+\lambda_{2}}+(n_{2}-e_{2})
(8.9) λ2\displaystyle\lambda_{2} =(e0n1)λ2e0+(e1n2)λ2λ1+λ2+e2.\displaystyle=(e_{0}-n_{1})\frac{\lambda_{2}}{e_{0}}+(e_{1}-n_{2})\frac{\lambda_{2}}{\lambda_{1}+\lambda_{2}}+e_{2}.

By summing up both sides of (8.7), (8.8) and (8.9), we get

λ0+λ1+λ2\displaystyle\lambda_{0}+\lambda_{1}+\lambda_{2} =e0.\displaystyle=e_{0}.

By summing up both sides of (8.8) and (8.9), we get

λ1+λ2\displaystyle\lambda_{1}+\lambda_{2} =(e0n1)λ1+λ2e0+(e1n2)+(n2e2)+e2\displaystyle=(e_{0}-n_{1})\frac{\lambda_{1}+\lambda_{2}}{e_{0}}+(e_{1}-n_{2})+(n_{2}-e_{2})+e_{2}

and thus

λ1+λ2\displaystyle\lambda_{1}+\lambda_{2} =e0e1n1.\displaystyle=\frac{e_{0}e_{1}}{n_{1}}.

By plugging-in this back into (8.9), we get

λ2\displaystyle\lambda_{2} =e0e1e2n1n2\displaystyle=\frac{e_{0}e_{1}e_{2}}{n_{1}n_{2}}

which completes the proof.

Acknowledgment

We would like to thank Henning Hohnhold and Kelvin Wu for many insightful discussions.

The event review processes described here are operationalized in partnership with multiple cross-functional teams. We would like to especially thank Vinit Arondekar, Jesse Breazeale, Han Hui, Jeffy Johns, Sidlak Malaki, Kentaro Takada, Ally Yang and Frank Zong for their collaboration.

References

  • [1] Peter J Bickel and Kjell A Doksum. Mathematical statistics: basic ideas and selected topics, volumes I, 2nd Edition. Chapman and Hall/CRC, 2015.
  • [2] Robert J Buehler. Confidence intervals for the product of two binomial parameters. Journal of the American Statistical Association, 52(280):482–493, 1957.
  • [3] Chin-Shan Chuang and Tze Leung Lai. Hybrid resampling methods for confidence intervals. Statistica Sinica, pages 1–33, 2000.
  • [4] William G Cochran. Sampling Techniques, 3rd edition. John Wiley & Sons, 1977.
  • [5] Arthur P Dempster, Nan M Laird, and Donald B Rubin. Maximum likelihood from incomplete data via the em algorithm. Journal of the Royal Statistical Society: Series B (Methodological), 39(1):1–22, 1977.
  • [6] Bradley Efron and Raoul LePage. Introduction to bootstrap. Wiley & Sons, New York, 1992.
  • [7] Michael P Fay and Eric J Feuer. Confidence intervals for directly standardized rates: a method based on the gamma distribution. Statistics in medicine, 16(7):791–801, 1997.
  • [8] Michael P Fay and Sungwook Kim. Confidence intervals for directly standardized rates using mid-p gamma intervals. Biometrical Journal, 59(2):377–387, 2017.
  • [9] Erich L Lehmann and George Casella. Theory of point estimation. Springer Science & Business Media, 2006.
  • [10] Hon Keung Tony Ng, Giovanni Filardo, and Gang Zheng. Confidence interval estimating procedures for standardized incidence rates. Computational statistics & data analysis, 52(7):3501–3516, 2008.
  • [11] Wolfgang A Rolke, Angel M Lopez, and Jan Conrad. Limits and confidence intervals in the presence of nuisance parameters. Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, 551(2-3):493–503, 2005.
  • [12] SAE. Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles - J3016 (April 2021). \urlhttps://www.sae.org/standards/content/j3016_202104.
  • [13] Matthew Schwall, Tom Daniel, Trent Victor, Francesca Favaro, and Henning Hohnhold. Waymo public road safety performance data. Technical report, Waymo LLC, 2020. \urlhttps://www.waymo.com/safety.
  • [14] Bodhisattva Sen, Matthew Walker, and Michael Woodroofe. On the unified method with nuisance parameters. Statistica Sinica, pages 301–314, 2009.
  • [15] Pei Sun, Henrik Kretzschmar, Xerxes Dotiwalla, Aurelien Chouard, Vijaysai Patnaik, Paul Tsui, James Guo, Yin Zhou, Yuning Chai, Benjamin Caine, Vijay Vasudevan, Wei Han, Jiquan Ngiam, Hang Zhao, Aleksei Timofeev, Scott Ettinger, Maxim Krivokon, Amy Gao, Aditya Joshi, Yu Zhang, Jonathon Shlens, Zhifeng Chen, and Dragomir Anguelov. Scalability in perception for autonomous driving: Waymo open dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
  • [16] Michael Bruce Swift. A simulation study comparing methods for calculating confidence intervals for directly standardized rates. Computational statistics & data analysis, 54(4):1103–1108, 2010.
  • [17] Ram C Tiwari, Limin X Clegg, and Zhaohui Zou. Efficient interval estimation for age-adjusted cancer rates. Statistical methods in medical research, 15(6):547–569, 2006.
  • [18] DJ Venzon and SH Moolgavkar. A method for computing profile-likelihood-based confidence intervals. Journal of the Royal Statistical Society: Series C (Applied Statistics), 37(1):87–94, 1988.
  • [19] Weizhen Wang. A note on bootstrap confidence intervals for proportions. Statistics & Probability Letters, 83(12):2699–2702, 2013.
  • [20] N. Webb, D. Smith, C. Ludwick, T.W. Victor, Q. Hommes, F. Favaro, G. Ivanov, and T. Daniel. Waymo’s safety methodologies and safety readiness determinations. Technical report, Waymo LLC, 2020. \urlhttps://www.waymo.com/safety.
  • [21] Yannis G Yatracos. A small sample optimality property of the mle. Sankhyā: The Indian Journal of Statistics, Series A, pages 90–101, 1998.
  • [22] Ekim Yurtsever, Jacob Lambert, Alexander Carballo, and Kazuya Takeda. A survey of autonomous driving: Common practices and emerging technologies. IEEE access, 8:58443–58469, 2020.