This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Some performance considerations when using multi-armed bandit algorithms in the presence of missing data

Xijin Chen E-mail: [email protected] MRC Biostatistics Unit, University of Cambridge, Cambridge, UK Kim May Lee Institute of Psychiatry, Psychology & Neuroscience, King’s College London, UK Sofia S. Villar MRC Biostatistics Unit, University of Cambridge, Cambridge, UK David S. Robertson MRC Biostatistics Unit, University of Cambridge, Cambridge, UK
Abstract

When comparing the performance of multi-armed bandit algorithms, the potential impact of missing data is often overlooked. In practice, it also affects their implementation where the simplest approach to overcome this is to continue to sample according to the original bandit algorithm, ignoring missing outcomes. We investigate the impact on performance of this approach to deal with missing data for several bandit algorithms through an extensive simulation study assuming the rewards are missing at random. We focus on two-armed bandit algorithms with binary outcomes in the context of patient allocation for clinical trials with relatively small sample sizes. However, our results apply to other applications of bandit algorithms where missing data is expected to occur. We assess the resulting operating characteristics, including the expected reward. Different probabilities of missingness in both arms are considered. The key finding of our work is that when using the simplest strategy of ignoring missing data, the impact on the expected performance of multi-armed bandit strategies varies according to the way these strategies balance the exploration-exploitation trade-off. Algorithms that are geared towards exploration continue to assign samples to the arm with more missing responses (which being perceived as the arm with less observed information is deemed more appealing by the algorithm than it would otherwise be). In contrast, algorithms that are geared towards exploitation would rapidly assign a high value to samples from the arms with a current high mean irrespective of the level observations per arm. Furthermore, for algorithms focusing more on exploration, we illustrate that the problem of missing responses can be alleviated using a simple mean imputation approach.

1 Introduction

In healthcare, rapid progress in machine learning is enabling opportunities for improved clinical decision support [20, 11, 56]. In this context, Multi-Armed Bandit Problems (MABPs), which can be traced back to proposals for allocating patients in two-armed clinical trials [64], have seen a surge of interest in recent years [67, 23]. The goal in MABPs is to achieve a balance between allocating resources to the current best choice (exploitation) or to alternatives that could potentially be a better choice (exploration). MABPs are a powerful framework for algorithms that make decisions over time under uncertainty, arising in a variety of application domains in recent decades, such as ad placement, optimizing seller’s prices (also known as dynamic pricing or learn-and-earn), and internet routing and congestion control.

In the context of clinical trials, bandit algorithms fall within the larger class of response-adaptive designs, which allow the allocation of patients to depend on the previously observed responses in order to achieve experimental aims such as allocating more patients in the best arm. This paper will use the clinical trial setting as our main focus. Jacko [34] illustrates the formulated terminology of MABPs in various disciplines, and we follow the terminology that is widely used in the machine learning and clinical trial communities. For instance, ‘algorithm’ and ‘arms’ in MABPs correspond to ‘design’ and ‘treatments’ in biostatistics (see Table 1 in [34]).

A fixed (equal) randomization scheme that does not change with patient responses is still the most traditional and well-accepted allocation rule in clinical trial designs. However, there has been a surge of recent interest in response-adaptive designs [52]. The use of bandit algorithms is thus increasingly being considered in the context of adaptive clinical trials [22], potentially in combination with other adaptive features such as early stopping or sample size re-estimation [59]. There is a growing methodological literature discussing implementations of bandit models in various types or phases of clinical trial designs [63, 49, 19, 67, 6, 58]. However, it is difficult to point to examples where bandit algorithms have been used in clinical trials [67], except for algorithms based on a variant of the Thompson Sampling (TS) algorithm. Such examples in oncology include the well-known BATTLE trial [76] and the I-SPY 2 trial [7].

A key problem with the use of multi-arm bandits in practice is that the response information required to update the allocation of subsequent patients might not be available for many reasons. In clinical trials, enrolled patients may not complete the study and drop out without further measurements due to adverse reactions, death, or a variety of other reasons. In some healthcare settings, malfunctions in electronic health record systems commonly occur. It is common that the consequent missing data reduces the statistical power of a study and produces biased parameter estimates, leading to a reduction in efficiency and invalid conclusions [43].

The simplest analysis approach to missing data problems in clinical trials is to use only those cases that do not have any data missing in the variable(s) we are concerned with. This approach is referred to as ‘complete case analysis’ or ‘listwise deletion’ [55]. In response-adaptive designs, it is far more challenging to handle this problem than in traditional fixed designs as a result of the fully sequential nature and the cumulative impact of the allocation procedure [44]. In other words, it is not only an analysis problem at the end of a trial (as in a fixed design), but also a design problem in terms of the allocation procedure. The naive approach of ignoring missing data is thus suboptimal in terms of performance in the context of response-adaptive designs since the original intention of placing more patients in the best arm could be violated in this case.

To the best of our knowledge, this challenging problem has received limited attention in the context of response-adaptive designs [4], including bandit algorithms. One exception is Biswas and Rao [14], who implement regression imputation for a covariate-adjusted response adaptive procedure based on the available responses and associated covariates. A doctoral dissertation of Ma [44] concentrates on the doubly-adaptive biased coin designs in the presence of missing responses. A common approach in the literature is to assume that data are ‘Missing at random’ (MAR) [55], meaning that the probability of data being missing depends on the observed data but not the missing data. For responses that are MAR, a likelihood-based approach was proposed by Ma [44] to incorporate the information of covariates and prevent inconsistent parameter estimates and undesirable allocation results. Sverdlov et al. [62] evaluate the impact of missing data on the proposed Longitudinal Covariate-Adjusted Response-Adaptive Randomization procedures with continuous responses, in terms of the targeted allocation proportions and the accuracy of final estimations of parameters of interest. Numerous simulation studies results of three different missing data analysis approaches show that the performance gets worse as the percentage of missing data increases. Finally, Williamson and Villar [71] propose using online imputation for the Forward-Looking Gittins Index rule [68] with normally distributed outcomes.

The issue of delayed outcomes in response-adaptive trials is closely related to the problem of missing responses. There have been a few proposals for dealing with delayed outcomes that simply treat those late-onset outcomes as missing data. A Bayesian data augmentation method has been proposed to impute those missing outcomes [35, 75]. Nevertheless, it is more common to distinguish between delayed responses and missing responses in most cases. Kim et al. [40] study the impact of delay in the outcome variables on the performance of doubly-adaptive biased coin designs with binary outcomes. It was shown that the corresponding performance could be improved by imputation based on a short-term predictor. Williamson et al. [70] identify the existing gap between the theory and clinical practice for Bayesian response-adaptive procedures by considering whether responses are intermediately available or delayed, respectively. However, these findings may not apply to the missing data problem in the context of response-adaptive designs, since missing values can be viewed as a very extreme form of delay where the outcomes would never be available [48]. In other words, the problem of missing data is distinct and has not received as much attention as the problem of delayed outcomes.

In the machine learning community, researchers have also realized that the key assumption of immediate reward after an action is taken may not hold in practice. For instance, Bouneffouf et al. [16] investigate contextual bandits with missing rewards, which reflect the outcome of the selected action in clinical trials, or whether an ad is clicked or not in recommender systems. They propose a modified version of the Upper Confidence Bound (UCB) algorithm, which imputes the missing rewards based on the available rewards from similar contexts. The setting of contextual bandits in the machine learning community corresponds to the covariate-adjusted response-adaptive designs in clinical practice, where the problem of missing data has been discussed the most together with associated covariates. For this reason, the proposed framework allows the use of available context information for future decision-making when some of the outcomes are missing. Missing data problem in the ‘context’ rather than in reward is more widely discussed, which is referred to as ‘corrupted contextual bandits’ [15]. Additionally, the problem with delays is also common in the field of machine learning, which might not be applicable in the extreme case of missing data problems as already mentioned above [36, 24, 3, 27, 33, 77].

Here we revisit the problem of missing responses for the following reasons: (a) Most response-adaptive methods in clinical research are limited to response-adaptive randomization, which does not cover the more general case of deterministic algorithms favored in the machine learning community from the perspective of quality of care or patients benefit; (b) There is a gap in investing the impact of missing data for finite sample sizes since related research in machine learning applications focuses more on some asymptotic properties; (c) Related research is limited to a limited range of response-adaptive designs or bandit algorithms. Besides, the fact that MABP algorithms and response-adaptive designs are usually not compared makes it harder to make broad comparisons or useful recommendations. This indicates that further systematic investigation based on the fundamentals of bandit algorithms (i.e., the exploration-exploitation trade-off) is required; (d) Related work seldom accommodates different probabilities of missingness in different arms, indicating a more extensive evaluation is in need; (e) Covariate-adjusted RAR and contextual bandits use side information, and it is unclear what the impact of missing data is on algorithms that do no use any covariates; (f) The closely related issue of delayed outcomes in trial designs has attracted a lot of attention and various solutions have followed. However, most of these proposed solutions do not necessarily apply to the problem of missing data. Our main contribution is to perform a comprehensive investigation concerning the fundamentals behind bandit algorithms, specifically, the exploration-exploitation trade-off, in the presence of missing responses. This goes beyond the few examples that have been looked at in isolation for specific response-adaptive designs or bandit models.

In Section 2, we first introduce the notations used in this paper, followed by a description of the allocation procedure in clinical trials in the presence of missing data. We then introduce some well-known bandit algorithms in the framework of MABPs. In Section 3, we consider different probabilities of missingness in the two arms, presenting a comprehensive simulation study of allocation results considering various scenarios either under the null or the alternative hypothesis. In addition, other patient-outcome metrics are also considered for illustration. In Section 4, we implement the mean imputation approach for missing binary responses, and discuss the impact of biased estimates in MABP on imputation. We conclude this paper with a discussion in Section 5.

2 General framework

We consider a general two-armed bandit model referring to the setting with binary responses in the presence of missing data. The two-armed bandit problem naturally arises as a sub-problem in some multi-armed generalizations and serves as a starting point for introducing additional problem features. Assuming responses are missing at random (MAR), the probability of being missing is the same within groups defined by the observed data (in our setting, the two treatment groups). Patients sequentially enrolled in a clinical trial will be assigned to the kk competing arms. In terms of the performance of different algorithms, we evaluate metrics of patient outcomes, including the proportion of patients assigned to the experimental arm (pp*) and the observed number of success (ONS), which we define below. Specifically, our main focus is on pp^{*} since the performance of different algorithms is more clearly seen. We evaluate the performance of different algorithms via simulation studies, which are performed using the R programming language [50].

2.1 Allocation procedure of bandit algorithms in the presence of missing data

The two-armed trial setting is a very common one in clinical trial designs. We denote the control arm by the subscript k=0k=0 and the experimental arm by k=1k=1, respectively. Assigned patients have missing responses with a probability pkm0.5p_{k}^{\text{m}}\leq 0.5, given that it is rare that the majority of patients would be missing in the setting of healthcare. The total number of missing responses, successes and failures on arm kk at time tt are denoted by Mk,tM_{k,t}, Sk,tS_{k,t} and Fk,tF_{k,t}.

The observed outcomes (Sk,t,Fk,t)(S_{k,t},\ F_{k,t}), with a true probability of success pkp_{k}, in turn have an impact on the next allocation due to the response-adaptive nature in bandit algorithms, while missing responses Mk,tM_{k,t} do not. The Bayesian feature is introduced to the algorithm by a uniform prior (sk,0=1,fk,0=1s_{k,0}=1,\ f_{k,0}=1), which is the initial state that enables the first patients to be assigned when there are no observations. The uniform prior implies that an equal allocation (as in the fixed design) is initially used, but a different prior could also be used if appropriate.

The current state xk,t=(sk,0+Sk,t,fk,0+Fk,t)x_{k,t}=(s_{k,0}+S_{k,t},\ f_{k,0}+F_{k,t}) after having observed Sk,tS_{k,t} and Fk,tF_{k,t} determines the subsequent patient allocation. Note that for each arm kk, the number of assigned patients in arm kk at time tt accounts for both the missing and observed data, namely, Nk,t=Sk,t+Fk,t+Mk,tN_{k,t}=S_{k,t}+F_{k,t}+M_{k,t}. The fixed total trial size is expressed as nn. The allocation procedure with bandit algorithms in the presence of missing data is attached in Appendix A. The operating characteristics we evaluate are defined as the proportion of patients assigned to the experimental arm (p=N1,nnp^{*}=\frac{N_{1,n}}{n}) and the observed total number of success (ONS=E[S0,n+S1,n]\text{ONS}=E[S_{0,n}+S_{1,n}]). In the clinical trial context, we will typically be interested in testing a null hypothesis H0:p0=p1H_{0}:p_{0}=p_{1} against a one-sided alternative hypothesis HA:p0<p1H_{A}:p_{0}<p_{1}.

2.2 Bandit algorithms

We classify several well-known bandit algorithms into three categories according to their exploration-exploitation trade-off. Details of these algorithms are summarized in Table 1, with further details given below.

  • Randomized algorithms (TTS, RTS, and RPW): The patient at time tt is randomly assigned to arm kk with probability πk,t\pi_{k,t}.

  • Deterministic algorithms (CB, GI, and UCB): The patient at time tt is deterministically assigned to the arm kk with the larger index values Ik,tI_{k,t}, indicating the existence of a ‘priority’ value for sampling from one of the arms [47].

  • Semi-randomized algorithms (RandUCB, RBI, and RGI): The patient at time tt is deterministically assigned to the arm with the larger index values Ik,tI_{k,t} as in deterministic algorithms. However, a stochastic element is involved in the computation of the index values Ik,tI_{k,t}.

Table 1: Bandit algorithms defined by allocation probability πk,t\pi_{k,t} or index value Ik,tI_{k,t}.
Algorithm πk,t\pi_{k,t} or Ik,tI_{k,t} Details Initial value
  TTS πk,t=P(maxtpt=pk)ck=1KP(maxtpt=pk)c\pi_{k,t}=\frac{P(\text{max}_{t}\ p_{t}=p_{k})^{c}}{\sum_{k=1}^{K}P(\text{max}_{t}\ p_{t}=p_{k})^{c}} c=t2nc=\frac{t}{2n} πk,0=0.5\pi_{k,0}=0.5
RTS πk,t=P(maxtpt=pk)ck=1KP(maxtpt=pk)c\pi_{k,t}=\frac{P(\text{max}_{t}\ p_{t}=p_{k})^{c}}{\sum_{k=1}^{K}P(\text{max}_{t}\ p_{t}=p_{k})^{c}} c=1c=1 πk,0=0.5\pi_{k,0}=0.5
RPW πk,t\pi_{k,t} is urn-based initial urn with one ball for each arm kk πk,0=0.5\pi_{k,0}=0.5
  CB Ik,t=μ^k,tBI_{k,t}=\hat{\mu}^{B}_{k,t} μ^k,tB=sk,0+Sk,tsk,0+fk,0+Sk,t+Fk,t\hat{\mu}^{B}_{k,t}=\frac{s_{k,0}+S_{k,t}}{s_{k,0}+f_{k,0}+S_{k,t}+F_{k,t}} Ik,0=0.5I_{k,0}=0.5
GI Ik,t=μ^k,tGI_{k,t}=\hat{\mu}^{G}_{k,t} μ^k,tG=G[sk,0+Sk,t][fk,0+Fk,t]\hat{\mu}^{G}_{k,t}=G[s_{k,0}+S_{k,t}][f_{k,0}+F_{k,t}] Ik,0=0.8699I_{k,0}=0.8699
UCB Ik,t=μ^k,tB+βλk(t)I_{k,t}=\hat{\mu}^{B}_{k,t}+\beta\cdot\lambda_{k}(t) β=2log(t),λk(t)=1sk,0+fk,0+Sk,t+Fk,t\beta=\sqrt{2\log(t)},\ \lambda_{k}(t)=\frac{1}{\sqrt{s_{k,0}+f_{k,0}+S_{k,t}+F_{k,t}}} Ik,0=0.5I_{k,0}=0.5
  RandUCB Ik,t=μ^k,tB+Ztλk(t)I_{k,t}=\hat{\mu}^{B}_{k,t}+Z_{t}\cdot\lambda_{k}(t) ZtfX(x),λk(t)=1sk,0+fk,0+Sk,t+Fk,tZ_{t}\sim f_{X}(x),\ \lambda_{k}(t)=\frac{1}{\sqrt{s_{k,0}+f_{k,0}+S_{k,t}+F_{k,t}}}
RBI Ik,t=μ^k,tB+Ztλk(t)I_{k,t}=\hat{\mu}^{B}_{k,t}+Z_{t}\cdot\lambda_{k}(t) ZtExp(1/K),λk(t)=Ksk,0+fk,0+Sk,t+Fk,tZ_{t}\sim\text{Exp}(1/K),\,\ \lambda_{k}(t)=\frac{K}{s_{k,0}+f_{k,0}+S_{k,t}+F_{k,t}}
RGI Ik,t=μ^k,tG+Ztλk(t)I_{k,t}=\hat{\mu}^{G}_{k,t}+Z_{t}\cdot\lambda_{k}(t) ZtExp(1/K),λk(t)=Ksk,0+fk,0+Sk,t+Fk,tZ_{t}\sim\text{Exp}(1/K),\,\ \lambda_{k}(t)=\frac{K}{s_{k,0}+f_{k,0}+S_{k,t}+F_{k,t}}

In the allocation procedure, the first assignment is based on the initial value (πk,0\pi_{k,0} or Ik,0I_{k,0}) and is an equal allocation for all algorithms, which is based on the choice of prior. This is achieved by allocation of the first patient via an allocation probability fixed at πk,0=0.5\pi_{k,0}=0.5 in the case of randomized algorithms, or via equal index values Ik,0I_{k,0} in both arms in the case of deterministic or semi-randomized algorithms.

  • Tuned Thompson Sampling (TTS): is not the original version of TS but a variant that has been more generally implemented [63]. TTS randomizes each patient to a treatment kk with a probability that is proportional to the posterior probability that treatment kk is the best arm given the accrued data. The best arm refers to the arm with the largest pkp_{k}, and it is assumed that in the case of a tie (p0=p1p_{0}=p_{1}), the arm (p0p_{0}) would be considered as the best. The positive tuning parameter c=t2nc=\frac{t}{2n} is time-varying.

  • Raw Thompson Sampling (RTS): is the ‘raw’ version of TS in its first proposal [64]. RTS is the most commonly used version in the machine learning community, which is different from the commonly used version, TTS, in the biostatistical community in terms of the tuning parameter [63]. The tuning parameter of TTS is fixed at c=1c=1 in RTS. When compared to the traditional fixed randomization (FR) with a fixed tuning parameter c=0c=0, RTS and TTS have different tuning parameters cc.

  • Randomized Play-the-Winner (RPW): is not a bandit algorithm but an urn-based model [69]. We include it since RPW has a long history in clinical trials methodology, and because its use helps explain why few response-adaptive clinical trials have occurred in practice. Indeed, the infamous Extracorporeal Circulation in Neonatal Respiratory Failure (ECMO) trial [9] used RPW to allocate patients. Due to the extreme treatment imbalance and highly controversial interpretation, the ECMO trial is regarded as a key example against the use of response-adaptive designs in clinical trials [53, 18]. The initial urn composition in this paper is the extreme case of one ball for each treatment. One would expect the allocation proportions to be less extreme when the initial urn composition is increased, as the urn will not favor the better treatment as highly.

  • Current Belief (CB): allocates each patient to the treatment arm with the highest immediate estimate of reward μ^k,tB\hat{\mu}^{B}_{k,t}, which is defined as the current highest mean posterior probability of success. With the only aim of exploitation, this myopic algorithm will tend to ‘select’ an arm before the trial is over. Once one arm is selected, all of the next enrolled patients will be assigned to this arm, and the index value of the unselected arm will not change anymore. One issue with CB is that the algorithm might make a wrong selection and make many allocations to the inferior arm. An early selection within the first few patients could be anticipated when the success probability pkp_{k} is large.

  • Gittins Index (GI): recovers the optimal solution to an infinite (discounted) MABP as obtained by Dynamic programming [67]. There is an upward adjustment in the index value Ik,tI_{k,t} in GI when compared to CB, referring to the uncertainty about the prospects of obtaining rewards from the arm [29]. This uncertainty corresponds to a decreasing exploration component, which decreases with more observations in the corresponding arm [30]. The value of the initial state Ik,0I_{k,0} of GI is larger than that of CB as a result of the upward adjustment. It is fixed at Ik,0=0.8699I_{k,0}=0.8699 for GI in this paper, corresponding to a particular discount factor d=0.99d=0.99, the widely used value in the related bandit literature [67].

  • Randomized Belief Index (RBI): is a semi-randomized approach with an index-based part μ^k,tB\hat{\mu}^{B}_{k,t} for exploitation and a random perturbation part Ztλk(t)Z_{t}\cdot\lambda_{k}(t) for exploration. ZtZ_{t} is a random variable and λk(t)\lambda_{k}(t) is a decreasing sequence of positive constants. This means that most patients are assigned to the (current) superior arm, but some patients will still be assigned to the inferior arm in order to achieve exploration [12]. In other words, there is no aggressive determinism in RBI when compared to CB as a result of the additional perturbation part. The overall pattern of its index values Ik,tI_{k,t} is decreasing as a result of the decreasing exploitation component, accompanied by some fluctuations representing the random exploration part.

  • Randomized Gittins Index (RGI): is a semi-randomized approach, which applies the random perturbation idea to the GI rule [31, 12]. Similar to RBI, the overall index values Ik,tI_{k,t} and the exploration component of the index values decrease over time.

  • Upper Confidence Bound Index (UCB): is based on the principle of ‘optimism in the face of uncertainty’ [5]. It explores the arm with higher uncertainty because of higher information gain represented by the term βλk(t)\beta\cdot\lambda_{k}(t), where β=2log(t)\beta=\sqrt{2\log(t)} and λk(t)=1sk,0+fk,0+Sk,t+Fk,t\lambda_{k}(t)=\sqrt{\frac{1}{s_{k,0}+f_{k,0}+S_{k,t}+F_{k,t}}} reflects the confidence interval corresponding to the standard deviation in the estimation of reward μ^k,tB\hat{\mu}^{B}_{k,t}. The more times a specific arm has been engaged before in the past, the greater the confidence boundary reduces towards the point estimate. If there are no observations in the kk-th term, then μ^k,tB=0\hat{\mu}^{B}_{k,t}=0 and βλk(t)\beta\cdot\lambda_{k}(t)\to\infty, thus each arm is selected at least once. As evidence accumulates and the term βλk(t)\beta\cdot\lambda_{k}(t) vanishes, the index values have an initial rapid increase for the first few patients but then decrease in both arms. There are many variants of UCB targeting various objectives in different disciplines [38].

  • Randomized Upper Confidence Bound (RandUCB): is similar to RBI and RGI in terms of being a semi-randomized algorithm. RandUCB takes random perturbations [12, 31] for the computation of its index values. RandUCB is different from the UCB algorithm in terms of the exploration term, which is constructed by a random variable ZtZ_{t} sampling from a discrete distribution fX(x)f_{X}(x) [66]. In addition, exploration in the arm with higher uncertainty is realized by accounting for the confidence interval or standard deviation λk(t)\lambda_{k}(t) as in the UCB. The discrete distribution is supported on MM points over the interval [L,U][L,U]. Let α1=L\alpha_{1}=L…, αM=U\alpha_{M}=U denote MM equally spaced points in [L,U][L,U]. We recover the UCB if M=1M=1 and [L,U]=[β,β][L,U]=[\beta,\beta] hold. RandUCB with MM\to\infty approaches optimistic TS (one of the different versions of TS) with a Gaussian prior and posterior. In other words, both UCB and TS can be recovered as special cases of RandUCB. In this paper, we take M=20M=20 and [L,U]=[0,1][L,U]=[0,1] as an example for illustration. More details are discussed in Vaswani et al. [66]. Actually, the idea of RandUCB is the same as that used in RBI and RGI, which can be traced back to around 40 years ago [12].

In general, there is more than one strategy to solve the exploration-exploitation dilemma inherent in the MABP. For algorithms in this framework, the exploitation goal could be achieved with an immediate reward based on current observations. For the aim of exploration, one of the classic strategies is to account for a confidence level in the estimate, namely, optimism in the face of uncertainty [1]. The UCB algorithm is an example using this strategy that has attracted a lot of attention in the field of machine learning. Another common device in optimization problems is randomized allocation, which is widely used in the context of clinical trials. This includes the first well-known attempt at response-adaptive design, RTS. Besides, this stochastic setting has also motivated the ‘Follow the Perturbed Leader’ algorithms, which add random perturbations to the estimates of rewards of each arm prior to computing the current ‘best arm’ [2, 39, 41]. The semi-randomized algorithms, RBI and RGI [67], operate in the same way through randomly perturbing arms that would be selected by a greedy algorithm, which disregards any advantages of exploring. Thus, it allows mixtures of alternative decisions of the selected arms. RandUCB uses randomization to trade off exploration and exploitation, like RBI and RGI, and simultaneously it maintains optimism in the face of uncertainty in a similar way to UCB.

2.3 Sampling behaviour without missing data

Bandit algorithms work in different ways according to their exploration-exploitation trade-off. Figure 1 demonstrates the sampling behaviour using the results of a single simulation, for a specific scenario under the null. The trial size is fixed at n=200n=200, indicating a relatively small confirmatory clinical trial. In the case of a larger trial, the simulation results reflect what could happen at its beginning. The true probabilities of success are fixed at p0=p1=0.7p_{0}=p_{1}=0.7, which might arise when testing two similarly efficacious vaccines in clinical trials. The assumption of identical arms also is relevant to some other machine learning applications, such as the problem of online allocation of homogeneous tasks to a pool of agents; a problem faced by many online platforms and matching markets. The regret performance of a bandit algorithm in terms of its arm-sampling characteristics are of great interest, not only in the ‘large gap’ (i.e., ‘well-separated’) instance but also the ‘small gap’ instance (i.e., ‘worst-case’), where the gap corresponds to the small difference between the top two arm mean rewards [37]. The large treatment effects in the example in Figure 1 might not be common in general, but it can provide some intuitive thinking and theoretical perspective. We illustrate a counterpart for a scenario under the alternative in Appendix E because, in general, research in machine learning illustrates bandit performance under the alternative but the null case is rarely displayed.

In Figure 1, randomized algorithms (i.e., the first row) allocate patients with a probability πk,t\pi_{k,t}, while the others allocate patients to the arm with a larger Ik,tI_{k,t} value. Note that the scales on the y-axis of Ik,tI_{k,t} of deterministic and semi-randomized algorithms are different, and it is the relative difference between the Ik,tI_{k,t} in the two arms rather than the absolute value that determines allocation results.

Refer to caption
Figure 1: Allocation procedure of bandit algorithms. Performance of different bandit algorithms over a single simulation (with p0=0.7p_{0}=0.7, p1=0.7p_{1}=0.7, and n=200n=200) under the null. The two colors reflect different arms that could be regarded as the same under the null.

0

The allocation probabilities πk,t\pi_{k,t} of TTS show that it tends to ‘select’ an arm (i.e., allocate to that arm with a much higher probability) by the end of a large cohort of patients, even when there is no treatment difference between these two arms. The random selection of one of these equal arms under the null has been discussed in the related discussion about bandit algorithms [72, 60]. A theoretical explanation of this imbalanced behaviour for Thompson Sampling under the null (i.e., despite the arms being statistically identical), also known as the ‘incomplete learning’ phenomenon, is provided in [37]. When compared with TTS, such a ‘selection’ under RTS happens earlier than that of TTS. RTS is also relatively more ‘deterministic’ than TTS due to the fixed tuning parameter, even though both are randomized algorithms. The urn-based algorithm RPW also illustrates some degree of ‘determinism’ in the first few patients. That is, one arm tends to be ‘selected’ at the beginning of the trial, as a result of the small initial urn size in our setting.

These randomized algorithms have a smaller chance of an early selection and never conduct a real selection in the finite sample size settings. In contrast, deterministic algorithms are more likely to truly select an arm, with patients always assigned to the selected arm from then on. An allocation skew could be anticipated if an early selection occurs. For CB and RandUCB, early selection is particularly noticeable, especially for CB. However, GI, although defined to be a deterministic rule, continues to explore in this instance where CB does not, as shown by the index values of both arms continuing to overlap throughout the trial. Meanwhile, RBI, RGI, and UCB are quite similar in a sense that the index values in both arms are close to to each other, although accompanied by fluctuations as a result of accounting for the uncertainty of the point estimate. This sampling behaviour is also referred to as ‘complete learning’, which was explained theoretically by [37], where UCB is taken as an example. In particular, UCB is able to discern statistical indistinguishability of the arm-means, and induce a ‘balanced’ allocation under the null when there is no treatment difference.

Since the results in Figure 1 are based on a single simulation under the null, a selection is neither incorrect nor correct as both arms are the same and there is no inferior or superior arm (all other costs being equal). A similar illustration of allocation probabilities πk,t\pi_{k,t} or index values Ik,tI_{k,t} in a scenario under the alternative is given in Appendix E. The illustration in Figure 1 gives an insight into the potential consequences of the presence of missing data as shown in Section 3.1. An extensive simulation involving various scenarios follows in Section 3.3.

3 Impact of missing data

In this section, we investigate the impact of missing data on the expected proportion of patients assigned to the experimental arm p=N1,nnp^{*}=\frac{N_{1,n}}{n}. On average, half of the patients (E[p]=0.5E[p^{*}]=0.5) will be assigned to both arms under the null H0H_{0}, and more than half of the patients (E[p]>0.5E[p^{*}]>0.5) will be assigned to the experimental arm under the alternative HAH_{A}. This aligns with the motivation for using response-adaptive designs in clinical practice. With a fixed trial size nn, pp^{*} is determined by the number of patients assigned to the experimental arm N1,nN_{1,n}. Thus, the metric pp^{*} reflects how the allocation is affected in the presence of missing data.

In Section 3.1, we extend the discussion about Figure 1 taking missing data into account. That is, we explore how missing data in one arm affects πk,t\pi_{k,t} or Ik,tI_{k,t} and hence the allocation results. In Section 3.2, we illustrate the impact of different probabilities of missingness in the two arms (p0mp_{0}^{\text{m}} and p1mp_{1}^{\text{m}}) for an example under the null. In Section 3.3, we present a comprehensive simulation study under the null and the alternative.

3.1 Impact of missing data in one arm

For simplicity, suppose that missing data only occurs in arm kk and not arm k~\tilde{k}. If the outcome at time t1t_{1} in arm kk is missing, the next allocation to arm kk at time t2t_{2} cannot be updated and remains what it was for the last patient (Ik,t2=Ik,t1I_{k,t_{2}}=I_{k,t_{1}} or πk,t2=πk,t1\pi_{k,t_{2}}=\pi_{k,t_{1}}). By contrast, since there is no missing data at time t2t_{2} in arm k~\tilde{k}, the value of Ik~,t2I_{\tilde{k},t_{2}} or πk~,t2\pi_{\tilde{k},t_{2}} would be updated as it was supposed to do. Consequently, the subsequent allocations could be affected due to the missing data.

In addition, the impact of missing data on πk,t\pi_{k,t} or Ik,tI_{k,t} is expected to differ across algorithms. For randomized and semi-randomized algorithms, the impact due to missing responses is relatively low as a result of exploration. This applies to both randomized algorithms (i.e., TTS and RTS) and semi-randomized algorithms (i.e., RBI and RGI). Similarly, algorithms taking uncertainty into account for allocations (i.e., UCB and GI) are also less affected by missing data due to their continued exploration.

By contrast, heavily exploitative algorithms (i.e., CB and RandUCB) are expected to be largely affected by missing data in terms of the selection. In the example in Figure 1 with large true probabilities of success (p0=p1=0.7p_{0}=p_{1}=0.7), the arm with missing outcomes is less likely to be selected. The reason is that the other arm without missing outcomes is more likely to have successful outcomes under such a large pkp_{k}, and Ik,tI_{k,t} of this arm will be updated to a larger value for the next allocation. As a consequence, these greedy or deterministic algorithms are less robust to the impact of missing data when compared with algorithms that are more explorative.

Besides, the changing patterns of πk,t\pi_{k,t} and Ik,tI_{k,t} provide some additional insights on the impact of missing data. On the one hand, randomized algorithms (TTS and RTS) keep learning and πk,t\pi_{k,t} of both arms change across the trial – starting from πk,0=0.5\pi_{k,0}=0.5, πk,t\pi_{k,t} values in these two arms become more distinguishable from each other with more enrolled patients and are always symmetric around 0.5. Consequently, the arm with larger πk,t\pi_{k,t} values is more likely to be selected even if there is missing data in this arm. On the other hand, Ik,tI_{k,t} values generally decrease as the immediate reward μ^k,tB\hat{\mu}^{B}_{k,t} or μk,tG\mu^{G}_{k,t} decreases with more patients. Once the selection of an arm occurs, Ik,tI_{k,t} of the other arm will remain unchanged as the arm will not be selected anymore, see Figure 1. Consequently, the arm with missing data is more likely to be selected.

Considering both the impact of missing data in one arm and the changing patterns of πk,t\pi_{k,t} or Ik,tI_{k,t} across a trial of different algorithms, it could infer the impact of missing data in one arm on allocation results, as shown in the simulation results in Section 3.2. Intuitively, for algorithms geared towards exploration, missing data in one arm slows down the progress of learning or exploration in randomized algorithms. More patients will thus be assigned to the arm with missing data to continue to explore. For algorithms geared towards exploitation, missing data in one arm impedes the decreasing trend in Ik,tI_{k,t} of the corresponding arm, which may thus have a larger index value than the other arm without missing data. Consequently, the arm with missing data is less likely to be selected, and fewer patients will be assigned to it.

3.2 An example of the impact of different types of missing data

In this section, we investigate the impact of having different probabilities of missingness in the two arms. Simulation results are based on 10410^{4} independent replications of each algorithm, except for TTS, which has 10310^{3} replications for computational reasons. We consider 36 combinations of missingness probabilities for the two arms, where p0mp_{0}^{\text{m}} and p1mp_{1}^{\text{m}} follow a sequence of discrete values ranging from 0 to 50% (0, 0.1, 0.2, 0.3, 0.4, 0.5). Figure 2 shows the simulation results for E[p]E[p^{*}] of three representative algorithms from Table 1.

The value of E[p]E[p^{*}] in each cell of Figure 2 corresponds to one of the combinations of missingness probabilities. The trial size is fixed at n=200n=200, and the true probabilities of success in this example (p0=p1=0.9p_{0}=p_{1}=0.9) are unrealistic in a healthcare setting, but are chosen to show the impact of missing data on allocations in an extreme case. A more comprehensive simulation study is given in Section 3.3, including scenarios under the alternative.

Refer to caption
Figure 2: Simulation results of E[p]E[p^{*}] for TTS, CB and UCB under the null. Expectations of 10410^{4} replications are taken for CB and UCB and 10310^{3} replications for TTS under different combinations of missingness probabilities, with p0=p1=0.9p_{0}=p_{1}=0.9 and n=200n=200.

As a comparison, fixed randomization always allocates approximately half of the patients to the experimental arm (E[p]=0.5E[p^{*}]=0.5), regardless of the missingness probabilities. Similarly, the values of E[p]E[p^{*}] for TTS under different probabilities of missingness are only slightly affected, as shown in Figure 2. However, this is not the case for CB and UCB, where allocations will be largely affected in the presence of missing data. Even though our assumption of MAR means that the missing data is non-informative, the corresponding impact on allocation results seems to be predictable if we know which arm has more available responses (equally fewer missing responses). To be specific, CB selects the experimental arm when this provides relatively more responses. Approximate 63% patients are assigned to the arm k=1k=1 when (p0m,p1m)=(0, 0.5)(p_{0}^{m},\ p_{1}^{m})=(0,\ 0.5). In contrast, UCB assigns approximate 34% patients to the arm k=1k=1 in the same case with (p0m,p1m)=(0, 0.5)(p_{0}^{m},\ p_{1}^{m})=(0,\ 0.5). One concern is that under the null, this can falsely promote the idea that the experimental arm (e.g., a new drug) is better because more patients are assigned to it and hence it will have relatively more observations (or vice-versa).

3.3 A comprehensive simulation study

3.3.1 Simulation settings

To investigate the impact of different probabilities of missingness, five scenarios under the null (H0:p0=p1H_{0}:\ p_{0}=p_{1}) and seven scenarios under the alternative (HA:p0<p1H_{A}:\ p_{0}<p_{1}) are investigated, as shown in Table 2. Scenarios under the null correspond to settings where the new treatment offers no benefit over control, which is important for clinicians and drug regulators to consider. In machine learning applications, the increase in performance goal fosters a focus on the alternative. In the context of clinical trials and other healthcare applications, simulation studies should consider both the null and alternative scenarios.

The twelve scenarios are taken from Rosenberger et al. [54], which uses the sample size that yields a power of 90% (under the alternative) when using equal allocation. Note that this consideration of power would not imply the same level of power for different response-adaptive rules. The setting of a fixed sample size may not be consistent with machine learning applications, which typically requires a large sample size. However, this setting is common for trial designs in clinical practice.

Simulation results are based on 10410^{4} independent replications (except for TTS, which uses 10310^{3} replications for computational reasons). We investigate 16 combinations of missingness probabilities, namely: (a) Equal probabilities of missingness in each arm (p0m=p1m=pmp_{0}^{\text{m}}=p_{1}^{\text{m}}=p^{\text{m}}), corresponding to the diagonal in Figure 2; (b) Missingness only occurring in the control arm (p0m{0,0.1,0.2,0.3,0.4,0.5},p1m=0p_{0}^{\text{m}}\in\{0,0.1,0.2,0.3,0.4,0.5\},\ p_{1}^{\text{m}}=0), corresponding to the leftmost column in Figure 2; (c) Missingness only occurring in the experimental arm (p1m{0,0.1,0.2,0.3,0.4,0.5},p0m=0p_{1}^{\text{m}}\in\{0,0.1,0.2,0.3,0.4,0.5\},\ p_{0}^{\text{m}}=0), corresponding to the bottom row in Figure 2. The problem of both differential and equal rates of missing data in two arms has attracted attention [13]. The impact of the other types of combinations of missing rates in two arms would correspond to interpolating between scenarios (a), (b), and (c).

Table 2: Simulation scenarios.
Scenario (p0p_{0}, p1p_{1}) Trial size nn
S1 (0.10, 0.10) 200
S2 (0.30, 0.30) 200
S3 (0.50, 0.50) 200
S4 (0.70, 0.70) 200
S5 (0.90, 0.90) 200
S6 (0.10, 0.20) 526
S7 (0.10, 0.30) 162
S8 (0.10, 0.40) 82
S9 (0.40, 0.60) 254
S10 (0.60, 0.90) 82
S11 (0.70, 0.90) 162
S12 (0.80, 0.90) 526

3.3.2 Under the null

Figure 3 show the simulation results for E[p]E[p^{*}] under the null. Each column corresponds to a bandit algorithm as described in Table 1. Simulation results of fixed randomization (FR) are also included as a reference. The five rows correspond to five scenarios under the null (H0H_{0}), as indicated in Table 2. The different line colors indicate results of E[p]E[p^{*}] under different kinds of missing data: Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness only in the control arm; Red lines correspond to missingness only in the experimental arm. Note that even though these scenarios are under the null, we distinguish the two arms (k=0k=0 and k=1k=1) with different colors.

When there is no missing data (i.e., when the probability of missingness is equal to zero), half of the patients are assigned to the experimental arm across all algorithms and all scenarios. With equal missingness, there is almost no impact on E[p]E[p^{*}] under the null. In contrast, the impact due to missing data occurring in one arm (p0m>0p_{0}^{\text{m}}>0 or p1m>0p_{1}^{\text{m}}>0) can be large, as shown in Figure 3 and described in more detail below.

Refer to caption
Figure 3: Simulation results under the null. Simulation results of E[p]E[p^{*}] under the null for different missing data combinations. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm.
  • TTS and RTS: these randomized algorithms are hardly affected by the impact of missing data occurring in one arm (p0m>0p_{0}^{\text{m}}>0, or p1m>0p_{1}^{\text{m}}>0), which is similar to the results of the standard FR. For RTS, the more exploitative nature introduced by the fixed tuning parameter c=1c=1 makes the allocation results slightly more affected by missing data under extreme scenarios (i.e., S1 (0.1, 0.1) and S5 (0.9, 0.9)) than TTS.

  • GI, UCB, RBI and RGI: are explorative algorithms as a result of a randomized term or the uncertainty taken into account during allocations. As mentioned in Section 3.1, the index values Ik,tI_{k,t} may stop decreasing in the arm with missing responses. As a result, higher Ik,tI_{k,t} values in the arm with missing data can be expected, and more patients will be assigned to this arm for the purpose of exploration. This conclusion applies to all scenarios under the null.

    GI is an exception where the magnitude of the impact of missing data varies across different scenarios, since uncertainty towards the future varies under different pkp_{k} values. Specifically, exploration dominates exploitation in most cases, while more exploitation and less exploration could be expected under larger pkp_{k} since GI is a deterministic algorithm. The allocation results are thus hardly affected by missing data under S5 (0.9, 0.9), where there is a balance of exploration and exploitation with the existence of missing data.

    By contrast, another deterministic algorithm that achieves exploration by taking uncertainty into account, UCB, performs differently from GI with the existence of missing data. Specifically, the goal of exploration is always dominant in UCB, and we can see consistent results under different scenarios. As a consequence, UCB assigns more patients to the arm with missing data. The allocation results can thus be strongly affected by missing data (steeper lines in Figure 3) when compared with semi-randomized algorithms (i.e., RBI and RGI) as well as randomized algorithms (i.e., TTS and RTS).

  • RPW, CB and RandUCB: are algorithms focusing more on exploitation. This exploitative feature is not that obvious with a small value of pkp_{k}, as there is a relatively long period before selection. By contrast, an early selection occurs with large values of pkp_{k}. In the case of small pkp_{k}, more allocations are required to the arm with missing values to decide on the selection. In the other case of large pkp_{k} with early selections, if missing data occurs in an arm, the algorithm may opt to select the other arm instead. Consequently, the arm with missing data is less likely to be selected, and thus patients are less likely to be assigned to the corresponding arm.

    Note that under scenario S3 (0.5, 0.5), E[p]E[p^{*}] is hardly affected by missingness. This is because of the balance between the impact of missing data before selecting the superior arm (exploration phase) and after selecting the superior arm (exploitation phase). For RandUCB, the values of E[p]E[p^{*}] depend heavily on the parameter MM, and as MM changes, the results would become similar to either TTS or UCB [66]. In the example displayed in Figure 3, the choice of M=20M=20 means that RandUCB is more exploitative, which is more clear in Figure 1.

3.3.3 Under the alternative

Results for scenarios under the alternative are shown in Figure 4, where the experimental arm (k=1k=1) has larger true probabilities of success pkp_{k} than the control arm (k=0k=0). More than half of the patients will be assigned to the superior arm when there is no missing data (E[p]>0.5E[p^{*}]>0.5), except for the standard FR. This is one of the motivations for implementing response-adaptive designs in clinical trials. In most of the scenarios, RTS, GI, and RandUCB assign more than 80% of patients to the experimental arm. Similarly, TTS, UCB, RBI, and RGI are algorithms with relatively robust results across different scenarios. Approximately 70%-80% of patients are assigned to the experimental arm, regardless of how large pkp_{k} are. The allocation results of CB depend heavily on the values of pkp_{k}. For example, more than 80% of patients are assigned to the experimental arm in S6 (0.1, 0.2), while approximately 60% of patients are assigned to the experimental arm in S12 (0.8, 0.9). This variation across different scenarios can also be seen for RPW, which allocates relatively more patients to the experimental arm in a scenario with small pkp_{k} values.

In the presence of missing data, fewer patients will be assigned to the superior arm in general, when compared with when there is no missing data. Namely, we could observe a generally decreasing trend in all three lines corresponding to different kinds of missing data in most cases in Figure 4. This is because the presence of treatment differences always requires more patients to make the right decision regarding assignments, namely, to select the better arm. More details are given below.

Refer to caption
Figure 4: Simulation results under the alternative. Simulation results of E[p]E[p^{*}] under the alternative for different missing data combinations. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm.
  • TTS and RTS: are slightly affected by missing data in a consistent way under the alternative scenarios. That is, fewer patients will be assigned to the superior arm no matter if the missing data occurs in both arms or only one of these arms. This feature allows the impact of missing data on TTS and RTS to be predictable in practice.

  • GI, UCB, RBI and RGI: are largely affected by missing data in a consistent way under the alternative scenarios. Specifically, more missing data in the inferior arm (p0m0p_{0}^{m}\geq 0) results in more allocations to this arm and fewer allocations to the superior arm, as indicated by decreasing blue lines. For UCB, which is most affected by missing data, a high probability of missing data occurring in the control arm (p0m=0.5p^{\text{m}}_{0}=0.5) can even reduce the proportion of patients assigned to the experimental arm to approximately half of the patients under S6 (0.1, 0.2). In comparison, RBI and RGI are always less affected by missingness in the control arm. GI seems to be relatively robust in terms of the impact due to missing data, with more than 80% of patients assigned to the experimental arm in all scenarios.

    In terms of missingness in the experimental arm, the impact is less obvious than that in the control. It is even hard to observe the impact of the missingness in the experimental arm in most cases in Figure 4. The exception is that UCB allocates more patients to the experimental arm with a larger p1mp_{1}^{\text{m}}. Lastly, the impact of equal missingness in both arms is in-between the impact of missingness in only one arm. This indicates that equal probabilities of missing data is only ignorable when there is no treatment difference.

  • CB, RPW and RandUCB: perform differently across the scenarios since their deterministic nature depends on pkp_{k}. In scenarios with small pkp_{k} (S6 (0.1. 0.2) – S8 (0.1, 0.4)), even though there is a slight reduction in E[p]E[p^{*}] due to the impact of missing data in the control arm, more than 80% of patients are still assigned to the experimental arm. However, in scenarios with large pkp_{k} (S10 (0.6, 0.9) – S12 (0.8, 0.9)), where an early selection occurs as a result of exploitation, these algorithms perform worse than in the scenarios with small pkp_{k}. In the extreme cases of CB and RPW, less than 50% of patients are assigned to the experimental arm when there is a large probability of missing data in the experimental arm in S12 (0.8, 0.9). That is, the inferior arm might be incorrectly selected at the beginning of a trial, while the experimental arm never gets a chance to be explored after this incorrect selection.

In general, bandit algorithms are more likely to be affected by missing data than the standard FR. The existence of missing data can slow the progress of exploration in the corresponding arm. This means that more patients will ultimately be assigned to the arm with missing data, such that this arm could still be explored. In this case, a smaller impact could be expected with algorithms that have a greater focus on exploration. This corresponds to the fact that randomized algorithms (i.e., TTS and RTS) are less affected than semi-randomized algorithms (i.e., UCB, RBI, and RGI) and deterministic algorithms (i.e., CB, RandUCB, and GI). In contrast, for algorithms that are more exploitative rather than explorative, an early selection of an arm can occur under large pkp_{k}. Consequently, fewer patients will be assigned to the experimental arm if missing data occurs in that arm. This fact warrants careful attention when such algorithms are implemented under large pkp_{k}. E[p]E[p^{*}] values of exploitative algorithms could be unexpectedly small because there are fewer observations in the experimental arm.

3.4 Observed number of successes

In clinical practice, another key metric of interest is the expected number of successes (ENS), namely, the oracle version based on the observed as well as the missing responses. Since in practice, ENS is the expected value of a quantity that will never be fully observed in the presence of missing data, the actual observed number of successes (ONS) might be more relevant to explore. Appendix F and Appendix G illustrate the simulation results for ONS under the twelve scenarios in Section 3.3. Since large numbers of missing values result in fewer successes, equal missingness always results in lower ONS values than missingness that occurred in one arm.

In scenarios under the alternative, the impact of missing data in the two arms is distinguishable from each other, and missing data in the experimental arm (p1m>0p_{1}^{\text{m}}>0) is a more serious problem. Across different algorithms in a given scenario, the impact of missingness in the experimental arm (the red lines in Appendix G) on the ONS varies across the algorithms. For the explorative algorithms, including randomized (i.e., TTS and RTS), semi-randomized (i.e., RBI and RGI), and deterministic algorithms (i.e., GI and UCB), more assignments to the arm with missing data are expected as discussed above. However, the ONS decreases dramatically with larger probabilities of missing data. In extreme cases with p1m=0.5p_{1}^{\text{m}}=0.5, we could even see smaller ONS in these algorithms than in FR. This indicates that assigning more patients to the experimental arm (due to missing data in this arm) cannot compensate for the missed observations. On the contrary, exploitative algorithms (i.e., RPW, CB, and RandUCB) will assign fewer patients to the arm with missing data, which is less likely to be incorrectly selected as the superior arm. It is interesting to see that the trend of decline of ONS with larger values of p1mp_{1}^{\text{m}} is far less dramatic than the explorative algorithms and even FR. This is reasonable considering that incorrect selection of the inferior arm as a result of the missing data could also lead to successful outcomes when p0p_{0} is large. In the extreme case with p1m=0.5p_{1}^{\text{m}}=0.5, we could even see ONS in these algorithms larger than that in FR. This again encourages us to carefully consider the implementation of bandit algorithms in the presence of missing data.

4 Imputation results for the missing responses

To mitigate the impact due to missing data, we consider the use of mean imputation. Mean imputation is a kind of single imputation that replaces missing values with the mean value of that variable [28, 45]. Considering outcomes are missing at random, we intend to investigate how a simple imputation technique would change the impact of missingness in terms of the allocation results of bandit algorithms. For each patient at time tt, we compute an estimate p^k,t=Sk,tSk,t+Fk,t\hat{p}_{k,t}=\frac{S_{k,t}}{S_{k,t}+F_{k,t}} of the true probability of success pkp_{k}. A random value is then drawn from a Bernoulli distribution with estimated probability of success p^k,t\hat{p}_{k,t} to replace the missing outcome. In the case of Sk,t+Fk,t=0S_{k,t}+F_{k,t}=0, the corresponding p^k,t\hat{p}_{k,t} is not available. We instead use a default initial value p^k,0=0.5\hat{p}_{k,0}=0.5, indicating that the treatment has an a-priori success probability of 50% in this case. The allocation procedure with bandit algorithms after the missing data is imputed is given in Appendix B.

4.1 Under the null

Figure 5 shows the simulation results of E[p]E[p^{*}] for scenarios under the null both with and without mean imputation for missing data. We use solid lines to represent the results without mean imputation, while dashed lines represent the results with mean imputation. For equal probabilities of missing data per arm, the results of E[p]E[p^{*}] are hardly affected by missing data. That is, approximately half of the patients are assigned to the experimental arm on average, and the mean imputation approach does not make any differences. In the cases of missingness in only one arm, imputation works differently in terms of the inherent exploitation-exploration trade-off and specific scenarios, as discussed below.

Refer to caption
Figure 5: Imputation results under the null. Imputation results of E[p]E[p^{*}] under the null for different missing data combinations with initial value p^k,0=0.5\hat{p}_{k,0}=0.5. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.
  • Exploration: more explorative algorithms, including randomized algorithms (i.e., TTS and RTS), semi-randomized algorithms (i.e., RBI and RGI), and deterministic algorithms (i.e., UCB), assign more patients to the arm with missing data. Mean imputation could largely solve this problem, and this applies across all scenarios (S1 (0.1, 0.1) – S5 (0.9, 0.9)), with the exception of RTS in S5 (0.9, 0.9), where the exploitative nature dominates the explorative nature as explained in Section 3.3.

  • Exploitation: In scenarios with small pkp_{k} (i.e., S1 (0.1, 0.1) and S2 (0.3, 0.3)), there is no early selection. In this case, we observe a similar but limited impact of missingness as in the case of explorative algorithms. That is, there are more allocations in the arm with missing data. Consequently, mean imputation could help with allocation skew in the way it works for the explorative algorithms. However, this is not the case under large pkp_{k} (i.e., S4 (0.7, 0.7) and S5 (0.9, 0.9)), where an early selection occurs and the algorithms exhibit great determinism in the allocation procedure. In this case, the impact due to missing data cannot be mitigated by mean imputation, and sometimes it even exaggerates this impact. For instance, GI assigns fewer patients to the experimental arm with imputation for p1mp_{1}^{\text{m}} in S5 (0.9, 0.9), which is even worse than without imputation.

4.2 Under the alternative

Figure 6 illustrates simulation results of E[p]E[p^{*}] for scenarios under the alternative both with and without mean imputation for missing data. In general, since more patients are required to make the right allocation decision (i.e., select the superior arm) under the alternative, relatively fewer patients are assigned to the experimental arm in the presence of missing data when compared with results in a same scenario without missing data (as discussed in Section 3.3.3).

  • Exploration: the impact of missing data could largely be mitigated for randomized (i.e., TTS and RTS), semi-randomized (i.e., RBI and RGI) and deterministic algorithms (i.e., UCB) as for scenarios under the null in Figure 5. One exception is RTS in scenario S12 (0.8, 0.9), where it exhibits a more exploitative nature due to the fixed tuning parameter.

  • Exploitation: mean imputation does not work well to mitigate the impact due to missing data, especially when there is an early selection with larger pkp_{k}. In this case, missing data occurring in the experimental arm is a serious problem with a substantial reduction of assigned patients to this arm. GI is a special case that is less likely to be affected by missing data across all of these scenarios. A particular concern with GI is that mean imputation in a scenario with large pkp_{k} (S10 (0.6, 0.9) - S12 (0.8, 0.9)) could even greatly exaggerate the impact of missing data. In S12 (0.8, 0.9), with the existence of a large proportion of missing data (pm=0.5p_{m}=0.5 or p1m=0.5p_{1}^{m}=0.5), approximately 90% of patients will be assigned to the experimental arm, while this drops to only 80% after imputation.

    Refer to caption
    Figure 6: Imputation results under the alternative. Imputation results of E[p]E[p^{*}] under the alternative for different missing data combinations initial value p^k,0=0.5\hat{p}_{k,0}=0.5. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.

4.3 The impact of biased estimates on mean imputation

The implemented mean imputation approach is based on the currently estimated probabilities of success p^k,t\hat{p}_{k,t}, which is known to be biased in the context of adaptive sampling [67, 17, 46, 57]. Villar et al.[67] empirically demonstrate the presence of negative bias of the sample means for a number of multi-armed bandit algorithms in a finite sample setting via a simulation study. Nie et al. [46] formally provide two conditions called ‘Exploit’ and ‘Independence of Irrelevant Options’ (IIO) under which the negative bias of sampling mean exist. Shin et al.[57] provides a more general and comprehensive characterization of the sign of the bias of the sample mean in multi-armed bandits, which discuss the same context as ours (adaptive sampling in finite sample) and some other cases (adaptive stopping and adaptive selection).

To interpret the results of using mean imputation of missing data in this setting, the magnitude of the bias matters more than the sign of the bias. Bowden and Trippa [17] derived an exact formula to quantify the bias for randomised data dependent rules in a finite sample as shown in Eq (1). This simple characterisation indicates that the bias increases in magnitude as the dependency of p^k,t\hat{p}_{k,t} and Nk,tN_{k,t} increases. For this reason, FR guarantying a zero covariance between these two is expected to have zero bias of sample mean estimate. In contrast, in the common case where optimistic adaptive randomization is used to direct more allocations towards arms that appear to work well, Nk,tN_{k,t} will be variable and positively correlated with p^k,t\hat{p}_{k,t}. Notice that the derivation of the bias in [17] applies to randomized algorithms, while [57] then extends this formula to more general cases of MABP (i.e. non-randomized algorithms).

Cov[Nk,t,p^k,t]E[Nk,t]=pkE[p^k,t]=Bias(p^k,t)\displaystyle\frac{Cov[N_{k,t},\ \hat{p}_{k,t}]}{E[N_{k,t}]}=p_{k}-E[\hat{p}_{k,t}]=-\text{Bias}(\hat{p}_{k,t}) (1)

Consequently, we could expect a larger negative bias under the exploitation feature where there is a large covariance between p^k,t\hat{p}_{k,t} and Nk,tN_{k,t} because of the rather myopically ‘optimistic’ sampling behaviour. In contrast, algorithms with more exploration features will see relatively smaller bias and be less reactive to current mean estimates. This aligns with the simulation results in Fig 4 in [67]. Similar simulation results for bias of the investigated algorithms in this paper are attached in Appendix L for scenarios under the null and in Appendix M for scenarios under the alternative. It is substantially more complicated to discuss the negative bias of treatment effect estimates accounting for the three different missing data combinations. However, the first intuitive conclusion we can make is that larger bias is expected in algorithms with a great amount of short term exploitation. For instance, CB and RandUCB show large negative bias in all scenarios. RTS and GI seem to be more biased when the treatment effect is larger (e.g., S4 (0.7, 0.7) – S5 (0.9, 0.9) and S9 (0.4, 0.6) – S12 (0.8, 0.9)). The magnitude of the negative bias is relatively small in TS, UCB, RBI, and RGI, which are relatively more explorative within the size of the experiment. Further investigation of how the bias varies between the control and treatment arms (under the alternative), as shown in Appendix M, is another interesting direction for research for in the future.

Consequently, for more explorative algorithms, a single success or failure does not have a large impact on the value of Ik,(t+1)I_{k,(t+1)} or πk,(t+1)\pi_{k,(t+1)} of the subsequent allocation. Besides, the negative bias of p^k,t\hat{p}_{k,t} is relatively small for these algorithms. Thus, mean imputation based on p^k,t\hat{p}_{k,t} could work to mitigate the impact due to missing data. In contrast, for the exploitative algorithms, an early selection could be substantially affected by a single success or failure. In addition, the negative bias of p^k,t\hat{p}_{k,t} is relatively large in these cases. As a consequence, imputation based on an inaccurate estimate p^k,t\hat{p}_{k,t} may fail to mitigate the impact of missing data on exploitation. In other words, mean imputation with missing data replaced by a single value does not work for more short term exploitative algorithms that are highly sensitive to the success or failure of each allocation.

We also tried conducting imputation in two settings: 1. with a large initial value p^k,0=0.9\hat{p}_{k,0}=0.9 (see Appendix H and Appendix I); and 2. only after the first observation (see Appendix J and Appendix K). In either of these cases, the problem of biased estimates lead to undesirable characteristics of E[p]E[p*].

Since the simple mean imputation based on biased estimates works differently under exploitation and exploration features, some adjustments of the balance of the exploration-exploitation trade-off might help. For instance, imputation for missing data might work for RandUCB with a different turning parameter (e.g., using a larger value of MM), which could introduce more ‘randomization’. Besides, the idea of an additional perturbed component to the current estimate of the reward provides a similar way of thinking about exploration. In this sense, a randomized version of deterministic algorithms (i.e., RBI to CB, RGI to GI and RandUCB to UCB) should be amenable to an imputation approach. Apart from this, taking uncertainty into account in the computation of Ik,tI_{k,t} also supports exploring both arms, and the idea of deterministic algorithms GI and UCB aligns with this when compared with the myopic and greedy algorithm CB. In these cases, a simple mean imputation could work for the impact of missing data. This idea of more exploration in deterministic algorithms has gained popularity in related domains [61, 2, 39].

Given the problem of negative bias of the treatment effect estimates, there have been a number of proposals in the literature showing to obtain unbiased or bias-reduced estimates. Nie et al. [46] explore data splitting and a conditional MLE as two approaches to reduce this bias, using considerable information about (and control over) the data generating process. Meanwhile, Deshpande et al. [25] use adaptive weights to form the ‘W-decorrelated’ estimator, a debiased version of OLS. They attempt to use the data at hand, along with coarse aggregate information on the exploration inherent in the data generating process. In addition, Dimakopoulou et al. [26] propose the Doubly-Adaptive Thompson Sampling, which uses the strengths of adaptive inference to ensure sufficient exploration in the initial stages of learning and the effective exploration-exploitation balance provided by the TS mechanism. In this paper, we did not try these novel debiasing approaches to obtain unbiased estimates for imputation. However, these proposals provide natural future directions to improve the performance of imputation in the presence of missing data.

In conclusion, mean imputation could reduce the problem of oversampling the arm with more missing data, which is a result of the explorative nature of some algorithms. However, it does not work for the problem of undersampling the arm with more missing data in the case of short term exploitative algorithms. The reason is that for these algorithms that perform differently in various scenarios, the impact of missing data varies according to this sampling behaviour. Since information about the true scenario is not accessible in reality, it is hard to use a simple mean or some other single imputation approach to mitigate the impact of missing data.

5 Conclusion

The problem of missing data in the implementation of machine learning algorithms in contexts such as healthcare applications is commonly overlooked. We demonstrate that bandit algorithms are not robust in the presence of missing data in terms of expected performance, e.g. in terms of expected patient outcomes. The omission of this practical issue, as well as the naive approach of ignoring it in related works, may prevent the realization of the potential advantages of widespread use of bandit algorithms in healthcare settings, including clinical trial designs. Through a simulation study, we have investigated the impact of missing data on randomized, semi-randomized, and deterministic bandit algorithms for scenarios under the null and the alternative. In terms of machine learning implementations, scenarios under the null are often ignored. However, these results under the null allow us to focus on the impact of missing data and isolate the impact of treatment differences on allocation results. In addition, it suggests that different scenarios need to be taken into account when choosing an algorithm in healthcare implementations. Specifically, for scenarios under the alternative, the potential consequences could be serious as the presence of missing data may reverse our preferences for the experimental or superior arm and opt for the control or inferior arm in extreme cases.

In this paper we focus on a practical setting of an experiment with a finite number of patients. The setting of a finite number of actions and outcomes has also gained attention in some other machine learning applications [10]. The investigation of different combinations of missing probabilities in two arms (i.e., equal missingness or missingness occurring in only one arm) should be given more attention. For scenarios under the null, the impact of equal missingness probabilities in both arms can be simply ignored. The practical implication is that if there is non-informative missing data (i.e., MAR) and even if the true success rates pkp_{k} are unknown, we can infer what the impact of missingness on different bandit algorithms will be. This applies to scenarios under the null and the alternative in general. Most importantly, there is a distinction between the results for explorative and exploitative algorithms, which corresponds to the inherent exploration-exploitation trade-off in a finite sample. Since deterministic algorithms are sometimes preferred over randomized algorithms from the perspective of quality of care [51, 65, 23], our findings show the need for careful thinking about the potential impact of missing data. More generally, incorporating randomization, perturbation, or accounting for uncertainty in deterministic algorithms for the aim of exploration can bring some advantages, see further discussions of optimism in the field of online learning [61, 2, 39].

The general rule guiding the impact of missing data on bandit algorithms, considering all the factors mentioned above, is in the way to balance exploration and exploitation. Explorative algorithms encourage more assignments in the arm with missing data (being optimistic in the face of uncertainty as less sample arms will have larger variance), while exploitative algorithms are less likely to select the arm with missing data (being more myopic and reacting more to the current mean than to the current variance). In the latter case of exploitative algorithms, this impact consequently depends on whether an early selection occurs and hence the success rates pkp_{k} in both arms. The impact of missing data on the use of machine learning in healthcare has not received much attention so far, especially in settings other than clinical trial designs. An important message of this paper is a call for actively considering the impact of missing data when selecting a bandit algorithm for a specific application in which missing data is likely to occur.

Instead of simply ignoring missing data, imputation replaces missing values in the response-adaptive procedure. Assuming outcomes are MAR, we have investigated conducting mean imputation based on the current estimated probability of success p^k,t\hat{p}_{k,t}. Our results show that the mean imputation approach supports the exploration procedure in heavily explorative algorithms, so that the impact of missing data on explorative algorithms could be mitigated by imputation. Specifically, there would be no additional assignments in the arm with missing data if imputation is implemented. This applies to the cases of explorative algorithms, including randomized algorithms (i.e., TTS and RTS), semi-randomized algorithms (i.e., RBI and RGI), and deterministic algorithms (i.e., UCB). By contrast, in the case of more exploitative algorithms, the impact of missing data is different for scenarios with small or large success rates (pkp_{k}). As a consequence, the problem of missing data could not be alleviated by a mean imputation approach based on current estimates p^k,t\hat{p}_{k,t}. Worse still, in some extreme cases, the impact of missing data could even be exaggerated. This is especially obvious with large treatment effects pkp_{k}. The negative bias that has been well recognized and proved for MABPs also helps to explain the imputation results. That is, exploitative algorithms are associated with a large covariance between the number of allocations in one specific arm and the corresponding estimated probability of success, and thus exhibit a large negative bias of less sampled arms. As a result, mean imputation can be based on biased estimates which does not alleviate the issues caused by the presence of missing data.

We note that the conclusions about imputation results are limited to the assumption of MAR, which is untestable. In cases where the MAR assumption does not hold, which is a common setting in machine learning applications, algorithms may not be able to avoid selection bias. This is of particular importance in many areas of machine learning applications. Hence, the assumption underlies our conclusions where fairness is a concern, which is an emerging area in machine learning of how to ensure that the outputs of a model do not depend on sensitive attributes in a way that is considered unfair [21, 8]. This leads to an important caveat of improper representation and exclusion of fairness issues in the implementation of machine learning algorithms. As future research, we could extend the current single imputation approach to more advanced methods of imputation [32, 16] or more complex missing data settings, such as when patient covariates are taken into account. In addition, rather than only considering a single approach to the missing data problem, appropriate forms of sensitivity analysis are important and necessary to assess the validity of these assumptions and the robustness of the results [62].

Finally, discussions about delayed outcomes have received increasing attention both in the community of machine learning [36, 42] and biostatistics [73, 74, 77] in recent years. In this paper, we assumed that all of the missing outcomes would never become available. This missing data problem could be viewed as an extreme form of delay, and an interesting avenue for future work would be to account for both missing data and delayed outcomes together.

References

  • [1] Abbasi-Yadkori, Y., Pál, D., and Szepesvári, C. Improved algorithms for linear stochastic bandits. Advances in neural information processing systems 24 (2011), 2312–2320.
  • [2] Abernethy, J., Lee, C., Sinha, A., and Tewari, A. Online linear optimization via smoothing. In Conference on Learning Theory (2014), PMLR, pp. 807–823.
  • [3] Agarwal, A., and Duchi, J. C. Distributed delayed stochastic optimization. In 2012 IEEE 51st IEEE Conference on Decision and Control (CDC) (2012), IEEE, pp. 5451–5452.
  • [4] Atkinson, A. C., and Biswas, A. Randomised response-adaptive designs in clinical trials. CRC Press, 2013.
  • [5] Auer, P., Cesa-Bianchi, N., and Fischer, P. Finite–time analysis of the multiarmed bandit problem. Machine learning 47, 2 (2002), 235–256.
  • [6] Aziz, M., Kaufmann, E., and Riviere, M.-K. On multi-armed bandit designs for dose-finding clinical trials. Journal of Machine Learning Research 22 (2021), 1–38.
  • [7] Barker, A., Sigman, C., Kelloff, G., Hylton, N., Berry, D., and Esserman, L. I-spy 2: an adaptive breast cancer trial design in the setting of neoadjuvant chemotherapy. Clinical Pharmacology & Therapeutics 86, 1 (2009), 97–100.
  • [8] Barocas, S., Hardt, M., and Narayanan, A. Fairness in machine learning. Nips tutorial 1 (2017), 2.
  • [9] Bartlett, R. H., Roloff, D. W., Cornell, R. G., Andrews, A. F., Dillon, P. W., and Zwischenberger, J. B. Extracorporeal circulation in neonatal respiratory failure: a prospective randomized study. Pediatrics 76, 4 (1985), 479–487.
  • [10] Bartók, G., Foster, D. P., Pál, D., Rakhlin, A., and Szepesvári, C. Partial monitoring—classification, regret bounds, and algorithms. Mathematics of Operations Research 39, 4 (2014), 967–997.
  • [11] Bastani, H., and Bayati, M. Online decision making with high-dimensional covariates. Operations Research 68, 1 (2020), 276–294.
  • [12] Bather, J. Randomized allocation of treatments in sequential experiments. Journal of the Royal Statistical Society: Series B (Methodological) 43, 3 (1981), 265–283.
  • [13] Bell, M. L., Kenward, M. G., Fairclough, D. L., and Horton, N. J. Differential dropout and bias in randomised controlled trials: when it matters and when it may not. Bmj 346 (2013).
  • [14] Biswas, A., and Rao, J. Missing responses in adaptive allocation design. Statistics & probability letters 70, 1 (2004), 59–70.
  • [15] Bouneffouf, D. Corrupted contextual bandits: Online learning with corrupted context. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2021), IEEE, pp. 3145–3149.
  • [16] Bouneffouf, D., Upadhyay, S., and Khazaeni, Y. Contextual bandit with missing rewards. arXiv preprint arXiv:2007.06368 (2020).
  • [17] Bowden, J., and Trippa, L. Unbiased estimation for response adaptive clinical trials. Statistical methods in medical research 26, 5 (2017), 2376–2388.
  • [18] Burton, P. R., Gurrin, L. C., and Hussey, M. H. Interpreting the clinical trials of extracorporeal membrane oxygenation in the treatment of persistent pulmonary hypertension of the newborn. In Seminars in Neonatology (1997), vol. 2, Elsevier, pp. 69–79.
  • [19] Chapelle, O., and Li, L. An empirical evaluation of thompson sampling. Advances in neural information processing systems 24 (2011), 2249–2257.
  • [20] Chen, I. Y., Joshi, S., Ghassemi, M., and Ranganath, R. Probabilistic machine learning for healthcare. Annual Review of Biomedical Data Science 4 (2020).
  • [21] Chien, I., Deliu, N., Turner, R., Weller, A., Villar, S., and Kilbertus, N. Multi-disciplinary fairness considerations in machine learning for clinical trials. In 2022 ACM Conference on Fairness, Accountability, and Transparency (2022), pp. 906–924.
  • [22] Chow, S.-C., and Chang, M. Adaptive design methods in clinical trials–a review. Orphanet journal of rare diseases 3, 1 (2008), 1–13.
  • [23] Demirel, I., Celik, A. A., and Tekin, C. Escada: Efficient safety and context aware dose allocation for precision medicine. arXiv preprint arXiv:2111.13415 (2021).
  • [24] Desautels, T., Krause, A., and Burdick, J. W. Parallelizing exploration-exploitation tradeoffs in gaussian process bandit optimization. Journal of Machine Learning Research 15 (2014), 3873–3923.
  • [25] Deshpande, Y., Mackey, L., Syrgkanis, V., and Taddy, M. Accurate inference for adaptive linear models. In International Conference on Machine Learning (2018), PMLR, pp. 1194–1203.
  • [26] Dimakopoulou, M., Ren, Z., and Zhou, Z. Online multi-armed bandits with adaptive inference. Advances in Neural Information Processing Systems 34 (2021).
  • [27] Dudik, M., Hsu, D., Kale, S., Karampatziakis, N., Langford, J., Reyzin, L., and Zhang, T. Efficient optimal learning for contextual bandits. arXiv preprint arXiv:1106.2369 (2011).
  • [28] Dziura, J. D., Post, L. A., Zhao, Q., Fu, Z., and Peduzzi, P. Strategies for dealing with missing data in clinical trials: from design to analysis. The Yale journal of biology and medicine 86, 3 (2013), 343.
  • [29] Gittins, J. A dynamic allocation index for the sequential design of experiments. Progress in statistics (1974), 241–266.
  • [30] Gittins, J., and Wang, Y.-G. The learning component of dynamic allocation indices. The Annals of Statistics (1992), 1625–1636.
  • [31] Glazebrook, K. On randomized dynamic allocation indices for the sequential design of experiments. Journal of the Royal Statistical Society: Series B (Methodological) 42, 3 (1980), 342–346.
  • [32] Graham, J. W. Missing data analysis: Making it work in the real world. Annual review of psychology 60 (2009), 549–576.
  • [33] Grover, A., Al-Shedivat, M., Gupta, J., Burda, Y., and Edwards, H. Learning policy representations in multiagent systems. In International conference on machine learning (2018), PMLR, pp. 1802–1811.
  • [34] Jacko, P. The finite-horizon two-armed bandit problem with binary responses: A multidisciplinary survey of the history, state of the art, and myths. arXiv preprint arXiv:1906.10173 (2019).
  • [35] Jin, I. H., Liu, S., Thall, P. F., and Yuan, Y. Using data augmentation to facilitate conduct of phase i–ii clinical trials with delayed outcomes. Journal of the American Statistical Association 109, 506 (2014), 525–536.
  • [36] Joulani, P., Gyorgy, A., and Szepesvári, C. Online learning under delayed feedback. In International Conference on Machine Learning (2013), PMLR, pp. 1453–1461.
  • [37] Kalvit, A., and Zeevi, A. A closer look at the worst-case behavior of multi-armed bandit algorithms. Advances in Neural Information Processing Systems 34 (2021).
  • [38] Kaufmann, E. On bayesian index policies for sequential resource allocation. The Annals of Statistics 46, 2 (2018), 842–865.
  • [39] Kim, B., and Tewari, A. On the optimality of perturbations in stochastic and adversarial multi-armed bandit problems. arXiv preprint arXiv:1902.00610 (2019).
  • [40] Kim, M.-O., Liu, C., Hu, F., and Lee, J. J. Outcome-adaptive randomization for a delayed outcome with a short-term predictor: imputation-based designs. Statistics in medicine 33, 23 (2014), 4029–4042.
  • [41] Kveton, B., Szepesvari, C., Ghavamzadeh, M., and Boutilier, C. Perturbed-history exploration in stochastic linear bandits. arXiv preprint arXiv:1903.09132 (2019).
  • [42] Lancewicki, T., Segal, S., Koren, T., and Mansour, Y. Stochastic multi-armed bandits with unrestricted delay distributions. arXiv preprint arXiv:2106.02436 (2021).
  • [43] Lee, K. M. Optimal designs in the presence of missing responses. PhD thesis, University of Southampton, 2016.
  • [44] Ma, Z. Missing data and adaptive designs in clinical studies. University of Virginia, 2013.
  • [45] Newman, D. A. Missing data: Five practical guidelines. Organizational Research Methods 17, 4 (2014), 372–411.
  • [46] Nie, X., Tian, X., Taylor, J., and Zou, J. Why adaptively collected data have negative bias and how to correct for it. In International Conference on Artificial Intelligence and Statistics (2018), PMLR, pp. 1261–1269.
  • [47] Niño-Mora, J. Dynamic priority allocation via restless bandit marginal productivity indices. Top 15, 2 (2007), 161–198.
  • [48] Pilarski, S., Pilarski, S., and Varro, D. Delayed reward bernoulli bandits: Optimal policy and predictive meta-algorithm pardi. IEEE Transactions on Artificial Intelligence (2021).
  • [49] Press, W. H. Bandit solutions provide unified ethical models for randomized clinical trials and comparative effectiveness research. Proceedings of the National Academy of Sciences 106, 52 (2009), 22387–22392.
  • [50] R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, 2021.
  • [51] Richards, S. Should the nhs strive to eradicate all unexplained variation? yes. BMJ 339 (2009).
  • [52] Robertson, D. S., Lee, K. M., Lopez-Kolkovska, B. C., and Villar, S. S. Response-adaptive randomization in clinical trials: from myths to practical considerations. arXiv preprint arXiv:2005.00564 (2020).
  • [53] Rosenberger, W. F., and Lachin, J. M. The use of response-adaptive designs in clinical trials. Controlled clinical trials 14, 6 (1993), 471–484.
  • [54] Rosenberger, W. F., Stallard, N., Ivanova, A., Harper, C. N., and Ricks, M. L. Optimal adaptive designs for binary response trials. Biometrics 57, 3 (2001), 909–913.
  • [55] Rubin, D. B. Inference and missing data. Biometrika 63, 3 (1976), 581–592.
  • [56] Scott, I., Carter, S., and Coiera, E. Clinician checklist for assessing suitability of machine learning applications in healthcare. BMJ Health & Care Informatics 28, 1 (2021).
  • [57] Shin, J., Ramdas, A., and Rinaldo, A. Are sample means in multi-armed bandits positively or negatively biased? Advances in Neural Information Processing Systems 32 (2019).
  • [58] Shrestha, S., and Jain, S. A bayesian-bandit adaptive design for n-of-1 clinical trials. Statistics in Medicine 40, 7 (2021), 1825–1844.
  • [59] Slivkins, A. Introduction to multi-armed bandits. arXiv preprint arXiv:1904.07272 (2019).
  • [60] Smith, A. L., and Villar, S. S. Bayesian adaptive bandit-based designs using the gittins index for multi-armed trials with normally distributed endpoints. Journal of applied statistics 45, 6 (2018), 1052–1076.
  • [61] Suggala, A., and Netrapalli, P. Follow the perturbed leader: Optimism and fast parallel algorithms for smooth minimax games. Advances in Neural Information Processing Systems 33 (2020), 22316–22326.
  • [62] Sverdlov, O. Modern adaptive randomized clinical trials: statistical and practical aspects, vol. 81. CRC Press, 2015.
  • [63] Thall, P. F., and Wathen, J. K. Practical bayesian adaptive randomisation in clinical trials. European Journal of Cancer 43, 5 (2007), 859–866.
  • [64] Thompson, W. R. On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 25, 3/4 (1933), 285–294.
  • [65] Tomson, C. R., and Van Der Veer, S. N. Learning from practice variation to improve the quality of care. Clinical medicine 13, 1 (2013), 19.
  • [66] Vaswani, S., Mehrabian, A., Durand, A., and Kveton, B. Old dog learns new tricks: Randomized ucb for bandit problems. arXiv preprint arXiv:1910.04928 (2019).
  • [67] Villar, S. S., Bowden, J., and Wason, J. Multi-armed bandit models for the optimal design of clinical trials: benefits and challenges. Statistical science: a review journal of the Institute of Mathematical Statistics 30, 2 (2015), 199.
  • [68] Villar, S. S., Wason, J., and Bowden, J. Response-adaptive randomization for multi-arm clinical trials using the forward looking gittins index rule. Biometrics 71, 4 (2015), 969–978.
  • [69] Wei, L., and Durham, S. The randomized play-the-winner rule in medical trials. Journal of the American Statistical Association 73, 364 (1978), 840–843.
  • [70] Williamson, S. F., Jacko, P., and Jaki, T. Generalisations of a bayesian decision-theoretic randomisation procedure and the impact of delayed responses. Computational Statistics & Data Analysis (2021), 107407.
  • [71] Williamson, S. F., and Villar, S. S. A response-adaptive randomization procedure for multi-armed clinical trials with normally distributed outcomes. Biometrics 76, 1 (2020), 197–209.
  • [72] Zhang, K., Janson, L., and Murphy, S. Inference for batched bandits. Advances in Neural Information Processing Systems 33 (2020), 9818–9829.
  • [73] Zhang, L., and Rosenberger, W. F. Response-adaptive randomization for clinical trials with continuous outcomes. Biometrics 62, 2 (2006), 562–569.
  • [74] Zhang, L., and Rosenberger, W. F. Response-adaptive randomization for survival trials: the parametric approach. Journal of the Royal Statistical Society: Series C (Applied Statistics) 56, 2 (2007), 153–165.
  • [75] Zhang, Y., Cao, S., Zhang, C., Jin, I. H., and Zang, Y. A bayesian adaptive phase i/ii clinical trial design with late-onset competing risk outcomes. Biometrics (2020).
  • [76] Zhou, X., Liu, S., Kim, E. S., Herbst, R. S., and Lee, J. J. Bayesian adaptive design for targeted therapy development in lung cancer—a step toward personalized medicine. Clinical Trials 5, 3 (2008), 181–193.
  • [77] Zhou, Z., Xu, R., and Blanchet, J. Learning in generalized linear contextual bandits with stochastic delays. Advances in Neural Information Processing Systems 32 (2019), 5197–5208.

6 Supporting information

Appendix A Algorithm Appendix.

The allocation procedure with bandit algorithms in the presence of missing data.

Input: trial size nn; probability of success pkp_{k}; probability of missingness pkmp_{k}^{\text{m}}.
Initialization: The initial prior xk,0=(sk,0,fk,0)x_{k,0}=(s_{k,0},\ f_{k,0}). while t<nt<n do
       Decision for allocation based on the current state xk,t=(sk,0+Sk,t,fk,0+Fk,t)x_{k,t}=(s_{k,0}+S_{k,t},\ f_{k,0}+F_{k,t});
       if observed (pkm<u,uU[0,1]p_{k}^{\text{m}}<u,\ u\sim U_{[0,1]}) then
             Mk,tM_{k,t} += 0.
             if failure (pk<u,uU[0,1]p_{k}<u,\ u\sim U_{[0,1]}) then
                   Fk,tF_{k,t} += 1;
                   Sk,tS_{k,t} += 0;
                  
            else
                   Sk,tS_{k,t} += 1;
                   Fk,tF_{k,t} += 0;
                  
             end if
            
      else
            Mk,tM_{k,t} += 1; Fk,tF_{k,t} += 0; Sk,tS_{k,t} += 0;
       end if
      Update xk,tx_{k,t} with (Sk,t,Fk,t)(S_{k,t},\ F_{k,t}) for the next allocation
end while
Output: (Sk,n,Fk,n,Mk,n)(S_{k,n},\ F_{k,n},\ M_{k,n})
Algorithm 1 Implementation of bandit algorithms with missing data

Appendix B Algorithm Appendix.

The allocation procedure with bandit algorithms and imputation for missing data.

Input: trial size nn; probabilities of success pkp_{k}; probabilities of missingness pkmp_{k}^{\text{m}}.
Initialization: The initial prior xk,0=(sk,0,fk,0)x_{k,0}=(s_{k,0},\ f_{k,0}).
while t<nt<n do
       Decision for allocation based on the current state xk,t=(sk,0+Sk,t,fk,0+Fk,t)x_{k,t}=(s_{k,0}+S_{k,t},\ f_{k,0}+F_{k,t});
       if observed (pkm<u,uU[0,1]p_{k}^{\text{m}}<u,\ u\sim U_{[0,1]}) then
             if failure (pk<u,uU[0,1]p_{k}<u,\ u\sim U_{[0,1]}) then
                   Fk,tF_{k,t} += 1;
                   Sk,tS_{k,t} += 0;
                  
            else
                   Sk,tS_{k,t} += 1;
                   Fk,tF_{k,t} += 0;
                  
             end if
            
      else
             Mk,tM_{k,t} += 1; Fk,tF_{k,t} += 0; Sk,tS_{k,t} += 0;
             Imputation based on p^k,t\hat{p}_{k,t};
             if failed (p^k,t<u,uU[0,1]\hat{p}_{k,t}<u,\ u\sim U_{[0,1]}) then
                   Fk,timpF^{\text{imp}}_{k,t} += 1;
                   Sk,timpS^{\text{imp}}_{k,t} += 0;
                  
            else
                   Sk,timpS^{\text{imp}}_{k,t} += 1;
                   Fk,timpF^{\text{imp}}_{k,t} += 0;
                  
             end if
            
       end if
      Update xk,tx_{k,t} with (Sk,t+Sk,timp,Fk,t+Fk,timp)(S_{k,t}+S^{\text{imp}}_{k,t},\ F_{k,t}+F^{\text{imp}}_{k,t}) for the next allocation
end while
Output: (Sk,n+Sk,nimp,Fk,n+Fk,nimp)(S_{k,n}+S^{\text{imp}}_{k,n},\ F_{k,n}+F^{\text{imp}}_{k,n})
Algorithm 2 Implementation of bandit algorithms with missing data and imputation

Appendix C Figure Appendix.

Simulation results of E[p]E[p^{*}] for TTS, CB and UCB under the null. Expectations of 10410^{4} replications are taken for CB and UCB and 10310^{3} replications for TTS under different combinations of missingness probabilities, with p0=p1=0.7p_{0}=p_{1}=0.7 and n=200n=200.

[Uncaptioned image]

Appendix D Figure Appendix.

Simulation results of E[p]E[p^{*}] for TTS, CB and UCB under the alternative. Expectations of 10410^{4} replications are taken for CB and UCB and 10310^{3} replications for TTS under different combinations of missingness probabilities, with p0=0.7p_{0}=0.7, p1=0.9p_{1}=0.9, and n=200n=200.

[Uncaptioned image]

Appendix E Figure Appendix.

Performance of different bandit algorithms over a single simulation under the alternative (p0=0.7p_{0}=0.7, p1=0.9p_{1}=0.9, and n=200n=200). Blue line represent πk,t\pi_{k,t} or Ik,tI_{k,t} in the experimental arm and red lines represent that in the control arm.

[Uncaptioned image]

Appendix F Figure Appendix.

Observed number of success in the presence of missing data under the null. The impact of missing data is similar under different algorithms under the null: the impact due to the equal probabilities of missing data per arm is larger than the impact of only having missingness in the control or experimental arm.

[Uncaptioned image]

Appendix G Figure Appendix.

Observed number of success in the presence of missing data under the alternative. The impact of missing data on ONS varies among different algorithms. Generally, bandit algorithms outperform FR when there is no missing data. Equal missingness is a more serious problem than missingness only occurring in the experimental arm, which in turn is a much more serious problem than missingness only occurring in the control arm.

[Uncaptioned image]

Appendix H Figure Appendix.

Imputation results under the null with initial value p^k,0=0.9\hat{p}_{k,0}=0.9. Imputation results of E[p]E[p^{*}] under the null for different missing data combinations with initial value p^k,0=0.9\hat{p}_{k,0}=0.9. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.

[Uncaptioned image]

Appendix I Figure Appendix.

Imputation results under the alternative with initial value p^k,0=0.9\hat{p}_{k,0}=0.9. Imputation results of E[p]E[p^{*}] under the alternative for different missing data combinations with initial value p^k,0=0.9\hat{p}_{k,0}=0.9. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.

[Uncaptioned image]

Appendix J Figure Appendix.

Imputation results under the null, with imputation starting after the first observation. Imputation results of E[p]E[p^{*}] under the null for different missing data combinations, with imputation starting after the first observation. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.

[Uncaptioned image]

Appendix K Figure Appendix.

Imputation results under the alternative, with imputation starting after the first observation. Imputation results of E[p]E[p^{*}] under the alternative for different missing data combinations, with imputation starting after the first observation. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.

[Uncaptioned image]

Appendix L Figure Appendix.

Bias of treatment effect after n=200n=200 patients. Simulation results of p^k,tpk,t\hat{p}_{k,t}-p_{k,t} under the null for different missing data combinations. We illustrate the result for one of the two equal arms. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm.

[Uncaptioned image]

Appendix M Figure Appendix.

Bias of treatment effect after n=200n=200 patients. Simulation results of p^k,tpk,t\hat{p}_{k,t}-p_{k,t} under the alternative for different missing data combinations. We illustrate the result for one of the two equal arms. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the bias in the treatment arm, while the dashed lines correspond to the bias in the control arm.

[Uncaptioned image]