Transfer learning of individualized treatment rules from experimental to real-world data
Abstract
Individualized treatment effect lies at the heart of precision medicine. Interpretable individualized treatment rules (ITRs) are desirable for clinicians or policymakers due to their intuitive appeal and transparency. The gold-standard approach to estimating the ITRs is randomized experiments, where subjects are randomized to different treatment groups and the bias is minimized to the extent possible. However, experimental data are limited in external validity because of their selection restrictions and therefore are not representative of the target real-world population. Conventional learning methods of optimal interpretable ITRs for a target population based only on experimental data are biased. On the other hand, real-world data (RWD) are becoming popular and provide a representative sample of the population. To learn the generalizable optimal interpretable ITRs, we propose an integrative transfer learning method based on weighting schemes to calibrate the covariate distribution of the experiment to that of the RWD. We show that the proposed ITR estimator has a theoretical guarantee of the risk consistency. We evaluate the transfer learner based on the finite-sample performance through simulation and apply it to a real data application of a job training program.
Keywords: Augmented inverse probability weighting, Classification error, Covariate shift, Cross-validation, Transportability, Weighting calibration
1 Introduction
A personalized recommendation, tailored to individual characteristics, is becoming increasingly popular in real life such as healthcare, education, e-commerce, etc. An optimal individualized treatment rule (ITR) is defined as maximizing the mean value function of an outcome of interest over the target population if all individuals in the whole population follows the given rule. Many machine learning approaches to estimating the optimal ITR are available, such as outcome-weighted learning with support vector machine (Zhao et al., 2012; Zhou et al., 2017), regression-based methods with Adaboost (Kang et al., 2014), regularized linear basis function (Qian and Murphy, 2011), K-nearest neighbor (Zhou and Kosorok, 2017) and generalized additive model (Moodie et al., 2014). However, the ITRs derived by machine learning methods can be too complex to extract practically meaningful insights for policymakers. For example, in healthcare, clinicians need to scrutinize the estimated ITRs for scientific validity but the black-box rules conceal the clear relationship between patient characteristics and treatment recommendations. In these cases, parsimonious and interpretable ITRs are desirable.
It is well known that randomized experiments are the gold standard for learning the optimal ITRs because randomization of treatments in the design stage can ensure no unmeasured confounding (Greenland, 1990). However, randomized experiments might lack external validity or representativeness (Rothwell, 2005) because of the selective criteria for individual eligibility in the experiments. Consequently, the distribution of the covariates in the randomized experiments might differ from that in the target population. Thus, a parsimonious ITR constructed from randomized experiments cannot directly generalize to a target population (Zhao et al., 2019). Alternatively, real-world data (RWD) such as population survey data usually involve a large and representative sample to the target population. A large body of literature in statistics has focused on transporting or generalizing the average treatment effects from the experiments to a target population or a different study population (e.g., Cole and Stuart, 2010; Hartman et al., 2015). One exception for the ITRs is Zhao et al. (2019); Mo et al. (2020) which estimates the optimal linear ITR from the experimental data by optimizing the worst-case quality assessment among all covariate distributions in the target population satisfying some moment conditions or distributional closeness, while we address the problem in a different situation where we already have a representative sample at hand.
In this article, we propose a general framework of estimating a generalizable interpretable optimal ITR from randomized experiments to a target population represented by the RWD. We propose transfer weighting that shifts the covariate distribution in the randomized experiments to that in the RWD. To correctly estimate the population optimal ITR, transfer weights are used in two places. First, the empirical estimator of the population value function is biased when considering only the experimental sample, and hence the transfer weights are used to correct the bias of the estimator of the value function. Second, the nuisance functions in the value function such as the propensity score and the outcome mean also require transfer weighting in order to estimate the population counterparts. Then we can obtain the weighted interpretable optimal ITR by reformulating the problem of maximizing the weighted value to that of minimizing the weighted classification error with flexible loss function choices. We consider different parametric and nonparametric approaches to learning transfer weights. Moreover, a practical issue arises regarding the choice of transfer weights. Weighting can correct for the bias but also reduce the effective sample size and increase the variance of the weighted value estimator. To address the bias-variance tradeoff, we propose a cross-validation procedure that maximizes the estimated population value function to choose among the weighting schemes. Lastly, we provide the theoretical guarantee of the risk consistency of the proposed transfer weighted ITRs.
The remaining of the article is organized as follows. We introduce our proposed method in Section 3. In Section 4, we study the asymptotic properties of the proposed estimator to guarantee its performance in the target population theoretically. In Section 5, we use various simulation settings to illustrate how this approach works. In Section 6, we apply the proposed method to the National Supported Work program (LaLonde, 1986) for improving the decisions of whether to recommend the job training program based on the individual characteristics over the target population. We conclude the article in Section 7.
2 Basic Setup
2.1 Notation
Let be a vector of pre-treatment covariates, the binary treatment, and the outcome of interest. Under the potential outcomes framework (Rubin, 1974), let denote the potential outcome had the individual received treatment . Define a treatment regime as a mapping from the space of individual characteristic to the treatment. Then we define the potential outcome under regime as , the value of regime as , where the expectation is taken over the distribution in the target population, and the optimal regime as the one satisfying As discussed in the introduction, in many applications such as healthcare, interpretable treatment regimes are desirable since they might bring more domain insights and be more trustworthy to clinicians compared with the complex black-box regimes. Thus, it is valuable to learn an optimal interpretable regime. Specifically, we focus on the regimes with linear forms, i.e, denoted by for simplicity.
We aim at learning ITRs which lead to the highest benefits for the target population with size , . Suppose we have access to two independent data sources: the experiment data with and the RWD with , where are samples’ index sets with size , respectively. In the experimental data, individuals are selected based on certain criteria and then are completely randomized to receive treatments. The experimental data might induce selection bias which leads to a different distribution of in experiments from that in the target population. On the other hand, we assume that individuals in the RWD are a random sample from the target population, thus the distribution of the covariates in the RWD can characterize that of the target population. Let be the binary indicator of the th individual participates in the randomized experiment: if and if . We denote as the propensity score of receiving the active treatment , as the conditional expectation of outcomes given covariates and treatments , and as the contrast function .
We make the following assumptions, which are standard and well-studied assumptions in the causal inference and treatment regime literature (Chakraborty, 2013).
Assumption 1.
2.2 Existing learning methods for ITRs
Identifying the optimal individualized treatment rules has been studied in a large body of statistical literature. One category is regression-based by directly estimating the conditional expectation of the outcome given patients’ characteristics and treatments received (Murphy, 2005). Both parametric and nonparametric models can be used to approximate . To encourage the interpretability, one strategy is to posit a parametric model as , and the estimated optimal ITR within the class indexed by is defined as .
Another category is value-based by first estimating the value of an ITRs and then searching the ITR that renders the highest value. One strategy is to estimate first, where one common way is to posit a parametric model as such as logistic regression or a constant, and then use inverse probability weighting (IPW; Horvitz and Thompson, 1952) to estimate the value of an ITR indexed by . This estimator is consistent to the true value only if is correctly specified and can be unstable if the estimate is close to zero. The augmented inverse probability weighting (AIPW; Zhang, Tsiatis, Laber and Davidian (2012)) approach is introduced to combine IPW and outcome regression estimator to estimate the value
(1) |
The AIPW estimator is doubly robust in the sense that it will be consistent to the value of the ITR if either the model for or is correctly specified, but not necessarily both.
Learning the optimal ITR by optimizing the value can also be formulated as a classification problem (Zhang, Tsiatis, Davidian, Zhang and Laber, 2012). The classification perspective has equivalent forms to all the regression-based and value-based methods above, and can also bring us additional optimization advantages over directly value searching.
Lemma 1.
The optimal treatment rule by maximizing in is equivalent to that by minimizing the risk
(2) |
in (Zhang, Tsiatis, Davidian, Zhang and Laber, 2012).
Lemma 1 reformulates the problem of estimating an optimal treatment rule as a weighted classification problem, where is regarded as the weight, the binary label, and the classification rule of interest. Let be a 0-1 loss, i.e., . The objective function (2) can also be rewritten as
|
(3) |
where , , is a certain class covering the parameters of the ITRs, i.e., .
Let be an estimator of . From Lemma 1, to estimate the optimal ITR, one can minimize the following empirical objective function:
(4) | |||||
In this classification framework, we can also consider the three different approaches to estimating , which have the equivalent forms and enjoy the same properties with the regression- and value-based estimator. For example, The AIPW estimator of is
The solution obtained by minimizing (4) in with replacing with is equivalent to that obtained by maximizing the AIPW value estimator in (1) and enjoys the doubly robust property.
When the experimental sample is representative of the target population, the above ITR learning methods can estimate the population optimal ITRs that leads to the highest value. To correct for the selection bias of the experimental sample, we propose an integrative strategy that leverages the randomization in the experimental sample and the representativeness of the RWD. As pointed out by Zhao et al. (2019), the optimal rule obtained from the experimental sample is also optimal for the target population if no restriction to the class of but such optimal rules are too complex. Therefore, we focus on learning interpretable and parsimonious ITRs and thus integrative methods are required to correct for the selection bias of the experimental sample.
3 Integrative transfer learning of ITRs
3.1 Assumptions for identifying population optimal ITRs
To identify population optimal ITRs, we assume the covariates can fully explain the selection mechanism of inclusion into the randomized experiment and the covariate distribution in the RWD is representative to the target population.
Assumption 2.
Combining Assumption 2.a and Assumption 1.b implies that the experimental sample is non-informative, i.e., the sampling score is . It requires that the observed covariates is rich enough such that it can decide whether or not the subject is selected into the experimental data.
Assumption 2.b requires that covariate distribution in the RWD is the same as that in the target population so that we can leverage the representativeness of the RWD. As a result, it implies that would be an unbiased estimator of , the covariate expectation in the target population for any . Assumption 2.a and Assumption 2.b are also adopted in the covariate shift problem in machine learning, a sub-area in domain adaptation, where standard classifiers do not perform well when the independent variables in the training data have a different covariate distribution from that in the testing data. Re-weighting individuals to correct the over- or under-representation is one way to cope with the covariate shift in classification problems (Kouw and Loog, 2018).
Assumption 2.c requires each subject has chance to be selected into the experimental data so as to guarantee the existence of .
The following proposition identifies the population value function using the inverse of sampling score as the transfer weight, and leads to an equivalent weighted classification loss function to identify the population optimal ITR.
Proposition 1.
Let the transfer weight be , under Assumption 1.b, Assumption 2.a and Assumption 2.c, the population value of an ITR is identified by . Then the population optimal ITR can be obtained by minimizing
Two challenges arise for using Proposition 1 to estimate the optimal ITR: (i) the weights are unknown and (ii) the estimators of discussed so far such as are biased based only on the experimental sample. In the following subsections, we present the estimation of the transfer weights and unbiased estimators of .
3.2 Estimation of transfer weights
We consider various parametric and nonparametric methods to estimate the transfer weights . Each method has its own benefits: parametric methods are easy to implement and efficient if the models are correctly specified, while nonparametric methods are more robust and less sensitive to model misspecification.
3.2.1 Parametric approach
For parametric approaches, we assume the target population size is known. Similar to modeling propensity score , we can posit a logistic regression model for ; i.e., , where . The typical maximum likelihood estimator of can be obtained by
(5) |
However, the second term in (5) is not feasible to calculate because we can not observe for all individuals in the population. The key insight is that this term can be estimated by based on the RWD. This strategy leads to our modified maximum likelihood estimator
Alternatively, Assumption 2.b leads to unbiased estimating equations of
for any . Therefore, we can use the following estimating equations
for , where . For simplicity, one can take as .
3.2.2 Nonparametric approach
The above parametric approaches require to be correctly specified and the population size to be known. These requirements may be stringent, because the sampling mechanism into the randomized experiment is unknown and the population size is difficult to obtain. We now consider the constrained optimization algorithm of Wang and Zubizarreta (2020) to estimate the transfer weights by
(6) |
where can be chosen as the first-, second- and higher order moments of the covariate distribution. The constants are the tolerate limits of the imbalances in . When all the ’s are taken 0, (6) becomes the exact balance, and the solution becomes the entropy balancing weights (Hainmueller, 2012). The choice of is related to the standard bias-variance tradeoff. On the one hand, if is too small, the weight distribution can have a large variability in order to satisfy the stringent constraints, and in some extreme scenarios the weights may not exist. On the other hand, if is too large, the covariate imbalances remain and therefore the resulting weights are not sufficient to correct for the selection bias. Following Wang and Zubizarreta (2020), we choose from a pre-specified set by the tuning algorithm described in Supplementary Materials.
3.3 Estimation of
The transfer weights are also required to estimate from the experimental data. In parallel to Section 2.2, we can obtain the weighted estimators of For example, the weighted AIPW estimator of
where can be estimated by weighted least square for linear functions with the estimated transfer weights, and is the weighted propensity score which can be obtained by the weighted regression model with the estimated transfer weights using the experimental data.
Let be a generic weighted estimator of . Proposition 1 implies an empirical risk function
(7) |
to estimate .
It is challenging to optimize the objective function (7) since it involves a non-smooth non-convex objective function of . One way is to optimize it directly using grid search or the genetic algorithm in Zhang, Tsiatis, Laber and Davidian (2012) for learning linear decision rules (implemented with rgenoud in R), but grid search becomes untenable for a higher dimension of , and the genetic algorithm cannot guarantee a unique solution. Alternatively, one can replace with a surrogate loss function, such as a hinge loss used in the support vector machine (Vapnik, 1998) and outcome weighted learning (Zhao et al., 2012), or a ramp loss (Collobert et al., 2006) which truncates the unbounded hinge loss by trading convexity for robustness to outliers (Wu and Liu, 2007). We adopt the smoothed ramp loss function proposed in Zhou et al. (2017), which retains the robustness and also gains computational advantages due to being smooth everywhere, defined in (8). can be decomposed into the difference of two smooth convex functions, , where is defined in (9).
(8) |
(9) |
Similarly to Zhou et al. (2017), we can apply the d.c. algorithm (Le Thi Hoai and Tao, 1997) to solve the non-convex minimization problem by first decomposing the original objective function into the summation of a convex component and a concave component and then minimizing a sequence of convex subproblems. Applying (9) to (7), our empirical risk objective becomes
(10) |
where . Let and
(11) |
for Therefore, the convex subproblem at th iteration in the d.c. algorithm is to
(12) |
which is a smooth unconstrained optimization problem and can be solved with many efficient algorithms such as L-BFGS (Nocedal, 1980). Algorithm 1 summarizes the proposed procedure to obtain the minimizer .
3.4 Cross-validation
We have discussed various approaches for the estimation of transfer weights and individual contrast function. Moreover, although transfer weights can correct the selection bias of the experimental data, they also increase the variance of the estimator compared to the unweighted counterpart. The effect of weighting on the precision of estimators can be quantified by the effective sample size originated in survey sampling (Kish and Stat, 1992). If the true contrast function can be approximated well by linear models, the estimated ITR with transfer weights may not outperform the one without weighting. As a result, for a given application, it is critical to develop a data-adaptive procedure to choose the best method among the aforementioned weighted estimators and whether or not using transfer weights.
Toward this end, we propose a cross-validation procedure. To proceed, we split the sample into two parts, one for learning the ITRs and the other for evaluating the estimated ITRs. We also adopt multiple sample splits in order to increase the robustness since single split might have dramatic varying results due to randomness. For evaluation of an ITR , we estimate the value for the given ITR by the weighted AIPW estimator in (13) with the estimated nonparametric transfer weights ,
(13) |
Since we are more interested in the comparisons among different weighting and unweighting methods instead of different contrast function estimation methods, thus we fix an estimation method for the contrast function at first, and then use cross-validation to choose among the set of candidate methods . This contains the proposed learning method with nonparametric transfer weights described in Section 3.2.2, and unweighted estimation method. If the target population size is available, we can also include the weighted estimators with the two parametric approaches introduced in Section 3.2.1 into . Then we can get the value estimates averaged over all the sample splits for each learning method in , and return the method with the highest value estimates. The more explicit description of the procedure is presented in Algorithm 2.
4 Asymptotic Properties
In this section, we provide a theoretical guarantee that the estimated ITR within a certain class is consistent for the true population optimal ITR within . Toward this end, we show that the expected risk of the estimated ITR is consistent for that of the true optimal ITR within . In particular, we consider the ITRs learned by the nonparametric transfer weights, weighted AIPW estimator with the logistic regression model and the linear function. The extensions to other cases with parametric transfer weights and regression and IPW estimators are straightforward.
We introduce more notation. Let and let the contrast function
(14) |
Denote the general smooth ramp loss function , where , is defined in (8), so it is easy to see that when , the general smooth ramp loss function will converge to the 0-1 loss function, i.e., . Note that the general smooth ramp loss function can also be written as the difference of two convex functions, , thus the d.c. algorithm in Algorithm 1 can be easily generalized to . Recalling the loss function and the corresponding -risk defined in (3), we replace with and obtain:
denoted as and respectively for simplicity. We define the minimal -risk as . Similarly, replacing in and with the general smooth ramp loss , we can obtain and , denoted as as and , respectively. Likewise, the minimal -risk with is defined as .
The goal is to show that the decision rule is -consistency (Bartlett et al., 2006), i.e.,
where , where , the estimated logistic regression coefficients which is converge in probability to as under regular conditions, and the estimated outcome regression coefficients , where , thus we have
in probability as . The convergences of and are implied from the weak law of large numbers and the result proved in Theorem 2 in Wang and Zubizarreta (2020). Let , thus we have
(15) |
It is easy to see that will converge to when therefore, to obtain -consistency, it is suffice to show that which is stated in the theorem below.
Theorem 1.
Under Assumptions S1–S10, we have
5 Simulation study
We evaluate the finite-sample performances of the proposed estimators with a set of simulation studies. We first generate a target population of a size . The covariates are generated by , and the potential outcome is generated by where i.i.d. follows for We consider three contrast functions: (I). (II). (III). .
In order to evaluate the necessity and performances of weighting methods, we consider two categories of data generating models. One is the scenarios where the true optimal policies are not in linear forms , i.e., the settings (I) and (II), while the other one is considering the true optimal policy is in a linear form as the setting (III). We start with the first two scenarios. From the data visualization, we observed that the covariates shifted between the experimental data and RWD to a varying extent in the two generative models: the covariates generated by (I) have similar distributions between the experimental data and RWD, while those generated by (II) have very different ones. Then we generate the RWD by randomly sample subjects from the target population , denoted as . To form the experimental sample, we generate the indicator of selection according to , where and which gives the average sample sizes of the experimental data over the simulation replicates round 1386, respectively. In the experimental data denoted as , the treatments are randomly assigned according to , and then the actual observed outcomes , .
For each setting, we compare the following estimators:
1. “”: our proposed weighted estimator with parametric weights learned by the modified maximum likelihood introduced in Section 3.2.1.
2. “”: our proposed weighted estimator with parametric weights learned by the estimating equations introduced in Section 3.2.1.
3. “cv”: our proposed weighted estimator using cross-validation procedure introduced in Section 3.4, and we set the number of sample splits as 10 in all the simulation studies.
4. “np”: our proposed weighted estimator with nonparametric weights.
5. “unweight”: only using the experimental data to learn the optimal linear ITRs.
6. “bm”: using the weighted estimator but we replace the estimated transfer weights with the normalized true weights , and replace the estimated contract functions with the true contrast values .
7. “Imai”: the method proposed in Imai et al. (2013). They construct the transfer weights by fitting a Bayesian additive regression trees (BART) model using covariates as predictors and whether to be in the experimental data as outcomes, and then use a variant of SVM to estimate the conditional treatment effect (implemented in the R package FindIt). Then given , we can assign treatment if its estimated conditional treatment effect is positive, and otherwise.
8. “DBN”: the method proposed in Mo et al. (2020), where they propose to use distributionally robust ITRs, that is, value functions are evaluated under all testing covariate distributions that are ”close” to the training distribution, and the worse-case takes a minimal one, and then the ITRs are learned by maximizing the defined worst-case value function.
For the estimators 1. – 5. in the above, we estimate the contrast function using with linear functions and constant propensity scores estimated via the proportion of in the experimental data, .
To study the impact of model misspecification of , we consider to estimate by fitting logistic regression wrongly: , where . And as described in the proposed methods, only estimators 1. – 3. need to specify the sampling model. To measure the performances of the different approaches, we use the mean value MSE defined as . Thus the lower the MSE, the better performances the approach has.
The results are summarized in Figure 1, the dark red point in each boxplot is the mean value MSE averaged over 200 Monte Carlo replicates. For the generative model (I) shown in Figure 1, unweighted and DBN rules perform worst in all the cases. The second parametric weighting method is relatively robust compare with the first parametric one when the sampling model is misspecified; this is reasonable since the first one uses maximum likelihood estimation which is more reliable on the correctness of the parametric model specification. When the training data and testing data are similarly distributed, the performances of parametric, nonparametric weighting method and the cross-validation procedure are similar; in this case, sampling model misspecification does not have much influence on the rule learning.

For the generative model (II) shown in Figure 1, among the performances of weighted and unweighted estimators, we can see the weighted estimators outperform the unweighted estimators when model correctly specified; the nonparametric weighted estimator outperforms other estimators and the performance of the parametric weighted estimators have dramatically decreased when the sampling model misspecified. These findings make sense, because when the more different the training and testing data are, the more crucial the bias correction is. Thus sampling model misspecification might have huge influence on the learned rules.
From the both generative models, we can see that the nonparametric weighting and cross-validation procedure are more robust when the sampling model is unknown and recommended in practice use. As for DBN estimator, it performs badly in terms of MSE magnitude or variability, one possible reason might be that the tuning parameters when measuring the ”closeness” of two distributions need to be chosen more carefully in practice Mo et al. (2020). As for the comparison with Imai also has relative good performance among all estimators and has relative small variability.
As for setting (III), we consider the correctly specified sample score model. As shown in Figure 2, unweighted method performances the best, close to the benchmark. This demonstrates that when the true optimal polices are in linear forms, then learning with only the experimental data performances the best. Thus weighting is not a necessity, and it even will induce the variance of the estimators. The other competitive estimator is using cross-validation. Therefore, when we have no idea of the true data generating models, using CV is a good choice in practice.

6 Real Data Application
We apply the proposed methods to personalized recommendations of a job training program on post-market earning. Our analysis is based on two data sources (downloaded from https://users.nber.org/~rdehejia/data/.nswdata2.html): an experimental dataset from the National Supported Work (NSW) program (LaLonde, 1986) and a non-experimental dataset from the Current Population Survey (CPS; LaLonde (1986)). The NSW dataset includes individuals who received the job training program () and individuals who did not receive the job training program (), and the CPS dataset includes individuals without the job training program. Let be a vector including an intercept and the pre-intervention covariates: age, education, Black (1 if black, 0 otherwise), Hispanic (1 if Hispanic, 0 otherwise), married (1 if married, 0 otherwise), whether having high school degree (1 if no degree, 0 otherwise), earning in 1974, and earning in 1975. The outcome variable is the earning in 1978. We transfer the earnings in the year 1974, 1975, and 1978 by taking the logarithm of the earnings plus one, denoted as “log.Re74”, “log.Re75” and “log.Re78”. We aim to learn a linear rule tailoring the recommendation of the job training program to an individual with characteristics .
The CPS sample is generally representative of the target population; while the participants to the NSW program may not represent the general population. Table 1 contrasts sample averages of each pre-intervention covariate in the NSW and CPS samples and shows that the covariates are highly imbalanced between the two samples. On average, the NSW sample is younger, less educated, more black people, fewer married, more without high school degree, fewer earnings in 1974 and 1975. Because of imbalance, ITRs cannot be learned based only on the NSW sample, thus we are motivated to use the proposed weighted transfer learning approaches to learn optimal linear ITRs which benefit the target population most.
intercept | Age | Educ | Black | Hisp | Married | Nodegr | log.Re74 | log.Re75 | |
---|---|---|---|---|---|---|---|---|---|
experiment NSW | 1 | 25.37 | 10.20 | 0.83 | 0.09 | 0.17 | 0.78 | 2.27 | 2.70 |
Non-experiment CPS | 1 | 33.23 | 12.03 | 0.07 | 0.07 | 0.71 | 0.30 | 8.23 | 8.29 |
Unweighted learning | -1 | 0.03 | 0.07 | 0.81 | 2.31 | 1.89 | -0.62 | -0.60 | 0.47 |
Weighted transfer learning | -1 | 0.02 | 0.05 | 0.66 | 0.36 | 0.14 | 0.28 | -0.27 | 0.15 |
We consider the unweighted and weighted transfer learning methods. For the weighted transfer learning method, we consider the nonparametric weighting approach, because the parametric approach requires the target population size which is unknown in this application. To estimate the weighted and unweighted decision rules, we use the AIPW estimator with linear Q functions since it outperforms other competitors in the simulation study in Section 5. We also implement the cross-validation procedure to choose between the unweighted and weighted transfer learning methods.
Table 1 presents the estimated linear coefficients of each corresponding variable in the linear decision rule from the weighted transfer and unweighted learning methods, respectively. Moreover, from the cross-validation procedure, the weighted transfer learning method produces a smaller population risk than the unweighted learning method and therefore is chosen as the final result. Compared with the learned decision rule with only the experiment data NSW which oversamples the low-income individuals, individuals without high school degrees (i.e. Nodegr = 1) are more likely to have higher outcomes if offered the job training program in the target population.
To assess the performances of the learned rules for the target population, following Kallus (2017), we create a test sample that is representative of the target population. We sample individuals at random from the CPS sample to form a control group and then find their closest matches in the NSW treated sample to form the treatment group (use pairmatch in R package optmatch). The test sample thus contains the matched pairs, mimicking a experimental sample where binary actions are randomly assigned with equal probabilities. We use random forest on the test data to estimate function , denoted the estimator as , then we can use to estimate the value of an ITR , where is the sample size of CPS. Besides the weighted linear rule and the unweighted linear rule learned from the above, we also compare the Imai method introduced in Imai et al. (2013): first to estimate the conditional probability of being selected into NSW given from the total sample which consists of NSW and CSP using BART model, and then to estimate the conditional treatment effect weighted by the inverse of the sampling probability. Then the Imai rule will recommend the job training program if the conditional treatment effect is positive and will not recommend it otherwise. We replicate this sampling test data procedure 100 times, estimate values of the three ITRs for each replicate, and get the averaged estimated values (standard error) as follows: 8.395 (0.009), 6.106 (0.011), and 8.049 (0.006) for , , and Imai respectively, thus the transfer weighted rule has the highest value estimate. Although this evaluation method needs assumption of no unmeasured confounders for the non-experimental data CPS, which is not needed in our learning procedure, it can still bring us confidence, to some degree, that the weighted linear rule has better generalizability to the target population.
7 Discussion
We develop a general framework to learn the optimal interpretable ITRs for the target population by leveraging no unmeasured confounding in the experimental data and the representativeness of covariates in the RWD. The proposed procedure is easy to implement. By constructing transfer weights to correct the distribution of covariates in the experimental data, we can learn the optimal interpretable ITRs for the target population. Besides, we provide a data-adaptive procedure to choose among different weighting methods by modified cross-validation. Moreover, we establish the theoretical guarantee for the consistency of the proposed estimator of ITRs.
Some future directions worth considering. One is to quantify the uncertainty of the estimated ITRs and therefore to conduct statistical inference. Moreover, in some scenarios, beside the covariate information, comparable treatment and outcome information are also available in the RWD. Unlike the randomized experiment, the treatment assignment in the RWD is determined by the preference of physicians and patients. In this case, an interesting topic is to develop integrative methods that can leverage the strengths of both data sources for efficient estimation of the ITRs.
References
- (1)
- Bartlett et al. (2006) Bartlett, P. L., Jordan, M. I. and McAuliffe, J. D. (2006). Convexity, classification, and risk bounds, Journal of the American Statistical Association 101(473): 138–156.
- Chakraborty (2013) Chakraborty, B. (2013). Statistical methods for dynamic treatment regimes, Springer.
- Cole and Stuart (2010) Cole, S. R. and Stuart, E. A. (2010). Generalizing evidence from randomized clinical trials to target populations: the actg 320 trial, American journal of epidemiology 172(1): 107–115.
- Collobert et al. (2006) Collobert, R., Sinz, F., Weston, J. and Bottou, L. (2006). Trading convexity for scalability, Proceedings of the 23rd international conference on Machine learning, pp. 201–208.
- Greenland (1990) Greenland, S. (1990). Randomization, statistics, and causal inference, Epidemiology pp. 421–429.
- Hainmueller (2012) Hainmueller, J. (2012). Entropy balancing for causal effects: A multivariate reweighting method to produce balanced samples in observational studies, Political analysis pp. 25–46.
- Hartman et al. (2015) Hartman, E., Grieve, R., Ramsahai, R. and Sekhon, J. S. (2015). From sample average treatment effect to population average treatment effect on the treated: combining experimental with observational studies to estimate population treatment effects, Journal of the Royal Statistical Society: Series A (Statistics in Society) 178(3): 757–778.
- Horvitz and Thompson (1952) Horvitz, D. G. and Thompson, D. J. (1952). A generalization of sampling without replacement from a finite universe, Journal of the American statistical Association 47(260): 663–685.
- Imai et al. (2013) Imai, K., Ratkovic, M. et al. (2013). Estimating treatment effect heterogeneity in randomized program evaluation, The Annals of Applied Statistics 7(1): 443–470.
- Kallus (2017) Kallus, N. (2017). Recursive partitioning for personalization using observational data, Proceedings of the 34th International Conference on Machine Learning-Volume 70, JMLR. org, pp. 1789–1798.
- Kang et al. (2014) Kang, C., Janes, H. and Huang, Y. (2014). Combining biomarkers to optimize patient treatment recommendations, Biometrics 70(3): 695–707.
- Kish and Stat (1992) Kish, L. and Stat, J. O. (1992). Weighting for unequal pi.
- Kouw and Loog (2018) Kouw, W. M. and Loog, M. (2018). An introduction to domain adaptation and transfer learning, arXiv preprint arXiv:1812.11806 .
- LaLonde (1986) LaLonde, R. J. (1986). Evaluating the econometric evaluations of training programs with experimental data, The American economic review pp. 604–620.
- Le Thi Hoai and Tao (1997) Le Thi Hoai, A. and Tao, P. D. (1997). Solving a class of linearly constrained indefinite quadratic problems by dc algorithms, Journal of global optimization 11(3): 253–285.
- Mo et al. (2020) Mo, W., Qi, Z. and Liu, Y. (2020). Learning optimal distributionally robust individualized treatment rules, Journal of the American Statistical Association pp. 1–16.
- Moodie et al. (2014) Moodie, E. E., Dean, N. and Sun, Y. R. (2014). Q-learning: Flexible learning about useful utilities, Statistics in Biosciences 6(2): 223–243.
- Murphy (2005) Murphy, S. A. (2005). A generalization error for q-learning, Journal of Machine Learning Research 6(Jul): 1073–1097.
- Nocedal (1980) Nocedal, J. (1980). Updating quasi-newton matrices with limited storage, Mathematics of computation 35(151): 773–782.
- Qian and Murphy (2011) Qian, M. and Murphy, S. A. (2011). Performance guarantees for individualized treatment rules, Annals of statistics 39(2): 1180.
- Rothwell (2005) Rothwell, P. M. (2005). External validity of randomised controlled trials:“to whom do the results of this trial apply?”, The Lancet 365(9453): 82–93.
- Rubin (1974) Rubin, D. B. (1974). Estimating causal effects of treatments in randomized and nonrandomized studies., Journal of educational Psychology 66(5): 688.
- Vapnik (1998) Vapnik, V. N. (1998). Statistical Learning Theory, Wiley-Interscience.
- Wang and Zubizarreta (2020) Wang, Y. and Zubizarreta, J. R. (2020). Minimal dispersion approximately balancing weights: asymptotic properties and practical considerations, Biometrika 107(1): 93–105.
- Wu and Liu (2007) Wu, Y. and Liu, Y. (2007). Robust truncated hinge loss support vector machines, Journal of the American Statistical Association 102(479): 974–983.
- Zhang, Tsiatis, Davidian, Zhang and Laber (2012) Zhang, B., Tsiatis, A. A., Davidian, M., Zhang, M. and Laber, E. (2012). Estimating optimal treatment regimes from a classification perspective, Stat 1(1): 103–114.
- Zhang, Tsiatis, Laber and Davidian (2012) Zhang, B., Tsiatis, A. A., Laber, E. B. and Davidian, M. (2012). A robust method for estimating optimal treatment regimes, Biometrics 68(4): 1010–1018.
- Zhao et al. (2019) Zhao, Y.-Q., Zeng, D., Tangen, C. M. and LeBlanc, M. L. (2019). Robustifying trial-derived optimal treatment rules for a target population, Electronic journal of statistics 13(1): 1717.
- Zhao et al. (2012) Zhao, Y., Zeng, D., Rush, A. J. and Kosorok, M. R. (2012). Estimating individualized treatment rules using outcome weighted learning, Journal of the American Statistical Association 107(499): 1106–1118.
- Zhou and Kosorok (2017) Zhou, X. and Kosorok, M. R. (2017). Causal nearest neighbor rules for optimal treatment regimes, arXiv preprint arXiv:1711.08451 .
- Zhou et al. (2017) Zhou, X., Mayer-Hamblett, N., Khan, U. and Kosorok, M. R. (2017). Residual weighted learning for estimating individualized treatment rules, Journal of the American Statistical Association 112(517): 169–187.