This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Efficient Heterogeneous Treatment Effect Estimation With Multiple Experiments and Multiple Outcomes

Leon Yao    Caroline Lo    Israel Nir22footnotemark: 2    Sarah Tan22footnotemark: 2 Massachusetts Institute of TechnologyFacebook    Ariel Evnine22footnotemark: 2    Adam Lerer22footnotemark: 2    Alex Peysakhovich22footnotemark: 2
Abstract

Learning heterogeneous treatment effects (HTEs) is an important problem across many fields. Most existing methods consider the setting with a single treatment arm and a single outcome metric. However, in many real world domains, experiments are run consistently - for example, in internet companies, A/B tests are run every day to measure the impacts of potential changes across many different metrics of interest. We show that even if an analyst cares only about the HTEs in one experiment for one metric, precision can be improved greatly by analyzing all of the data together to take advantage of cross-experiment and cross-outcome metric correlations. We formalize this idea in a tensor factorization framework and propose a simple and scalable model which we refer to as the low rank or LR-learner. Experiments in both synthetic and real data suggest that the LR-learner can be much more precise than independent HTE estimation.

Introduction

Constructing personalized policies is increasingly of interest across many diverse communities ranging from internet companies to personalized medicine (Peysakhovich and Lada, 2016; Kent et al., 2018). Unfortunately, estimating heterogeneous effects, sometimes referred to as conditional average treatment effects (CATEs), requires much more statistical power than simply estimating the average treatment effect (Gelman, 2018). In this paper, we introduce the LR-learner, a method of combining multiple experiments and multiple outcome metrics to improve the precision of CATE estimation across all of them. Even if an analyst is only interested in estimating CATEs for a single experiment, adding data from past experiments allows a more efficient estimation.

At its core, the goal of heterogeneous estimation is to output an estimate of the treatment effect for a unit whose features are specified by the vector xx. Under the potential outcomes framework (Rubin, 1974), we denote the potential outcomes under treatment and control as Y1Y_{1} and Y0Y_{0}, and their conditional means as μ1(x)=E[Y1|X=x]\mu_{1}(x)=E[Y_{1}|X=x] and μ0(x)=E[Y0|X=x]\mu_{0}(x)=E[Y_{0}|X=x], respectively. Our estimand of interest is the

CATE(x)=τ(x)=μ1(x)μ0(x).\text{CATE}(x)=\tau(x)=\mu_{1}(x)-\mu_{0}(x).

Much of the literature on experimental analysis focuses on the case of a single experiment with a single treatment and a single outcome (e.g. a medical trial of a drug with survival as an outcome). Here there is a large body of work on efficiently estimating both the average effect of this experiment as well as finding subgroups which are more or less affected by the treatment (Imbens and Rubin, 2015; Wager and Athey, 2018; Künzel et al., 2019).

In many environments of interest, however, we are not considering a single experiment/outcome in isolation. For example, the A/B testing that is common in modern internet companies is different than the classical case in a few ways. First, large companies can run hundreds of tests per day (bez, 2014; Xie and Aurisset, 2016), measuring and trading off multiple outcome measures. For example, a movie streaming company can conduct A/B tests that run the gamut from tweaking movie recommendation algorithm parameters, to testing landing page layout changes. Each of these experiments can be evaluated on a set of metrics ranging from user retention to server CPU usage. Decision makers then trade off between metrics to make a deployment or non-deployment decision. The effect of each intervention is likely to be heterogeneous across treated units. Second, unlike in many classical applications, the set of unit level features that analysts have access to is very high dimensional. Thus a key problem is one of representation learning: finding a good, lower dimensional, feature description for efficiently learning the heterogeneity of treatment effects.

Our key insight is that pooling information across multiple experiments and multiple outcomes and learning a single feature representation for treated units allows us to greatly improve the precision with which we estimate heterogeneous effects over standard methodologies. The potential to pool experiments is increasingly viable in the online A/B testing setting. In addition, the idea of multiple experiments with high dimensional feature spaces is becoming applicable to other domains e.g. pooling many clinical trials or experimental work in systems biology.

For the one experiment/one outcome case, a simple methodology for learning CATEs is the T-learner (Künzel et al., 2019), where we train a model, μ^1(x)\hat{\mu}_{1}(x), on data from units in the treatment group and train a model, μ^0(x)\hat{\mu}_{0}(x), on data from units in the control group. The estimated effect can be computed as τ^(x)=μ^1(x)μ^0(x)\hat{\tau}(x)=\hat{\mu}_{1}(x)-\hat{\mu}_{0}(x).

Here, we build on the T-learner and consider the case of multiple experiments and multiple outcomes. Every unit may be subject to multiple experiments, and all outcome metrics are observed. Our goal is to estimate the treatment effect for all units across all experiments. This task can be modeled as a tensor factorization problem and can be done efficiently when the tensor is low rank. We refer to our method as the Low Rank Learner, or LR-learner.

For each unit, the LR-learner learns a dd-dimensional unit embedding using a small neural network over input unit level features. The LR-learner also learns an embedding for each experiment treatment and control. Finally, for each outcome, a linear operator transforms the experiment and unit embeddings so that the dot product of these transformed embeddings predicts the desired outcome. The neural network, experiment embeddings and operators are learned jointly using the observed data.

The LR-learner takes advantage of the low dimensional bottleneck of the embeddings and linear operators, which is what allows us to pool information across experiments and outcome types. In practice, true CATEs are not observed (since we do not observe any unit in both treatment and control), so we first evaluate the LR-learner on fully synthetic and semi-synthetic data. We show that pooling information across experiments and outcomes greatly improves efficiency relative to independent learning.

Finally, we evaluate the LR-learner on a real dataset of 5353 product experiments run at a large internet company, focusing on 55 real key performance indicators used by teams to make product decisions. We show that there is substantial information in heterogeneous effects shared across experiments and metrics, and that the LR-learner outperforms the independent level T-learner. In addition, we discuss other real world advantages of the LR-learner including scalability and the ability to use the LR-learner as a feature extractor negating the need for constant retraining when new experiments are run.

Related Work

There exist many methods for estimating CATEs on data from a single experiment (Peysakhovich and Lada, 2016; Powers et al., 2018; Wager and Athey, 2018; Künzel et al., 2019; Kennedy, 2020). Our main contribution over this large literature is the consideration of combining data from multiple experiments and multiple metrics into a single model. Although there are many methods that aim to directly estimate the treatment effect (Athey et al., 2018b), in this paper we consider only the Two-Model approach (Gutierrez and Gérardy, 2017), because it naturally extends to the multiple experiment case. Possible extensions of our ideas to other “meta-learners,” like the X-learner (Künzel et al., 2019) are an interesting future direction though beyond the scope of this paper.

There is also a lot of recent literature on matrix completion methods on panel data (Agarwal et al., 2021a, b; Athey et al., 2018a; Bai and Ng, 2021; Fernández-Val et al., 2021). They consider the setting in which units are observed over multiple time periods. In each period, a unit can be exposed to some binary treatment, and their outcome for that period is recorded. These papers assume some sort of sparsity structure for when units receive treatment or when outcomes are observed. Our methodology does not assume any sparsity structure, so is not reliant on panel data, but can also be applied to this setting.

The model architecture employed by the LR-learner is inspired by the RESCAL model used for tensor factorization (Lerer et al., 2019; Nickel et al., 2011). Typically, these methods are applied to construct low dimensional embeddings of objects such as social or knowledge graphs (Lerer et al., 2019). So, a conceptual contribution of our paper is to make a link between two concepts. First, we frame the causal inference problem as a missing data problem (Ding and Li, 2018). Second, to solve the missing data problem, we use efficient low dimensional methods to fill in unobserved tensor values. We note that unlike in the case of a matrix, where the singular value decomposition provides the “canonical” way to factorize, there are many ways to factorize a tensor (Welling and Weber, 2001). Exploring other tensor factorization methods is an interesting avenue for future work.

Potential Outcomes as Tensor Factorization

We consider treatment assigned at the level of an individual (aka. unit), and consider the presence of multiple outcome metrics. Let nn be the number of units, mm be the number of observed pre-treatment features, KK be the number of experiments, and JJ be the number of observed outcome metrics. We denote by YijktY_{ijkt} the potential outcome of unit ii on metric jj when assigned to experiment kk with treatment arm tt. We denote t=0t=0 to be the ‘control’ arm for any experiment.

We make the following assumptions:

Assumption 1 (SUTVA (Rubin, 1978))

Every observed outcome for unit ii is independent of the actual treatment assignment of other units jij\neq i.

Assumption 2 (Random Assignment)

In any data we observe, units are assigned to treatments (or control) unconditional on any observed or unobserved feature values.

The individual treatment effect (ITE) for individual ii for metric jj in experiment kk for treatment tt is denoted

ITEijk=YijktYijk0.\text{ITE}_{ijk}=Y_{ijkt}-Y_{ijk0}.

In this framework, a typical ‘classical’ randomized trial (e.g. a medical trial) would have a single arm, a single outcome, and a feature vector for each unit. In this case, the ITE would simply be denoted by a vector with one value (i.e treatment effect estimate) per unit ii.

If we move to the case of a standard trial with multiple outcome measures the ITE can be represented as a matrix with rows corresponding to individuals and columns corresponding to outcome measures. Finally, moving to the fully general multi-experiment, multi-treatment, multi-outcome case the ITE becomes a tensor. This brings us to our final key assumption:

Assumption 3

The ITE tensor has a low rank structure.

Assuming the ITE tensor has low rank structure means that there is information about a given entry of interest in other entries. This is the core of the proposed methodology. In practice, we do not observe any ITEs since no individual is ever in both a treatment and a control simultaneously. Instead, we will focus on a relative of the ITE which instead featurizes each unit, ii, using a vector, xix_{i}. The CATE for treatment arm, tt, in experiment, kk, is defined as

CATEijkt(Xi)=E[YijktXi]E[Yijk0Xi].\text{CATE}_{ijkt}(X_{i})=E[Y_{ijkt}\mid X_{i}]-E[Y_{ijk0}\mid X_{i}].

If the ITE tensor is low rank, then observing CATEs for experiment kk and metric jj tells us something about CATEs in other experiments/metrics/treatments. This is precisely what the LR-learner will take advantage of.

LR-Learner for Heterogeneous Effects

We now discuss how to train the LR-learner as a tensor factorization. Our factorization has several components: We have a sample of units assigned to experiments/treatments and their features denoted by 𝒟={yijkt,xi}\mathcal{D}=\{y_{ijkt},x_{i}\}. We train a featurized tensor factorization model which includes

  1. 1.

    A dd dimensional representation of each unit ii, which is a function of xix_{i}, called v(xi)Rdv(x_{i})\in R^{d}. This is expressed as a simple neural network which inputs xix_{i} and outputs a dd dimensional feature vector viv_{i}.

  2. 2.

    A dd dimensional representation of each experiment treatment arm tt and control ektRde^{t}_{k}\in R^{d} and ek0Rde^{0}_{k}\in R^{d}.

  3. 3.

    A d×dd\times d transformation matrix for each outcome AjA_{j}.

  4. 4.

    The parameters are trained with an objective similar to RESCAL (Nickel et al., 2011)

    minv,A,e(yijkt,xi)𝒟(yv(xi)Ajekt)2.\min_{v,A,e}\sum_{(y_{ijkt},x_{i})\in\mathcal{D}}(y-v(x_{i})^{\prime}A_{j}e^{t}_{k})^{2}.

The LR-learner structure is graphically summarized in the figure 1.

Given the trained model, the predicted potential outcome for a unit ii in experiment jj in arm tt for outcome kk is given by:

μ^jkt(xi)=v(xi)Ajekt\hat{\mu}^{t}_{jk}(x_{i})=v(x_{i})^{\prime}A_{j}e^{t}_{k}

This means the estimate of the CATE is just a dot product between v(xi)v(x_{i}) (transformed for the particular outcome) and the difference between the treatment and control embeddings. Formally:

CATE^jkt(xi)=v(xi)Aj(ektek0)\hat{\textrm{CATE}}_{jkt}(x_{i})=v(x_{i})^{\prime}A_{j}(e^{t}_{k}-e^{0}_{k})

The fact that the CATE is fully linear is useful in real world systems where new experiments are constantly run. We will explore this idea in detail in section 6.

Refer to caption
Figure 1: Visual description of the LR-learner model. Red boxes indicate inputs transformed to give an output of the form yijkt=v(xi)Akekty_{ijkt}=v(x_{i})^{\prime}A^{k}e^{t}_{k}.

Evaluation

How to Evaluate?

Let ITE^\hat{\text{ITE}} be the predicted ITE tensor from some model. Ideally, we would be able to calculate the loss between ITE^\hat{\text{ITE}} and ITE, directly. This is often called the precision in estimating heterogeneous effects (PEHE) loss (Hill, 2011). Formally, we write this as

PEHE=1|ITE|ijk(ITE^ijkITEijk)2.\text{PEHE}=\dfrac{1}{|ITE|}\sum_{ijk}(\hat{\text{ITE}}_{ijk}-\text{ITE}_{ijk})^{2}.

Unfortunately, the true ITE is not observed (since any unit is ever only assigned to treatment or control in any experiment). Thus, there is an active area of research for proxy metrics which can be used instead of the true PEHE (Schuler et al., 2018).

We now describe two such candidates. Let 𝒱\mathcal{V} be a held out test set of unit, experiment, treatment, measure level outcomes with generic element (ijkt)𝒱.(ijkt)\in\mathcal{V}. Let Y^\hat{Y} be the full outcome tensor that is predicted by the model.

The first candidate is straightforward: it is the supervised learning objective for how well the model predicts the observed outcomes in the held out set. This is called the μ\mu-risk in the literature (Schuler et al., 2018). Formally this is given by

μ-risk^=1|𝒱|(ijkt)𝒱(y^ijktyijkt)2\widehat{\mu\textrm{-risk}}=\frac{1}{|\mathcal{V}|}\sum_{(ijkt)\in\mathcal{V}}(\hat{y}_{ijkt}-y_{ijkt})^{2}

While μ\mu-risk is often used in practice, there are certainly situations where improvements in μ\mu-risk does not imply improvements in PEHE (for example, the simple case where base outcomes are predicatble but the treatment effect is constant).

The second metric, called τ\tau-risk, is an unbiased proxy (Nie and Wager, 2020; Schuler et al., 2018) for the ITE. It is defined as follows:

τ-risk^=1|𝒱|xi,yi,ti𝒱((yimˇ(xi))(tipˇ(xi))τ^(xi))2,\widehat{\tau\textrm{-risk}}=\frac{1}{|\mathcal{V}|}\sum_{x_{i},y_{i},t_{i}\in\mathcal{V}}((y_{i}-\check{m}(x_{i}))-(t_{i}-\check{p}(x_{i}))\hat{\tau}(x_{i}))^{2},

where τ^(xi)\hat{\tau}(x_{i}) is the model estimate of τ\tau.

Here mˇ\check{m} is an estimate of E[Y|X]E[Y|X], and pˇ\check{p} is the estimate of the propensity score both trained using data in the validation set. In our case both mˇ\check{m} and pˇ\check{p} are trained using linear models. In expectation τ\tau-risk is minimized when τ^\hat{\tau} is the true CATE function. However, for a full derivation we point the interested reader to Nie and Wager (2020) and Schuler et al. (2018).

Both τ\tau-risk and μ\mu-risk are proxy metrics for the unobservable ITE. Some work argues that τ\tau-risk is a better proxy than μ\mu-risk for evaluating heterogeneous treatment effect models, though the debate is not settled (Nie and Wager, 2020; Schuler et al., 2018).

In the following sections, we always report the μ\mu-risk and report PEHE when ITEs are observed (in synthetic and semi-synthetic data) and the τ\tau-risk when PEHE is not observed.

Baseline - Independent T-learners

Algorithm 1 Independent T-learners
MLML is your favorite machine learning model
XX\leftarrow features for training set
YY\leftarrow outcomes from training set
for each experiment, kk do
     for each metric, jj do
         Xj,k1X^{1}_{j,k}\leftarrow features for units in treatment
         Xj,k0X^{0}_{j,k}\leftarrow features for units in control
         Yj,k1Y^{1}_{j,k}\leftarrow outcomes for units in treatment
         Yj,k0Y^{0}_{j,k}\leftarrow outcomes for units in control
         Learn μ^j,k1ML(Yj,k1Xj,k1)\hat{\mu}^{1}_{j,k}\leftarrow ML(Y^{1}_{j,k}\sim X^{1}_{j,k})
         Learn μ^j,k0ML(Yj,k0Xj,k0)\hat{\mu}^{0}_{j,k}\leftarrow ML(Y^{0}_{j,k}\sim X^{0}_{j,k})
         CATE^j,k(X)μ^j,k1(X)μ^j,k0(X)\hat{\text{CATE}}_{j,k}(X)\leftarrow\hat{\mu}^{1}_{j,k}(X)-\hat{\mu}^{0}_{j,k}(X)
     end for
end for

As the baseline, we use the standard independent T-learner approach. Algorithm 1 provides pseudocode for training independent T-learners for each metric-treatment combination.

Experiment 1: Synthetic Data

We evaluate the LR-learner on synthetic experiment data with feature and outcomes generated as shown in algorithm 2.

Algorithm 2 Generating synthetic data
nn\leftarrow num samples per exp. condition
dd\leftarrow latent feature dimension
mm\leftarrow observed feature dimension
for each metric, jj do
     Ajrandn(d,d)A_{j}\leftarrow\texttt{randn}(d,d)
end for
for each experiment, kk do
     ek1e^{1}_{k}\leftarrow l2_norm(randn(d))
     ek0e^{0}_{k}\leftarrow l2_norm(randn(d))
     vv\leftarrow fro_norm(randn(n, d))
     for each metric, jj do
         Yj,k1vAjek1Y^{1}_{j,k}\leftarrow v^{\prime}A_{j}e^{1}_{k}
         Yj,k0vAjek0Y^{0}_{j,k}\leftarrow v^{\prime}A_{j}e^{0}_{k}
         ITEj,k=Yj,k1Yj,k0ITE_{j,k}=Y^{1}_{j,k}-Y^{0}_{j,k}
     end for
     Lrandn(d,m)L\leftarrow\texttt{randn}(d,m)
     XkvLX_{k}\leftarrow v\cdot L
     TkBernn(0.5)T_{k}\sim\text{Bern}_{n}(0.5)
end for

Using algorithm 2, we generate 50 synthethic data experiments with m=128m=128 and 5 outcome metrics. Notice that our synthetic data assumes each unit is only a part of a single experiment, but this is not necessary for training the LR-learner.

For this data set we use an LR-learner without a non-linearity in the neural network with hidden dimension d=32d=32. We also use linear regressions for our independent T-learner models.

Since the data generating process (DGP) is linear by construction, both the independent T-learner and the LR-learner are able to learn the true DGP. The only difference between the two is that the LR-learner pools data across experiments and outcome measures.

In our analysis, we train both models on datasets of various sizes of per experiment data (n=n= 25, 50, 100, 250, 500, 1000, 2500, 500, 10000, 50000 per treatment condition). We average over 5 independent replications.

PEHE μ-risk\mu\textrm{-risk}
n
Ind.
T-learners
LR-learner
Ind.
T-learners
LR-learner
25 0.26792 0.02014 0.13407 0.01008
50 0.07956 0.02004 0.03962 0.01003
100 0.03292 0.02001 0.01651 0.01001
250 0.02307 0.02001 0.01154 0.01001
500 0.02144 0.02000 0.01072 0.01001
1000 0.02069 0.02000 0.01035 0.01001
2500 0.02028 0.02000 0.01014 0.01001
5000 0.02014 0.02000 0.01007 0.01001
10000 0.02007 0.02000 0.01004 0.01001
50000 0.02001 0.02000 0.01001 0.01000
Table 1: The PEHE and μ-risk\mu\textrm{-risk} averaged over all 50 synthethic data experiments and 5 outcome metrics evaluated on a validation set. For each value of nn, data is drawn 5 times without replacement.

Table 1 shows that the LR-learner performs better than the Independent T-learners on PEHE as well as μ-risk\mu\textrm{-risk} metrics, with 200×200\times less data.

Experiment 2: Semi-Synthetic Data

We now consider evaluating the LR-learner on a semi-synthetic dataset. A major issue with real datasets is that we never observe the true ITEs as each unit is either in treatment or control. Thus, the true PEHE error is never observable.

For this reason, using semi-synthetic data has become a popular methodology in theoretical causal inference (Dorie et al., 2019). Semi-synthetic datasets have features that are generated from real data but have potential outcomes that are constructed via a known generated mapping,

To generate a dataset with multiple potential outcomes, we use a technique employed in the multi-armed bandit literature to transform a multi-class classification problem into a multi-outcome contextual bandit problem (Dudík et al., 2014; Agarwal et al., 2014).

We start with an original multi-class dataset, and train a classifier to predict the classes. Then, for each sample ii, we use the predicted class jj logit as the potential outcome yijy_{ij}. For the features for the semi-synthetic causal inference task, we can use the original features or some transformation thereof. For example, if the original classifier is a multi-layer neural network, we can use activations from an intermediate layer.

We use CIFAR100 as our base dataset (Krizhevsky et al., 2009). For our base classifier, we use the simple neural network setup used in the PyTorch computer vision tutorial111We follow the setup here except we use CIFAR100 instead of CIFAR10 and use 400 hidden units for the intermediate linear layers. Otherwise, the code we use is identical..

Refer to caption
Refer to caption
Figure 2: The correlation matrix between the ITEs for each of the 99 treatments with hierchical clustering done on row/column ordering to emphasize the pattern as well as the singular value plot for the ITE matrix. We see that the ITE matrix induced by the CIFAR100 dataset is effectively low rank.

For features, we use the 2nd2^{nd} to last linear layer. For the semi-synthetic per-unit potential outcomes, we use the logit outputs for each class. This means that the real function, mapping features to logits, is a non-linear function.

Using this method on the CIFAR100 dataset gives us a semi-synthetic dataset with 400 features and 100 outcomes with 50000 units in our training set and 10000 units in our test set. We treat the first outcome as the control experiment and the other 99 experiments as the treatments.

Figure 2 shows the correlation matrix between the ITEs for each of the 99 treatments as well as the singular value plot for the ITE matrix. We see that the ITE matrix is effectively quite low rank.

We randomly allocate 10 percent of the units to each experiment, so that on average each unit is a part of 10 experiments. Note that this semi-synthetic dataset only has one outcome metric.

We train a LR-learner model with a 2-layer neural network with ReLU nonlinearity and hidden dimension 128. As baseline, we train an Independent T-learner model using both a simple linear regression and XGBoost (Chen and Guestrin, 2016). We tested other popular model classes (e.g. neural networks) for the independent T-learners but found that XGBoost performed best and required the least hyperparameter tuning to reach that level of performance.

PEHE μ-risk\mu\textrm{-risk}
LR-learner 0.5277 0.2560
XGBoost 0.5787 0.2696
Linear Model 0.6118 0.2897
Table 2: The PEHE and μ-risk\mu\textrm{-risk} for the LR-learner model and XGBoost and linear model independent T-learner baselines. Values are averaged over all 100 semi-synthethic data experiments on a single outcome metrics evaluated on the test set.

Table 2 shows that the LR-learner performs better than the Independent T-learners (for both XGBoost and linear model baselines) on PEHE as well as μ-risk\mu\textrm{-risk} metrics.

Experiment 3: Large Scale Real Data

We now turn to evaluating the LR-learner on a real dataset. We analyze a set of 53 randomized experiments run at Facebook during a 40 day period. These experiments are from various applications and are largely unrelated. In our sampled dataset, for each experiment we study 50,00050,000 deidentified users per treatment and control group in our training set, with an additional 10,00010,000 deidentified users in each group reserved for each of the test and validation sets.

We consider 5 main outcomes, or key performance indicators (KPIs), core to Facebook. These experiments are not necessarily designed to move these outcome metrics. Each user may be in multiple experiments, but vast majority of users are in only a single experiment.

As our XiX_{i} we use a set of 142 features that include user embeddings trained for Facebook recommender systems, pre-treatment values of the KPIs, and user demographics. Users’ pre-treatment features are taken before the 40 day period, and all post-treatment outcomes are recorded at the end of the 40 day period.

We define v(xi)v(x_{i}) as the output of a two-layered fully-connected neural network. The first layer takes in the set of 142 features as input and outputs a dimension dd representation, using a ReLU activation. The second layer outputs a dimension dd user embedding. The LR-learner is then trained over 250 epochs of the training data.

This model is hyperparameter tuned on learning rate, weight decay and the model hidden dimension, dd. We select the model with the lowest average risk across all 5 outcome measures computed on a validation dataset. When optimized for τ-risk\tau\textrm{-risk}, we use a learning rate of 5e-4, weight decay of 5e-3, and d=32d=32, and when optimized for μ-risk\mu\textrm{-risk}, we use a learning rate of 1e-4, weight decay of 1e-4, and d=64d=64.

We compare our LR-learner to a baseline of individual T-learners trained using XGBoost (Chen and Guestrin, 2016). As above, we tested other popular model classes (e.g. neural networks) for the independent T-learners but found that XGBoost performed best and required the least hyper parameter tuning to reach that level of performance.

The T-learners are trained on the same 142 feature set and are tuned with respect to the τ-risk\tau\textrm{-risk} or μ-risk\mu\textrm{-risk} on a held out datatest. Unlike the LR-learner where we use a single set of parameters with all experimental data pooled, we give the T-learner baseline an advantage by tuning each model for a given experiment and metric separately on the following parameters: learning rate, learning rate decay, max tree depth, min child weight, gamma regularization, and fraction of features/observations subsampled while training each tree.

Results

Refer to caption
Refer to caption
Figure 3: The correlations between treatments of independently estimated ITE matrices for each of the 55 outcome metrics and singular values of each ITE matrix suggest that the real ITE tensor is low rank.

The key assumption that makes our procedure useful is that there is information that can be gained about CATEs across metrics and across experiments; i.e., that the underlying ITE tensor is low-rank.

In our synthetic and semi-synthetic data, we could observe the ITEs/CATEs so we could verify this assumption in practice. In the real data, ground truth is unobserved, so we can assess the validity of the assumption by looking at the estimated ITE tensor from the independent T-learner procedure optimized for μ-risk\mu\textrm{-risk}.

We take each slice of this tensor for each metric. Let MjM^{j} denote such a slice. This is a matrix where a row is a unit and a column is an experiment and an entry is the unit’s estimated ITE on that metric.

Each column is estimated via an independent T-learner. Therefore, any correlation found in the column space is highly suggestive of an underlying correlation in true ITEs.

We evaluate the rank of each MjM^{j}. The plot of each matrix’s singular values (Figure 3) indeed suggests that the matrices appear to be effectively low rank. To quantify how low rank these matrices are, we use the bi-cross-validation procedure (Owen and Perry, 2009) to estimate the effective rank of each matrix. Using Wold-style 5-fold cross-validation, we find that the effective rank for each outcome measure is 3,5,3,4,33,5,3,4,3, respectively, confirming that all MjM^{j} are indeed low-rank relative to the number of experiments.

Having established that the ITE tensor is low rank, we ultimately wish to quantify how the low-rank structure improves CATE estimation for the LR-learner.

In Tables 3 and 4, we calculate the τ-risk\tau\textrm{-risk} and μ-risk\mu\textrm{-risk} averaged over all 53 experiments, for each metric, using models optimized for τ-risk\tau\textrm{-risk} and μ-risk\mu\textrm{-risk}, respectively. We notice that the LR-learner model optimized for μ-risk\mu\textrm{-risk} achieve relatively similar values of τ-risk\tau\textrm{-risk}, but has vastly improved μ-risk\mu\textrm{-risk}, which suggests that μ-risk\mu\textrm{-risk} is the more preferrable metric to optimize for. In table 3, the LR-learner outperforms the independent T-learners for both evaluation metrics, for all but one outcome.

Note that the difference in μ\mu-risks (and in prior experiments PEHE) has a straightforward interpretation (when outcomes are scaled to variance 11). The μ\mu-risk is just the out of sample percent of variance explained. By contrast, τ\tau-risk does not have such a simple interpretation since it is an unbiased but noisy proxy for the ITE. So, it is harder to say when there is a ‘big’ or ‘small’ difference in τ\tau-risk between two models.

τ-risk\tau\textrm{-risk} μ-risk\mu\textrm{-risk}
Ind.
T-learners
LR-learner
Ind.
T-learners
LR-learner
Metric 1 0.3747 0.3741 0.3637 0.3535
Metric 2 0.5566 0.5553 0.5515 0.5243
Metric 3 0.3219 0.3207 0.3270 0.3000
Metric 4 0.3977 0.3971 0.3749 0.3617
Metric 5 0.5357 0.5393 0.5354 0.4881
Table 3: The τ-risk\tau\textrm{-risk} and μ-risk\mu\textrm{-risk} averaged over all 53 experiments for each metric evaluated on a validation set. Models are parameter-tuned for μ-risk\mu\textrm{-risk}.
τ-risk\tau\textrm{-risk} μ-risk\mu\textrm{-risk}
Ind.
T-learners
LR-learner
Ind.
T-learners
LR-learner
Metric 1 0.3743 0.3741 0.3734 0.3792
Metric 2 0.5557 0.5552 0.5741 0.5524
Metric 3 0.3212 0.3205 0.3496 0.3237
Metric 4 0.3974 0.3970 0.3897 0.4009
Metric 5 0.5310 0.5328 0.5701 0.5205
Table 4: The τ-risk\tau\textrm{-risk} and μ-risk\mu\textrm{-risk} averaged over all 53 experiments for each metric evaluated on a validation set. Models are parameter-tuned for τ-risk\tau\textrm{-risk}.

Real World Deployment

So far, we have considered the case of a fixed dataset 𝒟\mathcal{D}, and an analyst interested in heterogeneous effects in all or a subset of the experiments/outcomes. We can also consider a scenario where the analyst already has access to an LR-learner trained on a large set of existing experiments. When a new experiment is run, the analyst may want to estimate HTEs for the new experiment. Call this new data 𝒟new\mathcal{D}_{\text{new}}. For simplicity, consider the case where there is just a single outcome metric of interest, and potentially this outcome measure is not one that is considered in past experiments.

One potential solution is to add the new experiment to the data, call this combined dataset 𝒟c\mathcal{D}_{c} and re-train the LR-learner on the complete dataset. However, when 𝒟\mathcal{D} is extremely large relative to 𝒟new\mathcal{D}_{\text{new}}, then the addition of the new data will not change the shared parameters across all experiments much. Thus, we can simply fix the parameters of the neural network, v𝒟()v_{\mathcal{D}}(\cdot), to the ones of the model trained on the original data. In other words, we can find new experiment treatment/control vectors enewe_{\text{new}} for the outcome in question by solving:

minenew(yit,xi)𝒟new(yv𝒟(xi)enewt)2.\min_{e_{\text{new}}}\sum_{(y_{it},x_{i})\in\mathcal{D}_{\text{new}}}(y-v_{\mathcal{D}}(x_{i})^{\prime}e^{t}_{\text{new}})^{2}.

In other words, we simply construct an independent T-learner on the new experiment. However, instead of using the original feature space xx and possibly a complex model, we use the pre-trained LR-learner as a feature extractor and a linear regression as the underlying model. This reduces the complexity of the learning problem substantially.

This is very similar to the idea of fine-tuning supervised (or unsupervised) models in other areas of machine learning such as vision or NLP where an embedding layer of a model trained on a large dataset is used as the feature space for a simpler model in a new dataset (Mikolov et al., 2013; Mahajan et al., 2018).

In practice, if an LR-learner were to be deployed in an organization, we would suggest the following process: Retrain the full model on a pool of experiment data intermittently. Save the model, and in particular, save the feature extractor v()v(\cdot). In daily analyst practice, when estimating heterogeneous treatment effects for a new experiment, reuse the saved v()v(\cdot) to train linear independent T-learners.

Conclusion

In this paper, we devise the LR-learner, an approach for simultaneous CATE estimation across multiple experiments and multiple outcomes. Our tensor factorization approach assumes a low rank structure in the ITE tensor across experiments and outcome metrics. We show that there is indeed a low-rank structure for CATEs across experiments and outcomes, which allows the LR-learner to be 200×200\times more sample efficent compared to individual CATE models. We show using real data from Facebook that the LR-learner outperforms the Independent T-learner approach in both τ-risk\tau\textrm{-risk} and μ-risk\mu\textrm{-risk} evaluation metrics.

Our approach can be especially useful in domains such as internet companies where experiments are continuously being run, potentially across various unrelated applications, and outcome metrics are inexpensive to obtain. This is in contrast to many fields like medicine in which panel data is more common, and in which experiments may be even more related, but multiple outcome metrics can be difficult to obtain. One avenue for future work is to evaluate the LR-learner on panel data and compare it to other causal matrix completion methods (Agarwal et al., 2021a; Athey et al., 2018a).

We have focused purely on the case of A/B tests, but another fruitful direction is to combine the LR-learner architecture with datasets that are either wholly or partially observational (Shalit et al., 2017; Lada et al., 2019). The architecture naturally extends to estimating treatment assigment probability as well as outcomes, so a possible future direction is incorporating either X-learner (Künzel et al., 2019) or doubly-robust (Kennedy, 2020) methods to the LR-learner.

We establish the contribution of the LR-Learner as a method for estimating CATEs, and we hope that future work explores potential applications for the learned embeddings. The user, metric, and experiment embeddings have many potential downstream uses, such as finding or comparing similar users or experiments. The learned embeddings may also inform the design of future experiments.

References

  • (1)
  • bez (2014) 2014. Amazon 2014 Annual Report. https://www.sec.gov/Archives/edgar/data/1018724/000119312514137753/d702518dex991.htm.
  • Agarwal et al. (2021a) Anish Agarwal, Munther Dahleh, Devavrat Shah, and Dennis Shen. 2021a. Causal Matrix Completion. arXiv:2109.15154 [econ.EM]
  • Agarwal et al. (2014) Alekh Agarwal, Daniel Hsu, Satyen Kale, John Langford, Lihong Li, and Robert Schapire. 2014. Taming the monster: A fast and simple algorithm for contextual bandits. In International Conference on Machine Learning. PMLR, 1638–1646.
  • Agarwal et al. (2021b) Anish Agarwal, Devavrat Shah, and Dennis Shen. 2021b. Synthetic Interventions. arXiv:2006.07691 [econ.EM]
  • Athey et al. (2018a) Susan Athey, Mohsen Bayati, Nikolay Doudchenko, Guido Imbens, and Khashayar Khosravi. 2018a. Matrix Completion Methods for Causal Panel Data Models. Working Paper 25132. National Bureau of Economic Research. https://doi.org/10.3386/w25132
  • Athey et al. (2018b) Susan Athey, Julie Tibshirani, and Stefan Wager. 2018b. Generalized Random Forests. arXiv:1610.01271 [stat.ME]
  • Bai and Ng (2021) Jushan Bai and Serena Ng. 2021. Matrix Completion, Counterfactuals, and Factor Analysis of Missing Data. arXiv:1910.06677 [econ.EM]
  • Chen and Guestrin (2016) Tianqi Chen and Carlos Guestrin. 2016. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 785–794.
  • Ding and Li (2018) Peng Ding and Fan Li. 2018. Causal inference: A missing data perspective. Statist. Sci. 33, 2 (2018), 214–237.
  • Dorie et al. (2019) Vincent Dorie, Jennifer Hill, Uri Shalit, Marc Scott, and Dan Cervone. 2019. Automated versus do-it-yourself methods for causal inference: Lessons learned from a data analysis competition. Statist. Sci. 34, 1 (2019), 43–68.
  • Dudík et al. (2014) Miroslav Dudík, Dumitru Erhan, John Langford, and Lihong Li. 2014. Doubly robust policy evaluation and optimization. Statist. Sci. 29, 4 (2014), 485–511.
  • Fernández-Val et al. (2021) Iván Fernández-Val, Hugo Freeman, and Martin Weidner. 2021. Low-Rank Approximations of Nonseparable Panel Models. arXiv:2010.12439 [econ.EM]
  • Gelman (2018) Andrew Gelman. 2018. You need 16 times the sample size to estimate an interaction than to estimate a main effect. https://statmodeling.stat.columbia.edu/2018/03/15/need-16-times-sample-size-estimate-interaction-estimate-main-effect/.
  • Gutierrez and Gérardy (2017) Pierre Gutierrez and Jean-Yves Gérardy. 2017. Causal Inference and Uplift Modelling: A Review of the Literature. In Proceedings of The 3rd International Conference on Predictive Applications and APIs (Proceedings of Machine Learning Research, Vol. 67), Claire Hardgrove, Louis Dorard, Keiran Thompson, and Florian Douetteau (Eds.). PMLR, 1–13. https://proceedings.mlr.press/v67/gutierrez17a.html
  • Hill (2011) Jennifer L Hill. 2011. Bayesian nonparametric modeling for causal inference. Journal of Computational and Graphical Statistics 20, 1 (2011), 217–240.
  • Imbens and Rubin (2015) Guido W Imbens and Donald B Rubin. 2015. Causal inference in statistics, social, and biomedical sciences. Cambridge University Press.
  • Kennedy (2020) Edward H Kennedy. 2020. Optimal doubly robust estimation of heterogeneous causal effects. arXiv preprint arXiv:2004.14497 (2020).
  • Kent et al. (2018) David M Kent, Ewout Steyerberg, and David van Klaveren. 2018. Personalized evidence based medicine: predictive approaches to heterogeneous treatment effects. Bmj 363 (2018).
  • Krizhevsky et al. (2009) Alex Krizhevsky, Geoffrey Hinton, et al. 2009. Learning multiple layers of features from tiny images. (2009).
  • Künzel et al. (2019) Sören R Künzel, Jasjeet S Sekhon, Peter J Bickel, and Bin Yu. 2019. Metalearners for estimating heterogeneous treatment effects using machine learning. Proceedings of the national academy of sciences 116, 10 (2019), 4156–4165.
  • Lada et al. (2019) Akos Lada, Alexander Peysakhovich, Diego Aparicio, and Michael Bailey. 2019. Observational data for heterogeneous treatment effects with application to recommender systems. In Proceedings of the 2019 ACM Conference on Economics and Computation. 199–213.
  • Lerer et al. (2019) Adam Lerer, Ledell Wu, Jiajun Shen, Timothee Lacroix, Luca Wehrstedt, Abhijit Bose, and Alex Peysakhovich. 2019. Pytorch-biggraph: A large-scale graph embedding system. arXiv preprint arXiv:1903.12287 (2019).
  • Mahajan et al. (2018) Dhruv Mahajan, Ross Girshick, Vignesh Ramanathan, Kaiming He, Manohar Paluri, Yixuan Li, Ashwin Bharambe, and Laurens Van Der Maaten. 2018. Exploring the limits of weakly supervised pretraining. In Proceedings of the European conference on computer vision (ECCV). 181–196.
  • Mikolov et al. (2013) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems 26 (2013).
  • Nickel et al. (2011) Maximilian Nickel, Volker Tresp, and Hans-Peter Kriegel. 2011. A three-way model for collective learning on multi-relational data. In Icml.
  • Nie and Wager (2020) Xinkun Nie and Stefan Wager. 2020. Quasi-Oracle Estimation of Heterogeneous Treatment Effects. arXiv:1712.04912 [stat.ML]
  • Owen and Perry (2009) Art B. Owen and Patrick O. Perry. 2009. Bi-cross-validation of the SVD and the nonnegative matrix factorization. The Annals of Applied Statistics 3, 2 (Jun 2009). https://doi.org/10.1214/08-aoas227
  • Peysakhovich and Lada (2016) Alexander Peysakhovich and Akos Lada. 2016. Combining observational and experimental data to find heterogeneous treatment effects. arXiv preprint arXiv:1611.02385 (2016).
  • Powers et al. (2018) Scott Powers, Junyang Qian, Kenneth Jung, Alejandro Schuler, Nigam H Shah, Trevor Hastie, and Robert Tibshirani. 2018. Some methods for heterogeneous treatment effect estimation in high dimensions. Statistics in medicine 37, 11 (2018), 1767–1787.
  • Rubin (1974) Donald B. Rubin. 1974. Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology 66, 5 (1974), 688–701.
  • Rubin (1978) Donald B Rubin. 1978. Bayesian inference for causal effects: The role of randomization. The Annals of statistics (1978), 34–58.
  • Schuler et al. (2018) Alejandro Schuler, Michael Baiocchi, Robert Tibshirani, and Nigam Shah. 2018. A comparison of methods for model selection when estimating individual treatment effects. arXiv:1804.05146 [stat.ML]
  • Shalit et al. (2017) Uri Shalit, Fredrik D Johansson, and David Sontag. 2017. Estimating individual treatment effect: generalization bounds and algorithms. In International Conference on Machine Learning. PMLR, 3076–3085.
  • Wager and Athey (2018) Stefan Wager and Susan Athey. 2018. Estimation and inference of heterogeneous treatment effects using random forests. J. Amer. Statist. Assoc. 113, 523 (2018), 1228–1242.
  • Welling and Weber (2001) Max Welling and Markus Weber. 2001. Positive tensor factorization. Pattern Recognition Letters 22, 12 (2001), 1255–1261.
  • Xie and Aurisset (2016) Huizhi Xie and Juliette Aurisset. 2016. Improving the sensitivity of online controlled experiments: Case studies at netflix. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 645–654.