Information Theoretic Counterfactual Learning from Missing-Not-At-Random Feedback
Abstract
Counterfactual learning for dealing with missing-not-at-random data (MNAR) is an intriguing topic in the recommendation literature since MNAR data are ubiquitous in modern recommender systems. Missing-at-random (MAR) data, namely randomized controlled trials (RCTs), are usually required by most previous counterfactual learning methods for debiasing learning. However, the execution of RCTs is extraordinarily expensive in practice. To circumvent the use of RCTs, we build an information-theoretic counterfactual variational information bottleneck (CVIB), as an alternative for debiasing learning without RCTs. By separating the task-aware mutual information term in the original information bottleneck Lagrangian into factual and counterfactual parts, we derive a contrastive information loss and an additional output confidence penalty, which facilitates balanced learning between the factual and counterfactual domains. Empirical evaluation on real-world datasets shows that our CVIB significantly enhances both shallow and deep models, which sheds light on counterfactual learning in recommendation that goes beyond RCTs.
1 Introduction
A surge of research shows that the real-world logging policy often collects missing-not-at-random (MNAR) data (or selective labels) [25]. For example, users tend to reveal ratings for items they like, thus the observed users’ feedback, usually described by click-through-rate (CTR), can be substantially higher than those not observed yet. Consider a mini system with two users and three items in Fig. 1, when ignoring the unobserved events, the estimated average CTR from the observed outcomes is . However, if the rest unobserved outcomes were , the true CTR would be . This gap between the factual and counterfactual ratings in MNAR situation further exaggerates due to path-dependence that the learned policy tends to overestimate on the observed events outcomes [32].
Item 1 | Item 2 | Item 3 | |
---|---|---|---|
User A | ✓ | ? | ? |
User B | ✗ | ? | ✓ |
Vast majority of existing works in the recommendation literature neglect the MNAR effect, as they mainly focus on designing novel architectures and training techniques for improving model performance on the observed events [8, 12, 24], where the objective function is designed in principle of empirical risk minimization (ERM) as
(1) |
is an event composed of user and item ; indicates whether the outcome of an event is observed (i.e., whether item is presented to user ); is the loss function taking true outcome and predicted outcome as its inputs; and and are the numbers of observed and unobserved events, respectively. However, is not an unbiased estimate of the true risk [32]:
(2) |
which indicates that the naive ERM-based method cannot guarantee model’s generalization ability on the counterfactuals.

As the distribution of depends on the deployed recommendation policy at past, as shown in Fig. 1, we can regard it a representative of the policy bias, which influences the distribution of factual events and then the learned policy. In order to alleviate the policy bias, there are a series of works emphasizing on employing randomized controlled trials (RCTs) [6] to collect the so-called unbiased dataset. By adopting a uniform policy that randomly displays items to users, the logged feedback can be regarded as missing at random (MAR), which is consistent with the underlying joint distribution . Therefore, one can either evaluate the model’s true generalization ability on MAR data [23], or utilize MAR data to debias learning via importance sampling [30]. Besides, Rosenfeld et al. [29] and Bonner and Vasile [5] proposed to employ domain adaptation from MNAR data to MAR data, in order to balance the predictive capability of the learned model over the factuals and counterfactuals. However, RCTs are extraordinarily expensive to be executed in a real-world recommender system. It is tricky because no solid theoretical definition of how large RCTs would be representative enough. It is questionable if small RCTs, compared with the enormous quantity of possible events, can lead to a proper estimate of the users’ true preference distribution.
Considering the above challenges, we try to mitigate MNAR sampling bias via an RCT-free paradigm. Different from the domain adaption based methods, we propose an information theoretic approach for learning a balanced model based on Information Bottleneck (IB) [36, 35], which is promising for learning a representation that is informative to labels and robust against input noise [2]. In particular, we extend the original IB to counterfactual learning with a variational approximation, termed Counterfactual Variational Information Bottleneck (CVIB). The principle of CVIB is to encourage the learned representation to be equally informative to both the factual and counterfactual outcomes, by specifying a contrastive information in order to improve the learned embeddings generalization ability on the counterfactual domain. Besides, the minimality term in CVIB contributes to pushing the embeddings independent of the policy bias. In summary, our contributions are as follows:
-
•
We establish a novel CVIB framework adapted from IB, which suggests new avenues in counterfactual learning from MNAR data in an RCT-free paradigm.
-
•
A novel solution is proposed to handle the optimization of CVIB on MNAR data, which specifies a contrastive information term associated with a minimality term.
-
•
We empirically investigate our method’s advantages on real-world datasets for correcting MNAR bias without the need of acquiring the RCTs.111Code is available at https://github.com/RyanWangZf/CVIB-Rec.
2 Counterfactual Variational Information Bottleneck
In this section, we start from the conception about the information bottleneck. Then, we propose our novel CVIB objective function by separating events into factual and counterfactual domains. After that, we provide insights in the interpretation of the minimality term in our CVIB.
2.1 Problem Setup
Embedding is a conceptually classical approach for modeling the user and item in rating prediction. For example, in collaborative filtering [22], it represents an event by concatenation as , then generates outcome prediction by . In deep learning models [13], there are multiple hidden layers sequentially processing the embedding , such as and for . In this work, we view the feedforward neural network layers as a Markov chain of successive representations, indicated with
(3) |
where is the true feedback returned by users (may be unknown to our learning algorithm). According to Tishby et al. [36], a standard information bottleneck has the following form
(4) |
where denotes mutual information of two random variables, and is a Lagrangian multiplier. Minimizing encourages the representation to be compressed, i.e., being minimal to . And, maximizing the second term encourages to be sufficient to task , such that can be predictive of . Essentially, in Eq. (4) can be regarded as a supervised loss term plus an additional information regularization term on the representation [1]. Moreover, there is another random variable that affects the appearance of events , but is not informative to the true outcomes , i.e., . In this scenario, we would like to be independent of , thus being free of policy bias.
The main challenge in adopting IB for optimization is that the mutual information in is cumbersome for calculation. Although previous works try to derive computable proxy for specific tasks [3], they are not suitable for MNAR data. Since we only have partial feedback about , i.e., the majority of events are counterfactuals, we have no access to their true outcomes. Next, we will turn to derive a new objective function addressing this challenge.
2.2 Building Contrastive Information Regularizer
For conciseness, we focus on a simple model with only the embedding layer , namely , and extend it to multi-layer scenario in Section 2.3. Specifically, the second term in Eq. (4) is mutual information between embedding and target task . We separate the embeddings into two parts: and , which represent factual and counterfactual embeddings, respectively, i.e., and .

As shown in Fig. 2, we factorize the original mutual information term by . We postulate that and are independent, therefore according to the chain rule of mutual information:
(5) |
However, as the outcomes of the counterfactuals are unknown, we have to identify another refined solution. Specifically, we cast Eq. (5) to
(6) |
from which we derive a contrastive term between and . This characterization is helpful for us to introduce a hyperparameter to control the degree of this contrastive penalty. We then rewrite the original IB loss to
(7) |
as our new objective function, where we propose an information theoretic regularization on and . Intuitively, minimizing this term corresponds to encouraging and to be equally informative to the targeted task variable , thus resulting in a more balanced model. More importantly, it does not need access to the counterfactual outcomes, which will be specified in Section 3.2.
2.3 Minimal Embedding Insensitive to Policy Bias
Aside from the task-aware mutual information in Eq. (7), corresponds to the minimality of the learned embedding. Recall that we assume that the event follows the generative process , where influences the appearance of . Because is independent to task , we hope the predicted outcome is not influenced by , or the learned embedding should contain low information about . In this viewpoint, following the practice by Achille and Soatto [1], we identify that the minimality term is actually beneficial for embedding’s insensitivity against the nuisance .
Proposition 1 (Minimal Representation Insensitive to Policy Bias).
With the Markov chain assumption defined by Eq. (3), for any hidden embedding , we can derive the upper bound of the
(8) |
where the last term is a constant with respect to the training process.
Please refer to Appendix A for a proof. Above proposition implies that an embedding is insensitive to the policy bias , by simply reducing . Meanwhile, by maximizing in IB Lagrangian, embedding will be forced to retain minimum information from that is pertinent to the task . In deep models, according to the Data Processing Inequality (DPI) [9], minimizing also works for controlling the policy bias of the successive layers.
3 Tractable Optimization Framework
The proposed is still intractable for optimization. In this section, we attempt to find a tractable solution for three terms in , respectively. And, we present our algorithm of learning debiased embeddings by at last.
3.1 Minimality Term
The minimality term in Eq. (7) can be measured with a Kullback-Leibler (KL) divergence as
(9) |
To avoid operating on the marginal , we use a variational approximation of as the marginal , which renders
(10) |
Suppose the posterior is a Gaussian distribution, where is the encoded embedding of input event and indicates a diagonal matrix with elements . In other words, we assume the embedding is generated by
(11) |
If we fix , then would default to a deterministic embedding. Moreover, by considering a standard Gaussian variational marginal , the KL term reduces to
(12) |
which means for a deterministic embedding, minimizing term is equivalent to directly applying -norm regularization on the embedding vector.
3.2 Contrastive Mutual Information Term
The mutual information , where demotes the entropy of . The first entropy term is constant, which means maximizing is equivalent to minimizing the second term . We then further derive the lower bound of as222Here we slightly abuse notation by denoting .
(13) |
The term is an estimate of with a classifier parameterized by , e.g., weight matrices in deep networks or embedding parameters. We can use the cross entropy as a proxy for the mutual information in the IB objective [1, 15], because as shown above. Therefore, replacing the contrastive term in Eq. (7) with obtains
(14) |
Since we assume and are independent, the generative process can be written as . In order to make the term tractable, we here approximate by ,333We may also pick as the approximation, which renders . However, it has less operational meaning and its last term cancels out with another task-aware term . then cast Eq.(14) to
(15) |
This term further goes to our final results as
(16) |
where the first term of the right hand side is the cross entropy between and . We identify that the second term is in line with the maximum entropy principle [17, 4] and the confidence penalty proposed by Pereyra et al. [27] that is imposed on the output distribution of deep neural networks. While out of our derivation, the confidence penalty is only imposed on the factual output . We hence propose to restrict the distance between factual and counterfactual posterior, which provably strengthens model’s generalization over the underlying users’ true preference distribution.
3.3 Task-aware Mutual Information Term
We next omit the superscript of in for simplicity in the following. Similar to the operation of task-aware mutual information in Eq. (13), we use an approximation to substitute
(17) |
which indicates that this term is a proxy of the cross entropy loss between the true outcome and the prediction .
3.4 Algorithm
Taking all the above derivations together, we conclude to the final objective function , which encompasses four terms:
(18) |
where term denotes the cross entropy between and ; term is cross entropy between and ; and term is entropy of . For computing this objective, we need to draw from the factual set and from the counterfactual set in one iteration. Specifically, term is supervised loss on the factual outcomes; terms , are contrastive regularization for balancing between factual and counterfactual domains; and term is minimality loss to improve the model’s robustness against policy bias. The whole optimization process over the is hence summarized in Algorithm 1, where we should specify the values of , and to balance these terms in optimization.
4 Empirical Evaluation
Aiming to validate CVIB’s effectiveness, we perform experiments on real-world datasets in this section. We elaborate on the experimental setups, and report the comparison results between CVIB and other baselines, which substantiate its all-round superiority in counterfactual learning.
4.1 Datasets
In order to evaluate the learned model’s generalization ability on the underlying groundtruth distribution , rather than the only on the logged feedback, i.e., the factual domain, we usually need an additional MAR test set, where the users are served with uniformly displayed items, namely RCTs. As far as we know, there are only two open datasets that satisfy this requirement:
Yahoo! R3 Dataset [25]. This is a user-song rating dataset, where there are over 300K ratings self-selected by 15,400 users in its training set, hence it is MNAR data. Besides, they collect an additional MAR test set by asking 5,400 users to rate 10 randomly displayed songs.
Coat Shopping Dataset [30]. This dataset consists of 290 users and 300 items. Each user rates 24 items by themselves, and is asked to rate 16 uniformly displayed items as the MAR test set.
We simplify the underlying rating prediction problem to binary classification in our experiments, by making rating which is or higher be positive feedback and those lower than as negative.
4.2 Baselines
Our CVIB is model-agnostic, thus applicable for most models in recommendation that take embeddings to encode the events, including the shallow and deep models. In our experiments, we pick matrix factorization (MF) [21] as shallow backbone and neural collaborative filtering (NCF) [13] as the deep backbone. Both of these methods project users and items into a shared space and represent them with unique vectors, namely embeddings.
The most popular technique to debias MNAR data by involving RCTs is the inverse propensity score (IPS) method [30]. Its variants, the self-normalized IPS (SNIPS) [33], doubly robust (DR) [18] and joint learning doubly robust (DRJL) [37] are also widely used. In our experiments, we take of test data to learn the propensity scores via a naive Bayes estimator [30]
(19) |
for IPS, SNIPS, DR, and DRJL. Applying these methods to both shallow and deep backbones, we involve muliple baselines in comparison with our MF-CVIB and NCF-CVIB. Note that only CVIB is RCT-free among all methods.
4.3 Experimental Protocol
We implement all the methods on PyTorch [26]. For both the MF and NCF, we fix the embedding size of both users and items to be because in our experiments, we find when embedding size gets larger, the performance of all methods on the MAR test set decays, which may be caused by overfitting. We randomly draw data from the training set for validation, on which we apply a grid search for hyperparameters to pick the best configuration. Adam [20] is utilized as the optimizer for fast convergence during training, with its learning rate in , weight decay in , and batch size in . For NCF, we set an additional hidden layer with width . Specifically for CVIB, we set the hyperparameters , and . Since we already set weight decay for Adam, we do not apply the -norm term on the embeddings. After finding out the best configuration on the validation set, we evaluate the trained models on the MAR test set.
COAT | YAHOO | |||
---|---|---|---|---|
MSE | AUC | MSE | AUC | |
MF | 0.2451 | 0.7020 | 0.2493 | 0.6767 |
+IPS [30] | 0.2299 | 0.7156 | 0.2260 | 0.6793 |
+SNIPS [33] | 0.2374 | 0.6960 | 0.1945 | 0.6810 |
+DR [18] | 0.2357 | 0.7058 | 0.2108 | 0.6883 |
+DRJL [37] | 0.2423 | 0.6915 | 0.2745 | 0.6892 |
+CVIB (ours) | 0.2189 | 0.7218 | 0.1671 | 0.7198 |
NCF | 0.2030 | 0.7688 | 0.3313 | 0.6772 |
+IPS [30] | 0.2008 | 0.7708 | 0.1777 | 0.6708 |
+SNIPS [33] | 0.1922 | 0.7695 | 0.1699 | 0.6880 |
+DR [18] | 0.2161 | 0.7514 | 0.1698 | 0.6886 |
+DRJL [37] | 0.2097 | 0.7579 | 0.2789 | 0.6820 |
+CVIB (ours) | 0.2017 | 0.7713 | 0.2820 | 0.6989 |
COAT | MF | IPS | SNIPS | DR | DRJL | CVIB |
---|---|---|---|---|---|---|
nDCG@5 | 0.589 | 0.633 | 0.603 | 0.622 | 0.608 | 0.663 |
nDCG@10 | 0.667 | 0.689 | 0.676 | 0.693 | 0.679 | 0.721 |
YAHOO | MF | IPS | SNIPS | DR | DRJL | CVIB |
nDCG@5 | 0.633 | 0.636 | 0.635 | 0.659 | 0.652 | 0.734 |
nDCG@10 | 0.762 | 0.760 | 0.762 | 0.774 | 0.770 | 0.820 |




4.4 Results
The evaluation results are reported in Table 2. There are three main observations:
(1) Even if they utilize additional RCTs, the IPS, SNIPS, DR and DRJL methods sometimes work worse than the naive model. By contrast, we identify that our RCT-free CVIB method is capable of enhancing both the shallow and deep models significantly in all experiments. It indicates that the contrastive regularization on the task-aware information, contained in embeddings of factual and counterfactual events, results in improvement of model generalization ability. To further evaluate our method in terms of ranking quality, we report the results of nDCG in Table 3. CVIB shows more significant gain over the baselines than on AUC.
(2) We perform repeated experiments to quantify our CVIB’s sensitivity to the balancing term weight and confident penalty term weight in Eq. (18). From Fig. 3 we identify that influences results significantly on COAT, while dose not make much difference on YAHOO. In general, increasing enhances the test AUC, which is aligned with our argument that contrastive information term benefits in balancing model between factual and counterfactual domains and then leads to better generalization ability. The confidence penalty plays less role in accuracy, but we should not set it too high to avoid underfitting.
(3) One would notice that the performance of NCF-CVIB in terms of MSE is relatively weak, while we argue that good recommendation does not necessarily rely on low MSE prediction. For instance, for the true outcomes , the predictions have much lower MSE than , although the former is obviously a worse prediction than the latter in terms of ranking. It turns out that NCF-CVIB overestimates the outcomes on the test set of YAHOO, nevertheless it still reaches the best ranking quality measured by AUC. In practice of recommender systems, instead of low MSE, we would rather appreciate high ranking quality, namely AUC, which demonstrates the model’s capability of finding out the positive examples.
5 Related Work
Counterfactual Learning from MNAR data. Propensity score based methods have been widely used in causal inference [16], in order to study the treatment effect between different policies. The nature of recommender systems decides that there is large gap between the factual and counterfactual events, i.e., MNAR data by non-uniform policy. In order to remove this sampling bias, inverse propensity score (IPS) method was adopted [28], then followed by doubly robust (DR) [11, 18], weighted DR [34], joint learning DR [37] and self-normalized IPS (SNIPS) [33], while all of these methods require additional RCTs for calculating propensity scores. Besides, recent works proposed to balance factual and counterfactual domains based on RCTs with domain adaptation [19], transfer learning [29], or multi-task paradigm [5]. On contrast, our method can alleviate policy bias by leveraging an information theoretic constraint on representation’s information on factual and counterfactual outcomes, which gets rid of the dependence on RCTs.
Information Theoretic Representation Learning. Information bottleneck (IB) is conceptually classical in information theory [36], and it has been used in explaining and guiding deep learning practice [31, 3]. The IB garners interests in its application for learning disentangled [14, 2], informative [7, 15], or compressed representation [10]. Above works are extension of the original IB objective loss that is suitable for a specific task. To the best of our knowledge, there is little work on adopting IB in recommendation, especially handling MNAR data. The proposed CVIB loss is based on contrastive information term adapted from the IB loss, which endows its capability of balancing learning with unknown counterfactual outcomes.
6 Conclusion
In this work, we proposed an information theoretic counterfactual learning framework in recommender systems. By separating the task-aware sufficiency term in the original information bottleneck objective into factual and counterfactual parts, we derived a contrastive information regularizer and a output confidence penalty, thus obtaining a novel CVIB objective function. The paradigm of information contraction encourages the learned model to be balanced between the factual and counterfactual domains, therefore improves its generalization ability significantly, which is validated by our empirical experiments. Our CVIB provides insight in utilizing information theoretic representation learning methods in recommendation, and sheds light on debiasing learning under MNAR data without performing expensive RCTs.
Acknowledgements
The research of Shao-Lun Huang was supported in part by the Natural Science Foundation of China under Grant 61807021, in part by the Shenzhen Science and Technology Program under Grant KQTD20170810150821146, and in part by the Innovation and Entrepreneurship Project for Overseas High-Level Talents of Shenzhen under Grant KQJSCX20180327144037831.
Broader Impact
In this work, we proposed a novel method for dealing with missing-not-at-random feedback in recommendation system. It has been criticized that current recommender systems are prone to overestimating the observed outcomes while underestimating those unobserved yet. As a result, users are restricted to a narrow scope of recommendation results. We would like to propose our method by considering unobserved events, namely counterfactuals, for counterfactual learning on missing-not-at-random feedback. We believe this method brings potential to develop a better recommender system on accuracy as well as diversity and fairness.
References
- Achille and Soatto [2018a] Alessandro Achille and Stefano Soatto. Emergence of invariance and disentanglement in deep representations. The Journal of Machine Learning Research, 19(1):1947–1980, 2018a.
- Achille and Soatto [2018b] Alessandro Achille and Stefano Soatto. Information dropout: Learning optimal representations through noisy computation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(12):2897–2905, 2018b.
- Alemi et al. [2016] Alexander A. Alemi, Ian Fischer, Joshua V. Dillon, and Kevin Murphy. Deep variational information bottleneck. arXiv preprint arXiv:1612.00410, 2016.
- Berger et al. [1996] Adam L. Berger, Vincent J. Della Pietra, and Stephen A. Della Pietra. A maximum entropy approach to natural language processing. Computational Linguistics, 22(1):39–71, 1996.
- Bonner and Vasile [2018] Stephen Bonner and Flavian Vasile. Causal embeddings for recommendation. In ACM Conference on Recommender Systems, pages 104–112, 2018.
- Chalmers et al. [1981] Thomas C. Chalmers, Harry Smith Jr., Bradley Blackburn, Bernard Silverman, Biruta Schroeder, Dinah Reitman, and Alexander Ambroz. A method for assessing the quality of a randomized control trial. Controlled Clinical Trials, 2(1):31–49, 1981.
- Chen et al. [2016] Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets. In Advances in Neural Information Processing Systems, pages 2172–2180, 2016.
- Cheng et al. [2016] Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, et al. Wide & deep learning for recommender systems. In Workshop on Deep Learning for Recommender Systems, pages 7–10, 2016.
- Cover and Thomas [2012] Thomas M. Cover and Joy A. Thomas. Elements of information theory. John Wiley & Sons, 2012.
- Dai et al. [2018] Bin Dai, Chen Zhu, and David Wipf. Compressing neural networks using the variational information bottleneck. arXiv preprint arXiv:1802.10399, 2018.
- Dudík et al. [2011] Miroslav Dudík, John Langford, and Lihong Li. Doubly robust policy evaluation and learning. arXiv preprint arXiv:1103.4601, 2011.
- Guo et al. [2017] Huifeng Guo, Ruiming Tang, Yunming Ye, Zhenguo Li, and Xiuqiang He. DeepFM: A factorization-machine based neural network for CTR prediction. arXiv preprint arXiv:1703.04247, 2017.
- He et al. [2017] Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. Neural collaborative filtering. In International Conference on World Wide Web, pages 173–182, 2017.
- Higgins et al. [2017] Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. Beta-VAE: Learning basic visual concepts with a constrained variational framework. International Conference on Learning Representations, 2(5):6, 2017.
- Hu et al. [2019] Jie Hu, Rongrong Ji, ShengChuan Zhang, Xiaoshuai Sun, Qixiang Ye, Chia-Wen Lin, and Qi Tian. Information competing process for learning diversified representations. In Advances in Neural Information Processing Systems, pages 2175–2186, 2019.
- Imbens and Rubin [2015] Guido W. Imbens and Donald B. Rubin. Causal inference in statistics, social, and biomedical sciences. Cambridge University Press, 2015.
- Jaynes [1957] Edwin T. Jaynes. Information theory and statistical mechanics. Physical Review, 106(4):620, 1957.
- Jiang and Li [2015] Nan Jiang and Lihong Li. Doubly robust off-policy value evaluation for reinforcement learning. arXiv preprint arXiv:1511.03722, 2015.
- Johansson et al. [2016] Fredrik Johansson, Uri Shalit, and David Sontag. Learning representations for counterfactual inference. In International Conference on Machine Learning, pages 3020–3029, 2016.
- Kingma and Ba [2014] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Koren [2008] Yehuda Koren. Factorization meets the neighborhood: A multifaceted collaborative filtering model. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 426–434, 2008.
- Koren et al. [2009] Yehuda Koren, Robert Bell, and Chris Volinsky. Matrix factorization techniques for recommender systems. Computer, (8):30–37, 2009.
- Lakkaraju et al. [2017] Himabindu Lakkaraju, Jon Kleinberg, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. The selective labels problem: Evaluating algorithmic predictions in the presence of unobservables. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, page 275–284, 2017. doi: 10.1145/3097983.3098066.
- Lian et al. [2018] Jianxun Lian, Xiaohuan Zhou, Fuzheng Zhang, Zhongxia Chen, Xing Xie, and Guangzhong Sun. xDeepFm: Combining explicit and implicit feature interactions for recommender systems. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 1754–1763, 2018.
- Marlin and Zemel [2009] Benjamin M. Marlin and Richard S. Zemel. Collaborative prediction and ranking with non-random missing data. In ACM Conference on Recommender Systems, pages 5–12. ACM, 2009.
- Paszke et al. [2019] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems, pages 8024–8035, 2019.
- Pereyra et al. [2017] Gabriel Pereyra, George Tucker, Jan Chorowski, Łukasz Kaiser, and Geoffrey Hinton. Regularizing neural networks by penalizing confident output distributions. arXiv preprint arXiv:1701.06548, 2017.
- Rosenbaum and Rubin [1983] Paul R. Rosenbaum and Donald B. Rubin. The central role of the propensity score in observational studies for causal effects. Biometrika, 70(1):41–55, 1983.
- Rosenfeld et al. [2017] Nir Rosenfeld, Yishay Mansour, and Elad Yom-Tov. Predicting counterfactuals from large historical data and small randomized trials. In International Conference on World Wide Web Companion, pages 602–609, 2017.
- Schnabel et al. [2016] Tobias Schnabel, Adith Swaminathan, Ashudeep Singh, Navin Chandak, and Thorsten Joachims. Recommendations as treatments: Debiasing learning and evaluation. arXiv preprint arXiv:1602.05352, 2016.
- Shwartz-Ziv and Tishby [2017] Ravid Shwartz-Ziv and Naftali Tishby. Opening the black box of deep neural networks via information. arXiv preprint arXiv:1703.00810, 2017.
- Steck [2013] Harald Steck. Evaluation of recommendations: rating-prediction and ranking. In ACM Conference on Recommender Systems, pages 213–220. ACM, 2013.
- Swaminathan and Joachims [2015] Adith Swaminathan and Thorsten Joachims. The self-normalized estimator for counterfactual learning. In Advances In Neural Information Processing Systems, pages 3231–3239, 2015.
- Thomas and Brunskill [2016] Philip Thomas and Emma Brunskill. Data-efficient off-policy policy evaluation for reinforcement learning. In International Conference on Machine Learning, pages 2139–2148, 2016.
- Tishby and Zaslavsky [2015] Naftali Tishby and Noga Zaslavsky. Deep learning and the information bottleneck principle. In IEEE Information Theory Workshop, pages 1–5. IEEE, 2015.
- Tishby et al. [2000] Naftali Tishby, Fernando C. Pereira, and William Bialek. The information bottleneck method. arXiv preprint physics/0004057, 2000.
- Wang et al. [2019] Xiaojie Wang, Rui Zhang, Yu Sun, and Jianzhong Qi. Doubly robust joint learning for recommendation on data missing not at random. In International Conference on Machine Learning, volume 97, pages 6638–6647. PMLR, 2019.
Appendix A Proof of Proposition 1
Proposition 1 (Minimal Representation Insensitive to Policy Bias).
With the Markov chain assumption defined by Eq. (3), for any hidden embedding , we can derive the upper bound of the
(A.1) |
where the last term is a constant with respect to the training process.
Proof.
From the Data Processing Inequality (DPI) [9], in this Markov chain, we can obtain
(A.2) |
For the second term , suppose and are independent, we can further factorize it and derive
(A.3) | ||||
(A.4) | ||||
(A.5) | ||||
(A.6) |
As we assume that is sufficient, we have = . Plugging above result back into Eq.(A.2) yields
(A.7) | ||||
(A.8) | ||||
(A.9) |
which indicates that bounds . And for any hidden embeddings , according to DPI, we have
(A.10) |
which yields the final result. ∎