This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Full Law Identification in Graphical Models of Missing Data:
Completeness Results

Razieh Nabi    Rohit Bhattacharya    Ilya Shpitser    Razieh Nabi    Rohit Bhattacharya    Ilya Shpitser

Supplementary Materials For:
Full Law Identification In Graphical Models Of Missing Data:
Completeness Results

Razieh Nabi    Rohit Bhattacharya    Ilya Shpitser    Razieh Nabi    Rohit Bhattacharya    Ilya Shpitser
Abstract

Missing data has the potential to affect analyses conducted in all fields of scientific study including healthcare, economics, and the social sciences. Several approaches to unbiased inference in the presence of non-ignorable missingness rely on the specification of the target distribution and its missingness process as a probability distribution that factorizes with respect to a directed acyclic graph. In this paper, we address the longstanding question of the characterization of models that are identifiable within this class of missing data distributions. We provide the first completeness result in this field of study – necessary and sufficient graphical conditions under which, the full data distribution can be recovered from the observed data distribution. We then simultaneously address issues that may arise due to the presence of both missing data and unmeasured confounding, by extending these graphical conditions and proofs of completeness, to settings where some variables are not just missing, but completely unobserved.

Missing Data, Identification, Missing Not At Random, Causality, Graphical Models, Selection Bias
Missing Data, Identification, Missing Not At Random, Causality, Graphical Models, Selection Bias

1 Introduction

Missing data has the potential to affect analyses conducted in all fields of scientific study, including healthcare, economics, and the social sciences. Strategies to cope with missingness that depends only on the observed data, known as the missing at random (MAR) mechanism, are well-studied (Dempster et al., 1977; Cheng, 1994; Robins et al., 1994; Tsiatis, 2006). However, the setting where missingness depends on covariates that may themselves be missing, known as the missing not at random (MNAR) mechanism, is substantially more difficult and under-studied (Fielding et al., 2008; Marston et al., 2010). MNAR mechanisms are expected to occur quite often in practice, for example, in longitudinal studies with complex patterns of dropout and re-enrollment, or in studies where social stigma may prompt non-response to questions pertaining to drug-use, or sexual activity and orientation, in a way that depends on other imperfectly collected or censored covariates (Robins & Gill, 1997; Vansteelandt et al., 2007; Marra et al., 2017).

Previous work on MNAR models has proceeded by imposing a set of restrictions on the full data distribution (the target distribution and its missingness mechanism) that are sufficient to yield identification of the parameter of interest. While there exist MNAR models whose restrictions cannot be represented graphically (Tchetgen Tchetgen et al., 2018), the restrictions posed in several popular MNAR models such as the permutation model (Robins & Gill, 1997), the block-sequential MAR model (Zhou et al., 2010), the itemwise conditionally independent nonresponse (ICIN) model (Shpitser, 2016; Sadinle & Reiter, 2017), and those in (Daniel et al., 2012; Thoemmes & Rose, 2013; Martel García, 2013; Mohan et al., 2013; Mohan & Pearl, 2014; Saadati & Tian, 2019) are either explicitly graphical or can be interpreted as such.

Despite the popularity of graphical modeling approaches for missing data problems, characterization of the class of missing data distributions identified as functionals of the observed data distribution has remained an open question (Bhattacharya et al., 2019). Several algorithms for the identification of the target distribution have been proposed (Mohan & Pearl, 2014; Shpitser et al., 2015; Tian, 2017; Bhattacharya et al., 2019). We show that even the most general algorithm currently published (Bhattacharya et al., 2019) still retains a significant gap in that there exist target distributions that are identified which the algorithm fails to identify. We then present what is, to our knowledge, the first completeness result for missing data models representable as directed acyclic graphs (DAGs) – a necessary and sufficient graphical condition under which the full data distribution is identified as a function of the observed data distribution. For any given field of study, such a characterization is one of the most powerful results that identification theory can offer, as it comes with the guarantee that if these conditions do not hold, the model is provably not identified.

We further generalize these graphical conditions to settings where some variables are not just missing, but completely unobserved. Such distributions are typically summarized using acyclic directed mixed graphs (ADMGs) (Richardson et al., 2017). We prove, once again, that our graphical criteria are sound and complete for the identification of full laws that are Markov relative to a hidden variable DAG and the resulting summary ADMG. This new result allows us to address two of the most critical issues in practical data analyses simultaneously, those of missingness and unmeasured confounding.

Finally, in the course of proving our results on completeness, we show that the proposed graphical conditions also imply that all missing data models of directed acyclic graphs or acyclic directed mixed graphs that meet these conditions, are in fact sub-models of the MNAR models in (Shpitser, 2016; Sadinle & Reiter, 2017). This simple, yet powerful result implies that the joint density of these models may be identified using an odds ratio parameterization that also ensures congenial specification of various components of the likelihood (Chen, 2007; Malinsky et al., 2019). Our results serve as an important precondition for the development of score-based model selection methods for graphical models of missing data, as an alternative to the constraint-based approaches proposed in (Strobl et al., 2018; Gain & Shpitser, 2018; Tu et al., 2019), and directly yield semi-parametric estimators using results in (Malinsky et al., 2019).

2 Preliminaries

A directed acyclic graph (DAG) 𝒢(V){\mathcal{G}}(V) consists of a set of nodes VV connected through directed edges such that there are no directed cycles. We will abbreviate 𝒢(V){\mathcal{G}}(V) as simply 𝒢,{\mathcal{G}}, when the vertex set is clear from the given context. Statistical models of a DAG 𝒢{\mathcal{G}} are sets of distributions that factorize as p(V)=ViVp(Vipa𝒢(Vi))p(V)=\prod_{V_{i}\in V}p(V_{i}\mid\operatorname{pa}_{\mathcal{G}}(V_{i})), where pa𝒢(Vi)\operatorname{pa}_{\mathcal{G}}(V_{i}) are the parents of ViV_{i} in 𝒢{\mathcal{G}}. The absence of edges between variables in 𝒢,{\mathcal{G}}, relative to a complete DAG entails conditional independence facts in p(V).p(V). These can be directly read off from the DAG 𝒢{\mathcal{G}} by the well-known d-separation criterion (Pearl, 2009). That is, for disjoint sets X,Y,ZX,Y,Z, the following global Markov property holds: (Xd-sepYZ)𝒢(XYZ)p(V).(X\perp\!\!\!\perp_{\text{d-sep}}Y\mid Z)_{\mathcal{G}}\implies(X\perp\!\!\!\perp Y\mid Z)_{p(V)}. When the context is clear, we will simply use XYZX\perp\!\!\!\perp Y\mid Z to denote the conditional independence between XX and YY given Z.Z.

In practice, some variables on the DAG may be unmeasured or hidden. In such cases, the distribution p(VU)p(V\cup U) is Markov relative to a hidden variable DAG 𝒢(VU),{\mathcal{G}}(V\cup U), where variables in UU are unobserved. There may be infinitely many hidden variable DAGs that imply the same set of conditional independences on the observed margin. Hence, it is typical to utilize a single acyclic directed mixed graph (ADMG) consisting of directed and bidirected edges that entails the same set of equality constraints as this infinite class (Evans, 2018). Such an ADMG 𝒢(V){\mathcal{G}}(V) is obtained from a hidden variable DAG 𝒢(VU){\mathcal{G}}(V\cup U) via the latent projection operator (Verma & Pearl, 1990) as follows. ABA\rightarrow B exists in 𝒢(V){\mathcal{G}}(V) if there exists a directed path from AA to BB in 𝒢(VU){\mathcal{G}}(V\cup U) with all intermediate vertices in U.U. An edge ABA\leftrightarrow B exists in 𝒢(V){\mathcal{G}}(V) if there exists a collider-free path (i.e., there are no consecutive edges of the form \rightarrow\circ\leftarrow) from AA to BB in 𝒢(VU){\mathcal{G}}(V\cup U) with all intermediate vertices in U,U, such that the first edge on the path is an incoming edge into AA and the final edge is an incoming edge into B.B.

Given a distribution p(VU)p(V\cup U) that is Markov relative to a hidden variable DAG 𝒢(V,U),{\mathcal{G}}(V,U), conditional independence facts pertaining to the observed margin p(V)p(V) can be read off from the ADMG 𝒢(V){\mathcal{G}}(V) by a simple analogue of the d-separation criterion, known as m-separation (Richardson, 2003), that generalizes the notion of a collider to include mixed edges of the form ,,\rightarrow\circ\leftrightarrow,\leftrightarrow\circ\leftarrow, and .\leftrightarrow\circ\leftrightarrow.

3 Missing Data Models

A missing data model is a set of distributions defined over a set of random variables {O,X(1),R,X}\{O,X^{(1)},R,X\}, where OO denotes the set of variables that are always observed, X(1)X^{(1)} denotes the set of variables that are potentially missing, RR denotes the set of missingness indicators of the variables in X(1)X^{(1)}, and XX denotes the set of the observed proxies of the variables in X(1).X^{(1)}. By definition missingness indicators are binary random variables; however, the state space of variables in X(1)X^{(1)} and OO are unrestricted. Given Xi(1)X(1)X^{(1)}_{i}\in X^{(1)} and its corresponding missingness indicator RiRR_{i}\in R, the observed proxy XiX_{i} is defined as XiXi(1)X_{i}\equiv X^{(1)}_{i} if Ri=1R_{i}=1, and Xi=?X_{i}=? if Ri=0R_{i}=0. Hence, p(XR,X(1))p(X\mid R,X^{(1)}) is deterministically defined. We call the non-deterministic part of a missing data distribution, i.e, p(O,X(1),R)p(O,X^{(1)},R), the full law, and partition it into two pieces: the target law p(O,X(1))p(O,X^{(1)}) and the missingness mechanism p(RX(1),O)p(R\mid X^{(1)},O). The censored version of the full law p(O,R,X),p(O,R,X), that the analyst actually has access to is known as the observed data distribution.

Following the convention in (Mohan et al., 2013), let 𝒢(V){\mathcal{G}}(V) be a missing data DAG, where V={OX(1)RX}.V=\{O\cup X^{(1)}\cup R\cup X\}. In addition to acyclicity, edges of a missing data DAG are subject to other restrictions: outgoing edges from variables in RR cannot point to variables in {X(1),O}\{X^{(1)},O\}, each XiXX_{i}\in X has only two parents in 𝒢,{\mathcal{G}}, i.e., RiR_{i} and Xi(1)X^{(1)}_{i} (these edges represent the deterministic function above that defines XiX_{i}, and are shown in gray in all the figures below), and there are no outgoing edges from XiX_{i} (i.e., the proxy XiX_{i} does not cause any variable on the DAG, however the corresponding full data variable Xi(1)X^{(1)}_{i} may cause other variables.) A missing data model associated with a missing data DAG 𝒢{\mathcal{G}} is the set of distributions p(O,X(1),R,X)p(O,X^{(1)},R,X) that factorizes as,

ViOX(1)Rp(Vipa𝒢(Vi))XiXp(XiXi(1),Ri).\displaystyle\prod_{V_{i}\in O\cup X^{(1)}\cup R}\ p(V_{i}\mid\operatorname{pa}_{\mathcal{G}}(V_{i}))\prod_{X_{i}\in X}\ p(X_{i}\mid X^{(1)}_{i},R_{i}).

By standard results on DAG models, conditional independences in p(X(1),O,R)p(X^{(1)},O,R) can still be read off from 𝒢{\mathcal{G}} by the d-separation criterion (Pearl, 2009). For convenience, we will drop the deterministic terms of the form p(XiXi(1),Ri)p(X_{i}\mid X^{(1)}_{i},R_{i}) from the identification analyses in the following sections since these terms are always identified by construction.

As an extension, we also consider a hidden variable DAG 𝒢(VU){\mathcal{G}}(V\cup U), where V={O,X(1),R,X}V=\{O,X^{(1)},R,X\} and variables in UU are unobserved, to encode missing data models in the presence of unmeasured confounders. In such cases, the full law would obey the nested Markov factorization (Richardson et al., 2017) with respect to a missing data ADMG 𝒢(V){\mathcal{G}}(V), obtained by applying the latent projection operator (Verma & Pearl, 1990) to the hidden variable DAG 𝒢(VU).{\mathcal{G}}(V\cup U). As a result of marginalization of latents U,U, there might exist bi-directed edges (to encode the hidden common causes) between variables in VV (bi-directed edges are shown in red in all the figures below). It is straightforward to see that a missing data ADMG obtained via projection of a hidden variable missing data DAG follows the exact same restrictions as stated in the previous paragraph (i.e., no directed cycles, pa𝒢(Xi)={Xi(1),Ri}\operatorname{pa}_{\mathcal{G}}(X_{i})=\{X^{(1)}_{i},R_{i}\}, every XiXX_{i}\in X is childless, and there are no outgoing edges from RiR_{i} to any variables in {X(1),O}\{X^{(1)},O\}.)

3.1 Identification in Missing Data Models

The goal of non-parametric identification in missing data models is twofold: identification of the target law p(O,X(1))p(O,X^{(1)}) or functions of it f(p(O,X(1))),f(p(O,X^{(1)})), and identification of the full law p(O,X(1),R),p(O,X^{(1)},R), in terms of the observed data distribution p(O,R,X).p(O,R,X).

A compelling reason to study the problem of identification of the full law in and of itself, is due to the fact that many popular methods for model selection or causal discovery, rely on the specification of a well-defined and congenial joint distribution (Chickering, 2002; Ramsey, 2015; Ogarrio et al., 2016). A complete theory of the characterization of missing data full laws that are identified opens up the possibility of adapting such methods to settings involving non-ignorable missingness, in order to learn not only substantive relationships between variables of interest in the target distribution, but also the processes that drive their missingness. This is in contrast to previous approaches to model selection under missing data that are restricted to submodels of a single fixed identified model (Strobl et al., 2018; Gain & Shpitser, 2018; Tu et al., 2019). Such an assumption may be impractical in complex healthcare settings, for example, where discovering the factors that lead to missingness or study-dropout may be just as important as discovering substantive relations in the underlying data.

Though the focus of this paper is on identification of the full law of missing data models that can be represented by a DAG (or a hidden variable DAG), some of our results naturally extend to identification of the target law (and functionals therein) due to the fact that the target law can be derived from the full law as Rp(O,X(1),R).\sum_{R}p(O,X^{(1)},R).

Remark 1.

By chain rule of probability, the target law p(O,X(1))p(O,X^{(1)}) is identified if and only if p(R=1O,X(1))p(R=1\mid O,X^{(1)}) is identified. The identifying functional is given by

p(O,X(1))=p(O,X(1),R=1)p(R=1O,X(1)).\displaystyle p(O,X^{(1)})=\frac{p(O,X^{(1)},R=1)}{p(R=1\mid O,X^{(1)})}.

(the numerator is a function of observed data by noting that X(1)=XX^{(1)}=X, and is observed when R=1R=1).

Remark 2.

The full law p(O,X(1),R)p(O,X^{(1)},R) is identified if and only if p(RO,X(1))p(R\mid O,X^{(1)}) is identified. According to Remark 1, the identifying functional is given by

p(O,X(1),R)=p(O,X(1),R=1)p(R=1O,X(1))×p(RO,X(1)).\displaystyle p(O,X^{(1)},R)=\frac{p(O,X^{(1)},R=1)}{p(R=1\mid O,X^{(1)})}\times p(R\mid O,X^{(1)}).

The rest of the paper is organized as follows. In Section 4, we explain, through examples, why none of the existing identification algorithms put forward in the literature are complete in the sense that there exist missing data DAGs whose full law and target law are identified but these algorithms fail to derive an identifying functional for them. In Section 5, we provide a complete algorithm for full law identification. In Section 6, we further extend our identification results to models where unmeasured confounders are present. We defer all proofs to the Appendix.

4 Incompleteness of Current Methods

In this section, we show that even the most general methods proposed for identification in missing data DAG models remain incomplete. In other words, we show that there exist identified MNAR models that are representable by DAGs, however all existing algorithms fail to identify both the full and target law for these models. For brevity, we use the procedure proposed in (Bhattacharya et al., 2019) as an exemplar. However, as it is the most general procedure in the current literature, failure to identify via this procedure would imply failure by all other existing ones. For each example, we also provide alternate arguments for identification that eventually lead to the general theory in Sections 5 and 6.

The algorithm proposed by (Bhattacharya et al., 2019) proceeds as follows. For each missingness indicator Ri,R_{i}, the algorithm tries to identify the distribution p(Ri|pa𝒢(Ri))|R=1,p(R_{i}|\operatorname{pa}_{\mathcal{G}}(R_{i}))|_{R=1}, sometimes referred to as the propensity score of Ri.R_{i}. It does so by checking if RiR_{i} is conditionally independent (given its parents) of the corresponding missingness indicators of its parents that are potentially missing. If this is the case, the propensity score is identified by a simple conditional independence argument (d-separation). Otherwise, the algorithm checks if this condition holds in post-fixing distributions obtained through recursive application of the fixing operator, which roughly corresponds to inverse weighting the current distribution by the propensity score of the variable being fixed (Richardson et al., 2017) (a more formal definition is provided in the Appendix.) If the algorithm succeeds in identifying the propensity score for each missingness indicator in this manner, then it succeeds in identifying the target law as Remark 1 suggests, since p(R=1|O,X(1))=RiRp(Ri|pa𝒢(R))|R=1.p(R=1|O,X^{(1)})=\prod_{R_{i}\in R}p(R_{i}|\operatorname{pa}_{\mathcal{G}}(R))|_{R=1}. Additionally, if it is the case that in the course of execution, the propensity score p(Ri|pa𝒢(Ri))p(R_{i}|\operatorname{pa}_{\mathcal{G}}(R_{i})) for each missingness indicator is also identified at all levels of its parents, then the algorithm also succeeds in identifying the full law (due to Remark 2).

In order to ground our theory in reality, we now describe a series of hypotheses that may arise during the course of a data analysis that seeks to study the link between the effects of smoking on bronchitis, through the deposition of tar or other particulate matter in the lungs. For each hypothesis, we ask if the investigator is able to evaluate the goodness of fit of the proposed model, typically expressed as a function of the full data likelihood, as a function of just the observed data. In other words, we ask if the full law is identified as a function of the observed data distribution. If it is, this enables the analyst to compare and contrast different hypotheses and select one that fits the data the best.

Setup. To start, the investigator consults a large observational database containing the smoking habits, measurements of particulate matter in the lungs, and results of diagnostic tests for bronchitis on individuals across a city. She notices however, that several entries in the database are missing. This leads her to propose a model like the one shown in Fig. 1(a), where X1(1),X2(1),X_{1}^{(1)},X_{2}^{(1)}, and X3(1)X_{3}^{(1)} correspond to smoking, particulate matter, and bronchitis respectively, and R1,R2,R_{1},R_{2}, and R3R_{3} are the corresponding missingness indicators.

For the target distribution p(X(1)),p(X^{(1)}), she proposes a simple mechanism that smoking leads to increased deposits of tar in the lungs, which in turn leads to bronchitis (X1(1)X2(1)X3(1)X_{1}^{(1)}\rightarrow X_{2}^{(1)}\rightarrow X_{3}^{(1)}). For the missingness process, she proposes that a suspected diagnosis of bronchitis is likely to lead to an inquiry about the smoking status of the patient (X3(1)R1X_{3}^{(1)}\rightarrow R_{1}), smokers are more likely to get tested for tar and bronchitis (X1(1)R2,X1(1)R3X_{1}^{(1)}\rightarrow R_{2},X_{1}^{(1)}\rightarrow R_{3}), and ordering a diagnostic test for bronchitis, increases the likelihood of ordering a test for tar, which in turn increases the likelihood of inquiry about smoking status (R1R2R3R_{1}\leftarrow R_{2}\leftarrow R_{3}).

We now show that for this preliminary hypothesis, if the investigator were to utilize the procedure described in (Bhattacharya et al., 2019) she may conclude that it is not possible to identify the full law. We go on to show that such a conclusion would be incorrect, as the full law is, in fact, identified, and provide an alternative means of identification.

X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}X3(1)X^{(1)}_{3}R1R_{1}R2R_{2}R3R_{3}X1X_{1}X2X_{2}X3X_{3}(a) 𝒢a{\mathcal{G}}_{a}X1X_{1}X2(1)X^{(1)}_{2}X3(1)X^{(1)}_{3}R1=1R_{1}=1R2R_{2}R3=1R_{3}=1X2X_{2}X3X_{3}(b) 𝒢b𝒢a(VR1){\mathcal{G}}_{b}\coloneqq{\mathcal{G}}_{a}(V\setminus R_{1})
Figure 1: (a) The missing data DAG used in scenario 1 (without the dashed edge X2(1)R3X^{(1)}_{2}\rightarrow R_{3}) and scenario 2 (with the dashed edge X2(1)R3X^{(1)}_{2}\rightarrow R_{3}) (b) Conditional DAG corresponding to the missing data DAG in (a) after fixing R1R_{1}, i.e., inverse weighting by the propensity score of R1R_{1}.

Scenario 1. Consider the missing data DAG model in Fig. 1(a) by excluding the edge X2(1)R3,X^{(1)}_{2}\rightarrow R_{3}, corresponding to the first hypothesis put forth by the investigator. The propensity score for R1R_{1} can be obtained by simple conditioning, noting that R1R3X3(1),R2R_{1}\perp\!\!\!\perp R_{3}\mid X_{3}^{(1)},R_{2} by d-separation. Hence, p(R1pa𝒢(R1))=p(R1X3(1),R2)=p(R1X3,R2,R3=1).p(R_{1}\mid\operatorname{pa}_{\mathcal{G}}(R_{1}))=p(R_{1}\mid X_{3}^{(1)},R_{2})=p(R_{1}\mid X_{3},R_{2},R_{3}=1).

Conditioning is not sufficient in order to identify the propensity score for R2,R_{2}, as R2⟂̸R1X1(1),R3R_{2}\not\perp\!\!\!\perp R_{1}\mid X_{1}^{(1)},R_{3}. However, it can be shown that in the distribution q(VR1R1=1)p(V)p(R1=1pa𝒢(R1))q(V\setminus R_{1}\mid R_{1}=1)\equiv\frac{p(V)}{p(R_{1}=1\mid\operatorname{pa}_{\mathcal{G}}(R_{1}))}, R2R1X1,R3=1R_{2}\perp\!\!\!\perp R_{1}\mid X_{1},R_{3}=1, since this distribution is Markov relative to the graph in Fig. 1(b) (see the Appendix for details). We use the notation q()q(\cdot\mid\cdot) to indicate that while qq acts in most respects as a conditional distribution, it was not obtained from p(V)p(V) by a conditioning operation. This implies that the propensity score for R2R_{2} (evaluated at R=1R=1) is identified as q(R2X1,R3=1,R1=1).q(R_{2}\mid X_{1},R_{3}=1,R_{1}=1).

Finally, we show that the algorithm in (Bhattacharya et al., 2019) is unable to identify the propensity score for R3.R_{3}. We first note that R3⟂̸R1X1(1)R_{3}\not\perp\!\!\!\perp R_{1}\mid X_{1}^{(1)} in the original problem. Furthermore, as shown in Fig. 1(b), fixing R1R_{1} leads to a distribution where R3R_{3} is necessarily selected on as the propensity score p(R1pa𝒢(R1))p(R_{1}\mid\operatorname{pa}_{\mathcal{G}}(R_{1})) is identified by restricting the data to cases where R3=1.R_{3}=1. It is thus impossible to identify the propensity score for R3R_{3} in this post-fixing distribution. The same holds if we try to fix R2R_{2} as identification of the propensity score for R2R_{2} required us to first fix R1,R_{1}, which we have seen introduces selection bias on R3.R_{3}.

Hence, the procedure in (Bhattacharya et al., 2019) fails to identify both the target law and the full law for the problem posed in Fig. 1(a). However, both these distributions are, in fact, identified as we now demonstrate.

A key observation is that even though the identification of p(R3X1(1))p(R_{3}\mid X^{(1)}_{1}) might not be so straightforward, p(R3X1(1),R2)p(R_{3}\mid X^{(1)}_{1},R_{2}) is indeed identified, because by d-separation R3R1X1(1),R2R_{3}\perp\!\!\!\perp R_{1}\mid X^{(1)}_{1},R_{2}, and therefore p(R3X1(1),R2)=p(R3X1,R2,R1=1).p(R_{3}\mid X^{(1)}_{1},R_{2})=p(R_{3}\mid X_{1},R_{2},R_{1}=1). Given that p(R3X1(1),R2)p(R_{3}\mid X^{(1)}_{1},R_{2}) and p(R2X1(1),R3=1)p(R_{2}\mid X^{(1)}_{1},R_{3}=1) are both identified (the latter is obtained through as described earlier), we consider exploiting an odds ratio parameterization of the joint density p(R2,R3pa𝒢(R2,R3))=p(R2,R3X1(1))p(R_{2},R_{3}\mid\operatorname{pa}_{{\mathcal{G}}}(R_{2},R_{3}))=p(R_{2},R_{3}\mid X^{(1)}_{1}). As we show below, such a parameterization immediately implies the identifiability of this density and consequently, the individual propensity scores for R2R_{2} and R3R_{3}.

Given disjoint sets of variables A,B,CA,B,C and reference values A=a0,B=b0,A=a_{0},B=b_{0}, the odds ratio parameterization of p(A,BC)p(A,B\mid C), given in (Chen, 2007), is as follows:

1Z×p(Ab0,C)×p(Ba0,C)×OR(A,BC),\displaystyle\frac{1}{Z}\times p(A\mid b_{0},C)\times p(B\mid a_{0},C)\times\text{OR}(A,B\mid C), (1)

where

OR(A=a,B=bC)\displaystyle\text{OR}(A=a,B=b\mid C)
=p(A=aB=b,C)p(A=a0B=b,C)×p(A=a0B=b0,C)p(A=aB=b0,C),\displaystyle\hskip 14.22636pt=\frac{p(A=a\mid B=b,C)}{p(A=a_{0}\mid B=b,C)}\times\frac{p(A=a_{0}\mid B=b_{0},C)}{p(A=a\mid B=b_{0},C)},

and ZZ is the normalizing term and is equal to

A,Bp(AB=b0,C)×p(BA=a0,C)×OR(A,BC).\displaystyle\displaystyle\sum_{A,B}p(A\mid B=b_{0},C)\times p(B\mid A=a_{0},C)\times\text{OR}(A,B\mid C).

Note that OR(A,BC)=OR(B,AC),{\small\text{OR}(A,B\mid C)=\text{OR}(B,A\mid C)}, i.e., the odds ratio is symmetric; see (Chen, 2007).

A convenient choice of reference value for the odds ratio in missing data problems is the value Ri=1.R_{i}=1. Given this reference level and the parameterization of the joint in Eq.  (1), we know that p(R2,R3X1(1))=1Z×p(R2R3=1,X1(1))×p(R3R2=1,X1(1))×OR(R2,R3X1(1)),p(R_{2},R_{3}\mid X^{(1)}_{1})=\frac{1}{Z}\times p(R_{2}\mid R_{3}=1,X^{(1)}_{1})\times p(R_{3}\mid R_{2}=1,X^{(1)}_{1})\times\text{OR}(R_{2},R_{3}\mid X^{(1)}_{1}), where ZZ is the normalizing term, and

OR(R2=r2,R3=r3X1(1))\displaystyle\text{OR}(R_{2}=r_{2},R_{3}=r_{3}\mid X^{(1)}_{1})
=p(R3=r3R2=r2,X1(1))p(R3=1R2=r2,X1(1))×p(R3=1R2=1,X1(1))p(R3=r3R2=1,X1(1)).\displaystyle\hskip 0.28436pt=\frac{p(R_{3}=r_{3}\mid R_{2}=r_{2},X^{(1)}_{1})}{p(R_{3}=1\mid R_{2}=r_{2},X^{(1)}_{1})}\times\frac{p(R_{3}=1\mid R_{2}=1,X^{(1)}_{1})}{p(R_{3}=r_{3}\mid R_{2}=1,X^{(1)}_{1})}.

The conditional pieces p(R2R3=1,X1(1))p(R_{2}\mid R_{3}=1,X^{(1)}_{1}) and p(R3R2=1,X1(1))p(R_{3}\mid R_{2}=1,X^{(1)}_{1}) are already shown to be functions of the observed data. To see that the odds ratio is also a function of observables, recall that R3R1R2,X1(1).R_{3}\perp\!\!\!\perp R_{1}\mid R_{2},X^{(1)}_{1}. This means that R1=1R_{1}=1 can be introduced into each individual piece of the odds ratio functional above, making it so that the entire functional depends only on observed quantities. Since all pieces of the odds ratio parameterization are identified as functions of the observed data, we can conclude that p(R2,R3X1(1))p(R_{2},R_{3}\mid X^{(1)}_{1}) is identified as the normalizing term is always identified if all the conditional pieces and the odds ratio are identified. This result, in addition to the fact that p(R1R2,X3(1))p(R_{1}\mid R_{2},X^{(1)}_{3}) is identified as before, leads us to the identification of both the target law and the full law, as the missingness process p(RX(1))p(R\mid X^{(1)}) is identified.

Scenario 2. Suppose the investigator is interested in testing an alternate hypothesis to see whether detecting high levels of particulate matter in the lungs, also serves as an indicator to physicians that a diagnostic test for bronchitis should be ordered. This corresponds to the missing data DAG model in Fig. 1(a) by including the edge X2(1)R3.X_{2}^{(1)}\rightarrow R_{3}. Since this is a strict super model of the previous example, the procedure in (Bhattacharya et al., 2019) still fails to identify the target and full laws in a similar manner as before.

However, it is still the case that both the target and full laws are identified. The justification for why the odds ratio parameterization of the joint density p(R2,R3pa𝒢(R2,R3))=p(R2,R3X1(1),X2(1))p(R_{2},R_{3}\mid\operatorname{pa}_{{\mathcal{G}}}(R_{2},R_{3}))=p(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) is identified in this scenario, is more subtle. We have,

p(R2,R3X1(1),X2(1))=1Z×p(R2R3=1,X1(1),X2(1))\displaystyle p(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2})=\frac{1}{Z}\times p(R_{2}\mid R_{3}=1,X^{(1)}_{1},X^{(1)}_{2})
×p(R3R2=1,X1(1),X2(1))×OR(R2,R3X1(1),X2(1)).\displaystyle\hskip 8.5359pt\times p(R_{3}\mid R_{2}=1,X^{(1)}_{1},X^{(1)}_{2})\times\text{OR}(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}).

Note that R2X2(1)R3,X1(1)R_{2}\perp\!\!\!\perp X^{(1)}_{2}\mid R_{3},X^{(1)}_{1}, and R3R1R2,X1(1),X2(1)R_{3}\perp\!\!\!\perp R_{1}\mid R_{2},X^{(1)}_{1},X^{(1)}_{2}. Therefore, p(R2R3=1,X1(1),X2(1))=p(R2R3=1,X1(1))p(R_{2}\mid R_{3}=1,X^{(1)}_{1},X^{(1)}_{2})=p(R_{2}\mid R_{3}=1,X^{(1)}_{1}) is identified the same way as described in Scenario 1, and p(R3R2=1,X1(1),X2(1))=p(R3R1=1,R2=1,X1,X2)p(R_{3}\mid R_{2}=1,X^{(1)}_{1},X^{(1)}_{2})=p(R_{3}\mid R_{1}=1,R_{2}=1,X_{1},X_{2}) is a function of the observed data and hence is identified. Now the identification of the joint density p(R2,R3X1(1),X2(1))p(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) boils down to identifiability of the odds ratio term. By symmetry, we can express the odds ratio OR(R2,R3X1(1),X2(1))\text{OR}(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) in two different ways,

OR(R2,R3X1(1),X2(1))\displaystyle\text{OR}(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2})
=p(R2R3,X1(1))p(R2=1R3,X1(1))×p(R2=1R3=1,X1(1))p(R2R3=1,X1(1))\displaystyle\!\!\!\!=\frac{p(R_{2}\mid R_{3},X^{(1)}_{1})}{p(R_{2}=1\mid R_{3},X^{(1)}_{1})}\times\frac{p(R_{2}=1\mid R_{3}=1,X^{(1)}_{1})}{p(R_{2}\mid R_{3}=1,X^{(1)}_{1})}
=p(R3|R2,X1(1),X2(1))p(R3=1|R2,X1(1),X2(1))×p(R3=1|R2=1,X1(1),X2(1))p(R3|R2=1,X1(1),X2(1)).\displaystyle\!\!\!\!=\frac{p(R_{3}|R_{2},X^{(1)}_{1},X^{(1)}_{2})}{p(R_{3}=1|R_{2},X^{(1)}_{1},X^{(1)}_{2})}\times\frac{p(R_{3}=1|R_{2}=1,X^{(1)}_{1},X^{(1)}_{2})}{p(R_{3}|R_{2}=1,X^{(1)}_{1},X^{(1)}_{2})}.

The first equality holds by d-separation (R2X2(1)R3,X1(1)R_{2}\perp\!\!\!\perp X^{(1)}_{2}\mid R_{3},X^{(1)}_{1}). This implies that OR(R2,R3X1(1),X2(1))\text{OR}(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) is not a function of X2(1).X^{(1)}_{2}. Let us denote this functional by f1(R2,R3,X1(1)).f_{1}(R_{2},R_{3},X^{(1)}_{1}). On the other hand, we can plug-in R1=1R_{1}=1 to pieces in the second equality since R3R1R2,X1(1),X2(1)R_{3}\perp\!\!\!\perp R_{1}\mid R_{2},X^{(1)}_{1},X^{(1)}_{2} (by d-separation.) This implies that OR(R2,R3X1(1),X2(1))\text{OR}(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) is a function of X1(1)X^{(1)}_{1} only through its observed values (i.e. X1X_{1}). Let us denote this functional by f2(R2,R3,X1,X2(1),R1=1).f_{2}(R_{2},R_{3},X_{1},X^{(1)}_{2},R_{1}=1). Since odds ratio is symmetric (by definition), then it must be the case that f1(R2,R3,X1(1))=f2(R2,R3,X1,X2(1),R1=1)f_{1}(R_{2},R_{3},X^{(1)}_{1})=f_{2}(R_{2},R_{3},X_{1},X^{(1)}_{2},R_{1}=1); concluding that f2f_{2} cannot be a function of X2(1)X^{(1)}_{2}, as the left hand side of the equation does not depend on X2(1)X^{(1)}_{2}. This renders f2f_{2} to be a function of only observed quantities, i.e. f2=f2(R2,R3,X1,R1=1)f_{2}=f_{2}(R_{2},R_{3},X_{1},R_{1}=1). This leads to the conclusion that p(R2,R3X1(1),X2(1))p(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) is identified and consequently the missingness process p(RX(1))p(R\mid X^{(1)}) in Fig. 1(a) is identified. According to Remarks 1 and 2, both the target and full laws are identified.

Adding any directed edge to Fig. 1(a) (including the dashed edge) allowed by missing data DAGs results in either a self-censoring edge (Xi(1)RiX_{i}^{(1)}\rightarrow R_{i}) or a special kind of collider structure called the colluder (Xj(1)RiRjX_{j}^{(1)}\rightarrow R_{i}\leftarrow R_{j}) in (Bhattacharya et al., 2019). We discuss in detail, the link between identification of missing data models of a DAG and the absence of these structures in Section 5.

Scenario 3. So far, the investigator has conducted preliminary analyses of the problem while ignoring the issue of unmeasured confounding. In order to address this issue, she first posits an unmeasured confounder U1,U_{1}, corresponding to genotypic traits that may predispose certain individuals to both smoke and develop bronchitis. She posits another unmeasured confounder U2,U_{2}, corresponding to the occupation of an individual, that may affect both the deposits of tar found in their lungs (for e.g., construction workers may accumulate more tar than an accountant due to occupational hazards) as well as limit an individual’s access to proper healthcare, leading to the absence of a diagnostic test for bronchitis.

The missing data DAG with unmeasured confounders, corresponding to the aforementioned hypothesis is shown in Fig. 2(a) (excluding the dashed edges). The corresponding missing data ADMG, obtained by latent projection is shown in Fig. 2(b) (excluding the dashed bidirected edge). A procedure to identify the full law of such an MNAR model, that is nested Markov with respect to a missing data ADMG, is absent from the current literature. The question that arises, is whether it is possible to adapt the odds ratio parameterization from the previous scenarios, to this setting.

We first note that by application of the chain rule of probability and Markov restrictions, the missingness mechanism still factorizes in the same way as in Scenario 2, i.e., p(RX(1))=p(R1R2,X3(1))×p(R2,R3X1(1),X2(1))p(R\mid X^{(1)})=p(R_{1}\mid R_{2},X^{(1)}_{3})\times p(R_{2},R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}) (Tian & Pearl, 2002). Despite the addition of the bidirected edges X1(1)X3(1)X_{1}^{(1)}\leftrightarrow X_{3}^{(1)} and X2(1)R3,X_{2}^{(1)}\leftrightarrow R_{3}, corresponding to unmeasured confounding, it is easy to see that the propensity score for R1R_{1} is still identified via simple conditioning. That is, p(R1pa𝒢(R1))=p(R1X3,R2,R3=1)p(R_{1}\mid\operatorname{pa}_{\mathcal{G}}(R_{1}))=p(R_{1}\mid X_{3},R_{2},R_{3}=1) as R1R3X3(1),R2R_{1}\perp\!\!\!\perp R_{3}\mid X_{3}^{(1)},R_{2} by m-separation. Furthermore, it can also be shown that the two key conditional independences that were exploited in the odds ratio parameterization of p(R2,R3X(1)),p(R_{2},R_{3}\mid X^{(1)}), still hold in the presence of these additional edges. In particular, R2X2(1)R3,X1(1)R_{2}\perp\!\!\!\perp X_{2}^{(1)}\mid R_{3},X_{1}^{(1)}, and R3R1R2,X1(1),X2(1),R_{3}\perp\!\!\!\perp R_{1}\mid R_{2},X_{1}^{(1)},X_{2}^{(1)}, by m-separation. Thus, the same odds ratio parameterization used for identification of the full law in Scenario 2, is also valid for Scenario 3. The full odds ratio parameterization of the MNAR models in Scenarios 2 and 3 is provided in Appendix B.

X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}X3(1)X^{(1)}_{3}R1R_{1}R2R_{2}R3R_{3}X1X_{1}X2X_{2}X3X_{3}U1U_{1}U3U_{3}U2U_{2}(a) 𝒢(V,U){\mathcal{G}}(V,U)X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}X3(1)X^{(1)}_{3}R1R_{1}R2R_{2}R3R_{3}X1X_{1}X2X_{2}X3X_{3}(b) 𝒢(V){\mathcal{G}}(V)
Figure 2: (a) The missing data DAG with unobserved confounders used in scenario 3 (without the dashed edges) and scenario 4 (with the dashed edges). (b) The corresponding missing data ADMGs obtained by applying the latent projection rules to the hidden variable DAG in (a).

Scenario 4. Finally, the investigator notices that a disproportionate number of missing entries for smoking status and diagnosis of bronchitis, correspond to individuals from certain neighborhoods in the city. She posits that such missingness may be explained by systematic biases in the healthcare system, where certain ethnic minorities may not be treated with the same level of care. This corresponds to adding a third unmeasured confounder U3,U_{3}, which affects the ordering of a diagnostic test for bronchitis as well as inquiry about smoking habits, as shown in Fig. 2(a) (including the dashed edges.) The corresponding missing data ADMG is shown in Fig. 2(b) (including the bidirected dashed edge.) Once again, we investigate if the full law is identified, in the presence of an additional unmeasured confounder U3U_{3}, and the corresponding bidirected edge R1R3.R_{1}\leftrightarrow R_{3}.

The missingness mechanism p(RX(1))p(R\mid X^{(1)}) in Fig. 2(b) (including the dashed edge) no longer follows the same factorization as the one described in Scenarios 2 and 3, due to the presence of a direct connection between R1R_{1} and R3.R_{3}. According to (Tian & Pearl, 2002), this factorization is given as p(RX(1))=p(R1R2,R3,X1(1),X2(1),X3(1))×p(R2R3,X1(1))×p(R3X1(1),X2(1))p(R\mid X^{(1)})=p(R_{1}\mid R_{2},R_{3},X^{(1)}_{1},X^{(1)}_{2},X^{(1)}_{3})\times p(R_{2}\mid R_{3},X^{(1)}_{1})\times p(R_{3}\mid X^{(1)}_{1},X^{(1)}_{2}). Unlike the previous scenarios, the propensity score of R1R_{1}, p(R1R2,R3,X1(1),X2(1),X3(1))p(R_{1}\mid R_{2},R_{3},X^{(1)}_{1},X^{(1)}_{2},X^{(1)}_{3}), includes X1(1),X2(1),X^{(1)}_{1},X^{(1)}_{2}, and R3R_{3} past the conditioning bar. Thus, the propensity score of R1R_{1} seems to be not identified, since there is no clear way of breaking down the dependency between R1R_{1} and X1(1).X^{(1)}_{1}. The problematic structure is the path X1(1)R3R1X^{(1)}_{1}\rightarrow R_{3}\leftrightarrow R_{1} which contains a collider at R3R_{3} that opens up when we condition on R3R_{3} in the propensity score of R1.R_{1}.

In light of the discussion in previous scenarios, another possibility for identifying p(RX(1))p(R\mid X^{(1)}) is through analysis of the odds ratio parameterization of the entire missingness mechanism. In Section 5, we provide a description of the general odds ratio parameterization on an arbitrary number of missingness indicators. For brevity, we avoid re-writing the formula here. We simply point out that the first step in identifying the missingness mechanism via the odds ratio parameterization is arguing whether conditional densities of the form p(RiRRi=1,X(1))p(R_{i}\mid R\setminus R_{i}=1,X^{(1)}) are identified, which is true if RiXi(1)RRi,X(1)Xi(1).R_{i}\perp\!\!\!\perp X^{(1)}_{i}\mid R\setminus R_{i},X^{(1)}\setminus X^{(1)}_{i}.

Such independencies do not hold in Fig. 2(b) (including the dashed edge) for any of the RRs, since there exist collider paths between every pair (Xi(1),Ri)(X^{(1)}_{i},R_{i}) that render the two variables dependent when we condition on everything outside Xi(1),RiX^{(1)}_{i},R_{i} (by m-separation). Examples of such paths are X1(1)R3R1X^{(1)}_{1}\rightarrow R_{3}\leftrightarrow R_{1} and X2(1)R3R1R2X^{(1)}_{2}\leftrightarrow R_{3}\leftrightarrow R_{1}\leftarrow R_{2} and X3(1)R1R3X^{(1)}_{3}\rightarrow R_{1}\leftrightarrow R_{3}.

In Section 6, we show that the structures arising in the missing data ADMG presented in Fig. 2(b) (including the dashed edge), give rise to MNAR models that are provably not identified without further assumptions.

5 Full Law Identification in DAGs

(Bhattacharya et al., 2019) proved that two graphical structures, namely the self-censoring edge (Xi(1)RiX_{i}^{(1)}\rightarrow R_{i}) and the colluder (Xj(1)RiRjX_{j}^{(1)}\rightarrow R_{i}\leftarrow R_{j}), prevent the identification of full laws in missing data models of a DAG. In this section we exploit an odds ratio parameterization of the missing data process to prove that these two structures are, in fact, the only structures that prevent identification, thus yielding a complete characterization of identification for the full law in missing data DAG models.

We formally introduce the odds ratio parameterization of the missing data process introduced in (Chen, 2007), as a more general version of the simpler form mentioned earlier in Eq. (1). Assuming we have KK missingness indicators, p(RX(1),O)p(R\mid X^{(1)},O) can be expressed as follows.

p(RX(1),O)=\displaystyle p(R\mid X^{(1)},O)= 1Z×k=1Kp(RkRk=1,X(1),O)\displaystyle\ \frac{1}{Z}\times\prod_{k=1}^{K}\ p(R_{k}\mid R_{-k}=1,X^{(1)},O)
×\displaystyle\times k=2KOR(Rk,RkRk=1,X(1),O),\displaystyle\prod_{k=2}^{K}\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)},O), (2)

where Rk=RRk,Rk={R1,,Rk1},Rk={Rk+1,,RK}R_{-k}=R\setminus R_{k},R_{\prec k}=\{R_{1},\ldots,R_{k-1}\},R_{\succ k}=\{R_{k+1},\ldots,R_{K}\}, and

OR(Rk,RkRk=1,X(1),O)\displaystyle\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)},O)
=p(RkRk=1,Rk,X(1),O)p(Rk=1Rk=1,Rk,X(1),O)\displaystyle\hskip 28.45274pt=\frac{p(R_{k}\mid R_{\succ k}=1,R_{\prec k},X^{(1)},O)}{p(R_{k}=1\mid R_{\succ k}=1,R_{\prec k},X^{(1)},O)}
×p(Rk=1Rk=1,X(1),O)p(RkRk=1,X(1),O).\displaystyle\hskip 42.67912pt\times\frac{p(R_{k}=1\mid R_{-k}=1,X^{(1)},O)}{p(R_{k}\mid R_{-k}=1,X^{(1)},O)}.

ZZ in Eq. (2) is the normalizing term and is equal to r{k=1Kp(rkRk=1,X(1),O)×k=2KOR(rk,rkRk=1,X(1),O)}\sum_{r}\{\prod_{k=1}^{K}\ p(r_{k}\mid R_{-k}=1,X^{(1)},O)\times\prod_{k=2}^{K}\text{OR}(r_{k},r_{\prec k}\mid R_{\succ k}=1,X^{(1)},O)\}.

Using the odds ratio reparameterization given in Eq. (2), we now show that under a standard positivity assumption, stating that p(RX(1),O)>δ>0p(R\mid X^{(1)},O)>\delta>0, with probability one for some constant δ\delta, the full law p(R,X(1),O)p(R,X^{(1)},O) of a missing data DAG is identified in the absence of self-censoring edges and colluders. Moreover, if any of these conditions are violated, the full law is no longer identified. We formalize this result below.

Theorem 1.

A full law p(R,X(1),O)p(R,X^{(1)},O) that is Markov relative to a missing data DAG 𝒢{\mathcal{G}} is identified if 𝒢{\mathcal{G}} does not contain edges of the form Xi(1)RiX_{i}^{(1)}\rightarrow R_{i} (no self-censoring) and structures of the form Xj(1)RiRjX_{j}^{(1)}\rightarrow R_{i}\leftarrow R_{j} (no colluders), and the stated positivity assumption holds. Moreover, the resulting identifying functional for the missingness mechanism p(RX(1),O)p(R\mid X^{(1)},O) is given by the odds ratio parameterization provided in Eq. 2, and the identifying functionals for the target law and full law are given by Remarks 1 and 2.

In what follows, we show that the identification theory that we have proposed for the full law in missing data models of a DAG is sound and complete. Soundness implies that when our procedure succeeds, the model is in fact identified, and the identifying functional is correct. Completeness implies that when our procedure fails, the model is provably not identified (non-parametrically). These two properties allow us to derive a precise boundary for what is and is not identified in the space of missing data models that can be represented by a DAG.

Theorem 2.

The graphical condition of no self-censoring and no colluders, put forward in Theorem 1, is sound and complete for the identification of full laws p(R,O,X(1))p(R,O,X^{(1)}) that are Markov relative to a missing data DAG 𝒢.{\mathcal{G}}.

We now state an important result that draws a connection between missing data models of a DAG 𝒢{\mathcal{G}} that are devoid of self-censoring and colluders, and the itemwise conditionally independent nonresponse (ICIN) model described in (Shpitser, 2016; Sadinle & Reiter, 2017). As a substantive model, the ICIN model implies that no partially observed variable directly determines its own missingness, and is defined by the restrictions that for every pair Xi(1),Ri,X_{i}^{(1)},R_{i}, it is the case that Xi(1)RiRi,Xi(1),O.X_{i}^{(1)}\perp\!\!\!\perp R_{i}\mid R_{-i},X_{-i}^{(1)},O. We utilize this result in the course of proving Theorem 2.

Lemma 1.

A missing data model of a DAG 𝒢{\mathcal{G}} that contains no self-censoring edges and no colluders, is a submodel of the ICIN model.

Xi(1)X^{(1)}_{i}RjR_{j}\cdotsRkR_{k}RiR_{i}Xi(1)X^{(1)}_{i}RjR_{j}\cdotsXk(1)X^{(1)}_{k}RiR_{i}Xi(1)X^{(1)}_{i}Xj(1)X^{(1)}_{j}\cdotsRkR_{k}RiR_{i}Xi(1)X^{(1)}_{i}Xj(1)X^{(1)}_{j}\cdotsXk(1)X^{(1)}_{k}RiR_{i}
Figure 3: All possible colluding paths between Xi(1)X_{i}^{(1)} and Ri.R_{i}. Each pair of dashed edges imply that the presence of either (or both) result in formation of a colluding path.

6 Full Law Identification in the Presence of Unmeasured Confounders

We now generalize identification theory of the full law to scenarios where some variables are not just missing, but completely unobserved, corresponding to the issues faced by the analyst in Scenarios 3 and 4 of Section 4. That is, we shift our focus to the identification of full data laws that are (nested) Markov with respect to a missing data ADMG 𝒢.{\mathcal{G}}.

Previously, we exploited the fact that the absence of colluders and self-censoring edges in a missing data DAG 𝒢,{\mathcal{G}}, imply a set of conditional independence restrictions of the form Xi(1)RiRi,Xi(1),OX^{(1)}_{i}\perp\!\!\!\perp R_{i}\mid R_{-i},X^{(1)}_{-i},O, for any pair Xi(1)X(1)X^{(1)}_{i}\in X^{(1)} and RiRR_{i}\in R (see Lemma 1). We now describe necessary and sufficient graphical conditions that must hold in a missing data ADMG 𝒢{\mathcal{G}} to imply this same set of conditional independences. Going forward, we ignore (without loss of generality), the deterministic factors p(XX(1),R),p(X\mid X^{(1)},R), and the corresponding deterministic edges in 𝒢,{\mathcal{G}}, in the process of defining this graphical criterion.

A colliding path between two vertices AA and BB is a path on which every non-endpoint node is a collider. We adopt the convention that ABA\rightarrow B and ABA\leftrightarrow B are trivially collider paths. We say there exists a colluding path between the pair (Xi(1),Ri)(X^{(1)}_{i},R_{i}) if Xi(1)X^{(1)}_{i} and RiR_{i} are connected through at least one non-deterministic colliding path i.e., one which does not pass through (using deterministic edges) variables in X.X.

We enumerate all possible colluding paths between a vertex Xi(1)X_{i}^{(1)} and its corresponding missingness indicator RiR_{i} in Fig. 3. Note that both the self-censoring structure and the colluding structure introduced in (Bhattacharya et al., 2019) are special cases of a colluding path. Using the m-separation criterion for ADMGs, it is possible to show that a missing data model of an ADMG 𝒢{\mathcal{G}} that contains no colluding paths of the form shown in Fig. 3, is also a submodel of the ICIN model in (Shpitser, 2016; Sadinle & Reiter, 2017).

Lemma 2.

A missing data model of an ADMG 𝒢{\mathcal{G}} that contains no colluding paths is a submodel of the ICIN model.

This directly yields a sound criterion for identification of the full law of missing data models of an ADMG 𝒢{\mathcal{G}} using the odds ratio parameterization as before.

Theorem 3.

A full law p(R,X(1),O)p(R,X^{(1)},O) that is Markov relative to a missing data ADMG 𝒢{\mathcal{G}} is identified if 𝒢{\mathcal{G}} does not contain any colluding paths and the stated positivity assumption in Section 5 holds. Moreover, the resulting identifying functional for the missingness mechanism p(RX(1),O)p(R\mid X^{(1)},O) is given by the odds ratio parametrization provided in Eq. 2.

We now address the question as to whether there exist missing data ADMGs which contain colluding paths but whose full laws are nevertheless identified. We show (see Appendix for proofs), that the presence of a single colluding path of any of the forms shown in Fig. 3, results in a missing data ADMG 𝒢{\mathcal{G}} whose full law p(X(1),R,O)p(X^{(1)},R,O) cannot be identified as a function of the observed data distribution p(X,R,O).p(X,R,O).

Lemma 3.

A full law p(R,X(1),O)p(R,X^{(1)},O) that is Markov relative to a missing data ADMG 𝒢{\mathcal{G}} containing a colluding path between any pair Xi(1)X(1)X_{i}^{(1)}\in X^{(1)} and RiRR_{i}\in R is not identified.

Revisiting our example in scenario 4, we note that every (Ri,Xi(1))(R_{i},X^{(1)}_{i}) pair is connected through at least one colluding path. Therefore, according to Lemma 3, the full law in Fig. 2(a) including the dashed edge, is not identified. It is worth emphasizing that the existence of at least one colluding path between any pair (Ri,Xi(1))(R_{i},X^{(1)}_{i}) is sufficient to conclude that the full law is not identified.

In what follows, we present a result on the soundness and completeness of our graphical condition that represents a powerful unification of non-parametric identification theory in the presence of non-ignorable missingness and unmeasured confounding. To our knowledge, such a result is the first of its kind. We present the theorem below.

Theorem 4.

The graphical condition of the absence of colluding paths, put forward in Theorem 3, is sound and complete for the identification of full laws p(X(1),R,O)p(X^{(1)},R,O) that are Markov relative to a missing data ADMG 𝒢.{\mathcal{G}}.

Throughout the paper, we have focused on identification of the full law which, according to Remark 1, directly yields identification for the target law. However, identification of the full law is a sufficient but not necessary condition for identification of the target law. In other words, the target law may still be identified despite the presence of colluding paths. Fig. 4(a) in (Bhattacharya et al., 2019) is an example of such a case where the full law is not identified due to the colluder structure at R2;R_{2}; however, as the authors argue the target law remains identified.

7 Conclusion

In this paper, we concluded an important chapter in the non-parametric identification theory of missing data models represented via directed acyclic graphs, possibly in the presence of unmeasured confounders. We provided a simple graphical condition to check if the full law, Markov relative to a (hidden variable) missing data DAG, is identified. We further proved that these criteria are sound and complete. Moreover, we provided an identifying functional for the missingness process, through an odds ratio parameterization that allows for congenial specification of components of the likelihood. Our results serve as an important precondition for the development of score-based model selection methods that consider a broader class of missing data distributions than the ones considered in prior works. An interesting avenue for future work is exploration of the estimation theory of functionals derived from the identified full data law. To conclude, we note that while identification of the full law is sufficient to identify the target law, there exist identified target laws where the corresponding full law is not identified. We leave a complete characterization of target law identification to future work.

Acknowledgements

This project is sponsored in part by the National Science Foundation grant 1939675, the Office of Naval Research grant N00014-18-1-2760, and the Defense Advanced Research Projects Agency under contract HR0011-18-C-0049. The content of the information does not necessarily reflect the position or the policy of the Government, and no official endorsement should be inferred.

For clearer presentation of materials in this supplement, we switch to a single-column format. In Appendix A, we provide an overview of the nested Markov model. We summarize the necessary concepts required in order to explain our proof of completeness for identification of the full law in missing data acyclic directed mixed graphs (ADMGs). These concepts draw on the binary parameterization of nested Markov models of an ADMG. In Appendix B, we provide a concrete example of the odds ratio parameterization. In Appendix C, we present proofs that were omitted from the main body of the paper for brevity.

A. Background: Fixing and Nested Markov Models of an ADMG

Given a DAG 𝒢(VU){\mathcal{G}}(V\cup U) where UU contains variables that are unobserved, the latent projection operator onto the observed margin produces an acyclic directed mixed graph 𝒢(V){\mathcal{G}}(V) that consists of directed and bidirected edges (Verma & Pearl, 1990). The bidirected connected components of an ADMG 𝒢(V),{\mathcal{G}}(V), partition the vertices VV into distinct sets known as districts. The district membership of a vertex ViV_{i} in 𝒢{\mathcal{G}} is denoted dis𝒢(Vi),\operatorname{dis}_{\mathcal{G}}(V_{i}), and the set of all districts in 𝒢{\mathcal{G}} is denoted 𝒟(𝒢).{\cal D}({\mathcal{G}}).

(Evans, 2018) showed that the nested Markov model (Richardson et al., 2017) of an ADMG 𝒢(V){\mathcal{G}}(V) is a smooth super model with fixed dimension, of the underlying latent variable model, that captures all equality constraints and avoids non-regular asymptotics arising from singularities in the parameter space (Drton, 2009; Evans, 2018). We use this fact in order to justify the use of nested Markov models of a missing data ADMG in order to describe full laws that are Markov relative to a missing data DAG with hidden variables. That is, the nested Markov model of a missing data ADMG 𝒢(V),{\mathcal{G}}(V), where V={O,X(1),R,X},V=\{O,X^{(1)},R,X\}, is a smooth super model of the missing data DAG model 𝒢(VU).{\mathcal{G}}(V\cup U). We also utilize nested Markov models of an ADMG 𝒢(VX(1)),{\mathcal{G}}(V\setminus X^{(1)}), corresponding to projection of the missing data ADMG 𝒢(V){\mathcal{G}}(V) onto variables that are fully observable. While such a model does not capture all equality constraints in the true observed law, it is still a smooth super model of it, thus providing an upper bound on the model dimension of the observed law.

CADMGs and Kernels

The nested Markov factorization of p(V)p(V) relative to an ADMG 𝒢(V){\mathcal{G}}(V) is defined with the use of conditional distributions known as kernels and their associated conditional ADMGs (CADMGs) that are derived from p(V)p(V) and 𝒢(V){\mathcal{G}}(V) respectively, via repeated applications of the fixing operator (Richardson et al., 2017). A CADMG 𝒢(V,W),{\mathcal{G}}(V,W), is an ADMG whose nodes can be partitioned into random variables VV and fixed variables W,W, with the restriction that only outgoing edges may be adjacent to variables in W.W. A kernel qV(VW)q_{V}(V\mid W) is a mapping from values in WW to normalized densities over VV i.e., vVqV(vw)=1\sum_{v\in V}q_{V}(v\mid w)=1 (Lauritzen, 1996). Conditioning and marginalization operations in kernels are defined in the usual way.

Fixing and Fixability

In Section 4 of the main paper, we provided an informal description of fixing as the operation of inverse-weighting by the propensity score of the variable being fixed; we now formalize this notion. A variable AVA\in V is said to be fixable if the paths AXA\rightarrow\cdots\rightarrow X and AXA\leftrightarrow\cdots\leftrightarrow X do not both exist for all XV{A}.X\in V\setminus\{A\}. Given a CADMG 𝒢(V,W){\mathcal{G}}(V,W) where AA is fixable, the graphical operator of fixing, denoted ϕA(𝒢),\phi_{A}({\mathcal{G}}), yields a new CADMG 𝒢(VA,WA){\mathcal{G}}(V\setminus A,W\cup A) with all incoming edges into AA being removed, and AA being set to a fixed value a.a. Given a kernel qV(VW),q_{V}(V\mid W), the corresponding probabilistic operation of fixing, denoted ϕA(qV;𝒢)\phi_{A}(q_{V};{\mathcal{G}}) yields a new kernel

qVA(VAWA)qV(VW)qV(Amb𝒢(A),W),\displaystyle q_{V\setminus A}(V\setminus A\mid W\cup A)\equiv\frac{q_{V}(V\mid W)}{q_{V}(A\mid\operatorname{mb}_{\mathcal{G}}(A),W)},

where mb𝒢(A)\operatorname{mb}_{\mathcal{G}}(A) is the Markov blanket of A,A, defined as the bidirected connected component (district) of AA (excluding AA itself) and the parents of the district of A,A, i.e., mb𝒢(A)dis𝒢(A)pa𝒢(dis𝒢(A)){A}.\operatorname{mb}_{\mathcal{G}}(A)\equiv\operatorname{dis}_{\mathcal{G}}(A)\cup\operatorname{pa}_{\mathcal{G}}(\operatorname{dis}_{\mathcal{G}}(A))\setminus\{A\}. It is easy to check that when 𝒢{\mathcal{G}} is a DAG, i.e., there are no bidirected edges, the denominator in the probabilistic operation of fixing, reduces to the familiar definition of a simple propensity score.

The notion of fixability can be extended to a set of variables SVS\subseteq V as follows. A set SS is said to be fixable if elements in SS can be ordered into a sequence σS=S1,S2,\sigma_{S}=\langle S_{1},S_{2},\dots\rangle such that S1S_{1} is fixable in 𝒢,{\mathcal{G}}, S2S_{2} is fixable in ϕS1(𝒢),\phi_{S_{1}}({\mathcal{G}}), and so on. This notion of fixability on sets of variables is essential to the description of the nested Markov model that we present in the following section.

Nested Markov Factorization

Given a CADMG 𝒢,{\mathcal{G}}, A set SVS\subseteq V is said to be reachable if there exists a valid sequence of fixing operations on vertices VS.V\setminus S. Further, SS is said to be intrinsic if it is reachable, and forms a single bidirected connected component or district in ϕσVS(𝒢),\phi_{\sigma_{V\setminus S}}({\mathcal{G}}), i.e., the CADMG obtained upon executing all fixing operations given by a valid fixing sequence σVS.\sigma_{V\setminus S}.

A distribution p(V)p(V) is said to obey the nested Markov factorization relative to an ADMG 𝒢(V){\mathcal{G}}(V) if for every fixable set S,S, and any valid fixing sequence σS,\sigma_{S},

ϕσS(p(V);𝒢)=D𝒟(ϕσS(𝒢))qD(DpaϕσS(𝒢)(D)),\displaystyle\phi_{\sigma_{S}}(p(V);{\mathcal{G}})=\prod_{D\in{\cal D}(\phi_{\sigma_{S}}({\mathcal{G}}))}q_{D}(D\mid\operatorname{pa}_{\phi_{\sigma_{S}}({\mathcal{G}})}(D)),

where all kernels appearing in the product above can be constructed by combining kernels corresponding to intrinsic sets i.e., {qI(Ipa𝒢(I))I is intrinsic in 𝒢}.\{q_{I}(I\mid\operatorname{pa}_{\mathcal{G}}(I))\mid I\text{ is intrinsic in }{\mathcal{G}}\}. Such a construction is made possible by the fact that all the sets DD quantified in the product are districts in a reachable graph derived from 𝒢.{\mathcal{G}}.

(Richardson et al., 2017) noted that when a distribution p(V)p(V) is nested Markov relative to an ADMG 𝒢,{\mathcal{G}}, all valid fixing sequences yield the same CADMG and kernel so that recursive applications of the fixing operator on a set VSV\setminus S can simply be denoted as ϕVS(𝒢)\phi_{V\setminus S}({\mathcal{G}}) and ϕVS(qV;𝒢)\phi_{V\setminus S}(q_{V};{\mathcal{G}}) without explicitly specifying any particular valid order. Thus, the construction of the set of kernels corresponding to intrinsic sets can be characterized as {qI(Ipa𝒢(I))I is intrinsic in 𝒢}={ϕVI(p(V;𝒢))I is intrinsic in 𝒢},\{q_{I}(I\mid\operatorname{pa}_{\mathcal{G}}(I))\mid I\text{ is intrinsic in }{\mathcal{G}}\}=\{\phi_{V\setminus I}(p(V;{\mathcal{G}}))\mid I\text{ is intrinsic in }{\mathcal{G}}\}, and the nested Markov factorization can be re-stated more simply as, for every fixable set SS we have,

ϕS(p(V;𝒢))=D𝒟(ϕS(𝒢))ϕVD(p(V);𝒢),\displaystyle\phi_{S}(p(V;{\mathcal{G}}))=\prod_{D\in{\cal D}\big{(}\phi_{S}({\mathcal{G}})\big{)}}\phi_{V\setminus D}(p(V);{\mathcal{G}}),

An important result from (Richardson et al., 2017) states that if p(VU)p(V\cup U) is Markov relative to a DAG 𝒢(VU),{\mathcal{G}}(V\cup U), then p(V)p(V) is nested Markov relative to the ADMG 𝒢(V){\mathcal{G}}(V) obtained by latent projection.

Binary Parameterization of Nested Markov Models

From the above factorization, it is clear that intrinsic sets given their parents form the atomic units of the nested Markov model. Using this observation, a smooth parameterization of discrete nested Markov models was provided by (Evans & Richardson, 2014). We now provide a short description of how to derive the so-called Moebius parameters of a binary nested Markov model.

For each district D𝒟(𝒢),D\in{\cal D}({\mathcal{G}}), consider all possible subsets SD.S\subseteq D. If SS is intrinsic (that is, reachable and bidirected connected in ϕVS(𝒢)\phi_{V\setminus S}({\mathcal{G}})), define the head HH of the intrinsic set to be all vertices in SS that are childless in ϕVS(𝒢),\phi_{V\setminus S}({\mathcal{G}}), and the tail TT to be all parents of the head in the CADMG ϕVS(𝒢),\phi_{V\setminus S}({\mathcal{G}}), excluding the head itself. More formally, H{ViSchϕ(𝒢)VS(Vi)=},H\equiv\{V_{i}\in S\mid\operatorname{ch}_{\phi_{{}_{V\setminus S}({\mathcal{G}})}}(V_{i})=\emptyset\}, and Tpaϕ(𝒢)VS(H)H.T\equiv\operatorname{pa}_{\phi_{{}_{V\setminus S}({\mathcal{G}})}}(H)\setminus H. The corresponding set of Moebius parameters for this intrinsic head and tail pair parameterizes the kernel qS(H=0T),q_{S}(H=0\mid T), i.e., the kernel where all variables outside the intrinsic set SS are fixed, and all elements of the head are set to zero given the tail. Note that these parameters are, in general, variationally dependent (in contrast to variationally independent in the case of an ordinary DAG model) as the heads and tails in these parameter sets may overlap. The joint density for any query p(V=v),p(V=v), can be obtained through the Moebius inversion formula; see (Lauritzen, 1996; Evans & Richardson, 2014) for details. For brevity, we will denote qS(H=0T)q_{S}(H=0\mid T) as simply q(H=0T),q(H=0\mid T), as it will be clear from the given context what variables are still random in the kernel corresponding to a given intrinsic set.

Binary Parameterization of Missing Data ADMGs

We use the parameterization described in the previous section in order to count the number of parameters required to parameterize the full law of a missing data ADMG and its corresponding observed law. We then use this to reason that if the number of parameters in the full law exceeds those in the observed law, it is impossible to establish a map from the observed law to the full law. This in turn implies that such a full law is not identified.

The binary parameterization of the full law of a missing data ADMG 𝒢(X(1),O,R,X){\mathcal{G}}(X^{(1)},O,R,X) is exactly the same as that of an ordinary ADMG, except that the deterministic factors p(XiRi,Xi(1)),p(X_{i}\mid R_{i},X_{i}^{(1)}), can be ignored, as Xi=Xi(1)X_{i}=X_{i}^{(1)} with probability one when Ri=1,R_{i}=1, and Xi=?X_{i}=? with probability one when Ri=0.R_{i}=0.

The observed law is parameterized as follows. First, variables in X(1)X^{(1)} are treated as completely unobserved, and an observed law ADMG 𝒢(X,O,R){\mathcal{G}}(X,O,R) is obtained by applying the latent projection operator to 𝒢(X(1),O,R,X).{\mathcal{G}}(X^{(1)},O,R,X). The Moebius parameters are then derived in a similar manner as before, with the additional constraint that if XiXX_{i}\in X appears in the head of a Moebius parameter, and the corresponding missingness indicator RiR_{i} appears in the tail, then the kernel must be restricted to cases where Ri=1.R_{i}=1. This is because when Ri=0,R_{i}=0, the probability of the head taking on any value, aside from those where Xi=?,X_{i}=?, is deterministically defined to be 0.0.

Note that parameterizing the observed law by treating variables in X(1)X^{(1)} as fully unobserved does not quite capture all equality constraints that may be detectable in the observed law, as these variables are, in fact, sometimes observable when their corresponding missingness indicators are set to one. Indeed, a smooth parameterization of the observed law of missing data models that captures all constraints implied by the model, is still an open problem. Nevertheless, parameterizing an observed law ADMG, such as the one mentioned earlier, provides an upper bound on the number of parameters required to parameterize the true observed law. This suffices for our purposes, as demonstrating that the upper bound on the number of parameters in the observed law is less than the number of parameters in the full law, is sufficient to prove that the full law is not identified.

B. Example: Odds Ratio Parameterization

X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}X3(1)X^{(1)}_{3}R1R_{1}R2R_{2}R3R_{3}X1X_{1}X2X_{2}X3X_{3}(a) 𝒢a{\mathcal{G}}_{a}X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}X3(1)X^{(1)}_{3}R1R_{1}R2R_{2}R3R_{3}X1X_{1}X2X_{2}X3X_{3}(b) 𝒢b{\mathcal{G}}_{b}
Figure 4: (a) The missing data DAG model used in Scenario 2. (b) the missing data ADMG model used in Scenario 3.

To build up a more concrete intuition for Theorems 1 and 3, we provide an example of the odds ratio parameterization for the missing data models used in Scenarios 2 and 3 of the main paper, reproduced here in Figs. 4(a, b). Utilizing the order R1,R2,R3R_{1},R_{2},R_{3} on the missingness indicators, the odds ratio parameterization of the missing data process for both models is as follows.

1Z×(k=13p(RiRi=1,X(1)))×OR(R1,R2,R3=1,X(1))×OR(R3,(R1,R2)X(1)).\displaystyle\frac{1}{Z}\times\bigg{(}\prod_{k=1}^{3}p(R_{i}\mid R_{-i}=1,X^{(1)})\bigg{)}\times\text{OR}(R_{1},R_{2},\mid R_{3}=1,X^{(1)})\times\text{OR}(R_{3},(R_{1},R_{2})\mid X^{(1)}). (3)

We now argue that each piece in Eq. 3 is identified. Note that, in the missing data DAG shown in Fig. 4(a), RiXi(1)Ri,Xi(1)R_{i}\perp\!\!\!\perp X_{i}^{(1)}\mid R_{-i},X_{-i}^{(1)} by d-separation. The same is true for the missing data ADMG in Fig. 4(b) by m-separation. Thus, in both cases, the product over conditional pieces of each RiR_{i} given the remaining variables is not a function Xi(1),X_{i}^{(1)}, and is thus a function of observed data. We now show that OR(R1,R2R3=1,X(1))\text{OR}(R_{1},R_{2}\mid R_{3}=1,X^{(1)}) is not a function of X1(1),X2(1)X_{1}^{(1)},X_{2}^{(1)} by utilizing the symmetry property of the odds ratio.

OR(R1,R2R3=1,X(1))\displaystyle\text{OR}(R_{1},R_{2}\mid R_{3}=1,X^{(1)}) =p(R1R2,R3=1,X2(1),X3(1))p(R1=1R2,R3=1,X2(1),X3(1))×p(R1=1R2=1,R3=1,X2(1),X3(1))p(R1R2=1,R3=1,X2(1),X3(1))\displaystyle=\frac{p(R_{1}\mid R_{2},R_{3}=1,X_{2}^{(1)},X_{3}^{(1)})}{p(R_{1}=1\mid R_{2},R_{3}=1,X_{2}^{(1)},X_{3}^{(1)})}\times\frac{p(R_{1}=1\mid R_{2}=1,R_{3}=1,X_{2}^{(1)},X_{3}^{(1)})}{p(R_{1}\mid R_{2}=1,R_{3}=1,X_{2}^{(1)},X_{3}^{(1)})}
=OR(R2,R1R3=1,X(1))\displaystyle=\text{OR}(R_{2},R_{1}\mid R_{3}=1,X^{(1)}) =p(R2R1,R3=1,X1(1),X3(1))p(R2=1R1,R3=1,X1(1),X3(1))×p(R2=1R1=1,R3=1,X1(1),X3(1))p(R2R1=1,R3=1,X1(1),X3(1)).\displaystyle=\frac{p(R_{2}\mid R_{1},R_{3}=1,X_{1}^{(1)},X_{3}^{(1)})}{p(R_{2}=1\mid R_{1},R_{3}=1,X_{1}^{(1)},X_{3}^{(1)})}\times\frac{p(R_{2}=1\mid R_{1}=1,R_{3}=1,X_{1}^{(1)},X_{3}^{(1)})}{p(R_{2}\mid R_{1}=1,R_{3}=1,X_{1}^{(1)},X_{3}^{(1)})}.

Thus, from the first equality, the odds ratio is not a function of X2(1)X_{2}^{(1)} as R1X1(1)R1,X1(1)R_{1}\perp\!\!\!\perp X_{1}^{(1)}\mid R_{-1},X_{-1}^{(1)} by d-separation in Fig. 4(a) and by m-separation in Fig. 4(b). A symmetric argument holds for X2(1)X_{2}^{(1)} and R2R_{2} as seen in the second and third equalities. Hence, the odds ratio is only a function of X3(1),X_{3}^{(1)}, which is observable, as the function is evaluated at R3=1.R_{3}=1.

We now utilize an identity from (Chen et al., 2015) in order to simplify the final term in Eq. 3. That is,

OR(R3,(R1,R2)X(1))\displaystyle\text{OR}(R_{3},(R_{1},R_{2})\mid X^{(1)}) =OR(R3,R2R1=1,X(1))OR(R3,R1R2,X(1))\displaystyle=\text{OR}(R_{3},R_{2}\mid R_{1}=1,X^{(1)})\ \text{OR}(R_{3},R_{1}\mid R_{2},X^{(1)})
=OR(R3,R2R1=1,X(1))OR(R3,R1R2=1,X(1))OR(R3,R1R2,X(1))OR(R3,R1R2=1,X(1))f(R1,R2,R3X(1)).\displaystyle=\text{OR}(R_{3},R_{2}\mid R_{1}=1,X^{(1)})\ \text{OR}(R_{3},R_{1}\mid R_{2}=1,X^{(1)})\ \underset{f(R_{1},R_{2},R_{3}\mid X^{(1)})}{\underbrace{\frac{\text{OR}(R_{3},R_{1}\mid R_{2},X^{(1)})}{\text{OR}(R_{3},R_{1}\mid R_{2}=1,X^{(1)})}}}.

The first two pairwise odds ratio terms are functions of observed data using an analogous argument that draws on the symmetry property of the odds ratio and the conditional independence RiXiRi,Xi(1),R_{i}\perp\!\!\!\perp X_{i}\mid R_{-i},X_{-i}^{(1)}, as before. The final term f(R1,R2,R3X(1))f(R_{1},R_{2},R_{3}\mid X^{(1)}), is a three-way interaction term on the odds ratio scale and can be expressed in three different ways as follows (Chen et al., 2015),

OR(R3,R1R2,X(1))OR(R3,R1R2=1,X(1))=OR(R2,R3R1,X(1))OR(R2,R3R1=1,X(1))=OR(R1,R2R3,X(1))OR(R1,R2R3=1,X(1)).\displaystyle\frac{\text{OR}(R_{3},R_{1}\mid R_{2},X^{(1)})}{\text{OR}(R_{3},R_{1}\mid R_{2}=1,X^{(1)})}=\frac{\text{OR}(R_{2},R_{3}\mid R_{1},X^{(1)})}{\text{OR}(R_{2},R_{3}\mid R_{1}=1,X^{(1)})}=\frac{\text{OR}(R_{1},R_{2}\mid R_{3},X^{(1)})}{\text{OR}(R_{1},R_{2}\mid R_{3}=1,X^{(1)})}.

From the first equality, we note by symmetry of the odds ratio and conditional independence that ff is not a function of X1(1),X3(1)X^{(1)}_{1},X^{(1)}_{3}. Similarly, from the second equality, we note that ff is not a function of X2(1),X3(1)X^{(1)}_{2},X^{(1)}_{3}. Finally, from the third equality, we note that ff is not a function of X1(1),X2(1)X^{(1)}_{1},X^{(1)}_{2}. Therefore, ff is not a function of X1(1),X2(1),X3(1)X^{(1)}_{1},X^{(1)}_{2},X^{(1)}_{3} and is identified.

The normalizing function Z,Z, is a function of all the pieces that we have already shown to be identified, and is therefore also identified. Thus, the missing data mechanisms p(RX(1)),p(R\mid X^{(1)}), and consequently, the full laws corresponding to the missing data graphs shown in Figs. 4(a,b) are identified by Remark 2.

C. Proofs

We first prove Lemmas 1 and 2 as we use them in the course of proving Theorems 1 and 3. We start with Lemma 2, as the proof for Lemma 1 simplifies to a special case.

Lemma 2  A missing data model of an ADMG 𝒢{\mathcal{G}} that contains no colluding paths is a submodel of the itemwise conditionally independent nonresponse model described in (Shpitser, 2016; Sadinle & Reiter, 2017).

Proof.

The complete Markov blanket of a vertex ViV_{i} in an ADMG 𝒢,{\mathcal{G}}, denoted mb𝒢c(Vi)\operatorname{mb}_{\mathcal{G}}^{c}(V_{i}) is the set of vertices such that ViVimb𝒢c(Vi)mb𝒢c(Vi)V_{i}\perp\!\!\!\perp V_{-i}\setminus\operatorname{mb}_{\mathcal{G}}^{c}(V_{i})\mid\operatorname{mb}_{\mathcal{G}}^{c}(V_{i}) (Pearl, 1988; Richardson, 2003). In ADMGs, this set corresponds to the Markov blanket of Vi,V_{i}, its children, and the Markov blanket of its children. That is,

mb𝒢c(Vi)mb𝒢(Vi)(Vjch𝒢(Vi)Vjmb𝒢(Vj)){Vi}.\displaystyle\operatorname{mb}_{\mathcal{G}}^{c}(V_{i})\equiv\operatorname{mb}_{\mathcal{G}}(V_{i})\cup\bigg{(}\bigcup_{V_{j}\in\operatorname{ch}_{\mathcal{G}}(V_{i})}V_{j}\cup\operatorname{mb}_{\mathcal{G}}(V_{j})\bigg{)}\setminus\{V_{i}\}.

Without loss of generality, we ignore the part of the graph involving the deterministic factors p(XX(1),R)p(X\mid X^{(1)},R) and the corresponding deterministic edges, in the construction of the Markov blanket and complete Markov blanket of variables in a missing data graph 𝒢(X(1),O,R).{\mathcal{G}}(X^{(1)},O,R). We now show that the absence of non-deterministic colluder paths between a pair Xi(1)X_{i}^{(1)} and RiR_{i} in 𝒢{\mathcal{G}} implies that Xi(1)mb𝒢c(Ri).X_{i}^{(1)}\notin\operatorname{mb}_{\mathcal{G}}^{c}(R_{i}).

  • Xi(1)X_{i}^{(1)} is not a parent of Ri,R_{i}, as Xi(1)RiX_{i}^{(1)}\rightarrow R_{i} is trivially a colluder path.

  • Xi(1)X_{i}^{(1)} is not in the district of Ri,R_{i}, as Xi(1)RiX_{i}^{(1)}\leftrightarrow\cdots\leftrightarrow R_{i} is also a colluder path.

These two points together imply that Xi(1)mb𝒢(Ri).X_{i}^{(1)}\notin\operatorname{mb}_{\mathcal{G}}(R_{i}). We now show that the union over children of RiR_{i} and their Markov blankets also exclude Xi(1).X_{i}^{(1)}.

  • Xi(1)X_{i}^{(1)} is not a child of Ri,R_{i}, as directed edges from RiR_{i} to variables in X(1)X^{(1)} are ruled out by construction in missing data graphs.

  • Xi(1)X_{i}^{(1)} is also not in the district of any children of Ri,R_{i}, as RiXi(1)R_{i}\rightarrow\cdots\leftrightarrow X_{i}^{(1)} is a colluding path.

  • Xi(1)X_{i}^{(1)} is also not a parent of the district of any children of Ri,R_{i}, as RiXi(1)R_{i}\rightarrow\cdots\leftarrow X_{i}^{(1)} is a colluding path.

These three points together rule out the possibility that Xi(1)X_{i}^{(1)} is present in the union over children and Markov blankets of children of Ri.R_{i}. Thus, we have shown that Xi(1)mb𝒢c(Ri).X_{i}^{(1)}\not\in\operatorname{mb}_{\mathcal{G}}^{c}(R_{i}). This implies the following,

RiV{Ri,mb𝒢c(Ri)}mb𝒢c(Ri)RiXi(1)mb𝒢c(Ri).\displaystyle R_{i}\perp\!\!\!\perp V\setminus\{R_{i},\operatorname{mb}_{\mathcal{G}}^{c}(R_{i})\}\mid\operatorname{mb}_{\mathcal{G}}^{c}(R_{i})\implies R_{i}\perp\!\!\!\perp X_{i}^{(1)}\mid\operatorname{mb}_{\mathcal{G}}^{c}(R_{i}).

By semi-graphoid axioms (see for example, (Lauritzen, 1996; Pearl, 2009)) this yields the conditional independence RiXi(1)Ri,Xi(1),O.R_{i}\perp\!\!\!\perp X_{i}^{(1)}\mid R_{-i},X^{(1)}_{-i},O.

The same line of reasoning detailed above can be used for all RiR,R_{i}\in R, which then gives us the set of conditional independences implied by the no self-censoring model. That is,

RiXi(1)Ri,Xi(1),O,RiR.\displaystyle R_{i}\perp\!\!\!\perp X^{(1)}_{i}\mid R_{-i},X^{(1)}_{-i},O,\quad\forall R_{i}\in R.

Lemma 1  A missing data model of a DAG 𝒢{\mathcal{G}} that contains no self-censoring edges and no colluders, is a submodel of the itemwise conditionally independent nonresponse model described in (Shpitser, 2016; Sadinle & Reiter, 2017).

Proof.

A DAG is simply a special case of an ADMG with no bidirected edges. Consequently the only two types of colluding paths, are self-censoring edges (Xi(1)RiX_{i}^{(1)}\rightarrow R_{i}) and colluder structures (Xi(1)RjRi).X_{i}^{(1)}\rightarrow R_{j}\leftarrow R_{i}). Thus, the absence of these two structures in a missing data DAG 𝒢,{\mathcal{G}}, rules out all possible colluding paths. The rest of the proof then carries over straightforwardly from Lemma 2. ∎

Theorem 1  A full law p(R,X(1),O)p(R,X^{(1)},O) that is Markov relative to a missing data DAG 𝒢{\mathcal{G}} is identified if 𝒢{\mathcal{G}} does not contain edges of the form Xi(1)RiX_{i}^{(1)}\rightarrow R_{i} (no self-censoring) and structures of the form Xj(1)RiRjX_{j}^{(1)}\rightarrow R_{i}\leftarrow R_{j} (no colluders), and the stated positivity assumption holds. Moreover, the resulting identifying functional for the missingness mechanism p(RX(1),O)p(R\mid X^{(1)},O) is given by the odds ratio parameterization provided in Eq. 2 of the main draft, and the identifying functionals for the target law and full law are given by Remarks 1 and 2.

Proof.

Given Eq. (2), we know that

p(RX(1),O)=\displaystyle p(R\mid X^{(1)},O)= 1Z×k=1Kp(RkRk=1,X(1),O)×k=2KOR(Rk,RkRk=1,X(1),O),\displaystyle\ \frac{1}{Z}\times\prod_{k=1}^{K}\ p(R_{k}\mid R_{-k}=1,X^{(1)},O)\times\prod_{k=2}^{K}\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)},O),

where Rk=RRk,Rk={R1,,Rk1},Rk={Rk+1,,RK}R_{-k}=R\setminus R_{k},R_{\prec k}=\{R_{1},\ldots,R_{k-1}\},R_{\succ k}=\{R_{k+1},\ldots,R_{K}\}, and

OR(Rk,RkRk=1,X(1),O)=p(RkRk=1,Rk,X(1),O)p(Rk=1Rk=1,Rk,X(1),O)×p(Rk=1Rk=1,X(1),O)p(RkRk=1,X(1),O),\displaystyle\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)},O)=\frac{p(R_{k}\mid R_{\succ k}=1,R_{\prec k},X^{(1)},O)}{p(R_{k}=1\mid R_{\succ k}=1,R_{\prec k},X^{(1)},O)}\times\frac{p(R_{k}=1\mid R_{-k}=1,X^{(1)},O)}{p(R_{k}\mid R_{-k}=1,X^{(1)},O)},

and ZZ is the normalizing term and is equal to r{k=1Kp(rkRk=1,X(1),O)×k=2KOR(rk,rkRk=1,X(1),O)}\sum_{r}\{\prod_{k=1}^{K}\ p(r_{k}\mid R_{-k}=1,X^{(1)},O)\times\prod_{k=2}^{K}\text{OR}(r_{k},r_{\prec k}\mid R_{\succ k}=1,X^{(1)},O)\}. If we can prove that all the pieces in this factorization are identified, then the missingness process is identified and so is the full law. We provide the proof in two steps. Our proof is similar to the identification proof of the no self-censoring model given in (Malinsky et al., 2019).

For each k3,,K,k\in 3,\dots,K, we can apply the following expansion to the odds ratio term. Without loss of generality we drop fully observed random variables OO for brevity,

OR(Rk,RkRk=1,X(1))=OR(Rk,Rk1R(k,k1)=1,X(1))×OR(Rk,Rk2Rk=1,Rk1,X(1)).\displaystyle\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)})=\text{OR}(R_{k},R_{k-1}\mid R_{-(k,k-1)}=1,X^{(1)})\times\text{OR}(R_{k},R_{\prec k-2}\mid R_{\succ k}=1,R_{k-1},X^{(1)}). (4)

This expansion can be applied inductively to the second term in the above product until OR(Rk,RkRk=1,X(1))\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)}) is expressed as a function of pairwise odds ratios and higher-order interaction terms. Applying the inductive expansion to each odds ratio term in k=2KOR(Rk,RkRk=1,X(1))\prod_{k=2}^{K}\text{OR}(R_{k},R_{\prec k}\mid R_{\succ k}=1,X^{(1)}) we can re-express the identifying functional as,

p(RX(1))=\displaystyle p(R\mid X^{(1)})= 1Z×k=1Kp(RkRk=1,X(1))\displaystyle\frac{1}{Z}\times\prod_{k=1}^{K}\ p(R_{k}\mid R_{-k}=1,X^{(1)})
×Rk,RlROR(Rk,RlR(k,l)=1,X(1))×Rk,Rl,RmRf(Rk,Rl,RmR(k,l,m)=1,X(1))\displaystyle\times\prod_{R_{k},R_{l}\in R}\text{OR}(R_{k},R_{l}\mid R_{-(k,l)}=1,X^{(1)})\times\prod_{R_{k},R_{l},R_{m}\in R}f(R_{k},R_{l},R_{m}\mid R_{-(k,l,m)}=1,X^{(1)})
×Rk,Rl,Rm,RnRf(Rk,Rl,Rm,RnR(k,l,m,n)=1,X(1))××f(R1,,RKX(1)),\displaystyle\times\prod_{R_{k},R_{l},R_{m},R_{n}\in R}f(R_{k},R_{l},R_{m},R_{n}\mid R_{-(k,l,m,n)}=1,X^{(1)})\times\dots\times f(R_{1},\dots,R_{K}\mid X^{(1)}), (5)

where ZZ is the normalizing constant as before, and each f(,X(1))f(\cdot\mid\cdot,X^{(1)}) are 3-way, 4-way, up to KK-way interaction terms. These interaction terms are defined as follows.

f(Ri,Rj,Rk|R(i,j,k)=1,X(1))=OR(Ri,Rj|Rk,R(i,j,k)=1,X(1))OR(Ri,Rj|Rk=1,R(i,j,k)=1,X(1)),f(R_{i},R_{j},R_{k}|R_{-(i,j,k)}=1,X^{(1)})=\frac{\textrm{OR}(R_{i},R_{j}|R_{k},R_{-(i,j,k)}=1,X^{(1)})}{\textrm{OR}(R_{i},R_{j}|R_{k}=1,R_{-(i,j,k)}=1,X^{(1)})},

and

f(Ri,Rj,Rk,Rl|R(i,j,k,l)=1,X(1))=OR(Ri,Rj|Rk,Rl,R(i,j,k,l)=1,X(1))OR(Ri,Rj|Rk=1,Rl,R(i,j,k,l)=1,X(1))×OR(Ri,Rj|Rk=1,Rl=1,R(i,j,k,l)=1,X(1))OR(Ri,Rj|Rk,Rl=1,R(i,j,k,l)=1,X(1)),f(R_{i},R_{j},R_{k},R_{l}|R_{-(i,j,k,l)}=1,X^{(1)})=\frac{\textrm{OR}(R_{i},R_{j}|R_{k},R_{l},R_{-(i,j,k,l)}=1,X^{(1)})}{\textrm{OR}(R_{i},R_{j}|R_{k}=1,R_{l},R_{-(i,j,k,l)}=1,X^{(1)})}\times\frac{\textrm{OR}(R_{i},R_{j}|R_{k}=1,R_{l}=1,R_{-(i,j,k,l)}=1,X^{(1)})}{\textrm{OR}(R_{i},R_{j}|R_{k},R_{l}=1,R_{-(i,j,k,l)}=1,X^{(1)})},

and so on, up to

f(R1,,RKX(1))\displaystyle f(R_{1},...,R_{K}\mid X^{(1)}) =OR(Ri,Rj|R(i,j),X(1))\displaystyle=\textrm{OR}(R_{i},R_{j}|R_{-(i,j)},X^{(1)})
×Rk,RlROR(Ri,Rj|R(k,l)=1,R(i,j,k,l),X(1))Rk,Rl,Rm,RnROR(Ri,Rj|R(k,l,m,n)=1,R(i,j,k,l,m,n),X(1))×RkROR(Ri,Rj|Rk=1,R(i,j,k),X(1))Rk,Rl,RmROR(Ri,Rj|R(k,l,m)=1,R(i,j,k,l,m),X(1))×.\displaystyle\times\frac{\displaystyle\prod_{R_{k},R_{l}\in R}\textrm{OR}(R_{i},R_{j}|R_{(k,l)}=1,R_{-(i,j,k,l)},X^{(1)})\prod_{R_{k},R_{l},R_{m},R_{n}\in R}\textrm{OR}(R_{i},R_{j}|R_{(k,l,m,n)}=1,R_{-(i,j,k,l,m,n)},X^{(1)})\times\cdots}{\displaystyle\prod_{R_{k}\in R}\textrm{OR}(R_{i},R_{j}|R_{k}=1,R_{-(i,j,k)},X^{(1)})\prod_{R_{k},R_{l},R_{m}\in R}\textrm{OR}(R_{i},R_{j}|R_{(k,l,m)}=1,R_{-(i,j,k,l,m)},X^{(1)})\times\cdots}.

Readers familiar with the clique potential factorization of Markov random fields may treat these interaction terms analogously (Malinsky et al., 2019). We now show that each term in the above factorization is identified.

Step 1.
We start off by looking at the conditional pieces p(RkRk=1,X(1),O)p(R_{k}\mid R_{-k}=1,X^{(1)},O). Given Lemma. 1, we know that RkXk(1)Rk,Xk(1),OR_{k}\perp\!\!\!\perp X^{(1)}_{k}\mid R_{-k},X^{(1)}_{-k},O. Therefore, p(RkRk=1,X(1),O)=p(RkRk=1,Xk(1),O),k,p(R_{k}\mid R_{-k}=1,X^{(1)},O)=p(R_{k}\mid R_{-k}=1,X^{(1)}_{-k},O),\forall k, is identified for all RkR.R_{k}\in R.

Step 2.
We now show that for any Rk,RlRR_{k},R_{l}\in R, the pairwise odds ratio OR(Rk,RlR{(k,l)}=1,X(1))\text{OR}(R_{k},R_{l}\mid R_{\{-(k,l)\}}=1,X^{(1)}) given in Eq. (5) is identified. We know that

OR(Rk,RlR(k,l)=1,X(1))=OR(Rk,RlR(k,l)=1,X(k,l),Xk(1),Xl(1)).\text{OR}(R_{k},R_{l}\mid R_{-(k,l)}=1,X^{(1)})=\text{OR}(R_{k},R_{l}\mid R_{-(k,l)}=1,X_{-(k,l)},X^{(1)}_{k},X^{(1)}_{l}).

Consequently, if we can show that the odds ratio is neither a function of Xk(1)X^{(1)}_{k} nor Xl(1)X^{(1)}_{l}, then we can safely claim that the odds ratio is only a function of observed data and hence is identified. We get to this conclusion by exploiting the symmetric notion in odds ratios.

OR(Rk,RlR(k,l)=1,X(1))\displaystyle\text{OR}(R_{k},R_{l}\mid R_{-(k,l)}=1,X^{(1)}) =p(RkRl,R(k,l)=1,X(1))p(Rk=1Rl,R(k,l)=1,X(1))×p(Rk=1Rk=1,X(1))p(RkRk=1,X(1))\displaystyle=\frac{p(R_{k}\mid R_{l},R_{-(k,l)}=1,X^{(1)})}{p(R_{k}=1\mid R_{l},R_{-(k,l)}=1,X^{(1)})}\times\frac{p(R_{k}=1\mid R_{-k}=1,X^{(1)})}{p(R_{k}\mid R_{-k}=1,X^{(1)})}
=p(RlRk,R(k,l)=1,X(1))p(Rl=1Rk,R(k,l)=1,X(1))×p(Rl=1Rl=1,X(1))p(RlRl=1,X(1))\displaystyle=\frac{p(R_{l}\mid R_{k},R_{-(k,l)}=1,X^{(1)})}{p(R_{l}=1\mid R_{k},R_{-(k,l)}=1,X^{(1)})}\times\frac{p(R_{l}=1\mid R_{-l}=1,X^{(1)})}{p(R_{l}\mid R_{-l}=1,X^{(1)})}

In the first equality, we can see that the odds ratio is not a function of Xk(1)X^{(1)}_{k} since RkXk(1)Rk,Xk(1)R_{k}\perp\!\!\!\perp X^{(1)}_{k}\mid R_{-k},X^{(1)}_{-k}. Similarly, from the second equality, we can see that the odds ratio is not a function of Xl(1)X^{(1)}_{l} since RlXl(1)Rl,Xl(1)R_{l}\perp\!\!\!\perp X^{(1)}_{l}\mid R_{-l},X^{(1)}_{-l}. Therefore, the pairwise odds ratios are all identified.

Finally we show that each of the higher-order interaction terms are identified. For each of these terms we need to show that they are not a function of missing variables with indices corresponding to indicators to the left of the conditioning bar. That is, we need to show that the 3-way interaction terms f(Rk,Rl,RmR(k,l,m)=1,X(1))f(R_{k},R_{l},R_{m}\mid R_{-(k,l,m)}=1,X^{(1)}) are not functions of X(k,l,m)(1),X_{(k,l,m)}^{(1)}, the 4-way interaction terms f(Rk,Rl,Rm,RnR(k,l,m,n)=1,X(1))f(R_{k},R_{l},R_{m},R_{n}\mid R_{-(k,l,m,n)}=1,X^{(1)}) are not functions of X(k,l,m,n)(1)X_{(k,l,m,n)}^{(1)}, and so on until finally the KK-way interaction term f(R1,,RKX(1))f(R_{1},\dots,R_{K}\mid X^{(1)}) is not a function of X(1).X^{(1)}.

Because of the way the odds ratio is defined, each f(,X(1))f(\cdot\mid\cdot,X^{(1)}) is symmetric in the kk arguments appearing to the left of the conditioning bar and can be rewritten in multiple equivalent ways. In particular, each kk-way interaction term can be rewritten in (k2)\binom{k}{2} ways for any choice of indices i,ji,j of the missingness indicators that appear to the left of the conditioning bar. Each such representation allows us to conclude that f(,X(1))f(\cdot\mid\cdot,X^{(1)}) is not a function of Xi(1),Xj(1).X_{i}^{(1)},X_{j}^{(1)}. Combining all these together allows us to conclude that the kk-way interaction term f(,X(1))f(\cdot\mid\cdot,X^{(1)}) is not a function of the missing variables corresponding to the indicators appearing on the left of the conditioning bar.

As a concrete example, consider the 3-way interaction f(R1,R2,R3R(1,2,3)=1,X(1)).f(R_{1},R_{2},R_{3}\mid R_{-(1,2,3)}=1,X^{(1)}). We can write it down in three different ways as follows.

f(Ri,Rj,RkR(1,2,3)=1,X(1))\displaystyle f(R_{i},R_{j},R_{k}\mid R_{-(1,2,3)}=1,X^{(1)})
=OR(R1,R2R(1,2,3)=1,R3,X(1))OR(R1,R2R(1,2,3)=1,R3=1,X(1))=OR(R1,R3R(1,2,3)=1,R2,X(1))OR(R1,R3R(1,2,3)=1,R2=1,X(1))=OR(R2,R3R(1,2,3)=1,R1,X(1))OR(R2,R3R(1,2,3)=1,R1=1,X(1))\displaystyle=\frac{\text{OR}(R_{1},R_{2}\mid R_{-(1,2,3)}=1,R_{3},X^{(1)})}{\text{OR}(R_{1},R_{2}\mid R_{-(1,2,3)}=1,R_{3}=1,X^{(1)})}=\frac{\text{OR}(R_{1},R_{3}\mid R_{-(1,2,3)}=1,R_{2},X^{(1)})}{\text{OR}(R_{1},R_{3}\mid R_{-(1,2,3)}=1,R_{2}=1,X^{(1)})}=\frac{\text{OR}(R_{2},R_{3}\mid R_{-(1,2,3)}=1,R_{1},X^{(1)})}{\text{OR}(R_{2},R_{3}\mid R_{-(1,2,3)}=1,R_{1}=1,X^{(1)})}

From the first equality, we note that ff is not a function of X1(1),X2(1)X^{(1)}_{1},X^{(1)}_{2}. From the second equality, we note that ff is not a function of X1(1),X3(1)X^{(1)}_{1},X^{(1)}_{3}. From the third equality, we note that ff is not a function of X2(1),X3(1)X^{(1)}_{2},X^{(1)}_{3}. Therefore, ff is not a function of X1(1),X2(1),X3(1)X^{(1)}_{1},X^{(1)}_{2},X^{(1)}_{3} and is identified. ∎

Theorem 2  The graphical condition of no self-censoring and no colluders, put forward in Theorem 1, is sound and complete for the identification of full laws p(R,O,X(1))p(R,O,X^{(1)}) that are Markov relative to a missing data DAG 𝒢.{\mathcal{G}}.

Proof.

Soundness is a direct consequence of Theorem 1. To prove completeness, it needs to be shown that in the presence of a self-censoring edge, or a colluder structure, the full law is no longer (non-parametrically) identified. A proof by counterexample of both these facts was provided in (Bhattacharya et al., 2019). However, this can also be seen from the fact that self-censoring edges and colluders are special cases of the colluding paths that we prove results in non-identification of the full law in Lemma 3. ∎

Theorem 3  A full law p(R,X(1),O)p(R,X^{(1)},O) that is Markov relative to a missing data ADMG 𝒢{\mathcal{G}} is identified if 𝒢{\mathcal{G}} does not contain any colluding paths and the stated positivity assumption in Section 5 holds. Moreover, the resulting identifying functional for the missingness mechanism p(RX(1),O)p(R\mid X^{(1)},O) is given by the odds ratio parametrization provided in Eq. 2 of the main draft.

Proof.

The proof strategy is nearly identical to the one utilized in Theorem 1, except the conditional independences RkXk(1)Rk,Xk(1),OR_{k}\perp\!\!\!\perp X_{k}^{(1)}\mid R_{-k},X_{-k}^{(1)},O come from Lemma 2 instead of Lemma 1. ∎

Lemma 3  A full law p(R,X(1),O)p(R,X^{(1)},O) that is Markov relative to a missing data ADMG 𝒢{\mathcal{G}} containing a colluding path between any pair Xi(1)X(1)X_{i}^{(1)}\in X^{(1)} and RiR,R_{i}\in R, is not identified.

X1(1)X^{(1)}_{1}R1R_{1}X1X_{1}(a)X1(1)X^{(1)}_{1}UUR1R_{1}X1X_{1}(b)R1R_{1}X1X_{1}(c)X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}R1R_{1}R2R_{2}X1X_{1}X2X_{2}(d)X1(1)X^{(1)}_{1}X2(1)X^{(1)}_{2}R1R_{1}R2R_{2}X1X_{1}X2X_{2}(e)R1R_{1}R2R_{2}X1X_{1}X2X_{2}(f)
Figure 5: (a, d, e) Examples of colluding paths in missing data models of ADMGs. (b) A DAG with hidden variable UU that is Markov equivalent to (a). (c) Projecting out X1(1)X^{(1)}_{1} from (a), (f) Projecting out X1(1)X^{(1)}_{1} and X2(1)X^{(1)}_{2} from (d) and (e).
Proof.

Proving the non-identifiability of missing data models of an ADMG 𝒢{\mathcal{G}} that contains a colluding path can be shown by providing two models 1{\cal M}_{1} and 2{\cal M}_{2} that disagree on the full law but agree on the observed law. Coming up with a single example of such a pair of models is sufficient for arguing against non-parametric identification of the full law. Therefore, for simplicity, we restrict our attention to binary random variables. We first provide an example of such a pair of models on the simplest form of a colluding path, a bidirected edge Xi(1)RiX_{i}^{(1)}\leftrightarrow R_{i} as shown in Fig. 5(a). According to Table 1, in order for the observed laws to agree, the only requirement is that the quantity ab+(1a)cab+(1-a)c remain equal in both models; hence we can come up with infinitely many counterexamples of full laws that are not the same but map to the same observed law.

Constructing explicit counterexamples are not necessary to prove non-identification as long as it can be shown that there exist at least two distinct functions that map two different full laws onto the exact same observed law. For instance, if the number of parameters in the full law is strictly larger than the number of parameters in the observed law, then there would exist infinitely many such functions. Consequently, we rely on a parameter counting argument to prove the completeness of our results. Since we are considering missing data models of ADMGs, we use the Moebius parameterization of binary nested Markov models of an ADMG described in Appendix A.

The nested Markov model of a missing data ADMG 𝒢(V),{\mathcal{G}}(V), where V={O,X(1),R,X},V=\{O,X^{(1)},R,X\}, is a smooth super model of the missing data DAG model 𝒢(VU),{\mathcal{G}}(V\cup U), and has the same model dimension as the latent variable model (Evans, 2018). We also utilize nested Markov models of an ADMG 𝒢(VX(1)),{\mathcal{G}}(V\setminus X^{(1)}), corresponding to projection of the missing data ADMG 𝒢(V){\mathcal{G}}(V) onto variables that are fully observable. While such a model does not capture all equality constraints in the true observed law, it is still a smooth super model of it, thus providing an upper bound on the model dimension of the observed law. This suffices for our purposes, as demonstrating that the upper bound on the number of parameters in the observed law is less than the number of parameters in the full law, is sufficient to prove that the full law is not identified. We first walk the reader through a few examples to demonstrate this proof strategy, and then provide the general argument.

UU p(U)p(U)
0 aa
11 1a1-a
R1R_{1} UU p(R1|U)p(R_{1}|U)
0 0 bb
11 0 1b1-b
0 11 cc
11 11 1c1-c
X1(1)X^{(1)}_{1} UU p(X1(1)|U)p(X^{(1)}_{1}|U)
0 0 dd
11 0 1d1-d
0 11 ee
11 11 1e1-e
R1R_{1} X1(1)X^{(1)}_{1} UU p(R1,X1(1),U)p(R_{1},X^{(1)}_{1},U)
0 0 0 abda*b*d
0 0 11 (1a)ce(1-a)*c*e
0 11 0 ab(1d)a*b*(1-d)
0 11 11 (1a)c(1e)(1-a)*c*(1-e)
11 0 0 a(1b)da*(1-b)*d
11 0 11 (1a)(1c)e(1-a)*(1-c)*e
11 11 0 a(1b)(1d)a*(1-b)*(1-d)
11 11 11 (1a)(1c)(1e)(1-a)*(1-c)*(1-e)
R1R_{1} X1(1)X^{(1)}_{1} p(Full Law) X1X_{1} p(Observed Law)
0 0 abd+(1a)cea*b*d+(1-a)*c*e ? ab+(1a)ca*b+(1-a)*c
1 ab(1d)+(1a)c(1e)a*b*(1-d)+(1-a)*c*(1-e)
1 0 a(1b)d+(1a)(1c)ea*(1-b)*d+(1-a)*(1-c)*e 0 a(1b)+(1a)(1c)a*(1-b)+(1-a)*(1-c)
1 a(1b)(1d)+(1a)(1c)(1e)a*(1-b)*(1-d)+(1-a)*(1-c)*(1-e) 1
Table 1: Construction of counterexamples for non-identifiablity of the full law in Fig. 5(a) using the DAG with hidden variable UU in Fig. 5(b) that is Markov equivalent to (a).
Moebius Parameterization of the Full Law in Fig. 5(d)
Districts Intrinsic Head/Tail Moebius Parameters Counts
{X1(1)}\{X^{(1)}_{1}\} {X1(1)},{}\{X^{(1)}_{1}\},\{\} q(X1(1)=0)q(X^{(1)}_{1}=0) 11
{R2}\{R_{2}\} {R2},{}\{R_{2}\},\{\} q(R2=0)q(R_{2}=0) 11
{R1,X2(1)}\{R_{1},X^{(1)}_{2}\} {R1},{}\{R_{1}\},\{\} q(R1=0)q(R_{1}=0) 11
{X2(1)},{X1(1)}\{X^{(1)}_{2}\},\{X^{(1)}_{1}\} q(X2(1)=0X1(1))q(X^{(1)}_{2}=0\mid X^{(1)}_{1}) 22
{R1,X2(1)},{X1(1)}\{R_{1},X^{(1)}_{2}\},\{X^{(1)}_{1}\} q(R1=0,X2(1)=0X1(1))q(R_{1}=0,X^{(1)}_{2}=0\mid X^{(1)}_{1}) 22
Total 77
Moebius Parameterization of the Full Law in Fig. 5(e)
Districts Intrinsic Head/Tail Moebius Parameters Counts
{R2}\{R_{2}\} {R2},{}\{R_{2}\},\{\} q(R2=0)q(R_{2}=0) 11
{R1,X1(1),X2(1)}\{R_{1},X^{(1)}_{1},X^{(1)}_{2}\} {R1},{}\{R_{1}\},\{\} q(R1=0)q(R_{1}=0) 11
{X1(1)},{}\{X^{(1)}_{1}\},\{\} q(X1(1)=0)q(X^{(1)}_{1}=0) 11
{X2(1)},{}\{X^{(1)}_{2}\},\{\} q(X2(1)=0)q(X^{(1)}_{2}=0) 11
{R1,X2(1)},{}\{R_{1},X^{(1)}_{2}\},\{\} q(R1=0,X2(1)=0)q(R_{1}=0,X^{(1)}_{2}=0) 11
{X1(1),X2(1)},{}\{X^{(1)}_{1},X^{(1)}_{2}\},\{\} q(X1(1)=0,X2(1)=0)q(X^{(1)}_{1}=0,X^{(1)}_{2}=0) 11
{R1,X1(1),X2(1)},{}\{R_{1},X^{(1)}_{1},X^{(1)}_{2}\},\{\} q(R1=0,X1(1)=0,X2(1)=0)q(R_{1}=0,X^{(1)}_{1}=0,X^{(1)}_{2}=0) 11
Total 77
Moebius Parameterization of the Observed Law in Fig. 5(f)
Districts Intrinsic Head/Tail Moebius Parameters Counts
R2R_{2} {R2},{}\{R_{2}\},\{\} q(R2=0)q(R_{2}=0) 11
{R1,X1,X2}\{R_{1},X_{1},X_{2}\} {R1},{}\{R_{1}\},\{\} q(R1=0)q(R_{1}=0) 11
{X1},{R1}\{X_{1}\},\{R_{1}\} q(X1=0R1)q(X_{1}=0\mid R_{1}) 11
{X2},{R2}\{X_{2}\},\{R_{2}\} q(X2=0R2)q(X_{2}=0\mid R_{2}) 11
{R1,X2},{R2}\{R_{1},X_{2}\},\{R_{2}\} q(R1=0,X2=0R2)q(R_{1}=0,X_{2}=0\mid R_{2}) 11
{X1,X2},{R1,R2}\{X_{1},X_{2}\},\{R_{1},R_{2}\} q(X1=0,X2=0R1,R2)q(X_{1}=0,X_{2}=0\mid R_{1},R_{2}) 11
Total 66
Table 2: Moebius Parameterization of the Full and Observed Laws of missing data ADMGs

Self-censoring through unmeasured confounding:

We start by reanalyzing the colluding path given in Fig. 5(a) and the corresponding projection given in Fig. 5(c). The Moebius parameters associated with the full law are q(X1(1)=0),q(R1=0),q(X1(1)=0,R1=1),q(X^{(1)}_{1}=0),q(R_{1}=0),q(X^{(1)}_{1}=0,R_{1}=1), for a total of 3 parameters. The Moebius parameters associated with the observed law in Fig 5(c) are q(R1=0),q(X1(1)=0R1=0),q(R_{1}=0),q(X^{(1)}_{1}=0\mid R_{1}=0), for a total of only 2 parameters. Since 2<32<3, we can construct infinitely many mappings, as it was shown in Table 1.

Simple colluding paths:

Consider the colluding paths given in Fig. 5(d, e) and the corresponding projection (which are identical in both cases) given in Fig. 5(f). The Moebius parameters associated with the full laws and observed law are shown in Table 2. Once again, since the number of parameters in the observed law is less than the number in the full law (6<7)(6<7), we can construct infinitely many mappings.

A general argument:

In order to generalize our argument, we first provide a more precise representation (that does not use dashed edges) in Figs. 6(a-d), of all possible colluding paths between Xi(1)X_{i}^{(1)} and Ri.R_{i}. Without loss of generality, assume that there are KK variables in X(1)X^{(1)} and there are SS variables that lie on the collider path between Xi(1)X^{(1)}_{i} and RiR_{i}, S{0,1,,2(K1)}S\in\{0,1,\dots,2*(K-1)\}. We denote the ssth variable on the collider path by VsV_{s}; Vs{X(1)Xi(1),RRi}.V_{s}\in\{X^{(1)}\setminus X^{(1)}_{i},R\setminus R_{i}\}. Note that VSV_{S} in Figs. 6(c, d) can only belong to {RRi}\{R\setminus R_{i}\} by convention. Fig. 6(e) illustrates the corresponding projections of figures (a) and (b), and Fig. 6(f) illustrates the corresponding projections of figures (c) and (d). In the projections shown in Figs. 6(e, f), V{XXi(1),RRi}.V^{*}\in\{X\setminus X^{(1)}_{i},R\setminus R_{i}\}.

Xi(1)X^{(1)}_{i}V1V_{1}\cdotsVSV_{S}RiR_{i}(a)Xi(1)X^{(1)}_{i}V1V_{1}\cdotsVSV_{S}RiR_{i}(b)Xi(1)X^{(1)}_{i}V1V_{1}\cdotsRSR_{S}RiR_{i}(c)Xi(1)X^{(1)}_{i}V1V_{1}\cdotsRSR_{S}RiR_{i}(d)XiX_{i}V1V^{*}_{1}\cdotsVSV^{*}_{S}RiR_{i}(e)XiX_{i}V1V^{*}_{1}\cdotsRSR_{S}RiR_{i}(f)
Figure 6: (a) Colluding paths (b) Projecting out X(1)X^{(1)}

We now go over each of these colluding paths and their corresponding latent projections, as if they appear in a larger graph that is otherwise completely disconnected. We count the number of Moebius parameters as a function of SS, and show that the full law always has one more parameter than the observed law. One can then imagine placing these colluding paths in a larger graph with arbitrary connectivity, and arguing that the full law is still not identified as a consequence of the parameter discrepancy arising from the colluding path alone. That is, if we show a fully disconnected graph containing a single colluding path is not identified, then it is also the case that any edge super graph (super model) is also not identified.

In the following proof we heavily rely on the following fact. Given a bidirected chain of length V1,,VK,V_{1}\leftrightarrow,\cdots,\leftrightarrow V_{K}, of length K,K, the number of Moebius parameters required to parameterize this chain is given by the sum of natural numbers 11 to K,K, i.e., K(K+1)2.\frac{K(K+1)}{2}. This can be seen from the fact that the corresponding Moebius parameters are given by the series,

  • q(V1=0),q(V1=0,V2=0),,q(V1=0,,VK=0)q(V_{1}=0),q(V_{1}=0,V_{2}=0),\dots,q(V_{1}=0,\dots,V_{K}=0) corresponding to KK parameters.

  • q(V2=0),q(V2,V3=0),,q(V2=0,,VK=0)q(V_{2}=0),q(V_{2},V_{3}=0),\dots,q(V_{2}=0,\dots,V_{K}=0) corresponding to K1K-1 parameters.

  • \dots

  • q(VK=0)q(V_{K}=0) corresponding to 11 parameter.

In counting the number of parameters for a disconnected graph (with the exception of the colluding path), we can also exclude the singleton (disconnected) nodes from the counting argument since they account for the same number of parameters in both the full law and observed law. In the full law they are either q(Rs=0)q(R_{s}=0) or q(Xs(1)=0)q(X^{(1)}_{s}=0) and the corresponding parameters in the observed law are q(Rs=0)q(R_{s}=0) or q(Xs=0Rs=1)q(X_{s}=0\mid R_{s}=1). The Moebius parameter counts for each of the colluding paths in Figs. 6(a-d) and their corresponding latent projections in Figs. 6(e,f) are as follows.

Figures a, b, and e

  1. 1.

    Number of Moebius parameters in Fig. 6(a) is (S+2)(S+3)2\frac{(S+2)(S+3)}{2}

    • A bidirected chain Xi(1),,RiX_{i}^{(1)}\leftrightarrow,\cdots,\leftrightarrow R_{i} of length S+2S+2, i.e., (S+2)(S+3)/2(S+2)*(S+3)/2 parameters.

  2. 2.

    Number of Moebius parameters in Fig. 6(b) is (S+2)(S+3)2\frac{(S+2)(S+3)}{2}

    • q(Xi(1)=0)q(X^{(1)}_{i}=0), i.e. 11 parameter,

    • A bidirected chain V2RiV_{2}\leftrightarrow\cdots\leftrightarrow R_{i} of length SS, i.e. S(S+1)/2S*(S+1)/2 parameters,

    • Intrinsic sets involving V1V_{1}, i.e., q(V1=0Xi(1)),q(V1=0,V2=0Xi(1)),q(V1=0,,Ri=0Xi(1))q(V_{1}=0\mid X_{i}^{(1)}),q(V_{1}=0,V_{2}=0\mid X_{i}^{(1)}),q(V_{1}=0,\dots,R_{i}=0\mid X_{i}^{(1)}) corresponding to 2(S+1)2*(S+1) parameters.

  3. 3.

    Number of Moebius parameters in Fig. 6(e) is (S+2)(S+3)21\frac{(S+2)(S+3)}{2}-1

    • Note that even though each proxy XsX_{s} that may appear in the bidirected chain has a directed edge from RsR_{s} pointing into it, the corresponding intrinsic head tail pair that involves both variables, will always have Ri=1.R_{i}=1. Hence, we may ignore these deterministic edges and count the parameters as if it were a bidirected chain V1RiV_{1}^{*}\leftrightarrow\cdots\leftrightarrow R_{i} of length S+1S+1, corresponding to (S+1)(S+2)/2(S+1)*(S+2)/2 parameters,

    • When enumerating intrinsic sets involving Xi,X_{i}, we note that {Xi,V1,VS}\{X_{i},V_{1}^{*},\dots V_{S}^{*}\} is not intrinsic as RiR_{i} is not fixable (due to the bidirected path between RiR_{i} and XiX_{i} and the edge RiXiR_{i}\rightarrow X_{i}). Thus, as there is one less intrinsic set involving Xi,X_{i}, the number of parameters required to parameterize all intrinsic sets involving XiX_{i} is one fewer, i.e., S+1S+1 (instead of S+2S+2) parameters.

Figures c, d, and f

  1. 1.

    Number of Moebius parameters in Fig. 6(c) is (S+2)(S+3)2\frac{(S+2)(S+3)}{2}

    • q(Ri=0)q(R_{i}=0), i.e. 11 parameter,

    • A bidirected chain Xi(1)VS1X_{i}^{(1)}\leftrightarrow\cdots\leftrightarrow V_{S-1} of length SS, i.e. S(S+1)/2S*(S+1)/2 parameters,

    • Intrinsic sets involving RSR_{S}, i.e., q(RS=0Ri),q(RS=0,VS1=0Ri),,q(RS=0,VS1=0,Xi(1)Ri),q(R_{S}=0\mid R_{i}),q(R_{S}=0,V_{S-1}=0\mid R_{i}),\dots,q(R_{S}=0,V_{S-1}=0\dots,X_{i}^{(1)}\mid R_{i}), corresponding to 2(S+1)2*(S+1) parameters.

  2. 2.

    Number of Moebius parameters in Fig. 6(d) is (S+2)(S+3)2\frac{(S+2)(S+3)}{2}

    • q(Xi(1)=0),q(Ri=0)q(X^{(1)}_{i}=0),q(R_{i}=0), i.e. 22 parameters,

    • A bidirected chain V2VS2V_{2}\leftrightarrow\cdots\leftrightarrow V_{S-2} of length S2S-2, i.e. (S2)(S1)/2(S-2)*(S-1)/2 parameters,

    • Intrinsic sets involving V1V_{1} and not RS,R_{S}, i.e., q(V1=0Xi(1)),q(V1=0,V2=0Xi(1)),,q(V1=0,V2=0,,VS1Xi(1)),q(V_{1}=0\mid X_{i}^{(1)}),q(V_{1}=0,V_{2}=0\mid X_{i}^{(1)}),\dots,q(V_{1}=0,V_{2}=0,\dots,V_{S-1}\mid X_{i}^{(1)}), corresponding to 2(S1)2*(S-1) parameters,

    • Intrinsic sets involving RSR_{S} and not V1V_{1}, i.e., q(RS=0Ri),q(RS=0,VS1=0Ri),,q(RS=0,VS1=0,,V2Ri)q(R_{S}=0\mid R_{i}),q(R_{S}=0,V_{S-1}=0\mid R_{i}),\dots,q(R_{S}=0,V_{S-1}=0,\dots,V_{2}\mid R_{i}) corresponding to 2(S1)2*(S-1) parameters.

    • The intrinsic set involving both V1V_{1} and RS,R_{S}, i.e., q(V1=0,V2=0,,RS=0Xi(1),Ri),q(V_{1}=0,V_{2}=0,\dots,R_{S}=0\mid X_{i}^{(1)},R_{i}), corresponding to 44 parameters.

  3. 3.

    Number of Moebius parameters in Fig. 6(f) is (S+2)(S+3)21\frac{(S+2)(S+3)}{2}-1

    • q(Ri=0)q(R_{i}=0), i.e. 11 parameter,

    • By the same argument as before, deterministic tails can be ignored. Hence, we have a bidirected chain XiVS1X_{i}\leftrightarrow\cdots\leftrightarrow V_{S-1} of length SS, i.e. S(S+1)/2S*(S+1)/2 parameters,

    • Intrinsic sets involving RSR_{S}, i.e., q(RS=0Ri),q(RS=0,VS1Ri),,q(RS,VS1,,V1Ri),q(R_{S}=0\mid R_{i}),q(R_{S}=0,V_{S-1}\mid R_{i}),\dots,q(R_{S},V_{S-1},\dots,V_{1}\mid R_{i}), corresponding to 2S2*S parameters, and the special intrinsic set which results in the observed law having one less parameter q(RS,VS1,,V1,XiRi=1){\color[rgb]{0,0,1}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,1}q(R_{S},V_{S-1},\dots,V_{1},X_{i}\mid R_{i}=1)} corresponding to just 11 parameter instead of 2 due to the presence of the proxy XiX_{i} in the head and the corresponding RiR_{i} in the tail.

Theorem 4  The graphical condition of the absence of colluding paths, put forward in Theorem 3, is sound and complete for the identification of full laws p(R,O,X(1))p(R,O,X^{(1)}) that are Markov relative to a missing data ADMG 𝒢.{\mathcal{G}}.

Proof.

Soundness is a direct consequence of Theorem 3 and completeness is a direct consequence of Lemma. 3. ∎

References

  • Bhattacharya et al. (2019) Bhattacharya, R., Nabi, R., Shpitser, I., and Robins, J. M. Identification in missing data models represented by directed acyclic graphs. In Proceedings of the 35th Conference on Uncertainty in Artificial Intelligence. AUAI Press, 2019.
  • Chen (2007) Chen, H. Y. A semiparametric odds ratio model for measuring association. Biometrics, 63:413–421, 2007.
  • Chen et al. (2015) Chen, H. Y., Rader, D. E., and Li, M. Likelihood inferences on semiparametric odds ratio model. Journal of the American Statistical Association, 110(511):1125–1135, 2015.
  • Cheng (1994) Cheng, P. E. Nonparametric estimation of mean functionals with data missing at random. Journal of the American Statistical Association, 89(425):81–87, 1994.
  • Chickering (2002) Chickering, D. M. Optimal structure identification with greedy search. Journal of Machine Learning Research, 3(Nov):507–554, 2002.
  • Daniel et al. (2012) Daniel, R. M., Kenward, M. G., Cousens, S. N., and De Stavola, B. L. Using causal diagrams to guide analysis in missing data problems. Statistical Methods in Medical Research, 21(3):243–256, 2012.
  • Dempster et al. (1977) Dempster, A. P., Laird, N. M., and Rubin, D. B. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society: Series B (Methodological), 39(1):1–22, 1977.
  • Drton (2009) Drton, M. Discrete chain graph models. Bernoulli, 15(3):736–753, 2009.
  • Evans (2018) Evans, R. J. Margins of discrete Bayesian networks. The Annals of Statistics, 46(6A):2623–2656, 2018.
  • Evans & Richardson (2014) Evans, R. J. and Richardson, T. S. Markovian acyclic directed mixed graphs for discrete data. The Annals of Statistics, pp.  1452–1482, 2014.
  • Fielding et al. (2008) Fielding, S., Fayers, P. M., McDonald, A., McPherson, G., Campbell, M. K., et al. Simple imputation methods were inadequate for missing not at random (MNAR) quality of life data. Health and Quality of Life Outcomes, 6(1):57, 2008.
  • Gain & Shpitser (2018) Gain, A. and Shpitser, I. Structure learning under missing data. In Proceedings of the 9th International Conference on Probabilistic Graphical Models, pp.  121–132, 2018.
  • Lauritzen (1996) Lauritzen, S. L. Graphical Models. Oxford, U.K.: Clarendon, 1996.
  • Malinsky et al. (2019) Malinsky, D., Shpitser, I., and Tchetgen Tchetgen, E. J. Semiparametric inference for non-monotone missing-not-at-random data: the no self-censoring model. arXiv preprint arXiv:1909.01848, 2019.
  • Marra et al. (2017) Marra, G., Radice, R., Bärnighausen, T., Wood, S. N., and McGovern, M. E. A simultaneous equation approach to estimating HIV prevalence with nonignorable missing responses. Journal of the American Statistical Association, 112(518):484–496, 2017.
  • Marston et al. (2010) Marston, L., Carpenter, J. R., Walters, K. R., Morris, R. W., Nazareth, I., and Petersen, I. Issues in multiple imputation of missing data for large general practice clinical databases. Pharmacoepidemiology and Drug Safety, 19(6):618–626, 2010.
  • Martel García (2013) Martel García, F. Definition and diagnosis of problematic attrition in randomized controlled experiments. Working paper. Available at SSRN 2302735, 2013.
  • Mohan & Pearl (2014) Mohan, K. and Pearl, J. Graphical models for recovering probabilistic and causal queries from missing data. In Proceedings of the 28th Conference on Advances in Neural Information Processing Systems, pp.  1520–1528. 2014.
  • Mohan et al. (2013) Mohan, K., Pearl, J., and Tian, J. Graphical models for inference with missing data. In Proceedings of the 27th Conference on Advances in Neural Information Processing Systems, pp.  1277–1285. 2013.
  • Ogarrio et al. (2016) Ogarrio, J. M., Spirtes, P. L., and Ramsey, J. D. A hybrid causal search algorithm for latent variable models. In Proceedings of the 8th International Conference on Probabilistic Graphical Models, pp.  368–379, 2016.
  • Pearl (1988) Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, 1988.
  • Pearl (2009) Pearl, J. Causality. Cambridge university press, 2009.
  • Ramsey (2015) Ramsey, J. D. Scaling up greedy causal search for continuous variables. arXiv preprint arXiv:1507.07749, 2015.
  • Richardson (2003) Richardson, T. S. Markov properties for acyclic directed mixed graphs. Scandinavian Journal of Statistics, 30(1):145–157, 2003.
  • Richardson et al. (2017) Richardson, T. S., Evans, R. J., Robins, J. M., and Shpitser, I. Nested Markov properties for acyclic directed mixed graphs. arXiv:1701.06686v2, 2017. Working paper.
  • Robins & Gill (1997) Robins, J. M. and Gill, R. D. Non-response models for the analysis of non-monotone ignorable missing data. Statistics in Medicine, 16(1):39–56, 1997.
  • Robins et al. (1994) Robins, J. M., Rotnitzky, A., and Zhao, L. P. Estimation of regression coefficients when some regressors are not always observed. Journal of the American Statistical Association, 89:846–866, 1994.
  • Saadati & Tian (2019) Saadati, M. and Tian, J. Adjustment criteria for recovering causal effects from missing data. In European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2019.
  • Sadinle & Reiter (2017) Sadinle, M. and Reiter, J. P. Itemwise conditionally independent nonresponse modelling for incomplete multivariate data. Biometrika, 104(1):207–220, 2017.
  • Shpitser (2016) Shpitser, I. Consistent estimation of functions of data missing non-monotonically and not at random. In Proceedings of the 30th Annual Conference on Neural Information Processing Systems. 2016.
  • Shpitser et al. (2015) Shpitser, I., Mohan, K., and Pearl, J. Missing data as a causal and probabilistic problem. In Proceedings of the 31st Conference on Uncertainty in Artificial Intelligence, pp.  802–811. AUAI Press, 2015.
  • Strobl et al. (2018) Strobl, E. V., Visweswaran, S., and Spirtes, P. L. Fast causal inference with non-random missingness by test-wise deletion. International Journal of Data Science and Analytics, 6(1):47–62, 2018.
  • Tchetgen Tchetgen et al. (2018) Tchetgen Tchetgen, E. J., Wang, L., and Sun, B. Discrete choice models for nonmonotone nonignorable missing data: Identification and inference. Statistica Sinica, 28(4):2069–2088, 2018.
  • Thoemmes & Rose (2013) Thoemmes, F. and Rose, N. Selection of auxiliary variables in missing data problems: Not all auxiliary variables are created equal. Technical report, R-002, Cornell University, 2013.
  • Tian (2017) Tian, J. Recovering probability distributions from missing data. In Proceedings of the Ninth Asian Conference on Machine Learning, 2017.
  • Tian & Pearl (2002) Tian, J. and Pearl, J. A general identification condition for causal effects. In Proceedings of the 18th National Conference on Artificial Intelligence, pp.  567–573, 2002.
  • Tsiatis (2006) Tsiatis, A. Semiparametric Theory and Missing Data. Springer-Verlag New York, 1st edition edition, 2006.
  • Tu et al. (2019) Tu, R., Zhang, C., Ackermann, P., Mohan, K., Kjellström, H., and Zhang, K. Causal discovery in the presence of missing data. In The 22nd International Conference on Artificial Intelligence and Statistics, pp.  1762–1770, 2019.
  • Vansteelandt et al. (2007) Vansteelandt, S., Rotnitzky, A., and Robins, J. M. Estimation of regression models for the mean of repeated outcomes under nonignorable nonmonotone nonresponse. Biometrika, 94(4):841–860, 2007.
  • Verma & Pearl (1990) Verma, T. and Pearl, J. Equivalence and synthesis of causal models. In Proceedings of the 6th Conference on Uncertainty in Artificial Intelligence, 1990.
  • Zhou et al. (2010) Zhou, Y., Little, R. J. A., and Kalbfleisch, J. D. Block-conditional missing at random models for missing data. Statistical Science, 25(4):517–532, 2010.