Marginal Independence and Partial Set Partitions
Abstract.
We establish a bijection between marginal independence models on random variables and split closed order ideals in the poset of partial set partitions. We also establish that every discrete marginal independence model is toric in cdf coordinates. This generalizes results of Boege, Petrovic, and Sturmfels [1] and Drton and Richardson [2], and provides a unified framework for discussing marginal independence models.
1. Introduction
Marginal independence models are often studied by associating a combinatorial structure to describe the independence relations between random variables. For instance, Drton and Richardson [2] used bidirected graphs to develop a family of marginal independence models for discrete random variables, and Boege, Petrovic, and Sturmfels [1] used simplicial complexes. The marginal independence models of [2] fit more broadly into the class of graphical models [7, 9]. Each of these structures describes a very different family of models, with very little overlap between them. In both cases it had been shown that when we restrict to discrete random variables, we obtain toric models after a linear change of coordinates. However, there are collections of marginal independence statements that cannot be expressed either as graphs or simplicial complexes. This motivates the proposal of a more general combinatorial structure which we use in our treatment of marginal independence.
In this paper, we study marginal independence using the poset of partial set partitions, . In Section 2, we give a combinatorial description of marginal independence models on random variables by showing a correspondence between the models and certain order ideals of .
In Section 3, we then restrict to discrete random variables and show that the resulting models are toric and smooth after a linear change of coordinates. In [1, 2, 9], the proof that the mentioned family of marginal independence models are toric uses a linear change of coordinates into Möbius coordinates, by using the Möbius transformation on an appropriately defined poset. In this paper, we describe a change of coordinates using the cumulative distribution function, which is classically used to describe marginal independence. This description has the advantage of tying the results to the classical description of marginal independence. We then move from the affine description of our models to the projective description by showing a parametrization of the homogeneous ideal corresponding to the model.
Finally, in Section 4 we discuss how marginal independence models associated to graphs and simplicial complexes can fit in the framework proposed in this paper. We then compare both classes of models and give an explicit characterization of the models that can be described using both simplicial complexes and graphs. We also count the number of distinct models in these three classes for up to random variables.
2. Marginal Independence and the Poset of Partial Set Partitions
In this section we discuss general marginal independence models, for completely general types of random variables. Marginal independence is naturally described via constraints on the joint cumulative distribution function of a collection of random variables. We will also see a combinatorial structure which can be used to capture all marginal independence models. This structure is called the poset of partial set partitions, and we will show how certain order ideals of this poset correspond to marginal independence models. We achieve this by introducing a closure operation on order ideals which is based on translating certain implications of marginal independence statements to the language of this poset.
Definition 2.1.
Let be random variables in . The joint cumulative distribution function (cdf) of , is the function defined on defined by the rule
Note that for a function to be a joint cumulative distribution function it must satisfy a few properties.
-
•
For all ,
-
•
.
-
•
should be nondecreasing in each coordinate: That is, if coordinate-wise, then .
-
•
should be right continuous in each coordinate: That is, for each , and all ,
-
•
For all , .
These conditions are not exhaustive; that is, there are functions that satisfy all of these properties but are not cdfs of a collection of random variables. However, we do not need to know a precise characterization of functions which are joint cdfs to derive our results.
Note that the joint cdf contains all the marginal cdfs as well. For example, with three random variables, we have
Marginal independence of collections of random variables can be expressed in terms of the joint cdf. Let denote the random vector, and be a subvector.
Definition 2.2.
Let , be disjoint subsets. We say that and are marginally independent (denoted , or just for short), if
for all such that for all and where
Example 2.3.
Suppose , and consider the marginal independence statement (or , or ). This translates into the condition:
Said another way, we have, for all ,
More generally, we can define marginal independence for a collection of sets.
Definition 2.4.
Let with for all . Then if, for all with for all , we have
where
Now that we have the notion of generalized marginal independence, we can discuss marginal independence models. A marginal independence model is a collection of marginal independence statements and the set of probability distributions that are compatible with those statements. We will describe a completely general combinatorial structure for representing all such marginal independence models. This depends on the structure of a certain poset called the poset of partial set partitions.
Definition 2.5.
A partial set partition of is an unordered list of nonempty sets such that each , and for all . The set of all partial set partitions of is denoted . A set in the partial set partition is called a block. The ground set, of is defined as . The set of partial set partitions where each has at least blocks is denoted .
Note the elements of correspond to marginal independence statements. Specifically corresponds to the independence statement .
We want to turn into a poset that is compatible with marginal independence implications. We will explore some implications between marginal independence statements that will lead naturally to a poset structure on .
Lemma 2.6.
If a cdf satisfies then,
-
(1)
satisfies for all such that .
-
(2)
satisfies where for all .
Proof.
First, we will show that for such that , if satisfies , then it satisfies . Let be such that if , in particular, if as well. Thus,
where
Hence,
We get an analogous result for .
Now, let be such that if . Then, it holds that
where
Hence,
where
Now we will show that if satisfies , then it satisfies whenever for all . Assume satisfies . Let be such that for all , then it follows that
where
which proves also satisfies . ∎
Proposition 2.7.
Suppose that the joint cdf satisfies the marginal independence statement . Let be another marginal independence statement such that for each , there is a set such that for all , and . Then also satisfies .
Proof.
This yields a partial order on by the same rule as in Proposition 2.7.
Definition 2.8.
Let and be two partial set partitions. Define a partial order on by declaring if for each , there is a set such that for all , and .
Note that the covering relations in are of two types. We have that covers if either
-
(1)
can be obtained from by adding a single element to one of the blocks of , or
-
(2)
can be obtained from by splitting a block of into two blocks.
Proposition 2.9.
In , if and only if and , where is the standard order in the lattice of set partitions of and .
Proof.
This follows directly from the covering relations. ∎
Remark.
Note that is not, in general, a lattice. Indeed, the elements and have two least upper bounds, namely and . This shows a key difference with the poset of set partitions , which is well known to be a lattice [8].
On the other hand, is a graded poset, with
This formula follows by counting the number of covering relations to get from the bottom element to a partial set partition .
Remark.
Note that contains the Boolean lattice as a subposet. The lattice of set partitions is not a subposet of , but rather its opposite is a subposet. In fact, contains the opposite of the lattice of set partitions of any .
For studying marginal independence models, it is more natural to consider the poset of partial set partitions where each partition has at least two blocks. We will see that marginal independence models correspond to certain types of order ideals in .
Definition 2.10.
Let , then
or, equivalently,
is a list of independence statements on random variables . A marginal independence model, , on these random variables is the set of distributions that satisfy all the statements in .
Definition 2.11.
Let and . We say splits if for some . We denote the splitting of by with .
Definition 2.12.
Let be an order ideal of . We say is split closed if for all such that splits .
Definition 2.13.
The split closure of is the smallest split closed order ideal containing .
Remark.
This allows us to make a precise connection between collections of marginal independence statements and certain order ideals in , namely, their split closure .
Proposition 2.14.
Let be a a split closed order ideal. If are such that for some , then where is the common refinement in the lattice of set partitions of .
Proof.
Let be such that for some . Observe that where for each , for some . Now consider the partition , by definition, either for some or . Observe further that we can write . If for all , we have that , then . On the other hand, if , clearly and since is a split closed order ideal, so will . Thus, ∎
We finish this section stating a general theorem about marginal independence models that makes the correspondence between split closed order ideals and models precise. We prove this theorem in Section 3.
Theorem 2.15.
Marginal independence models on random variables are in bijective correspondence to split closed order ideals .
In this level of generality, it is not clear how to prove the correspondence yet. One of the directions is straightforward, based on results we have already proven: for any marginal independence model, there exists a split closed order ideal by taking all implied marginal independence statements. For the other direction, it suffices to show the existence of probability distributions that satisfy exactly the independence statements in a given split closed order ideal, and no others. We will derive such result as a consequence of our study of marginal independence for discrete random variables in the following section.
3. Discrete marginal independence models
In this section, we will discuss marginal independence models restricting the class of random variables to have finitely many states to allow us to translate the independence statements into polynomial constraints. Past work has been done on certain classes of these models where the marginal independence statements are given by graphs or simplicial complexes [1, 2]. However, there are marginal independence models that cannot be represented with either of these structures, which is why we present a more general approach using split closed order ideals from the set of partial set partitions. In previous work [1, 2, 9], it has been shown these models are toric after a certain linear change of coordinates whose inverse could be found by applying Möbius inversion on a specific poset. In our treatment of marginal independence, we will use the cumulative distribution function as our coordinate system. This allows us to express marginal independence constraints in a very natural way while preserving the structure of the inverse map as a Möbius inversion on a poset isomorphic to the one used in [1].
We begin by defining what a single marginal independence statement looks like in terms of the probability tensor and then we will generalize it further to define our models. For this, we need to first discuss some notation that will be used.
Definition 3.1.
For finite random variables , with respective state spaces , , we use the shorthand to denote . Furthermore, when we discuss marginalizations of the probability tensor, we will allow to symbolize adding over all the states of the corresponding variable, i.e. if , then
For an index vector , we define the support of as .
Definition 3.2.
Let be finite random variables, we say that for disjoint , holds if and only if for a probability distribution indexed by we have that
for all , with , , , , and .
Equivalently, these constraints can be obtained from all the minors of the matrix resulting from marginalizing the probability tensor by summing out the indices not in and whose columns are indexed by the states of and the rows are indexed by the states of , where and are seen as random vectors.
Definition 3.3.
Let be a probability distribution on random variables with respective state spaces , with for . Then, we define .
Definition 3.4.
We define the ring of polynomials in probability coordinates, , for the probability tensor of format as
Definition 3.5.
The marginal independence ideal () associated to the split closed order ideal on discrete random variables is
where and is the -way tensor which is obtained by first marginalizing the tensor by summing out the indices not in , and then flattening the resulting tensor according to the set partition .
Alternatively, is the ideal resulting from imposing rank 1 conditions on the tensors , for all . This definition is a slight modification of the one in [1] to fit our class of models.
Example 3.6.
Let be a binary random variable and let be ternary randcom variables and consider the marginal independence statement . Then the equations that define are
for all . We can also obtain these equations from the minors of the matrix
Example 3.7.
Let and consider the model given by binary random variables such that . Then, the equations that define are
for all . These equations can again be obtained from the minors of the matrix
The ideal tends not to be prime when has more than one maximal element, which represents a challenge when studying these models. We get around this issue by using the cumulative distribution function as our new system of coordinates, since marginal independence constraints can be expressed in a simpler way using the cdf.
Definition 3.8 (cdf coordinates).
We define the linear change of coordinates , by transforming the tensor so that the entries of are cumulative distributions as follows:
Remark.
In the case of binary random variables, cdf coordinates become much simpler. We have that for any , either or , thus we can represent each cdf coordinate by indexing it with its support.
We can discuss this transformation, and in particular, its inverse, in terms of the poset of the index vectors of the tensors. Let be the usual direct product poset, where two index vectors if for all . Then, we can state the equation for cdf coordinates in Definition 3.8 as follows, for all ,
Using the Möbius inversion formula we have that for all ,
where
Example 3.9.
In the case when , we have that the poset would be as in Figure 1.
Then, we have that
Hence, using Möbius inversion on this poset yields
Marginal independence is usually stated in terms of the cumulative probability distribution, which is why this is a natural transformation to consider as it simplifies the equations for marginal independence.
Our goal now is to to explore the ideals of discrete marginal independence models in the cdf coordinates. To do this, we need to explore some properties of split closed order ideals.
Definition 3.10.
Let be a split closed order ideal of , then
-
(1)
We say that a nonempty set is disconnected (with respect to ) if there exists such that .
-
(2)
A connected set is a set that is not disconnected.
-
(3)
A connected set is maximal with respect to some disconnected set if for any set such that , then is disconnected.
-
(4)
We define the set
Proposition 3.11.
For any disconnected set with respect to , there exists a unique maximal such that
where all are maximal connected sets.
Proof.
Since is disconnected, there exists such that . If some block is disconnected, then there is some such that . Then, is the new decomposition, additionally, and . We can repeat this process on if any of its blocks are disconnected. This process terminates since every chain of a finite poset is finite.
To show uniqueness of , let us consider a different maximal decomposition . Since , and by assumption, , then, thus contradicting maximality of and . ∎
Let . Define the support of to be . Note that for cdf coordinates this will agree with our previous notion of support, since if in , it means we are assuming in our cdf calculations, a condition denoted by in .
Corollary 3.12.
Let be a split closed order ideal. A probability tensor lies in the model if and only if its cdf coordinates satisfy the following condition. For any that is disconnected by , and be index vectors where
-
•
is the decomposition of into maximal connected sets.
-
•
, , and
-
•
we have
(1) |
Proof.
We know lies in the model if and only if its corresponding cdf satisfies all of the marginal independence constraints. We know that the cdf must satisfy that for all ,
where
-
•
, ,
-
•
.
If we have a non maximal , then we can decompose it further using Proposition 3.11. Hence, the non maximal yield equations that are implied by maximal set partitions in the sense of Proposition 3.11. ∎
Remark.
Observe that in the case of binary random variables, we get that for any disconnected set with respect to with maximal connected sets ,
Definition 3.13.
We define the ring of polynomials in cdf coordinates for the cdf tensor of format as
Definition 3.14.
Let be a split closed order ideal on and such that is disconnected with respect to . Then, the associated maximal equation is given by
where
-
•
,
-
•
, and
-
•
is the decomposition of into maximal connected sets.
Definition 3.15.
The inhomogeneous marginal independence ideal () associated to the split closed order ideal on discrete random variables is
Example 3.16.
In Example 3.7, there are 16 cdf coordinates indexed by subsets of . The equations that define the ideal without indexing them with their support are
However, since all random variables are binary, we can instead index them by their support as follows:
This ideal is prime and therefore toric in cdf coordinates.
Definition 3.17.
The marginal independence ideal in cdf coordinates () associated to the split closed order ideal on discrete random variables is
where and is the -way tensor which is obtained by first marginalizing the tensor by making all the indices not in equal to their maximum value (this corresponds to summing out indices in Definition 3.5, as we are in cdf coordinates), and then flattening the resulting tensor according to the set partition .
Alternatively, is the ideal resulting from imposing rank 1 conditions on the tensors , for all .
Example 3.18.
Let us consider the same model as in Example 3.7. Then, is going to be given by imposing rank conditions on the tensors , and , all of which are matrices since they are all partial set partitions with 2 blocks. Following the construction in Definition 3.17, we obtain
Furthermore, since the random variables are binary, we can reindex the cdf coordinates by their support as follows:
Observe that all of the rank 1 conditions in and are found in the minors of . Thus, is given by the minors of
This ideal and its generators generally have a more complicated structure than . However, once we add a constraint that is imposed by the probability simplex, both ideals have the same generators, which allows us to work with the simpler ideal to study these models. To prove this result, we first need to prove the following:
Lemma 3.19.
Let be such that for some . Then, if and only if the minors involving vanish.
Proof.
The forward direction is direct. For the converse, without loss of generality (since rank is invariant under row/column operations), assume and the minors involving vanish. Let denote the -th row of . To show is rank 1, it suffices to show is a scalar multiple of , for any . We claim that scalar is . Let , note
which is true as these correspond to the vanishing minors involving . ∎
Proposition 3.20.
For a split closed order ideal , we have that .
Proof.
First observe that for any , there exist partial set partitions such that
these are given by
In other words, we can get any partial set partition by taking partitions with 2 blocks and applying the splitting operation to them. On the algebraic side, this means the maximal equations that generate can be obtained if we have all the equations corresponding to set partitions with 2 blocks in .
Note that the matrices with always have appearing. Furthermore, the minors that involve are of the form
where are such that
-
•
,
-
•
,
-
•
,
-
•
, and
-
•
Once we add the constraint , we observe that
Since any set partition in can be obtained by splitting set partitions with 2 blocks, we can keep splitting and until we get the maximal equation . Thus showing .
For the other direction, observe that for any , we know that any flattening of the tensor contains . Additionally, under the constraint , any such matrix contains a 1. Thus, a flattening of is rank 1 if and only if the minors involving the entry of vanish, by Lemma 3.19. However, these minors are all equations of the same form as above, all of which are in by construction. Hence showing the tensor is rank 1 if and only if its entries lie in , which in turn shows generates . ∎
Now we will exploit the combinatorial structure of the equations in to study these models. Observe that Corollary 3.12 gives us a unique factorization for all cdf coordinates and, in particular, it tells us that the cdf coordinates that appear in the defining equations of the variety are precisely those with disconnected support.
Theorem 3.21.
Every discrete marginal independence model is toric in cdf coordinates.
Proof.
By Corollary 3.12, we know that the ideal that generates the model is given by equations of the form
where
-
•
is disconnected.
-
•
, ,
-
•
and
-
•
is the decomposition of into maximal connected sets.
Now, let us consider the term order on defined by a lexicographic order with the condition that if , then . Therefore, for each generator of , we have that .
This is a Gröbner basis for since the leading terms of any pair of equations are coprime, so any -pair reduces to modulo the basis of difference of binomials. Thus, the initial ideal of , is generated by linear terms, so it is prime. This implies that is a prime ideal generated by differences of binomials, hence it is toric. ∎
The explicit description of the generators of also allows us to show that its associated affine variety is smooth.
Example 3.22.
Consider the split closed order ideal on 3 binary random variables generated by the independence statement . Then, by Corollary 3.12,
Hence, the Jacobian of is
Observe that is smooth everywhere since its Jacobian has full rank independent of the point of evaluation.
Theorem 3.23.
For a split closed order ideal , the affine variety that defines the model is smooth.
Proof.
It suffices to show that the Jacobian of is full rank. Observe that is generated by equations of the form
(2) |
where
-
•
, ,
-
•
and
-
•
is the decomposition of into maximal connected sets.
for all disconnected . Hence, the equations that appear on the generating set are precisely those indexed by such that is disconnected. In addition, observe that we have that for all index vectors and with disconnected support,
This follows because of the uniqueness of the decomposition of . Hence, the columns of the Jacobian contain some permutation of the columns of the identity matrix of size , which shows that has full rank everywhere. ∎
We are now ready to prove a result about existence of distributions in the binary case.
Proposition 3.24.
For each split closed order ideal , there exists a probability distribution in the binary model such that and satisfies all the marginal independence statements in and no other marginal independence statements.
Proof.
We want to show there is a point in the interior of the probability simplex that satisfies exactly the statements in and no others.
Since we are using binary random variables, the description of our generators becomes much simpler,
Take to be minimal, this corresponds to an equation where is connected. Let
Observe that . So that we still get a Gröbner basis for composed of linear terms and therefore, . Hence, we have a strict inclusion, , showing that . Since the uniform distribution , it intersects the interior of the probability simplex (in cdf coordinates). Finally, since we additionally know is smooth, we conclude that . ∎
Proof of Theorem 2.15.
Given a model , we obtain a unique split closed order ideal given by the closure of . Now let us start with a split closed order ideal . It suffices to show that for any , there exists a distribution that satisfies the statements in and does not satisfy , this follows directly from Proposition 3.24. ∎
We have now proven the bijective correspondence between split closed order ideals and marginal independence models, which provides us with a combinatorial description of this class of models.
So far, we have been describing our models and varieties in affine space. However, Proposition 3.11 and Corollary 3.12 allow us to give an explicit homogeneous parametrization of the ideals defining our models.
Corollary 3.25.
Let . A homogeneous parametrization of the model in cdf coordinates is given by the homomorphism between polynomial rings defined by
where is the decomposition of into maximal connected sets.
Remark.
In the binary case, we can lose the subscript on our parameters as a consequence of the reindexing of binary cdf coordinates with their support.
Definition 3.26.
The homogeneous marginal independence ideal () associated to the split closed order ideal on discrete random variables is given by
where is the parametrization given in Corollary 3.25.
Proposition 3.27.
For a split order model ideal , we have
where is the homogenization of .
Proof.
This follows from construction of the parametrization in Corollary 3.25 and the definition of . ∎
In general, toric varieties are defined by a monomial map with an associated integer matrix . This parametrization allows us to compute this integer matrix and the associated polytope given by the convex hull of the columns of .
Example 3.28.
Consider the model on 4 binary variables given by . Then, we have 16 cdf coordinates, and the equations that define are:
From this description, we get the generators for and consequently all coordinates with disconnected support. Hence, the parametrization described in Corollary 3.25 yields the following integer matrix:
Furthermore, we know that the ideal is going to be the kernel of the monomial map given by , which, in this case we compute using Macaulay2 [3] and find that there are 46 degree 2 generators of the ideal. They include homogenized equations corresponding to the generators of
However, we find more equations in the minimal generators of the homogeneous ideal, such as
which are not necessary in the de-homogenized cdf coordinates since they are implied by the original equations (when ).
4. Graphical and simplicial marginal independence models
Two particularly interesting classes of marginal independence models are discussed in [1] and in [2]. They are graphical models and what we will refer to as simplicial models. We aim to contrast these different classes of models and show how they fit in the framework presented above.
4.1. Graphical models
Graphical models have been studied extensively because of their many applications and interesting structure that comes with a graphical representation of the independence relations. Different types of graphs are used to express different relationships between variables: directed acyclic graphs are been used to study causality in Bayesian networks, undirected graphs on the other hand, are used to study conditional independence models. Following the work of Drton and Richardson in [2], we use bidirected graphs as a framework to discuss marginal independence. Hence, for this section, will be a bidirected graph with vertex set and a set of edges .
Definition 4.1.
Let be a bidirected graph, for a vertex we define the spouse of as the set of all vertices that share an edge with and itself, i.e., . For a subset , we define .
There are several different ways in which we can read different kinds of independent statements from a graph. The rules for reading a graph are called Markov properties, we are particularly interested in two sets of Markov properties: the global Markov property and the connected set Markov property.
Definition 4.2.
Let be a bidirected graph. A probability distribution is said to satisfy the global Markov property if for all pairwise disjoint such that separates and , we have that
Definition 4.3.
A probability distribution is said to satisfy the connected set Markov property associated to a bidirected graph if
where is a connected subgraph of .
The global Markov property contains all of the marginal statements given by the connected set Markov property since for any connected set , we have that is separated from by (in this case, the third set would be empty). Even though the global Markov property has conditional statements that are not given by the connected set property, we know that a probability distribution satisfies the global Markov property if and only if it satisfies the connected set property.
Additionally, since is connected, the connected set Markov rules imply that if do not have an edge between them, then . This is known as the pairwise Markov property, hence, both the global and connected set Markov properties imply all the statements in the pairwise Markov property.
Example 4.4.
Consider the chain with 4 vertices. The connected set Markov property yields and and the global Markov property gives us as well.
Definition 4.5 (Graphical model).
The marginal independence graphical model associated to , , is given by the set of distributions and discrete random variables that satisfy the marginal independence statements given by the connected set Markov property of .
Example 4.6.
Let be the graph on 3 vertices with the edge and binary random variables. Then, the marginal independence statements in this model are given by and the equations defining are
These are the same defining equations for the model as the ones in Example 3.7. However, they represent slightly different models since here we have 3 random variables instead of 4.
In general, given a graph , the collection of the marginal independence statements is defined by all the connected set Markov statements associated to the graph. This means that we can apply the results from Section 3 to this special class of models. Furthermore, the notion of connectedness relative to a collection of statements agrees with the notion of connectedness of subgraphs of the corresponding graph .
Proposition 4.7.
Let be a bidirected graph and the split closed order ideal defined by the connected set Markov statements. Then a set is disconnected with respect to if and only if the corresponding subgraph with vertices in is disconnected.
Proof.
If is disconnected with respect to , then there exists such that . Without loss of generality, we may assume that is connected with respect to and that it only has two blocks, so, . Thus, the subgraph induced by is disconnected.
Conversely, if the subgraph of is disconnected, then we can certainly find a connected component of , and take so that is disconnected with respect to our collection . ∎
Proposition 4.7 ensures that in this particular class of models, we can restate the results of Section 3 in terms of the graph theoretic sense of connectedness, this agrees with the results presented in [2] in the binary case and allows us to extend them to the general case.
Example 4.8.
Let be the model on 3 binary variables given by the marginal independence statements , , . The only graphical model on 3 binary variables that contains these marginal independence statements is given by the graph on 3 vertices that has no edges. However, because of the connected set Markov rule, this graphical model also has the statement , which cannot be inferred from . We will show in the following section that this model is a simplicial complex model.
4.2. Simplicial complex models
This class of models was proposed in [1] and it concerns a very different class of models. We obtain these models from a simplicial complex instead of a graph, and the way we read statements from simplices is that all of the random variables indexed by a face are completely independent. Furthermore, this can be achieved by imposing rank 1 constraints on the probability tensor indexed by any particular face, which consequently makes the factorization in cdf coordinates and the homogeneous parametrization much simpler.
Definition 4.9 (Simplicial complex model).
Let be a simplicial complex, then the model is defined by the distributions and finite random variables such that the random variables are completely independent for any face .
Example 4.10.
Let be the simplicial complex on 3 binary variables defined by the 3 cycle, i.e., the marginal independence relations would be exactly the same as in Example 4.8. This shows that there are simplicial models that are not graphical models.
Remark.
We can discuss this class of models in terms of general marginal independence models by defining a collection of marginal independence statements generated by . Thus, we can apply all of the results from Section 3 to simplicial complex models, i.e. we get a toric model in cdf coordinates and a parametrization in terms of the connected sets. By construction, the disconnected sets of are the faces of (except the vertices), and the connected sets are the non-faces and the vertices.
Next, we wish to compare graphical models and simplicial complex models. We need to be careful when comparing the two classes of models since the way we read the marginal independence statements from their corresponding structures is very different: in a simplicial complex, any face gives an marginal independence statement (in particular, edges represent marginal independence relations), whereas in graphs, any non-edge gives a marginal independence statement (and we get more from Markov rules). For this reason, when we are comparing them, it is useful to talk about the complementary graph of .
In general, we can associate a simplicial complex model to any graphical model such that the graphical model is a submodel of the simplicial complex model.
Proposition 4.11.
Let be a graph and be its complementary graph, where is the usual complement of . Then, the associated simplicial complex, , given by
gives us that .
Proof.
We need to show the containment of the ideals , which in turn requires us to show that any marginal independence statement from can be obtained from as well. The edges (1 dimensional faces) of are exactly the non-edges in and the faces of are the completely disconnected subgraphs of . Hence, every marginal independence statement from can be found in the model associated to . Thus proving that is a submodel of . ∎
Remark.
Observe that the construction in Proposition 4.11 for a given implies that the -skeleton of is the complimentary graph . The existence of a simplicial complex model that contains a given graphical model raises the question of when the simplicial complex models and graphical models are isomorphic. Consider the following example:
Example 4.12.
Let be the model associated to the binary 4-cycle.
This gives us the marginal independence relations and . As we can see in Figure 2, the simplicial complex associated to this graph gives us the same marginal independence statements as , i.e. .
Example 4.13.
Let be the model associated to the binary 4-chain.
This graph gives us the following marginal independence statements: , . Its associated simplicial complex, , is again a 4 chain that implies the following: , and . We cannot derive the marginal independence statements given by from this . Thus, in this case the models are not isomorphic (even though the 1-skeleton of and are isomorphic as graphs).
For a graphical model to be isomorphic to its simplicial complex counterpart, we would need that the marginal independence statements on are only statements of complete marginal independence of individual variables, since those are the only type of marginal independence statements that we can obtain from the simplicial complex model.
Proposition 4.14.
For a graph , if the complimentary graph of is a disjoint union of complete graphs.
Proof.
Let where each , , is a complete graph and they are disjoint. Then, observe that each corresponds to a totally disconnected subgraph of whose vertices are connected all vertices not in . Hence, each yields a complete marginal independence statement on the random variables indexed by the vertices of . Furthermore, by construction, these are all the statements that we are going to get from the graph . We know that those are precisely the statements that we get from , thus showing that the models are isomorphic. ∎
Remark.
We can restate Proposition 4.14 without using the complimentary graph since that condition on is equivalent to being a multipartite graph. Additionally, we can see from this proposition that the number of nontrivial isomorphic models is going to be the number of partitions of .
Example 4.15.
Let be finite random variables, and consider the model given by the marginal independence statements and . This model is not simplicial and not graphical, but it still is toric.
4.3. Counting models and computational results.
We counted all models on 3 and 4 random variables and counted up to symmetry of relabeling the random variables and obtained the following:
Total | Up to symmetry | Total | Up to symmetry | |
---|---|---|---|---|
Graphical | 8 | 4 | 64 | 11 |
Simplicial | 9 | 5 | 114 | 20 |
Graphical and simplicial | 5 | 3 | 18 | 5 |
General | 12 | 6 | 496 | 53 |
For the simplicial and graphical models, we used data from OEIS [4, 5, 6]. For the general case, we do not yet have an efficient way of counting split closed order ideals, so we used exhaustive computation checking split closures of order ideals.
Additionally, we computed the homogeneous ideals for all models up to labeling symmetry for and found the highest degree of the generators, degree and dimension of the projective variety that comes from and whether or not the models are simplicial or graphical.
Split order | Max degree | Degree | Dimension | Graphical | Simplicial |
---|---|---|---|---|---|
ideal | gens | ||||
1 | 15 | x | x | ||
2 | 2 | 14 | x | x | |
2 | 3 | 13 | x | ||
2 | 4 | 13 | x | x | |
2 | 4 | 12 | x | ||
2 | 4 | 12 | x | ||
3 | 5 | 12 | x | x | |
2 | 5 | 12 | x | ||
3 | 7 | 11 | x | ||
2 | 6 | 11 | x | ||
2 | 5 | 11 | |||
2 | 7 | 11 | |||
2 | 6 | 11 | x | ||
3 | 9 | 10 | x | ||
3 | 9 | 10 | |||
2 | 9 | 10 | |||
2 | 8 | 10 | x | ||
2 | 6 | 10 | |||
2 | 10 | 10 | x | ||
3 | 12 | 9 | x | ||
3 | 12 | 9 | |||
3 | 11 | 9 | x | ||
3 | 11 | 9 | |||
2 | 14 | 9 | |||
2 | 7 | 9 | |||
2 | 10 | 9 | |||
2 | 10 | 9 | |||
3 | 16 | 8 | x | ||
3 | 15 | 8 | |||
3 | 17 | 8 | |||
3 | 13 | 8 | x | ||
3 | 14 | 8 | |||
3 | 14 | 8 | |||
2 | 16 | 8 | |||
2 | 8 | 8 | x | ||
2 | 12 | 8 | |||
2 | 12 | 8 | x | ||
3 | 21 | 7 | |||
3 | 18 | 7 | x | ||
3 | 17 | 7 | |||
2 | 19 | 7 | |||
2 | 14 | 7 | |||
3 | 17 | 7 | |||
2 | 15 | 7 | |||
3 | 23 | 6 | |||
2 | 20 | 6 | x | ||
3 | 19 | 6 | |||
2 | 18 | 6 | |||
3 | 20 | 6 | x | ||
2 | 20 | 5 | x | ||
3 | 23 | 5 | x | ||
3 | 25 | 5 | |||
2 | 24 | 4 | x | x |
References
- [1] Tobias Boege, Sonja Petrovíc, and Bernd Sturmfels. Marginal independence models. Proceedings of the 2022 International Symposium on Symbolic and Algebraic Computation, 2021.
- [2] Mathias Drton and Thomas S. Richardson. Binary models for marginal independence. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 70(2):287–309, 2008.
- [3] Daniel R. Grayson and Michael E. Stillman. Macaulay2, a software system for research in algebraic geometry. Available at http://www2.macaulay2.com.
- [4] OEIS Foundation Inc. Entry A000088 in The On-Line Encyclopedia of Integer Sequences. https://oeis.org/A000088, 2023.
- [5] OEIS Foundation Inc. Entry A261005 in The On-Line Encyclopedia of Integer Sequences. https://oeis.org/A261005, 2023.
- [6] OEIS Foundation Inc. Entry A307249 in The On-Line Encyclopedia of Integer Sequences. https://oeis.org/A307249, 2023.
- [7] S.L. Lauritzen. Graphical Models. Oxford Statistical Science Series. Clarendon Press, 1996.
- [8] Richard P. Stanley. Enumerative Combinatorics, volume 1 of Cambridge Studies in Advanced Mathematics. Cambridge University Press, 2 edition, 2011.
- [9] Seth Sullivant. Algebraic statistics, volume 194. American Mathematical Soc., 2018.