This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Revisiting Differentiable Structure Learning: Inconsistency of 1\ell_{1} Penalty and Beyond

Kaifeng Jin
University of California San Diego
[email protected]
&Ignavier Ng11footnotemark: 1
Carnegie Mellon University
[email protected]
\ANDKun Zhang
Carnegie Mellon University
Mohamed bin Zayed University of Artificial Intelligence
[email protected]
&Biwei Huang
University of California San Diego
[email protected]
Equal contribution.
Abstract

Recent advances in differentiable structure learning have framed the combinatorial problem of learning directed acyclic graphs as a continuous optimization problem. Various aspects, including data standardization, have been studied to identify factors that influence the empirical performance of these methods. In this work, we investigate critical limitations in differentiable structure learning methods, focusing on settings where the true structure can be identified up to Markov equivalence classes, particularly in the linear Gaussian case. While Ng et al. (2024) highlighted potential non-convexity issues in this setting, we demonstrate and explain why the use of 1\ell_{1}-penalized likelihood in such cases is fundamentally inconsistent, even if the global optimum of the optimization problem can be found. To resolve this limitation, we develop a hybrid differentiable structure learning method based on 0\ell_{0}-penalized likelihood with hard acyclicity constraint, where the 0\ell_{0} penalty can be approximated by different techniques including Gumbel-Softmax. Specifically, we first estimate the underlying moral graph, and use it to restrict the search space of the optimization problem, which helps alleviate the non-convexity issue. Experimental results show that the proposed method enhances empirical performance both before and after data standardization, providing a more reliable path for future advancements in differentiable structure learning, especially for learning Markov equivalence classes.

1 Introduction

Probabilistic graphical models, such as Bayesian networks, are powerful tools for capturing complex probabilistic relationships in a concise way (PEARL, 1988; Koller & Friedman, 2009). Their graph structures, usually encoded as Directed Acyclic Graphs (DAGs), allow efficient representation of data dependencies and have become essential in fields like health (Tennant et al., 2021) and economy (Awokuse & Bessler, 2003). Traditionally, learning these structures involves discrete methodologies. Constraint-based methods, which leverage conditional independence tests, are one common approach (Spirtes & Glymour, 1991; Spirtes et al., 2001). Another popular technique involves score-based methods, where the search space of potential graphs is explored based on scoring functions (Koivisto & Sood, 2004; Singh & Moore, 2005; Cussens, 2011; Yuan & Malone, 2013; Chickering, 2002; Peters & Bühlmann, 2014). Given the combinatorial nature of the task, greedy search strategies have been commonly used (Chickering, 1996; Chickering et al., 2004).

In recent years, Zheng et al. (2018) introduced a continuous formulation for characterizing the acyclicity constraint, effectively converting the discrete nature of the structure learning problem into one that can be approached using gradient-based optimization techniques. Although this formulation still involves nonconvex optimization, it opened the door to applying efficient gradient-based methods. This formulation has since inspired a wide range of extensions, being adapted to deal with nonlinearity (Yu et al., 2019; Lachapelle et al., 2019; Zheng et al., 2020; Ng et al., 2022b; Kalainathan et al., 2022), latent confounding (Bhattacharya et al., 2021; Bellot & van der Schaar, 2021), interventional data (Brouillard et al., 2020; Faria et al., 2022), time series data (Pamfil et al., 2020; Sun et al., 2021), and missing data (Wang et al., 2020; Gao et al., 2022). Other applications include multi-task learning (Chen et al., 2021), and working with federated learning systems (Ng & Zhang, 2022; Gao et al., 2021), domain adaptation (Yang et al., 2021), and recommendation system (Wang et al., 2022).

This move toward continuous structure learning has prompted growing attention to both its theoretical underpinnings and practical performance. Researchers like Wei et al. (2020) and Ng et al. (2022a) have investigated the optimality and convergence properties of continuous, constrained optimization techniques (Zheng et al., 2018). Meanwhile, Deng et al. (2023) provided insight into how an appropriately designed optimization scheme can reach the global minimum for least squares objectives in simple cases. Further refinements have also been proposed, with Zhang et al. (2022) and Bello et al. (2022) highlighting challenges such as gradient vanishing in existing DAG constraints (Zheng et al., 2018; Yu et al., 2019) and suggesting potential improvements.

Recently, Ng et al. (2024) highlighted the non-convexity issues in differentiable structure learning methods, particularly in the linear Gaussian setting where the true structure can be identified up to Markov equivalence classes. While non-convexity poses major challenges in this context, we further examine another critical issue: 1\ell_{1}-penalized likelihood is inconsistent, even if the global optimum of the optimization problem can be found. To address these limitations, we propose a method that resolves the 1\ell_{1} inconsistency and enhances empirical performance, both before and after data standardization, even under non-convex conditions.

Contributions

In this work, we tackle fundamental challenges in differentiable structure learning, particularly in the linear Gaussian case, by focusing on the limitations of penalized likelihood approaches. Our contributions include:

  • We examine and demonstrate the inconsistency of using 1\ell_{1}-penalized likelihood in differentiable structure learning methods, even if the global optimum of the optimization problem can be found, particularly when learning Markov equivalence classes in the linear Gaussian case.

  • We develop a differentiable structure learning method that optimizes an 0\ell_{0}-penalized likelihood with hard acyclicity constraints and incorporates a moral graph estimation procedure, where the 0\ell_{0} penalty is approximated by differentiable techniques, such as Gumbel-Softmax. We call our method CALM (Continuous and Acyclicity-constrained L0-penalized likelihood with estimated Moral graph). CALM not only addresses 1\ell_{1} inconsistency, but also results in a solution much closer to the true structure or its Markov equivalent graphs. Our method provides a more reliable path for future advancements in differentiable structure learning, especially for learning Markov equivalence classes.

  • Our method performs consistently well both before and after data standardization, demonstrating its robustness.

2 Background

In this section, we introduce our problem setting and, while reviewing the hard and soft DAG constraints, we also revisit NOTEARS (Zheng et al., 2018) and GOLEM (Ng et al., 2020).

2.1 Problem Setting

Setup

In this work, we focus on linear Gaussian Structural Equation Models (SEMs), where the variables X=(X1,,Xd)X=(X_{1},\dots,X_{d}) follow linear relationships represented by a DAG. The model is expressed as X=BX+NX=B^{\top}X+N. Here, Bd×dB\in\mathbb{R}^{d\times d} is the weighted adjacency matrix encoding the relationships between variables. Specifically, an entry Bij0B_{ij}\neq 0 indicates a directed edge from XjX_{j} to XiX_{i}. The noise vector N=(N1,,Nd)N=(N_{1},\dots,N_{d}) consists of independent noise terms, each corresponding to a variable XiX_{i}. The noise terms are assumed to follow a normal distribution with diagonal covariance matrix Ω=diag(σ12,,σd2)\Omega=\text{diag}(\sigma_{1}^{2},\dots,\sigma_{d}^{2}), where σi2\sigma_{i}^{2} represents the variance of NiN_{i}. Given a DAG, a moral graph is an undirected graph obtained by removing the directions of the edges in the DAG and connecting all pairs of parents of common children.

Unlike many existing methods that assume equal noise variances (Zheng et al., 2018; Yu et al., 2019; Zhang et al., 2022; Bello et al., 2022), in this study, we focus on the general non-equal noise variance (NV) case, where the variances σ12,,σd2\sigma_{1}^{2},\dots,\sigma_{d}^{2} are not assumed to be equal. Our goal is to estimate the DAG GG or its Markov equivalence class (MEC) from the data matrix 𝐗n×d\mathbf{X}\in\mathbb{R}^{n\times d}, consisting of nn i.i.d. samples drawn from the distribution P(X)P(X).

2.2 Hard and Soft DAG Constraint

In the context of DAG learning, the DAG constraint, denoted as h(B)h(B), ensures that the learned structure is a DAG when h(B)=0h(B)=0. NOTEARS (Zheng et al., 2018) employs a hard DAG constraint, whereas GOLEM (Ng et al., 2020) introduces a soft DAG constraint.

NOTEARS and hard DAG constraint

NOTEARS solves the following constrained optimization problem

minBd×d(B;𝐗):=12n𝐗𝐗BF2+λB1subject toh(B)=0.\min_{B\in\mathbb{R}^{d\times d}}\ell(B;\mathbf{X}):=\frac{1}{2n}\|\mathbf{X}-\mathbf{X}B\|_{F}^{2}+\lambda\|B\|_{1}\quad\text{subject to}\quad h(B)=0.

Here, (B;X)\ell(B;X) is the least squares loss with 1\ell_{1} penalty, and h(B)=0h(B)=0 enforces the hard DAG constraint. The constrained optimization problem can be solved using standard algorithms such as augmented Lagrangian method (Wright, 2006), quadratic penalty method (Ng et al., 2022a), and barrier method (Bello et al., 2022).

GOLEM and soft DAG constraint

The GOLEM framework aims to maximize the likelihood of the observed data under the assumption of a linear Gaussian model. There are two formulations in GOLEM, one assuming equal noise variance across variables (GOLEM-EV), and the other allowing for non-equal noise variance (GOLEM-NV).

Unlike NOTEARS, GOLEM adopts the soft DAG constraint, making the problem unconstrained. In other words, GOLEM incorporates h(B)h(B) as an additional penalty term in the score function, controlled by the hyperparameter λ2\lambda_{2}. Here, we only review the the non-equal noise variance formulation, GOLEM-NV, which is the focus of this paper. Assuming that 𝐗\mathbf{X} is centered and that the sample covariance matrix Σ=1n𝐗𝐗\Sigma=\frac{1}{n}\mathbf{X}^{\top}\mathbf{X}, GOLEM-NV’s unconstrained optimization problem is

minBd×d(B;Σ)+λ1B1+λ2h(B),\displaystyle\min_{B\in\mathbb{R}^{d\times d}}\mathcal{L}(B;\Sigma)+\lambda_{1}\|B\|_{1}+\lambda_{2}h(B),
where (B;Σ)=12i=1dlog(((IB)Σ(IB))i,i)log|det(IB)|.\displaystyle\mathcal{L}(B;\Sigma)=\frac{1}{2}\sum_{i=1}^{d}\log\left(\left((I-B)^{\top}\Sigma(I-B)\right)_{i,i}\right)-\log\left|\det(I-B)\right|.

Here, (B;Σ)\mathcal{L}(B;\Sigma) is the likelihood function of linear Gaussian directed graphical models. This allows us to use BB and Σ\Sigma to express GOLEM-NV’s formulation.

3 Inconsistency of 1\ell_{1} penalty in Structure Learning

In this section, we explore the inconsistency of the 1\ell_{1} penalty in structure learning by comparing its behavior with 0\ell_{0} in linear Gaussian cases. We demonstrate this inconsistency through extensive experiments and conclude with a counterexample that highlights the issue.

3.1 0\ell_{0} vs 1\ell_{1} penalty in Linear Gaussian Cases

In structure learning for linear Gaussian cases, score-based methods, specifically with the BIC score (Schwarz, 1978; Chickering, 2002), often aim to recover the sparsest underlying DAG that best explains the observed data (Singh & Moore, 2005; Cussens, 2011; Yuan & Malone, 2013; Chickering, 2002). In the asymptotic case where the population covariance matrix, denoted by Σ\Sigma^{*}, is available, this can, loosely speaking, be formulated as

minB,ΩB0subject to(IB)Ω(IB)1=ΣandB is a DAG.\min_{B,\Omega}\;\|B\|_{0}\quad\text{subject to}\quad(I-B)^{-\top}\Omega(I-B)^{-1}=\Sigma^{*}\quad\text{and}\quad B\text{ is a DAG}.

Recall that BB is the weighted adjacency matrix representing the structure of the DAG and Ω\Omega is the diagonal matrix of noise variances. That is, the goal is to minimize the 0\ell_{0} norm of BB, i.e., the number of edges, while maximizing the likelihood by satisfying the covariance constraint and ensuring that BB is a valid DAG. In other words, the objective is to recover the sparsest DAG, B^\hat{B}, along with its corresponding Ω^\hat{\Omega}, that can generate the observed covariance matrix Σ\Sigma^{*} (i.e., (IB^)Ω^(IB^)1=Σ(I-\hat{B})^{-\top}\hat{\Omega}(I-\hat{B})^{-1}=\Sigma^{*}). Under the sparsest Markov faithfulness assumption (Raskutti & Uhler, 2018), the estimated B^\hat{B} will be Markov equivalent to the true graph BB^{*}.

Many previous work, including GOLEM, replace the 0\ell_{0} penalty with the more tractable 1\ell_{1} penalty. The corresponding optimization problem becomes

minB,ΩB1subject to(IB)Ω(IB)1=ΣandB is a DAG.\min_{B,\Omega}\;\|B\|_{1}\quad\text{subject to}\quad(I-B)^{-\top}\Omega(I-B)^{-1}=\Sigma^{*}\quad\text{and}\quad B\text{ is a DAG}. (1)

The 1\ell_{1} penalty encourages smaller edge weights but introduces a key inconsistency: it doesn’t guarantee true sparsity in the resulting structure. Minimizing the 1\ell_{1} norm may lead to a denser structure with more edges than the solution from minimizing the 0\ell_{0} norm. This occurs because 1\ell_{1} favors edges with small absolute values, even if they represent spurious edges. In some cases, the sum of the absolute values of more edges can be smaller than that of fewer, larger-magnitude edges, which leads 1\ell_{1}-based methods to include unnecessary edges. As a result, the learned structure may deviate from the true DAG or its Markov equivalence class. Taking GOLEM as an example, even if the covariance constraint (IB)Ω(IB)1=Σ(I-B)^{-\top}\Omega(I-B)^{-1}=\Sigma^{*} is satisfied in both 0\ell_{0}- and 1\ell_{1}-based formulations, the structural properties of the solution can differ significantly.

In Section 3.2, we demonstrate this with a large number of experiments, showing that a large proportion of DAGs B~\tilde{B} satisfying the covariance constraint (IB~)Ω(IB~)1=Σ(I-\tilde{B})^{-\top}\Omega(I-\tilde{B})^{-1}=\Sigma^{*} have smaller 1\ell_{1} norms than the true graph BB^{*}. Moreover, in these DAGs satisfying the covariance constraint and having smaller 1\ell_{1} norms than the true graph BB^{*}, we can always find ones with much larger 0\ell_{0} values than BB^{*}, meaning they do not correspond to the true graph BB^{*} or its Markov equivalence class, with the structural Hamming distance (SHD) computed over the true and estimated CPDAG (Completed Partially Directed Acyclic Graph), which we refer to as SHD of CPDAG in this paper, far from zero.

3.2 Experiment: Assessing the Inconsistency of 1\ell_{1} penalty

In this section, we demonstrate and evaluate the inconsistency of the 1\ell_{1} penalty in likelihood-based GOLEM through experiments. We generated 1,000 true DAGs BB^{*}, compute their corresponding covariance matrices Σ\Sigma^{*} under infinite sample conditions. For each BB^{*}, we use Cholesky decomposition to generate large number of DAGs B~\tilde{B} that can generate the same Σ\Sigma^{*}. Our goal is to identify the B~\tilde{B} with the minimum 1\ell_{1} norm among these B~\tilde{B}s, denoted as B1{B}_{\ell_{1}}, and compare it with BB^{*} in terms of 1\ell_{1} norm, 0\ell_{0} norm (i.e., edge count), and record its SHD of CPDAG. Additionally, we also record the proportion of B~\tilde{B}s that satisfy the covariance constraint (i.e., generate the same Σ\Sigma^{*}) but have a smaller 1\ell_{1} norm than BB^{*}, to give an intuitive sense of the extent of 1\ell_{1} inconsistency.

True DAG generation and covariance matrix computation

We generate 1000 8-node (d=8d=8) ER1 graphs BB^{*}s . The data is generated with a fixed noise ratio of 16, where the variances of two randomly selected noise variables are set to 1 and 16, respectively. The variances of the remaining noise variables are sampled uniformly from the range [1,16][1,16]. The edge weights are uniformly sampled from [2,0.5][0.5,2][-2,-0.5]\cup[0.5,2]. For each BB^{*}, we compute the corresponding population covariance matrix Σ\Sigma^{*} under infinite samples using the equation Σ=(IB)Ω(IB)1\Sigma^{*}=(I-B^{*})^{-\top}\Omega^{*}(I-B^{*})^{-1}.

Generating DAGs which meet covariance constraint

Following the sparsest permutation approach developed by Raskutti & Uhler (2018), for each true covariance matrix Σ\Sigma^{*}, we generate all possible d!d! permutations of its rows and columns. For each permuted covariance matrix, we apply Cholesky decomposition to find a DAG B~\tilde{B} that generates the permuted covariance matrix. After that, we restore B~\tilde{B} to the original variable order. This ensures that all B~\tilde{B} satisfies the covariance constraint (IB~)Ω~(IB~)1=Σ(I-\tilde{B})^{-\top}\tilde{\Omega}(I-\tilde{B})^{-1}=\Sigma^{*}, while remaining a valid DAG. After the above steps, for each of the 1,000 true BB^{*}, we identified d!d! different DAGs B~\tilde{B} that all generate Σ\Sigma^{*}.

Metrics and analysis

For each true DAG BB^{*}, we analyze the following metrics across all d!d! DAGs B~\tilde{B} that satisfy the covariance constraint: (1) 1\ell_{1} norm comparison: we calculate the 1\ell_{1} norm of each B~\tilde{B} and record the proportion of B~\tilde{B}s whose 1\ell_{1} norm is smaller than that of BB^{*}; (2) selecting B~\tilde{B} with the minimum 1\ell_{1} norm: among the d!d! B~\tilde{B}s, we select the one with the smallest 1\ell_{1} norm, denoted as B1{B}_{\ell_{1}}. We then compare B1{B}_{\ell_{1}} with BB^{*} in terms of 1\ell_{1}, and record its edge count and SHD of CPDAG (to test its distance to BB^{*} or its Markov equivalence class).

Experimental results

After running experiments for 1000 BB^{*}s, we summarized the result in Table 1. Table 1 shows a comparison between the true DAG BB^{*} and the B1{B}_{\ell_{1}} that generate the same covariance matrix Σ\Sigma^{*}. On average, for each true DAG BB^{*}, 77.86% of the d!d! DAGs B~\tilde{B} satisfying the covariance constraint have smaller 1\ell_{1} norms than BB^{*}. In the 1,000 runnings, the average 1\ell_{1} norm of B1{B}_{\ell_{1}} is 4.22, which is smaller than the average 1\ell_{1} norm of BB^{*}, which is 10.04. The average 0\ell_{0} norm (number of edges) of B1{B}_{\ell_{1}} is 22.74, which is larger than the 0\ell_{0} norm (number of edges) of BB^{*}, which is 8. The average SHD of CPDAG between B1{B}_{\ell_{1}} and BB^{*} is 19.97. In addition, in each running, B1{B}_{\ell_{1}} consistently has a smaller 1\ell_{1} norm than BB^{*}, a larger 0\ell_{0} norm (number of edges), and a SHD of CPDAG greater than zero, indicating that B1{B}_{\ell_{1}} is structurally different from BB^{*} and its Markov equivalence class. These results demonstrate two key points: (1) a significant proportion of DAGs B~\tilde{B} that satisfy the covariance constraint have smaller 1\ell_{1} norms than BB^{*}, and (2) in each running, we can find a counterexample (i.e., the B1{B}_{\ell_{1}}) where the 1\ell_{1} norm is smaller, but the 0\ell_{0} norm is larger, and the resulting DAG is not equivalent to BB^{*} or its Markov equivalence class. This supports the conclusion that 1\ell_{1}-based solutions are inconsistent in recovering the true structure.

Table 1: Comparison of B~\tilde{B}s which generate Σ\Sigma^{*} with True DAG BB^{*}. The results are averaged over 1,000 simulated BB^{*}s. The ”Proportion” column reflects the average percentage of DAGs B~\tilde{B} with 1\ell_{1} norms smaller than that of BB^{*} among d!d! B~\tilde{B}s per BB^{*}.
Metric BB^{*} B1{B}_{\ell_{1}} Proportion of B~\tilde{B} with B~1<B1\|\tilde{B}\|_{1}<\|B^{*}\|_{1}
Average 1\ell_{1} norm 10.04 4.22 77.86%
Average 0\ell_{0} norm (Number of Edges) 8.0 22.74
Average SHD of CPDAG 0 19.97

3.3 Case Study: A Specific Counterexample

In this section, we present a 3-node counterexample to demonstrate the inconsistency of the 1\ell_{1} penalty in differentiable structural learning. Specifically, we compare a true weighted adjacency matrix BB^{*} with an estimated adjacency matrix B~\tilde{B}, and show that although both matrices can generate the same covariance matrix, their 0\ell_{0} norm (edge count), 1\ell_{1} norm, and structural differences, measured by SHD of CPDAG, reveal the inconsistency of the 1\ell_{1} penalty.

The true adjacency matrix BB^{*} and its corresponding noise covariance matrices Ω\Omega^{*} are given as:

B=[0120000010],Ω=[1600040001].B^{*}=\begin{bmatrix}0&\frac{1}{2}&0\\ 0&0&0\\ 0&-1&0\end{bmatrix},\quad\Omega^{*}=\begin{bmatrix}16&0&0\\ 0&4&0\\ 0&0&1\end{bmatrix}.

The estimated adjacency matrix B~\tilde{B} and its corresponding noise covariance matrices Ω~\tilde{\Omega} are:

B~=[0121100015000],Ω~=[16000500045].\tilde{B}=\begin{bmatrix}0&\frac{1}{2}&\frac{1}{10}\\ 0&0&-\frac{1}{5}\\ 0&0&0\end{bmatrix},\quad\tilde{\Omega}=\begin{bmatrix}16&0&0\\ 0&5&0\\ 0&0&\frac{4}{5}\end{bmatrix}.

Both matrices BB^{*} and B~\tilde{B}, along with their respective noise covariance matrices, generate the same covariance matrix:

Σ=[1680891011].\Sigma^{*}=\begin{bmatrix}16&8&0\\ 8&9&-1\\ 0&-1&1\end{bmatrix}.

We have B0=2\|B^{*}\|_{0}=2 and and B~0=3\|\tilde{B}\|_{0}=3, indicating that BB^{*} is sparser than B~\tilde{B}.

However, when considering the 1\ell_{1} norm, we observe that: B1=32andB~1=45.\|B^{*}\|_{1}=\frac{3}{2}\quad\text{and}\quad\|\tilde{B}\|_{1}=\frac{4}{5}. That is, although B~\tilde{B} has a higher 0\ell_{0} norm, it achieves a lower 1\ell_{1} norm, highlighting the inconsistency between the two norms. Therefore, the optimization problem in Eq. equation 1 may return B~\tilde{B}, which is clearly not Markov equivalent to BB^{*}.

This counterexample demonstrates the inconsistency of the 1\ell_{1} penalty: it may lead to solutions with smaller total edge weights (resulting in a lower 1\ell_{1} norm), but these solutions may still have more edges (a higher 0\ell_{0} norm) and deviate from the true DAG structure and its Markov equivalence class, even if these solutions can generate the same covariance matrix as the ground truth DAG.

4 Continuous and Acyclicity-constrained 0\ell_{0}-penalized likelihood with estimated Moral graph

In Section 2.2, we reviewed the GOLEM-NV formulation proposed by Ng et al. (2020), which aims to maximize the data likelihood utilizes an 1\ell_{1} penalty and soft DAG constraint. We refer to this original model as GOLEM-NV-1\ell_{1} throughout this paper. The problem formulation can be expressed as

minBd×d(B;Σ)+λ1B1+λ2h(B).\min_{B\in\mathbb{R}^{d\times d}}\mathcal{L}(B;\Sigma)+\lambda_{1}\|B\|_{1}+\lambda_{2}h(B).

However, as pointed out by Ng et al. (2024), GOLEM-NV-1\ell_{1} suffers from significant non-convexity, often leading to suboptimal local minima with poor performance, both before and after data standardization. Moreover, as we demonstrated in section 3, the 1\ell_{1} penalty leads to inconsistent solutions. To address these limitations, we propose CALM (Continuous and Acyclicity-constrained L0-penalized likelihood with estimated Moral graph), a differentiable structure learning method that optimizes an 0\ell_{0}-penalized likelihood with hard DAG constraints and incorporates moral graphs. Our experiments demonstrate that CALM significantly improves performance compared to the original GOLEM-NV-1\ell_{1}, yielding results much closer to the true DAG or its Markov equivalence class.

In the following subsections, we will first introduce the implementation of CALM, followed by the experimental setup and an evaluation of CALM’s performance under various configurations.

4.1 Introduction to CALM

0\ell_{0} penalty and its approximation with Gumbel Softmax

CALM begins with applying an 0\ell_{0} penalty to regularize the likelihood, enforcing sparsity in the learned adjacency matrix. Inspired by Ng et al. (2022b); Kalainathan et al. (2022), we use Gumbel Softmax as an example to show how we achieve the approximation of 0\ell_{0} penalty in our approach, as it proved to be the most effective and robust 0\ell_{0} approximation among those we experimented with in section 4.5. When using Gumbel Softmax (Jang et al., 2017) to approximate 0\ell_{0} penalty, CALM starts by representing the learned adjacency matrix BB as an element-wise product of a learned mask, gτ(U)d×dg_{\tau}(U)\in\mathbb{R}^{d\times d}, which determines the presence of edges, and a learned parameter matrix, Pd×dP\in\mathbb{R}^{d\times d}, which learns the weights of the edges. The mask gτ(U)g_{\tau}(U) is generated using the Gumbel-Softmax approach. Here, Ui,jU_{i,j} represents the logits, and a logistic noise Gi,jLogistic(0,1)G_{i,j}\sim\text{Logistic}(0,1) is added to Ui,jU_{i,j}, producing gτ(Ui,j)=σ((Ui,j+Gi,j)/τ)g_{\tau}(U_{i,j})=\sigma((U_{i,j}+G_{i,j})/\tau), where τ\tau is the temperature that controls the smoothness of the Softmax, and σ()\sigma(\cdot) is the logistic sigmoid function. As the optimization process proceeds, the values of gτ(Ui,j)g_{\tau}(U_{i,j}) approach either 0 or 1, approximating an 0\ell_{0} penalty.

Incorporating the moral graph and hard DAG constraints

Furthermore, CALM incorporates a learned moral graph M{0,1}d×dM\in\{0,1\}^{d\times d} to restrict the optimization to edges within the moral graph, thus reducing the search space. Note that similar idea has been used in various existing works such as Loh & Bühlmann (2014); Nazaret et al. (2024). This moral graph acts as a filter over the Gumbel-Softmax mask, allowing only edges present in the moral graph. The final learned BB can be represented as B=Mgτ(U)PB=M\circ g_{\tau}(U)\circ P, incorporating both the sparsity from the Gumbel-Softmax mask and the structural constraints from the moral graph. Here, BB contains the edge weights for the DAG, and its structure is determined by the mask Mgτ(U)M\circ g_{\tau}(U).

CALM’s objective function, incorporating the Gumbel-Softmax mask, moral graph, and hard DAG constraints into the GOLEM-NV-1\ell_{1} formulation, is given by

minUd×d,Pd×d(Mgτ(U)P;Σ)+λ1Mgτ(U)1subject toh(Mgτ(U))=0.\min_{U\in\mathbb{R}^{d\times d},P\in\mathbb{R}^{d\times d}}\mathcal{L}(M\circ g_{\tau}(U)\circ P;\Sigma)+\lambda_{1}\|M\circ g_{\tau}(U)\|_{1}\quad\text{subject to}\quad h(M\circ g_{\tau}(U))=0.

where (Mgτ(U)P;Σ)\mathcal{L}(M\circ g_{\tau}(U)\circ P;\Sigma) is the likelihood term, and the λ1Mgτ(U)1\lambda_{1}\|M\circ g_{\tau}(U)\|_{1} term approximates the 0\ell_{0} penalty for sparsity. Here, both the 0\ell_{0} penalty for sparsity and the DAG constraints are applied to the final learned mask Mgτ(U)M\circ g_{\tau}(U), which determines the presence of edges.

4.2 EXPERIMENTAL SETUP

Across all experiments in section 4, we simulate Erdös–Rényi graphs (ERDdS & R&wi, 1959) with kdkd edges, denoted as ERk graphs, with edge weights uniformly sampled from [2,0.5][0.5,2][-2,-0.5]\cup[0.5,2]. For all experiments, the data is generated with a fixed noise ratio of 16. Specifically, the variances of two randomly selected noise variables are set to 1 and 16, respectively, while the variances of the remaining noise variables are sampled uniformly from the range [1,16][1,16]. This setting ensures a realistic variation in noise across the variables, aligning with the assumptions of non-equal noise variances (NV). Further implementation details of our experiments in section 4 are in Appendix A. From this point forward, unless otherwise specified, we use ”CALM” to refers to the specific version of the method where the 0\ell_{0}-penalty is approximated using Gumbel Softmax.

The following four sub-sections evaluate CALM’s performance: first, we compare soft vs. hard DAG constraints and moral graph vs. no moral graph; second, we assess the effect of data standardization; third, we test different 0\ell_{0} approximation methods. Finally, we compare CALM with baseline approaches. For each scenario, we conducted 10 experiments and calculated the mean SHD of CPDAG, precision of skeleton, recall of skeleton, and their standard errors.

4.3 Impact of Moral Graph and Soft/Hard DAG Constraints

Recall that NOTEARS (Zheng et al., 2018) adopts a hard DAG constraint while GOLEM (Ng et al., 2020) uses a soft DAG constraint. Here, we evaluate the impact of incorporating the moral graph and using either soft or hard DAG constraints on the results of the 0\ell_{0}-penalized likelihood optimization. We consider linear Gaussian models with 50 variables and ER1 graphs. Here, we focus on the nonconvexity aspect of the optimization problem, and thus set the sample size to infinity to eliminate finite sample errors (the way we achieve infinite samples is in Appendix A.4). The experiment results for finite samples are included in Section 4.6. Furthermore, following Reisach et al. (2021); Kaiser & Sipos (2022), we standardize the data. All experiments were conducted with the Gumbel-Softmax approach to approximate 0\ell_{0} penalty. The implementation details of Gumbel-Softmax-based 0\ell_{0} penalty and the hard DAG constraints is in Appendix A.2. The implementation details of soft DAG constraints is in Appendix A.3.

Table 2: Impact of moral graph and soft/hard DAG constraints for 50-node ER1 graphs under data standardization
SHD of CPDAG Precision of Skeleton Recall of Skeleton
Soft Constraints Without Moral 33.8 ± 2.7 0.98 ± 0.01 0.43 ± 0.05
Soft Constraints With Moral 7.6 ± 2.5 0.98 ± 0.01 0.95 ± 0.03
Hard Constraints Without Moral 16.7 ± 3.4 0.88 ± 0.03 0.97 ± 0.01
Hard Constraints With Moral (CALM) 5.5 ± 1.9 0.98 ± 0.01 0.99 ± 0.00

Comparison of soft and hard DAG constraints

From the results in Table 2, we observe that using hard DAG constraints leads to a lower SHD of CPDAG compared to soft DAG constraints, regardless of whether the moral graph is incorporated. This suggests that, even when both formulation with soft and hard DAG constraints converge to local optima, the hard DAG constraint results are closer to the true adjacency matrix BB or its Markov equivalence class.

One explanation for the improved performance of hard DAG constraints is the use of a quadratic penalty method (QPM) (Ng et al., 2022a). In this framework, the hyperparameter ρ\rho, which controls the weight of the DAG constraint, is progressively increased during optimization, with each ρ\rho value triggering a full subproblem optimization. This leads to a more refined optimization process. In contrast, the soft DAG constraint uses a fixed ρ\rho, resulting in only one subiteration and possibly worse convergence. Additionally, hard constraints ensure that the final graph is always a DAG, eliminating the need for post-processing, whereas soft constraints often require post-processing to enforce acyclicity (Ng et al., 2020), which may introduce errors and increase SHD of CPDAG.

Impact of including the moral graph.

Table 2 also shows that incorporating the moral graph improves performance in both soft and hard DAG constraint settings, with a notably lower SHD of CPDAG. The moral graph reduces the search space by focusing on edges within it, which is especially beneficial in higher-dimensional settings like our 50-node experiments, where the reduction in search space is more substantial. This significantly simplifies the optimization process and leads to better convergence towards the true adjacency matrix or its Markov equivalence class.

In summary, CALM, combining hard DAG constraints and the moral graph, delivers the best results.

4.4 Impact of Data Standardization

Ng et al. (2024) previously pointed out that the original GOLEM-NV-1\ell_{1} formulation performed poorly both before and after data standardization. To further evaluate the robustness of CALM, we conduct experiments to compare its performance with (CALM-Standardized) and without data standardization (CALM-Non-Standardized). Here, we use infinite samples to eliminate finite sample error and consider a 50-node linear Gaussian model with ER1 graphs. The implementation details of Gumbel-Softmax-based 0\ell_{0} penalty and the hard DAG constraints for CALM is in Appendix A.2. In Table 3, CALM shows consistently low SHD of CPDAG before and after data standardization, demonstrating its stability and robustness across both standardized and non-standardized data. Interestingly, this is in constrast with the observation by Reisach et al. (2021); Kaiser & Sipos (2022) that differentiable structure learning methods do not perform well after data standardization, which further validate the robustness of our method.

Table 3: Impact of Data Standardization on CALM for 50-node ER1 graphs
SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM-Non-Standardized 9.9 ± 3.4 0.95 ± 0.02 0.99 ± 0.01
CALM-Standardized 5.5 ± 1.9 0.98 ± 0.01 0.99 ± 0.00

4.5 Comparison of 0\ell_{0} Approximation Methods

We compare our method with three 0\ell_{0} approximations (CALM, CALM-STG, CALM-Tanh) and the original GOLEM-NV-1\ell_{1}. For all methods here, We used infinite samples to eliminate finite sample error and considered 50- node linear Gaussian model with ER1 graphs. Additional results for ER4 with 50 nodes and ER1 with 100 nodes are presented in Appendix B. Here, CALM maintains our definition, specifically referring to our method that employs the Gumbel-Softmax approximation for the 0\ell_{0} penalty. CALM-STG (Stochastic Gates) refers to our method that uses stochastic gates to approximate 0\ell_{0} penalty (Yamada et al., 2020), while CALM-Tanh (Hyperbolic Tangent) refers to our method that employs the smooth hyperbolic tangent function to approximate 0\ell_{0} penalty (Bhattacharya et al., 2021). The key parameter selection and implementation details for STG and Tanh in approximating the 0\ell_{0} penalty can be found in Appendix A.5.

Table 4: Comparison of our method using different 0\ell_{0} approximations and original GOLEM-NV-1\ell_{1} for 50-node ER1 graphs under both data standardization and no data standardization.
Standardized Non-Standardized
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 5.5 ± 1.9 0.98 ± 0.01 0.99 ± 0.00 9.9 ± 3.4 0.95 ± 0.02 0.99 ± 0.01
CALM-STG 5.9 ± 1.5 0.97 ± 0.01 0.99 ± 0.00 34.1 ± 3.6 0.79 ± 0.02 0.95 ± 0.01
CALM-Tanh 8.6 ± 1.6 0.95 ± 0.01 0.99 ± 0.01 51.7 ± 2.5 0.69 ± 0.02 0.91 ± 0.01
GOLEM-NV-1\ell_{1} 56.2 ± 3.5 0.60 ± 0.03 0.55 ± 0.05 121.1 ± 6.1 0.35 ± 0.02 0.76 ± 0.03

Table 4 shows that CALM-STG achieves competitive results after data standardization but underperforms without standardization. CALM-Tanh shows the weakest performance among the three 0\ell_{0} approximations methods. Only CALM performs well both with and without data standardization, highlighting its robustness, which is why we ultimately selected Gumbel-Softmax as the 0\ell_{0} approximation method as our representative implementation. Additionally, all three methods outperform GOLEM-NV-1\ell_{1}, underscoring the importance of 0\ell_{0} penalty approximation, as 1\ell_{1} penalty suffers from inconsistency (as discussed in Section 3). Moreover, this also shows that the effectiveness of combining 0\ell_{0} approximation with moral graphs and hard constraints.

4.6 Comparison with Baseline Methods

We finally compare the performance of CALM against several baseline methods, including the original GOLEM-NV-1\ell_{1}, NOTEARS, PC (Spirtes & Glymour, 1991), and FGES (Ramsey et al., 2017) (see Appendix A.6 for baseline methods’ implemention details). We evaluated the methods at two different sample sizes: n=1000n=1000 and n=106n=10^{6}. We considered a 50-node linear Gaussian model with ER1 and ER4 graphs, as well as a 100-node linear Gaussian model with ER1 graphs. Here, the moral graph in CALM is estimated from finite samples (see Appendix A.1 for how we estimate the moral graph). Specifically, for the 1000-sample experiments, the moral graph is estimated from 1000 samples, and for the 10610^{6}-sample experiments, the moral graph is estimated from 10610^{6} samples. All results presented here are from experiments with standardized data. The additional experimental results for data without standardization are presented in Appendix C.

Table 5: Comparison with baseline methods for 50-node ER1, 50-node ER4, and 100-node ER1 graphs under data standardization, using 1000 and 10610^{6} samples.
50-node ER1 graphs
1000 Samples 10610^{6} Samples
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 12.1 ± 2.7 0.93 ± 0.01 0.98 ± 0.00 7.0 ± 2.5 0.97 ± 0.01 0.98 ± 0.01
GOLEM-NV-1\ell_{1} 60.0 ± 3.9 0.58 ± 0.03 0.65 ± 0.07 55.6 ± 2.6 0.60 ± 0.02 0.63 ± 0.07
NOTEARS 46.3 ± 1.9 0.76 ± 0.01 0.81 ± 0.02 46.6 ± 1.9 0.75 ± 0.02 0.79 ± 0.02
PC 11.0 ± 1.4 0.98 ± 0.01 0.92 ± 0.01 2.8 ± 0.8 0.99 ± 0.01 0.99 ± 0.00
FGES 8.4 ± 2.4 0.94 ± 0.02 0.98 ± 0.00 1.3 ± 0.9 1.00 ± 0.00 1.00 ± 0.00
50-node ER4 graphs
1000 Samples 10610^{6} Samples
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 168.8 ± 8.3 0.62 ± 0.02 0.67 ± 0.02 139.4 ± 10.2 0.68 ± 0.02 0.75 ± 0.02
GOLEM-NV-1\ell_{1} 211.6 ± 4.2 0.59 ± 0.02 0.22 ± 0.02 211.0 ± 4.5 0.58 ± 0.03 0.22 ± 0.02
NOTEARS 209.5 ± 1.2 0.66 ± 0.02 0.15 ± 0.01 209.1 ± 1.3 0.65 ± 0.02 0.15 ± 0.01
PC 200.5 ± 2.1 0.61 ± 0.02 0.22 ± 0.01 231.0 ± 4.5 0.47 ± 0.01 0.33 ± 0.01
FGES 425.6 ± 23.2 0.33 ± 0.02 0.80 ± 0.01 750.0 ± 48.4 0.23 ± 0.02 1.00 ± 0.00
100-node ER1 graphs
1000 Samples 10610^{6} Samples
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 26.7 ± 3.6 0.90 ± 0.01 0.99 ± 0.00 17.0 ± 2.9 0.96 ± 0.01 0.99 ± 0.00
GOLEM-NV-1\ell_{1} 120.5 ± 6.7 0.55 ± 0.02 0.75 ± 0.06 115.7 ± 4.2 0.57 ± 0.02 0.68 ± 0.07
NOTEARS 87.4 ± 2.9 0.76 ± 0.01 0.74 ± 0.03 86.3 ± 2.7 0.75 ± 0.01 0.78 ± 0.03
PC 24.7 ± 1.6 0.95 ± 0.01 0.89 ± 0.01 4.2 ± 0.8 0.97 ± 0.01 1.00 ± 0.00
FGES 12.1 ± 1.7 0.95 ± 0.01 0.98 ± 0.00 1.0 ± 0.8 1.00 ± 0.00 1.00 ± 0.00

Table 5 summarizes the performance comparison between CALM and the baseline methods. The results clearly demonstrate that CALM consistently outperforms both NOTEARS and the original GOLEM-NV-1\ell_{1} across all graph structures and sample sizes. This highlights the effectiveness and robustness of incorporating the Gumbel-Softmax approximation to 0\ell_{0}, moral graph, and hard DAG constraints. It is observed that the results of CALM are not as competitive as those obtained by the PC and FGES methods for sparse graphs such as ER1 graphs. This outcome is expected, given that the continuous optimization in linear likelihood-based formulation struggles with such high levels of nonconvexity.

However, it is worth noting that for ER1 graphs, in the case of 1000 samples, the results of CALM are nearly identical to those of PC. This indicates that in practical scenarios with relatively small sample sizes, even in sparse graphs, CALM is able to compete with the performance of discrete methods like PC. This represents a significant breakthrough.

Furthermore, in more dense graphs, the 50-node ER4 graphs, CALM demonstrates superior performance compared to the PC and FGES methods. This result suggests that in higher-density graphs, CALM enables continuous optimization methods to outperform discrete methods.

5 Conclusion

Our work begins by identifying the inconsistency of 1\ell_{1}-penalized likelihood in differentiable structure learning for the linear Gaussian case. To address this and improve performance, we propose CALM, which optimizes an 0\ell_{0}-penalized likelihood with hard acyclicity constraints and incorporates moral graphs. Our results show that CALM, particularly with Gumbel-Softmax 0\ell_{0} approximation, significantly outperforms GOLEM-NV-1\ell_{1} and NOTEARS across various graph types and sample sizes. In sparse graphs like ER1, CALM’s performance rivals PC with smaller samples, while in dense graphs like ER4, it achieves the best results among all baseline methods. CALM also maintains robust performance both before and after data standardization. Future work includes extending CALM to nonlinear models and integrating advanced optimization techniques for further improvements in linear models.

Reproducibility Statement

Our code will be released publicly upon acceptance of this paper. The implementation details and parameter settings of the CALM-related experiments and baseline methods are mentioned in Section 4 and Appendix A. The details for the experiments proving the inconsistency of 1\ell_{1} penalty are provided in Section 3.2.

References

  • Awokuse & Bessler (2003) Titus O Awokuse and David A Bessler. Vector autoregressions, policy analysis, and directed acyclic graphs: an application to the us economy. Journal of Applied Economics, 6(1):1–24, 2003.
  • Bello et al. (2022) Kevin Bello, Bryon Aragam, and Pradeep Ravikumar. Dagma: Learning dags via m-matrices and a log-determinant acyclicity characterization. Advances in Neural Information Processing Systems, 35:8226–8239, 2022.
  • Bellot & van der Schaar (2021) Alexis Bellot and Mihaela van der Schaar. Deconfounded score method: Scoring dags with dense unobserved confounding. arXiv preprint arXiv:2103.15106, 2021.
  • Bhattacharya et al. (2021) Rohit Bhattacharya, Tushar Nagarajan, Daniel Malinsky, and Ilya Shpitser. Differentiable causal discovery under unmeasured confounding. In International Conference on Artificial Intelligence and Statistics, pp.  2314–2322. PMLR, 2021.
  • Brouillard et al. (2020) Philippe Brouillard, Sébastien Lachapelle, Alexandre Lacoste, Simon Lacoste-Julien, and Alexandre Drouin. Differentiable causal discovery from interventional data. Advances in Neural Information Processing Systems, 33:21865–21877, 2020.
  • Chen et al. (2021) Xinshi Chen, Haoran Sun, Caleb Ellington, Eric Xing, and Le Song. Multi-task learning of order-consistent causal graphs. Advances in Neural Information Processing Systems, 34:11083–11095, 2021.
  • Chickering (1996) David Maxwell Chickering. Learning bayesian networks is np-complete. Learning from data: Artificial intelligence and statistics V, pp.  121–130, 1996.
  • Chickering (2002) David Maxwell Chickering. Optimal structure identification with greedy search. Journal of machine learning research, 3(Nov):507–554, 2002.
  • Chickering et al. (2004) Max Chickering, David Heckerman, and Chris Meek. Large-sample learning of bayesian networks is np-hard. Journal of Machine Learning Research, 5:1287–1330, 2004.
  • Cussens (2011) James Cussens. Bayesian network learning with cutting planes. In 27th Conference on Uncertainty in Artificial Intelligence, pp.  153–160. AUAI Press, 2011.
  • Deng et al. (2023) Chang Deng, Kevin Bello, Pradeep Ravikumar, and Bryon Aragam. Global optimality in bivariate gradient-based dag learning. Advances in Neural Information Processing Systems, 36:17929–17968, 2023.
  • ERDdS & R&wi (1959) P ERDdS and A R&wi. On random graphs i. Publ. math. debrecen, 6(290-297):18, 1959.
  • Faria et al. (2022) Gonçalo Rui Alves Faria, Andre Martins, and Mário AT Figueiredo. Differentiable causal discovery under latent interventions. In Conference on Causal Learning and Reasoning, pp.  253–274. PMLR, 2022.
  • Gao et al. (2021) Erdun Gao, Junjia Chen, Li Shen, Tongliang Liu, Mingming Gong, and Howard Bondell. Feddag: Federated dag structure learning. arXiv preprint arXiv:2112.03555, 2021.
  • Gao et al. (2022) Erdun Gao, Ignavier Ng, Mingming Gong, Li Shen, Wei Huang, Tongliang Liu, Kun Zhang, and Howard Bondell. Missdag: Causal discovery in the presence of missing data with continuous additive noise models. Advances in Neural Information Processing Systems, 35:5024–5038, 2022.
  • Jang et al. (2017) Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparametrization with gumble-softmax. In International Conference on Learning Representations (ICLR 2017). OpenReview. net, 2017.
  • Kaiser & Sipos (2022) Marcus Kaiser and Maksim Sipos. Unsuitability of NOTEARS for causal graph discovery when dealing with dimensional quantities. Neural Processing Letters, 54:1–9, 06 2022.
  • Kalainathan et al. (2022) Diviyan Kalainathan, Olivier Goudet, Isabelle Guyon, David Lopez-Paz, and Michèle Sebag. Structural agnostic modeling: Adversarial learning of causal graphs. Journal of Machine Learning Research, 23(219):1–62, 2022.
  • Koivisto & Sood (2004) Mikko Koivisto and Kismat Sood. Exact bayesian structure discovery in bayesian networks. The Journal of Machine Learning Research, 5:549–573, 2004.
  • Koller & Friedman (2009) Daphne Koller and Nir Friedman. Probabilistic graphical models: principles and techniques. MIT press, 2009.
  • Lachapelle et al. (2019) Sébastien Lachapelle, Philippe Brouillard, Tristan Deleu, and Simon Lacoste-Julien. Gradient-based neural dag learning. arXiv preprint arXiv:1906.02226, 2019.
  • Loh & Bühlmann (2014) Po-Ling Loh and Peter Bühlmann. High-dimensional learning of linear causal networks via inverse covariance estimation. Journal of Machine Learning Research, 15(88):3065–3105, 2014.
  • Nazaret et al. (2024) Achille Nazaret, Justin Hong, Elham Azizi, and David Blei. Stable differentiable causal discovery. In Proceedings of the 41st International Conference on Machine Learning, 2024.
  • Ng & Zhang (2022) Ignavier Ng and Kun Zhang. Towards federated bayesian network structure learning with continuous optimization. In International Conference on Artificial Intelligence and Statistics, pp.  8095–8111. PMLR, 2022.
  • Ng et al. (2020) Ignavier Ng, AmirEmad Ghassami, and Kun Zhang. On the role of sparsity and dag constraints for learning linear dags. Advances in Neural Information Processing Systems, 33:17943–17954, 2020.
  • Ng et al. (2022a) Ignavier Ng, Sébastien Lachapelle, Nan Rosemary Ke, Simon Lacoste-Julien, and Kun Zhang. On the convergence of continuous constrained optimization for structure learning. In International Conference on Artificial Intelligence and Statistics, pp.  8176–8198. Pmlr, 2022a.
  • Ng et al. (2022b) Ignavier Ng, Shengyu Zhu, Zhuangyan Fang, Haoyang Li, Zhitang Chen, and Jun Wang. Masked gradient-based causal structure learning. In Proceedings of the 2022 SIAM International Conference on Data Mining (SDM), pp.  424–432. SIAM, 2022b.
  • Ng et al. (2024) Ignavier Ng, Biwei Huang, and Kun Zhang. Structure learning with continuous optimization: A sober look and beyond. In Causal Learning and Reasoning, pp.  71–105. PMLR, 2024.
  • Pamfil et al. (2020) Roxana Pamfil, Nisara Sriwattanaworachai, Shaan Desai, Philip Pilgerstorfer, Konstantinos Georgatzis, Paul Beaumont, and Bryon Aragam. Dynotears: Structure learning from time-series data. In International Conference on Artificial Intelligence and Statistics, pp.  1595–1605. Pmlr, 2020.
  • PEARL (1988) J PEARL. Probabilistic reasoning in intelligent systems; network of plausible inference. Morgan Kaufmann, 1988, 1988.
  • Peters & Bühlmann (2014) Jonas Peters and Peter Bühlmann. Identifiability of gaussian structural equation models with equal error variances. Biometrika, 101(1):219–228, 2014.
  • Ramsey et al. (2017) Joseph Ramsey, Madelyn Glymour, Ruben Sanchez-Romero, and Clark Glymour. A million variables and more: the fast greedy equivalence search algorithm for learning high-dimensional graphical causal models, with an application to functional magnetic resonance images. International journal of data science and analytics, 3:121–129, 2017.
  • Raskutti & Uhler (2018) Garvesh Raskutti and Caroline Uhler. Learning directed acyclic graph models based on sparsest permutations. Stat, 7(1):e183, 2018.
  • Reisach et al. (2021) Alexander Reisach, Christof Seiler, and Sebastian Weichwald. Beware of the simulated DAG! causal discovery benchmarks may be easy to game. In Advances in Neural Information Processing Systems, 2021.
  • Saboksayr et al. (2023) Seyed Saman Saboksayr, Gonzalo Mateos, and Mariano Tepper. Colide: Concomitant linear dag estimation. arXiv preprint arXiv:2310.02895, 2023.
  • Scheines et al. (1998) Richard Scheines, Peter Spirtes, Clark Glymour, Christopher Meek, and Thomas Richardson. The tetrad project: Constraint based aids to causal model specification. Multivariate Behavioral Research, 33(1):65–117, 1998.
  • Schwarz (1978) Gideon Schwarz. Estimating the dimension of a model. The annals of statistics, pp.  461–464, 1978.
  • Singh & Moore (2005) Ajit P Singh and Andrew W Moore. Finding optimal Bayesian networks by dynamic programming. Carnegie Mellon University. Center for Automated Learning and Discovery, 2005.
  • Spirtes & Glymour (1991) Peter Spirtes and Clark Glymour. An algorithm for fast recovery of sparse causal graphs. Social science computer review, 9(1):62–72, 1991.
  • Spirtes et al. (2001) Peter Spirtes, Clark Glymour, and Richard Scheines. Causation, prediction, and search. MIT press, 2001.
  • Sun et al. (2021) Xiangyu Sun, Guiliang Liu, Pascal Poupart, and Oliver Schulte. Nts-notears: Learning nonparametric temporal dags with time-series data and prior knowledge. arXiv preprint arXiv:2109.04286, 2021.
  • Tennant et al. (2021) Peter WG Tennant, Eleanor J Murray, Kellyn F Arnold, Laurie Berrie, Matthew P Fox, Sarah C Gadd, Wendy J Harrison, Claire Keeble, Lynsie R Ranker, Johannes Textor, et al. Use of directed acyclic graphs (dags) to identify confounders in applied health research: review and recommendations. International journal of epidemiology, 50(2):620–632, 2021.
  • Tsamardinos et al. (2003) Ioannis Tsamardinos, Constantin F Aliferis, Alexander R Statnikov, and Er Statnikov. Algorithms for large scale markov blanket discovery. In FLAIRS, volume 2, pp.  376–81, 2003.
  • Wang et al. (2020) Yuhao Wang, Vlado Menkovski, Hao Wang, Xin Du, and Mykola Pechenizkiy. Causal discovery from incomplete data: a deep learning approach. arXiv preprint arXiv:2001.05343, 2020.
  • Wang et al. (2022) Zhenlei Wang, Xu Chen, Zhenhua Dong, Quanyu Dai, and Ji-Rong Wen. Sequential recommendation with causal behavior discovery. arXiv preprint arXiv:2204.00216, 2022.
  • Wei et al. (2020) Dennis Wei, Tian Gao, and Yue Yu. Dags with no fears: A closer look at continuous optimization for learning bayesian networks. Advances in Neural Information Processing Systems, 33:3895–3906, 2020.
  • Wright (2006) Stephen J Wright. Numerical optimization, 2006.
  • Yamada et al. (2020) Yutaro Yamada, Ofir Lindenbaum, Sahand Negahban, and Yuval Kluger. Feature selection using stochastic gates. In International conference on machine learning, pp.  10648–10659. PMLR, 2020.
  • Yang et al. (2021) Shuai Yang, Kui Yu, Fuyuan Cao, Lin Liu, Hao Wang, and Jiuyong Li. Learning causal representations for robust domain adaptation. IEEE Transactions on Knowledge and Data Engineering, 2021.
  • Yu et al. (2019) Yue Yu, Jie Chen, Tian Gao, and Mo Yu. Dag-gnn: Dag structure learning with graph neural networks. In International conference on machine learning, pp.  7154–7163. PMLR, 2019.
  • Yuan & Malone (2013) Changhe Yuan and Brandon Malone. Learning optimal bayesian networks: A shortest path perspective. Journal of Artificial Intelligence Research, 48:23–65, 2013.
  • Zhang et al. (2022) Zhen Zhang, Ignavier Ng, Dong Gong, Yuhang Liu, Ehsan Abbasnejad, Mingming Gong, Kun Zhang, and Javen Qinfeng Shi. Truncated matrix power iteration for differentiable dag learning. Advances in Neural Information Processing Systems, 35:18390–18402, 2022.
  • Zheng et al. (2018) Xun Zheng, Bryon Aragam, Pradeep K Ravikumar, and Eric P Xing. Dags with no tears: Continuous optimization for structure learning. Advances in neural information processing systems, 31, 2018.
  • Zheng et al. (2020) Xun Zheng, Chen Dan, Bryon Aragam, Pradeep Ravikumar, and Eric Xing. Learning sparse nonparametric dags. In International Conference on Artificial Intelligence and Statistics, pp.  3414–3425. Pmlr, 2020.

Appendix A Further implementation details for section 4

This appendix provides additional implementation details for experiments in Section 4 to ensure clarity and reproducibility of the experiments. We describe our moral graph estimation algorithm; our implementation details of Gumbel-Softmax-based 0\ell_{0}-penalty and hard DAG constraints in our proposed CALM method and the experiments with hard DAG constraints in Section 4.3; our implementation details of Gumbel-Softmax-based 0\ell_{0}-penalty and soft DAG constraints in the experiments with soft DAG constraints in section 4.3; our way to achieve infinite samples; our implementation details for STG and Tanh; our implementation details for baseline methods. In some cases in Section 4, particularly when using soft constraints or certain baseline methods, the resulting structures are always not guaranteed to be DAGs after thresholding. To address this, we applied a postprocessing step where edges with the smallest absolute weights were iteratively removed until the structure formed a valid DAG.

A.1 Moral Graph Estimation

We estimate the moral graph MM using the Incremental Association Markov Blanket (IAMB) algorithm (Tsamardinos et al., 2003), which is modified from an implementation available on GitHub: https://github.com/wt-hu/pyCausalFS/blob/master/pyCausalFS/CBD/MBs/IAMB.py.

A.2 Implementation Details of Gumbel-Softmax-based 0\ell_{0}-Penalty and Hard DAG Constraints (CALM and Section 4.3 Hard DAG Constraint Experiments)

In this subsection, we describe the implementation details of the Gumbel-Softmax-based 0\ell_{0}-penalty and the hard DAG constraints applied in both the CALM method and the experiments that use hard DAG constraints in Section 4.3. We describe four components: (1) Gumbel-Softmax-based 0\ell_{0}-penalty’s initialization, parameter settings and final thresholding; (2) our DAG constraint formulation; and (3) our way to enforce hard DAG constraint and its parameter choices.

Gumbel-Softmax-based 0\ell_{0}-penalty

As introduced in Section 4.1, we use the Gumbel-Softmax mask gτ(U)g_{\tau}(U) to approximate the 0\ell_{0}-penalty. The logits matix UU is initialized as a zero matrix. The temperature parameter τ\tau is set to 0.5. The learned parameter matrix PP is initialized uniformly between -0.001 and 0.001, following a uniform distribution, to provide a small range of values for the initial weights. This ensures that the learned structure is unbiased at the start of optimization. The hyperparameter λ1\lambda_{1}, which controls the strength of the 0\ell_{0}-penalty in our method, is set to 0.005.

Additionally, for result thresholding, we follow the approach from Ng et al. (2022b). Specifically, After obtaining the learned logits matrix UU, we compute σ(U/τ)\sigma\left(U/\tau\right), filter it by the moral graph MM, and then apply a threshold of 0.5. The resulting matrix is used to compute SHD of CPDAG, precision of skeleton, and recall of skeleton.

DAG constraint formulation

The DAG constraint that we use is adapted from H(B)=Tr((I+1dBB)d)dH(B)=\text{Tr}\left(\left(I+\frac{1}{d}B\circ B\right)^{d}\right)-d, which proposed by Yu et al. (2019), due to its computational efficiency. In our method, we slightly modified it. Since the mask generated by gumbel softmax and a moral graph, Mgτ(U)M\circ g_{\tau}(U), is already non-negative (bounded between 0 and 1), we replace the element-wise multiplication BBB\circ B with a single mask Mgτ(U)M\circ g_{\tau}(U) (for experiments in Section 4.3 where the moral graph is not incorporated, we replace BBB\circ B with gτ(U)g_{\tau}(U) alone, which is also bounded between 0 and 1). Hence, the final DAG constraint we used is H(Mgτ(U))=Tr((I+1dMgτ(U))d)dH(M\circ g_{\tau}(U))=\text{Tr}\left(\left(I+\frac{1}{d}M\circ g_{\tau}(U)\right)^{d}\right)-d (for experiments in Section 4.3 without the moral graph, this becomes H(gτ(U))=Tr((I+1dgτ(U))d)dH(g_{\tau}(U))=\text{Tr}\left(\left(I+\frac{1}{d}g_{\tau}(U)\right)^{d}\right)-d). This simplifies the computation while still ensuring the final result B=Mgτ(U)PB=M\circ g_{\tau}(U)\circ P is a DAG when H(Mgτ(U))=0H(M\circ g_{\tau}(U))=0.

Enforcing the hard DAG constraint

Unlike NOTEARS(Zheng et al., 2018), which uses the augmented Lagrangian method (ALM), we use the quadratic penalty method (QPM) (Ng et al., 2022a) to enforce hard DAG constraints. As noted in Ng et al. (2022a), QPM always yields experimental results consistent with ALM. In our implementation of QPM, ρ\rho serves as a penalty parameter that increases iteratively across optimization steps, and optimization continues until h(B)h(B) falls below a predefined threshold. This ensures the final solution satisfies the DAG constraint in most of the cases, progressively tightening the constraint over iterations until convergence at a valid solution. In CALM and the experiments with hard constraints in Section 4.3, the ρ\rho starts at 10510^{-5} and is gradually increased by a factor of 3 after each subproblem iteration (a block of 40,000 iterations), continuing until the DAG constraint hh falls below a threshold of 10810^{-8}. For the optimizer, we choose the Adam optimizer with a learning rate of 0.001.

A.3 Implementation Details of Gumbel-Softmax-based 0\ell_{0}-Penalty and Soft DAG Constraints (Section 4.3 Soft DAG Constraint Experiments)

Here, we show implementation and parameters for experiments with soft constraint in Section 4.3. The implementation of the Gumbel-Softmax-based 0\ell_{0}-penalty and the DAG constraint formulation in the soft DAG constraint experiments follows the same parameter settings and implementation details as outlined in Appendix A.2, including the Gumbel-Softmax mask gτ(U)g_{\tau}(U), initialization of the logits matrix UU and parameter matrix PP, temperature τ\tau, hyperparameter λ1\lambda_{1}, result thresholding and DAG constraint formulation.

However, as these experiments use soft DAG constraints, we incorporate the DAG constraint as a penalty term in the objective function rather than enforcing it strictly as a hard constraint. In this case, the penalty weight for the DAG constraint, denoted as λ2\lambda_{2}, is set to 0.1. Unlike hard constraints, which are enforced using QPM, the soft constraint is optimized with a single run of 40,000 iterations. We use the Adam optimizer with a learning rate of 0.001.

A.4 The way to achieve infinity samples

A significant portion of our experiments is conducted under the assumption of infinite sample size. To achieve infinite samples, we use the true covariance matrix, which is calculated as

Σ=(IB)Ω(IB)1.{\Sigma}^{*}=(I-{B}^{*})^{-\top}{\Omega}^{*}(I-{B}^{*})^{-1}.

where B{B}^{*} is the true adjacency matrix (ground truth DAG) and Ω{\Omega}^{*} is the true noise variance matrix. The true covariance matrix Σ{\Sigma}^{*} is then substituted into the likelihood term of the objective function, allowing us to simulate the infinite sample case. Specifically, we substitute Σ{\Sigma}^{*} for Σ\Sigma in the likelihood term of our method’s objective function (Mgτ(U)P;Σ)\mathcal{L}(M\circ g_{\tau}(U)\circ P;\Sigma). When data standardization is required for infinite samples, we simply compute the standardized covariance matrix, Σstd\Sigma_{\text{std}}^{*} and substitute Σstd\Sigma_{\text{std}}^{*} for Σ\Sigma in (Mgτ(U)P;Σ)\mathcal{L}(M\circ g_{\tau}(U)\circ P;\Sigma).

A.5 Implementation details for STG and Tanh

In this section, we provide the implementation details and parameter selection for the STG (Stochastic Gates) and Tanh (Hyperbolic Tangent) methods used to approximate the 0\ell_{0} penalty in our experiments in section 4.5 and Appendix B.

The specific details for each method are as follows:

  • STG (Stochastic Gates) (Yamada et al., 2020): Following the approach of Yamada et al. (2020), we approximate the 0\ell_{0} penalty using a stochastic gate mechanism. We define Zd×dZ\in\mathbb{R}^{d\times d} as the matrix of gates zijz_{ij}, which represents the mask (similar to the role of gτ(U)g_{\tau}(U) in Section 4.1). zijz_{ij} is defined as a clipped, mean-shifted, Gaussian random variable zij=max(0,min(1,μij+ϵij))z_{ij}=\max(0,\min(1,\mu_{ij}+\epsilon_{ij})), where ϵij𝒩(0,σ2)\epsilon_{ij}\sim\mathcal{N}(0,\sigma^{2}) and σ\sigma is fixed during training. ZZ is then element-wise multiplied by the moral graph MM. The 0\ell_{0} penalty is approximated by the sum of probabilities that the gates are active, which is i,jΦ(μij/σ)Mij\sum_{i,j}\Phi\left(\mu_{ij}/\sigma\right)\cdot M_{ij}, where Φ()\Phi(\cdot) is the standard Gaussian CDF, and MijM_{ij} represents the elements of the moral graph. In our experiments, we initialize μ=0.5\mu=0.5 for all elements, as indicated in the pseudocode in Algorithm 1 of Yamada et al. (2020), and set σ=0.5\sigma=0.5.

  • Tanh (Hyperbolic Tangent) (Bhattacharya et al., 2021): Following Bhattacharya et al. (2021), the 0\ell_{0} penalty is approximated using the hyperbolic tangent function, given by i,jtanh(c|Bi,j|)\sum_{i,j}\tanh(c|B_{i,j}|), where Bi,jB_{i,j} represents the elements of the weighted adjacency matrix BB, which has already been restricted by a moral graph. In our experiments, we set the hyperparameter cc to 15.

A.6 Implementation Details for Baseline Methods

In this section, we provide implementation details for the baseline methods used in our experiments, as outlined in Section 4.6 and Appendix C. These include the GOLEM-NV-1\ell_{1}, NOTEARS, PC, and FGES. As noted by Ng et al. (2024) in their paper’s Section 5.1 Observation 2, using a high threshold for edge removal may lead to the wrongful removal of many true edges, causing a significant drop in recall. To mitigate this, we adopted a relatively small threshold of 0.1 in our experiments for the GOLEM-NV-1\ell_{1} and NOTEARS. Additionally, for cases where the resulting graph was not a DAG, we applied a post-processing step to remove edges with the smallest absolute values until the resulting graph became a valid DAG.

The specific details for each baseline method are as follows:

  • GOLEM-NV-1\ell_{1} (Ng et al., 2020): We use the parameters recommended by Ng et al. (2020) in their paper. The 1\ell_{1} sparsity penalty hyperparameter λ1\lambda_{1} was set to 2×1032\times 10^{-3}, and the soft DAG constraint hyperparameter λ2\lambda_{2} was set to 5.

  • NOTEARS (Zheng et al., 2018): We set the 1\ell_{1} penalty hyperparameter λ\lambda to 0.1, and use augmented Lagrangian method to enforce hard DAG constraints like the auther did. All other hyperparameter setting and implementation followed the default setting of the code of Zheng et al. (2018).

  • PC (Spirtes & Glymour, 1991) and FGES (Ramsey et al., 2017): Both PC and FGES were implemented using the py-causal package, a Python wrapper of the TETRAD project (Scheines et al., 1998). For PC, we employed the Fisher Z test, and for FGES, we adopted the BIC score (Schwarz, 1978) and set faithfulnessAssumed = False.

Appendix B Additional Comparison of 0\ell_{0} Approximation Methods for 50-node ER4 graphs and 100-node ER1 graphs

This section serves as a supplement to the results presented in Section 4.5, comparing the performance of different 0\ell_{0} approximation methods and the original GOLEM-NV-1\ell_{1} on 50-node ER4 graphs and 100-node ER1 graphs. For all methods here, We used infinite samples to eliminate finite sample error. As shown in Table 6, the results here are consistent with the findings in Section 4.5.

Table 6: Comparison of our method using different 0\ell_{0} approximations and original GOLEM-NV-1\ell_{1} for 50-node ER4 graphs and 100-node ER1 graphs under both data standardization and no data standardization.
50-node ER4 graphs
Standardized Non-Standardized
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 131.6 ± 12.3 0.70 ± 0.02 0.75 ± 0.02 139.4 ± 11.3 0.67 ± 0.02 0.77 ± 0.02
CALM-STG 140.8 ± 12.3 0.70 ± 0.02 0.71 ± 0.03 150.5 ± 8.5 0.69 ± 0.02 0.65 ± 0.02
CALM-Tanh 186.2 ± 4.5 0.69 ± 0.02 0.41 ± 0.03 251.6 ± 9.6 0.46 ± 0.02 0.45 ± 0.02
GOLEM-NV-1\ell_{1} 212.2 ± 4.6 0.58 ± 0.03 0.22 ± 0.02 294.7 ± 8.3 0.29 ± 0.01 0.20 ± 0.02
100-node ER1 graphs
Standardized Non-Standardized
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 16.0 ± 2.9 0.96 ± 0.01 0.99 ± 0.00 28.0 ± 3.6 0.92 ± 0.01 0.98 ± 0.01
CALM-STG 11.5 ± 2.2 0.97 ± 0.01 0.99 ± 0.00 79.6 ± 3.5 0.75 ± 0.01 0.95 ± 0.01
CALM-Tanh 26.2 ± 3.0 0.93 ± 0.01 0.97 ± 0.01 103.4 ± 3.2 0.67 ± 0.01 0.88 ± 0.02
GOLEM-NV-1\ell_{1} 109.2 ± 4.2 0.59 ± 0.02 0.60 ± 0.07 217.1 ± 8.3 0.36 ± 0.01 0.70 ± 0.03

We observe that our method using all three different 0\ell_{0} approximations—Gumbel-Softmax, STG, and Tanh—consistently outperform GOLEM-NV-1\ell_{1} in both 50-node ER4 graphs and 100-node ER1 graphs. This reinforces the effectiveness of combining 0\ell_{0} approximations, moral graph and hard constraints. Among the three 0\ell_{0} approximation methods, Gumbel-Softmax achieves the best performance across both graph structures, with strong results observed in both standardized and non-standardized settings. STG shows comparable results to Gumbel-Softmax in the standardized data, but its performance in non-standardized data lags behind that of Gumbel-Softmax, especially in the case of 100-node ER1 graphs.

Appendix C Comparison with baseline methods without data Data Standardization

This appendix serves as a supplement to the results presented in Section 4.6, where we compared the performance of CALM against several baseline methods after data standardization. Here, we provide a comparison of the same methods on 50-node ER1, 50-node ER4, and 100-node ER1 graphs before data standardization, using 1000 and 10610^{6} samples. In this section, the moral graph in CALM is estimated from finite samples. Specifically, for the 1000-sample experiments, the moral graph is estimated from 1000 samples, and for the 10610^{6}-sample experiments, the moral graph is estimated from 10610^{6} samples. This ensures consistency in the evaluation across different sample sizes.

Table 7: Comparison with baseline methods for 50-node ER1, 50-node ER4, and 100-node ER1 graphs without data standardization, using 1000 and 10610^{6} samples.
50-node ER1 graphs
1000 Samples 10610^{6} Samples
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 15.3 ± 3.4 0.90 ± 0.02 0.98 ± 0.00 10.7 ± 3.4 0.94 ± 0.02 0.98 ± 0.01
GOLEM-NV-1\ell_{1} 119.4 ± 5.6 0.35 ± 0.02 0.76 ± 0.02 117.3 ± 6.3 0.36 ± 0.02 0.75 ± 0.03
NOTEARS 25.7 ± 2.4 0.74 ± 0.02 0.98 ± 0.01 14.7 ± 2.1 0.86 ± 0.02 0.98 ± 0.01
PC 11.0 ± 1.4 0.98 ± 0.01 0.92 ± 0.01 2.8 ± 0.8 0.99 ± 0.01 0.99 ± 0.00
FGES 8.4 ± 2.4 0.94 ± 0.02 0.98 ± 0.00 1.3 ± 0.9 1.00 ± 0.00 1.00 ± 0.00
50-node ER4 graphs
1000 Samples 10610^{6} Samples
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 174.6 ± 7.8 0.60 ± 0.01 0.71 ± 0.02 151.4 ± 10.2 0.65 ± 0.02 0.75 ± 0.02
GOLEM-NV-1\ell_{1} 293.4 ± 8.6 0.29 ± 0.01 0.21 ± 0.02 291.2 ± 7.8 0.29 ± 0.01 0.20 ± 0.02
NOTEARS 191.5 ± 21.3 0.54 ± 0.03 0.91 ± 0.01 178.4 ± 14.6 0.55 ± 0.03 0.92 ± 0.01
PC 202.9 ± 2.0 0.61 ± 0.02 0.22 ± 0.01 233.1 ± 3.4 0.47 ± 0.01 0.32 ± 0.01
FGES 425.6 ± 23.2 0.33 ± 0.02 0.80 ± 0.01 750.0 ± 48.3 0.23 ± 0.02 0.99 ± 0.00
100-node ER1 graphs
1000 Samples 10610^{6} Samples
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 37.0 ± 4.0 0.86 ± 0.01 0.98 ± 0.01 29.3 ± 3.4 0.92 ± 0.01 0.98 ± 0.01
GOLEM-NV-1\ell_{1} 230.0 ± 7.9 0.35 ± 0.01 0.75 ± 0.02 215.1 ± 9.4 0.36 ± 0.01 0.69 ± 0.04
NOTEARS 74.7 ± 3.7 0.63 ± 0.01 0.99 ± 0.01 28.8 ± 5.2 0.84 ± 0.03 0.99 ± 0.00
PC 24.5 ± 1.5 0.95 ± 0.01 0.89 ± 0.01 4.2 ± 0.8 0.97 ± 0.01 1.00 ± 0.00
FGES 12.1 ± 1.7 0.95 ± 0.01 0.98 ± 0.00 1.0 ± 0.8 1.00 ± 0.00 1.00 ± 0.00

It is important to clarify that the results without data standardization are not as significant as those presented in Section 4.6, where data standardization was applied. Nonetheless, we include this comparison in Table 7 for completeness.

From Table 7, we can find: Firstly, even without data standardization, CALM continues to outperform GOLEM-NV-1\ell_{1} in all cases, demonstrating the robustness of our approach. In particular, the incorporation of the Gumbel-Softmax-based 0\ell_{0} approximation, hard DAG constraints, and moral graph still contributes to a substantial improvement over the original GOLEM-NV-1\ell_{1}.

Secondly, NOTEARS, which is designed specifically for cases with equal noise variance (EV), performs better before data standardization. This is because, although the noise ratio is set to 16 in the data, the non-equal variance is not as pronounced in the non-standardized data. In contrast, after standardization, the noise ratio becomes more extreme, emphasizing the non-equal variance nature of the data. This explains why NOTEARS performs better in non-standardized settings compared to its performance in standardized settings, even marginally outperforming the CALM in the 100-node ER1 graph with 10610^{6}samples. However, this advantage is not meaningful because NOTEARS is inherently based on the EV formulation, which does not align with the non-equal noise variance (NV) setting of our experiments.

Even so, CALM generally surpasses NOTEARS in the non-standardized setting, particularly for the majority of scenarios.

Compared to discrete methods, although the performance gap between CALM and PC slightly widens in the 1000-sample experiments without data standardization, this difference is not significant. One can always standardize the data, and thus, the results from Section 4.6 should be considered more relevant for real-world applications. The pre-standardization results provided here mainly offer insight into the robustness of our method across different settings.

Finally, just as in the results after data standardization in Section 4.6, in more dense graphs, the 50-node ER4 graphs, CALM demonstrates superior performance compared to the PC and FGES methods. This result suggests that in higher-density graphs, CALM enables continuous optimization methods to outperform discrete methods.

Appendix D Comparison of CALM and CoLiDE score functions

In this section, we compare two different objective functions for linear gaussian non-equal noise variance (NV) formulations: the likelihood-based objective used in GOLEM-NV and CALM and the objective proposed by Saboksayr et al. (2023), named CoLiDE-NV. For a fair comparison, we apply the Gumbel-Softmax approximation for 0\ell_{0} to CoLiDE-NV as well, incorporating hard DAG constraints and a moral graph, similar to CALM.

Originally, Saboksayr et al. (2023) propose CoLiDE-NV’s score function as

𝒮(B;Σ;Ω)=12Tr(Ω12(IB)Σ(IB))+12Tr(Ω12)+λ1B1.\mathcal{S}(B;\Sigma;\Omega)=\frac{1}{2}\text{Tr}\left(\Omega^{-\frac{1}{2}}(I-B)^{\top}\Sigma(I-B)\right)+\frac{1}{2}\text{Tr}(\Omega^{\frac{1}{2}})+\lambda_{1}\|B\|_{1}.

Unlike the GOLEM-NV model, CoLiDE-NV did not profile out the noise, so the Ω\Omega was kept in the score function. Here, Σ\Sigma is the sample covariance matrix. Also, CoLiDE-NV still used the 1\ell_{1} penalty. Since we have shown in Section 3 that 1\ell_{1} penalty often leads to inconsistent solutions, we substitute the 1\ell_{1} penalty in CoLiDE-NV with the 0\ell_{0} penalty, approximated by Gumbel-Softmax. Like CALM, we also incorporate hard DAG constraints and a moral graph to CoLiDE-NV as well. This yields the CoLiDE-0\ell_{0}-hard-moral formulation by defining 𝒢(B;Σ;Ω)=12Tr(Ω12(IB)Σ(IB))+12Tr(Ω12)\mathcal{G}(B;\Sigma;\Omega)=\frac{1}{2}\text{Tr}\left(\Omega^{-\frac{1}{2}}(I-B)^{\top}\Sigma(I-B)\right)+\frac{1}{2}\text{Tr}(\Omega^{\frac{1}{2}}),

minUd×d,Pd×d𝒢(Mgτ(U)P;Σ;Ω)+λ1Mgτ(U)1,subject toh(Mgτ(U))=0.\min_{U\in\mathbb{R}^{d\times d},P\in\mathbb{R}^{d\times d}}\mathcal{G}(M\circ g_{\tau}(U)\circ P;\Sigma;\Omega)+\lambda_{1}\|M\circ g_{\tau}(U)\|_{1},\quad\text{subject to}\quad h(M\circ g_{\tau}(U))=0.
Table 8: Comparison of CALM and CoLiDE-0\ell_{0}-hard-moral across 50-node ER1, 50-node ER4, and 100-node ER1 graphs under both No Standardization and Standardization.
50-node ER1 graphs
No Standardization Standardization
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 9.9 ± 3.4 0.95 ± 0.02 0.99 ± 0.01 5.5 ± 1.9 0.98 ± 0.01 0.99 ± 0.00
CoLiDE-0\ell_{0}-hard-moral 42.5 ± 2.6 0.73 ± 0.02 0.99 ± 0.00 56.2 ± 2.7 0.69 ± 0.02 0.98 ± 0.01
50-node ER4 graphs
No Standardization Standardization
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 139.4 ± 11.3 0.67 ± 0.02 0.77 ± 0.02 131.6 ± 12.3 0.70 ± 0.02 0.75 ± 0.02
CoLiDE-0\ell_{0}-hard-moral 157.2 ± 6.2 0.62 ± 0.01 0.82 ± 0.01 185.4 ± 3.4 0.67 ± 0.01 0.51 ± 0.01
100-node ER1 graphs
No Standardization Standardization
SHD of CPDAG Precision of Skeleton Recall of Skeleton SHD of CPDAG Precision of Skeleton Recall of Skeleton
CALM 28.0 ± 3.6 0.92 ± 0.01 0.98 ± 0.01 16.0 ± 2.9 0.96 ± 0.01 0.99 ± 0.00
CoLiDE-0\ell_{0}-hard-moral 88.8 ± 4.2 0.71 ± 0.01 0.99 ± 0.00 114.0 ± 3.5 0.66 ± 0.01 0.97 ± 0.01

We also use the quadratic penalty method (QPM) (Ng et al., 2022a) to enforce hard DAG constraints in CoLiDE-0\ell_{0}-hard-moral and the results are shown in Table 8. Table 8 summarizes the performance of CALM and CoLiDE-0\ell_{0}-hard-moral across linear gaussian model with 50-node ER1, 50-node ER4, and 100-node ER1 graphs under both data standardization and no data standardization. Here, we consider infinite samples. The results show that CALM consistently outperforms CoLiDE-0\ell_{0}-hard-moral in all cases.

This demonstrates that CALM’s likelihood-based objective is better suited for non-equal noise variance scenarios in the linear Gaussian case. The CoLiDE-0\ell_{0}-hard-moral, despite using the correct 0\ell_{0} penalty approximation, does not achieve as good results due to its alternative objective CoLiDE-NV.