# identifying_causal_direction_via_variational_bayesian_compression__a5618b77.pdf Identifying Causal Direction via Variational Bayesian Compression Quang-Duy Tran 1 Bao Duong 1 Phuoc Nguyen 1 Thin Nguyen 1 Telling apart the cause and effect between two random variables with purely observational data is a challenging problem that finds applications in various scientific disciplines. A key principle utilized in this task is the algorithmic Markov condition, which postulates that the joint distribution, when factorized according to the causal direction, yields a more succinct codelength compared to the anti-causal direction. Previous approaches approximate these codelengths by relying on simple functions or Gaussian processes (GPs) with easily evaluable complexity, compromising between model fitness and computational complexity. To address these limitations, we propose leveraging the variational Bayesian learning of neural networks as an interpretation of the codelengths. This allows the improvement of model fitness, while maintaining the succinctness of the codelengths, and the avoidance of the significant computational complexity of the GP-based approaches. Extensive experiments on both synthetic and real-world benchmarks in cause-effect identification demonstrate the effectiveness of our proposed method, showing promising performance enhancements on several datasets in comparison to most related methods. 1. Introduction Cause-effect identification or bivariate causal discovery the task of telling apart the cause and the effect between two random variables is a critical task across various scientific disciplines, including biology, economics, and sociology (Pearl, 2009). While randomized controlled trials (RCTs) are considered the most accurate method for identifying these types of causal relationships, especially in 1Deakin Applied Artificial Intelligence Initiative, Deakin University, Geelong, Australia. Correspondence to: Quang-Duy Tran . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). medical research (Guyon et al., 2019), they are often impractical due to resource constraints and ethical considerations. Studying passive observations offers a viable alternative for causal inference, despite requiring assumptions on the data generating process to detect the asymmetry between the causal and anti-causal directions. One intuitive interpretation of the causal asymmetry from an information-theoretic perspective is the independence postulate of algorithmic Markov kernels (Janzing & Sch olkopf, 2010), which states that the true causal direction must yield the lowest algorithmic Kolmogorov complexity factorization of the joint distribution. However, because of the incomputability of the Kolmogorov complexity (Li & Vit anyi, 2019), approximation methods via the principle of Minimum Description Length (MDL, Gr unwald, 2007; Marx & Vreeken, 2017; 2019a;b) are proposed to instantiate this complexity empirically. The two-part MDL aims to find the model that minimize two criteria: (1) the complexity of the data given that model and (2) the model complexity of the model used for modeling the data. The former complexity measures the model fitness, which is commonly evaluated through the log-likelihood of the data given the model. The options for estimating the latter model complexity are more diverse, which usually involve the number of available models and the parameters in each model. As these methods attempt to minimize the total codelength, they are also called compression-based methods. While these approaches have shown promising results, the estimation of the conditional distributions in these methods relies on traditional regression methods that offer easily evaluable model complexity. If the ground truth conditional models are more complex or too distinct from the predefined ones, this implementation for the model classes can result in lower fitness and higher complexity of data given the model, leading to suboptimal approximations for the algorithmic complexity. Dhir et al. (2024a) overcome this restriction by leveraging Gaussian processes (GPs), though this involves a significant trade-off between flexibility and computational complexity. To address the limitations of these previous compressionbased methods, we propose the leverage of neural networks for learning the conditional models, which are regarded Identifying Causal Direction via Variational Bayesian Compression as universal approximators (Hornik et al., 1989). Moreover, the algorithmic complexity of the networks can be approximated empirically via the concepts of bits-back coding (Hinton & van Camp, 1993; Wallace, 1990) and variational Bayesian coding (Honkela & Valpola, 2004; Louizos et al., 2017; Blier & Ollivier, 2018). In this study, we introduce COMIC a Bayesian COMpression-based approach to Identifying the Causal direction that improves the model fitness of compression-based methods without compromising the computability of model complexity and avoids the higher computational complexity of GP-based modeling. Correspondingly, through extensive empirical evaluation on benchmarks for the cause-effect identification task, our approach demonstrates promising results with performance improvements compared to a majority of complexity-based and maximum likelihood-based approaches on multiple benchmarks. Contributions The key contributions of this work can be outlined as follows: 1. We propose the utilization of Bayesian neural networks for modeling the conditional distributions to address the challenge of balancing flexibility and scalability, which is hindering previous complexity/compressionbased methods. By minimizing the variational Bayesian codelength, we can approximate the algorithmic complexity of neural networks. 2. From our proposed encoding scheme of the data given each causal direction, the causal identifiability can be proven. In particular, our models are nonseparable-compatible, which implies that given a sufficient amount of data, the causal direction can be identified via the complexity of our models. 3. The capability of our approach is assessed on both synthetic and real-world bivariate causal discovery benchmarks. In comparison to most related approaches, our method achieves performance improvements in several benchmarks, demonstrating the effectiveness of our approach in identifying the causal direction. 2. Related Works Despite being a well-defined task, the amount of information for determining the causal direction in the bivariate setting is limited, hindering further improvements in both theoretical and empirical results. In the following section, we provide an overview of recent related publications about this task, which includes two popular approaches. Functional Causal Models The earliest functional causal model (FCM, Pearl, 2009) being proposed is the additive noise model (ANM, Shimizu et al., 2006; Hoyer et al., 2008; B uhlmann et al., 2014; Peters et al., 2014), where the effect Y is generated from a function of the cause X and an independent noise EY (X EY ) by adding them as Y := f (X) + EY . In this model, the cause is assumed to only contribute to the mean, which is can be estimated by mean regression methods (Shimizu et al., 2006; Hoyer et al., 2008; B uhlmann et al., 2014; Peters et al., 2014). From the estimated models, CAM (B uhlmann et al., 2014) determines the causal direction by selecting the one with the greater maximum likelihood, whereas RESIT (Peters et al., 2014) quantifies the independence between the cause and the estimated noise with the Hilbert Schmidt Independence Criterion (HSIC, Gretton et al., 2005). Due to the strict assumption on model classes of ANMs, generalization approaches for the ANMs have been introduced to allow for more complex functions, including post nonlinear models (PNLs, Zhang & Hyv arinen, 2009) and heteroscedastic/location-scale noise models (LSNMs, Immer et al., 2023). Specifically, LSNMs assume that the cause not only contributes to the mean but also the scale through another function g (X), resulting in the effect Y := f (X)+g (X) EY . These models are more flexible than the ANMs due to their ability to also cover multiplicative noise models with the scale functions. LOCI (Immer et al., 2023) chooses the Gaussian likelihood to estimate the mean and scale functions, and recovers the noise from the fitted models. Similar to RESIT, LOCI also consider HSIC as a criterion in addition to the likelihood for predicting the causal direction. Since the Gaussian likelihood may not be robust to epistemic uncertainty, ROCHE (Tran et al., 2024a) suggested a more robust estimation for LSNMs by replacing the Gaussian likelihood with a likelihood based on Student s t-distribution. Principle of Independent Causal Mechanisms Beside the algorithmic complexity, there are other approaches for interpreting the principle of independent causal mechanisms (ICMs, Peters et al., 2017, Sec. 2.1), which assumes the independence between the marginal distribution of the cause and the conditional distribution of the effect given the cause. With the assumption of low noise levels and invertible causal mechanisms, IGCI (Daniuˇsis et al., 2010) formulates this independence using orthogonality in information space to distinguish cause and effect, which is implemented by the relative entropy distances. CDCI (Duong & Nguyen, 2022) postulates that due to this ICM principle, the shape of the conditional distributions will be invariant, and compute the variations in shape to find the causal directions. The algorithmic complexity-based methods (Marx & Vreeken, 2019a;b; Tagasovska et al., 2020) do not only consider the model fitness objective as in FCM-based methods but also examine the complexity of the model. SLOPE and SLOPER (Marx & Vreeken, 2017; 2019b) use a set of Identifying Causal Direction via Variational Bayesian Compression basis functions to regress the data globally and locally to account for both deterministic and non-deterministic functions, and compute the codelengths for encoding the data with two-part codes. SLOPPY (Marx & Vreeken, 2019a) is an improvement of RECI (Bl obaum et al., 2018) that utilizes regularized regressions of Identifiable Regressionbased Scoring Functions and find the one with the lowest regularized score to find the minimal model in each direction. QCCD (Tagasovska et al., 2020) uses non-parametric conditional quantile regression methods and encodes the data via each quantile model. Our work COMIC also belongs to this category where the conditional distributions of the data are modeled by neural networks and encoded by the variational Bayesian coding scheme. This approach allows for more flexibility compared to previous methods while allowing for the balance between model fitness and model complexity. Another study by Dhir et al. (2024a) interprets the principle of ICMs from the view of the Bayesian model selection and proposes using the marginal likelihoods estimated by latent variable Gaussian processes (GPLVM, Titsias & Lawrence, 2010) for identifying the causal direction. 3. Preliminaries In this work, the assumption of causal sufficiency is adopted, which is similar to previous publications (Immer et al., 2023; Marx & Vreeken, 2017; 2019a;b; Mooij et al., 2016; Peters et al., 2014; Tagasovska et al., 2020; Tran et al., 2024a). This means that we assume that there is no hidden confounder between the two random variables. In other words, given two variables X and Y , if they are not independent, either X or Y will be the cause of the other. As mentioned in Sec. 2, interpreting the principle of independent causal mechanisms is one of two major approaches for determining the direction of a causal relation. The algorithmic interpretation of this independence via the Kolmogorov complexity originates from the postulate about the algorithmic independence of conditionals (Janzing & Sch olkopf, 2010, Eq. 26) as follows: Postulate 1 (Algorithmic Independence of Conditionals, Janzing & Sch olkopf, 2010). Let G be causal hypothesis, represented by a directed acyclic graph (DAG), over a set of d variables X1, . . . , Xd with a joint density p (X1, . . . , Xd), which is lower semi-computable, that is, K (p (X1, . . . , Xd)) < . The causal hypothesis is only acceptable if the shortest description (i.e., the Kolmogorov complexity) of the joint density K (p (X1, . . . , Xd)) is equal to a concatenation of the shortest description of the Markov kernels up to an independent additive constant. This postulate can be described formally as K (p (X1, . . . , Xd)) += j=1 K (p (Xj | PAG,j)) , (1) where K ( ) is the Kolmogorov complexity, PAG,j denotes the parents of Xj in the causal hypothesis G, and += denotes equality up to a constant, which is independent of p ( ). In the bivariate setting with two random variables X and Y , if X causes Y (denoted as X Y ), Eq. (1) will become K (p (X, Y )) += K (p (X)) + K (p (Y | X)) . (2) Following the algorithmically independent conditionals, the algorithmic independence of Markov kernels has also been postulated by Janzing & Sch olkopf (2010); Mooij et al. (2010) as follows Postulate 2 (Algorithmic Independence of Markov Kernels, Janzing & Sch olkopf, 2010). If X Y , the marginal distribution of the cause p (X) and the conditional distribution of the effect given the cause p (Y | X) are algorithmically independent of each other. In other words, their algorithmic mutual information (IA) will be equal to zero up to an additive constant, IA (p (X) : p (Y | X)) += 0, (3) and this independence does not hold in the other direction. From these postulates, Mooij et al. (2010, Thm. 1) induce a rule for identifying the causal direction. Theorem 3.1 (Asymmetry in Complexities of Markov Kernels, Mooij et al., 2010). If X is the cause of Y and Pos. 2 holds, the description of the joint distribution K (p (X, Y )) via the description of the marginal distribution of the cause K (p (X)) and the description of the conditional distribution of the effect given the cause K (p (Y | X)) is the most succinct one, or formally, K (p (X)) + K (p (Y | X)) + K (p (Y )) + K (p (X | Y )) . (4) As a consequence of this rule, a causal indicator score can be obtained for the assumed causal directions of X Y as follows X Y := K (p (X)) + K (p (Y | X)) , (5) and vice versa for the remaining direction of Y X. From these indicator scores, we can infer that X Y if Y X X Y > 0 and Y X if Y X X Y < 0. The Kolmogorov complexity is not computable in practice (Li & Vit anyi, 2019). Hence, the causal indicators in the previous section are substituted by approximating approaches such as Minimum Message Length (MML, Wallace & Freeman, 1987) or Minimum Description Length (MDL, Rissanen, 1978) in previous information-theoretic methods (Mooij et al., 2010; Marx & Vreeken, 2017; Identifying Causal Direction via Variational Bayesian Compression 2019b;a). MML and MDL share a two-part coding principle where the complexity or codelength1 L2-p of the data D is computed via a model M M by combining the complexity (fitness) of the data given that model L1 (D | M) and the complexity of the model L2 (M) as L2-p M (D) := L1 (D | M) + L2 (M) . (6) The codelength with a model M that minimizes this equation is appointed as an instantiation for the algorithmic complexity. If there are multiple solutions for M , the one with the smallest model complexity L1 is selected to model the data. From this two-part code, we can attain an approximation for the causal indicator score in Eq. (5) as follows ˆ 2-p X Y := L2-p M X (X) + L2-p M Y |X (Y | X) , (7) where M X and M Y |X are models that minimize L2-p M X (X) and L2-p M Y |X (Y | X), respectively. The definitions of Kolmogorov complexity and algorithmic mutual information, as well as the discussion of the MDL-based instantiation of Kolmogorov complexity in the context of causal discovery, as referenced in this section, are provided in App. A. 4. COMIC: Bayesian Compression for Identifying Causal Direction 4.1. Classes of Models for the Conditionals For the conditional distribution, most previous MDL-related studies choose the set M of candidate functions by predefining a list of basis functions (Marx & Vreeken, 2017; 2019b) or using more advanced regression methods such as cubic spline regression (Marx & Vreeken, 2019a). Although these classes of functions have the model codelengths that are easily computable (e.g., through the number of bits of linear parameters in polynomial regressions), the fitness can be compromised when the ground truth models are more complex, leading to suboptimal codelengths. Dhir et al. (2024a) utilize GPLVM (Titsias & Lawrence, 2010) to improve the fitness; however, this substantially increases the computational complexity due to the poor scalability of GPs. Neural networks are one class of models that can overcome these limitations thanks to their universality in approximations (Hornik et al., 1989) and better scalability. Additionally, the complexity of neural networks has also been well studied and implemented with different encoding methods (Blier & Ollivier, 2018; Louizos et al., 2017; Voita & Titov, 2020). Hence, neural network can be a viable class of 1Since algorithmic complexity is defined based on the length of the code, or codelength, we use the terms of complexity and codelength interchangeably in this work. models for computing the codelengths of conditional distributions. Moreover, due to their flexibility and capability of approximating a wide range of functions, neural networks allow us to only focus on this class of models where complexity scores are determined solely by their parameters. The prequential code (Dawid, 1984) and the variational Bayesian code (Honkela & Valpola, 2004) are two effective MDL approaches for encoding the conditional distribution modeled by neural networks (Blier & Ollivier, 2018; Gr unwald, 2007; Voita & Titov, 2020). The former encodes the model implicitly through sequential data transmission, as seen in applications such as time series. The latter involves predefining the priors over the parameters and utilizes variational inference to learn the posteriors from the samples. In contrast to the online prequential code, this approach is aligned with the two-part code. Despite these differences in coding strategies, both methods yield consistent results, as noted by Voita & Titov (2020). We opted for the variational Bayesian code to assess the codelengths because it explicitly captures the model complexity and does not necessitate a specific order of data transmission. 4.2. Variational Bayesian Code for Evaluating Complexity of Neural Networks The problem of encoding a conditional distribution p (Y | X) is often defined via a transmitting perspective. Alice has a dataset DN := x(i), y(i) N i=1 that needs to be transported to Bob, who has already got the input samples of this dataset x(i) N i=1. The most efficient method is to encode the conditional distribution p (Y | X) so that Bob can predict the remaining output part y(i) N i=1 of DN. The variational Bayesian code is a two-part code in MDL where both Alice and Bob first designate a class of model M = {p (y | x, θ) | θ Θ}, where θ represents the parameters of the conditional probability density function p (y | x, θ), and a prior distribution for the parameters with the probability density function p (θ). The corresponding two-part codelength2 for this setting can be computed as L2-p p(θ) y(1:N) | x(1:N) := log p y(1:N) | x(1:N), θ log p (θ) , (8) where the former term corresponds to L1, the latter corresponds L2 in Eq. (6), and p y(1:N) | x(1:N), θ := QN i=1 p y(i) | x(i), θ . The variational Bayesian code is 2In most information-theoretic literature, the amount of information is measured in bits. As a result, the logarithm with the base of 2 is commonly chosen. In this work, for convenience, we will use the natural logarithm and measure the amount of information in nats (natural units of information, Gr unwald, 2007). The amount in nats can easily be converted to the corresponding value in bits by dividing it by log 2. Identifying Causal Direction via Variational Bayesian Compression based on the bits-back coding scheme (Wallace, 1990; Hinton & van Camp, 1993). In this scheme, Alice employs a codelength with redundant code and computes the codelength with respect to that auxiliary information (i.e., the redundant code). After that, Bob will perform the same learning process and observe the choice made by this process to retrieve the auxiliary information and the transmitted data (Honkela & Valpola, 2004). By applying this scheme, instead of finding a point estimate of the parameters θ that minimize Eq. (8), we introduce redundant code by choosing the parameters from a variational distribution qϕ (θ) and compute the expected codelength on this distribution. In this scenario, the amount of redundant code to encode qϕ (θ) is its entropy H (qϕ (θ)) = Eqϕ(θ) [ log qϕ (θ)] (Honkela & Valpola, 2004). We can obtain the amount of original information being transmitted by deducting the excessive entropy H (qϕ (θ)) from the expectation codelength as follows Lvar qϕ(θ) y(1:N) | x(1:N) := Eqϕ(w) h L2-p y(1:N) | x(1:N), θ i H (qϕ (θ)) (9) := Eqϕ(w) h log p y(1:N) | x(1:N), θ i + KL (qϕ (θ) || p (θ)) , (10) where the first term corresponds to the fitness of the data given the model and the second term corresponds to the complexity of the model of the two-part MDL principle (Honkela & Valpola, 2004; Louizos et al., 2017). Lvar qϕ(θ) y(1:N) | x(1:N) is known as the evidence lower bound (ELBO) in the variational inference problem (Blei et al., 2017). From this perspective, by minimizing Eq. (10), we can expect to achieve the negative logarithm of the evidence in the Bayesian inference problem, which is also known as the Bayesian codelength in MDL coding (Gr unwald, 2007) as follows LBayes p(θ) y(1:N) | x(1:N) := log Z p y(1:N) | x(1:N), θ p (θ) dθ. (11) The gap between the optimal codelength LBayes p(θ) y(1:N) | x(1:N) and its upper bound Lvar qϕ(θ) y(1:N) | x(1:N) can also be formulated from the variational perspective as follows Lvar qϕ(θ) y(1:N) | x(1:N) LBayes p(θ) y(1:N) | x(1:N) = KL qϕ (θ) || p θ | x(1:N), y(1:N) , (12) where p θ | x(1:N), y(1:N) is the posterior distribution of the parameters given the observed samples. From this formulation, it is obvious that we can retrieve the Bayesian codelength iff. qϕ (θ) converges to p θ | x(1:N), y(1:N) . The selection of the priors p (θ) and the variational distributions qϕ (θ) over the parameters is a necessary step in Bayesian learning. In this work, we employ Gaussian distributions as the family for both the priors p (θ) and the mean-field variational posteriors qϕ (θ). The details on these priors and variational posteriors are provided in App. B. 4.3. Identifying Causal Direction via the Codelengths From the formulation of variational Bayesian codelength of the conditional distribution described in the previous section, we can approximate the conditional codelength for the assumed causal direction X Y using Eq. (10) with the chosen likelihood distribution being the Gaussian distribution as p y(1:N) | x(1:N), θ i=1 N y(i) | µ x(i); θ , σ2 x(i); θ , (13) where µ ( ; θ) and σ ( ; θ) are modeled via a neural network f Y : R Θ R2 with parameters θ Θ. In particular, we implement a neural network with one hidden layer and two output nodes f Y,1 ( ; θ) and f Y,2 ( ; θ), and compute the parameters of the likelihood as follows µ ( ; θ) = f Y,1 ( ; θ) and σ ( ; θ) = ζ (f Y,2 ( ; θ)) , (14) where ζ ( ) is a positive link function, such as the exponential or softplus function, for ensuring the standard deviation values being positive. Both x(1:N) and y(1:N) are standardized with respect to their corresponding sample means and standard deviations to avoid the scale-based bias on the identifiability (Reisach et al., 2021). Regarding the codelength of the assumed cause X, we adopt standard Gaussian distribution N (x | 0, 1) to encode the data using the marginal codelength, which is a common choice in previous methods (Mooij et al., 2016; Immer et al., 2023). We denote this marginal codelength LN x(1:N) . The approximated causal indicator score for this direction is the sum of the two codelengths ˆ var X Y DN := LN x(1:N) + Lvar qϕ (θ) y(1:N) | x(1:N) , (15) where qϕ (θ) is the model that minimizes the variational Bayesian codelength. The corresponding score ˆ Y X for the reversed direction Y X is estimated by a similar procedure. Once the scores in both directions are obtained, the difference between them provides a final score: ˆ var DN := ˆ var Y X DN ˆ var X Y DN , (16) Identifying Causal Direction via Variational Bayesian Compression which indicates the inferred causal direction with its absolute value reflecting the confidence of the inference. If the optimized variational distribution qϕ (θ) converges to the posterior distribution p θ | x(1:N), y(1:N) and N (0, 1) is the ground truth distribution of p (X), we can expect ˆ X Y to converge to the Bayesian causal indicator score Bayes X Y DN := LN x(1:N) + LBayes p(θ) y(1:N) | x(1:N) (17) := log p DN | MX Y , (18) where p DN | MX Y is the marginal likelihood of the dataset DN factorized in accordance with the causal model MX Y . Conversely, Bayes Y X DN can be achieved if Y N (0, 1) and the minimized variational codelength in this case also converges to the Bayesian codelength. 4.4. Causal Identifiability The identifiability of our approach is closely related to the identifiability of Bayesian causal models via marginal likelihoods. First, we introduce the definition of separablecompatibility by Dhir et al. (2024a), which is a necessary condition of two Bayesian causal models being unidentifiable via their marginal likelihoods, regardless of the dataset DN. Definition 4.1 (Separable-Compatibility of Bayesian Causal Models, informally restated from Dhir et al., 2024a). Two causal models MX Y and MY X are separablecompatible if the anti-causal factorizations of MX Y and MY X respectively belong to the same classes of distributions as MY X and MX Y , and the priors of these anti-causal factorizations can also be factorized with respect to the priors of MY X and MX Y . In this definition, the anti-causal factorization of the causal model MX Y refers to the factorization of the joint distribution p (X, Y ) into p (Y ) and p (X | Y ) with respect to MX Y . Similarly, the anti-causal factorization of the causal model MY X involves factorizing p (X, Y ) into p (X) and p (Y | X) according to MY X. If the anticausal factorization the causal model MX Y results in the same distributions as the causal factorization o f MY X, or vice versa, the two causal models are said to be separablecompatible. As a result of our Bayesian coding scheme, the identifiability of our method is verifiable through an orthogonal perspective of marginal likelihoods (Dhir et al., 2024a). Let us assume that the approximated scores in Eq. (15) would converge to the Bayesian scores in Eq. (17), our results on the non-separable-compatibility of Bayesian causal models employed in our method can be presented as follows: Proposition 4.2 (Non-Separable-Compatibility of Our Causal Models). Let the two Bayesian causal models MX Y and MY X respectively be factorized into the marginal densities p (x | MX Y ) and p (y | MY X) and the conditional densities p (y | x, θY , MX Y ) and p (x | y, θX, MY X). The marginal densities are standard Gaussian N ( | 0, 1). The conditional densities are Gaussian likelihoods N | µ ( ; θ) , σ2 ( ; θ) , where µ ( ; θ) and σ ( ; θ) are parametrized by neural networks as described in Eq. (14), Sec. 4.3, and the parameters θ follow Gaussian priors specified in App. B. Under these conditions, the Bayesian causal models are not separable-compatible. Corollary 4.3 (Causal Identifiability of Our Causal Models). Let p ( | MX Y ) and p ( | MY X) respectively be the marginal likelihoods of the data given the causal models MX Y and MY X. Because of the non-separablecompatibility in Prop. 4.2, there exists a dataset DN whose causal direction is identifiable via the difference between Bayesian indicator scores: Bayes DN := Bayes Y X DN Bayes X Y DN . (19) Cor. 4.3 implies that in a large sample limit, the Bayesian causal indicator scores defined in Eq. (17) can distinguish the cause and effect from observational data. Details on the causal identifiability via marginal likelihoods and the proof of Prop. 4.2 are further discussed in App. C. 5. Experiments Throughout this section, the empirical performance of our COMIC approach is evaluated in comparison with stateof-the-art complexity-based and regression-based methods for bivariate causal discovery. Implementation details and descriptions of the benchmarks related to this section are provided in App. D and E, respectively. 5.1. Experimental Settings Benchmarks Following previous works (Immer et al., 2023; Marx & Vreeken, 2019a; Tagasovska et al., 2020; Tran et al., 2024a), the experiments in this section utilizes both synthetic and real-world data for evaluation. The synthetic data consists of 12 common benchmarks. The first collection of synthetic datasets, including AN, AN-s, LS, LS-s, and MN-U, are proposed by Tagasovska et al. (2020). The next group of simulated benchmarks, comprising SIM, SIM-c, SIM-G, and SIM-ln, is introduced by Mooij et al. (2016). The remaining synthetic benchmarks consist of the CE-Multi, CE-Net, and CE-Cha datasets described by Goudet et al. (2018). For real-world data, we choose the T ubingen cause-effect pairs (Mooij et al., 2016). Baselines We compare our work against various baseline ICM-based and FCM-based methods. Methods based Identifying Causal Direction via Variational Bayesian Compression SIM-c SIM-G SIM-c SIM-G (a) COMIC (Ours) SIM-c SIM-G SIM-c SIM-G (b) SLOPPY (solid: AIC, dashed: BIC) SIM-c SIM-G SIM-c SIM-G (c) SLOPE (solid: SLOPER, dashed: SLOPE) SIM-c SIM-G SIM-c SIM-G SIM-c SIM-G SIM-c SIM-G (e) IGCI (solid: Uniform, dashed: Gaussian) SIM-c SIM-G SIM-c SIM-G SIM-c SIM-G SIM-c SIM-G SIM-c SIM-G SIM-c SIM-G Figure 1. Performance on synthetic and real-world benchmarks. Higher accuracy and bidirectional AUROC (Bi-AUROC) values are preferable. Our COMIC method is compared against SLOPPY (with the AIC and BIC variants, Marx & Vreeken, 2019a), SLOPER (Marx & Vreeken, 2019b) and SLOPE (Marx & Vreeken, 2017), QCCD (Tagasovska et al., 2020), IGCI (with uniform and Gaussian reference measures, Daniuˇsis et al., 2010), GPLVM (Dhir et al., 2024a), LOCI (Immer et al., 2023), and CAM (B uhlmann et al., 2014). The marginal likelihood-based objective in COMIC achieves promising results with enhanced performance compared to methods with the maximum likelihood-based objectives in LOCI and CAM, especially on the real-world T ubingen benchmark. In comparison to most complexity-based methods, COMIC also achieves better performance on multiple synthetic datasets and obtains comparable results on real-world data. Identifying Causal Direction via Variational Bayesian Compression COMIC (Ours) QCCD IGCI-Uniform IGCI-Gaussian COMIC (Ours) QCCD IGCI-Uniform IGCI-Gaussian Figure 2. Overall performance of COMIC and the baseline methods on all benchmarks. The accuracy and bidirectional AUROC (Bi-AUROC) scores are averaged over all datasets. The results indicate that our COMIC approach achieves better and more consistent overall results in comparison to a majority of baselines. on the principle of ICMs include SLOPPY (Marx & Vreeken, 2019a), SLOPE (Marx & Vreeken, 2017; 2019b), QCCD (Tagasovska et al., 2020), and IGCI (Daniuˇsis et al., 2010), and GPLVM (Dhir et al., 2024a). For FCM-based methods, we choose CAM (B uhlmann et al., 2014) as a representative for ANM-based methods, and LOCI (Immer et al., 2023) with maximum likelihood scoring to represent LSNM-based methods. Some of the ICM-based methods have different variants, which we also include in the evaluations. Evaluation Metrics The identification result of each pair in a dataset is considered as a sample in a binary classification problem of that dataset. Hence, the accuracy score and the area under receiver operating characteristic curve (AUROC) are commonly utilized as evaluation metrics for the task of bivariate causal discovery. As the ground truth directions in the benchmarks are imbalanced, we compute the bidirectional AUROC (Bi-AUROC, Guyon et al., 2019, Sec. 2.4.3), which is the average of the forward AUROC and the backward AUROC corresponding to X Y and Y X, respectively. 5.2. Results & Discussion The experimental results on all aforementioned benchmarks are visualized in Fig. 1 and 2. The overall results in Fig. 2 demonstrate that our COMIC approach achieves secondbest performance in comparison to complexity-based and maximum likelihood-based methods. Our method is also more consistent across all benchmarks, as indicated by the smaller standard deviation of the accuracy and Bi-AUROC scores compared to most baseline methods. Furthermore, COMIC outperforms the maximum likelihood-based methods, i.e., LOCI and CAM, suggesting that its marginal likelihood-based objective delivers promising results with improved performance relative to methods with the maximum likelihood-based objectives. GPLVM is the baseline that attains the best overall results. However, it is important to note that GPLVM also requires the most substantial computational resources, with execution on GPUs (as noted in App. D), compared to COMIC and other baselines, which can be executed on CPUs. This high resource demand stems from the computational complexity of Gaussian processes with latent variables and the extensive use of random restarts for hyperparameter selection. Performance on Synthetic Benchmarks On synthetic AN, AN-s, LS, LS-s, and MN-U benchmarks, COMIC perfectly determines the causal directions in every case. QCCD, GPLVM, and LOCI also achieve equivalent accuracy and Bi-AUROC scores on these datasets. These findings align with theoretical expectation that models identifiable via maximum likelihood are also identifiable via marginal likelihood (Dhir et al., 2024a). IGCI with Gaussian reference measure is another method that excels in this group of datasets. The remaining baseline approaches do not perform as well on the more challenging LS, LS-s, and MN-U sets. In particular, the predictions of SLOPER, SLOPE, IGCI with uniform reference measure, and CAM on these datasets are noticeably suboptimal compared to other baselines. On the more challenging SIM and SIM-c datasets, our COMIC approach demonstrates performance comparable to complexity-based methods, such as SLOPPY with AIC, SLOPER, and SLOPE. A majority of baseline methods perform decently on SIM-G and SIM-ln, which are more manageable compared to the previous two sets. COMIC achieves the second-highest scores on SIM-ln, with slightly lower accuracy and Bi-AUROC than GPLVM. One unexpected outlier among the baseline models is IGCI, which exhibits the lowest accuracy and Bi-AUROC on SIM-ln, despite its intended focus on telling apart causal directions in low-noise scenarios. Moreover, SLOPPY with AIC, SLOPER, Slope, and QCCD underperform relative to COMIC on the Identifying Causal Direction via Variational Bayesian Compression SIM-G dataset with nearly Gaussian distributions over the causes. The diverse causal modeling of CE-Multi results in varying performance across benchmarks. Information-theoretic approaches, including COMIC, SLOPPY with BIC, SLOPER, SLOPE, IGCI with Gaussian reference measure, and GPLVM, tend to perform well on this benchmark. QCCD and uniform-referenced IGCI, while also belonging to this category, perform less effectively. Since CAM is designed for additive noise models, it exhibits inefficacious performance on this dataset. A majority of methods featuring regressions in the learning process can adequately predict causal relations of the CE-Net benchmark. Hence, IGCI is the only baseline tested that underperforms on this benchmark. Due to the difficulty of CE-Cha, the results on this set resemble those on SIM, where most methods struggle, except for GPLVM. Performance on Real-World Benchmark On the realworld T ubingen benchmark, complexity-based methods, including our COMIC approach, GPLVM, SLOPPY with AIC, SLOPE, and QCCD, achieve comparable accuracy and Bi-AUROC scores. This highlights the effectiveness of compression-based methods, whose Bi-AUROC scores are notably higher than the remaining maximum likelihoodbased methods that focus solely on model fitness. As noted by Marx & Vreeken (2019a), SLOPPY with AIC offers greater flexibility for more complex datasets, such as this one, and performs better than the BIC variant. A similar pattern appears in IGCI, where the uniform reference measure variant obtains higher accuracy and Bi-AUROC. Although SLOPER outperforms its respective variant SLOPE on previous benchmarks, it is surpassed by SLOPE on this benchmark. 6. Conclusion In this work, we have proposed COMIC a neural network compression-based approach for determining the cause and effect via the variational Bayesian code where a more universal and scalable class of neural networks is utilized for modeling the conditionals to improve fitness and induce better codelengths. With the variational Bayesian coding scheme, the algorithmic complexity of these networks can be assessed empirically to approximate the theoretical Kolmogorov complexity, and its identifiability can also be scrutinized from a marginal likelihood-based perspective. The effectiveness of COMIC has been validated through comprehensive experiments, delivering promising results and demonstrating enhanced performance compared to most related methods based on the compression and FCM regression objectives across multiple benchmarks. Limitations & Future Work One persistent limitation of our current work is the non-convexity of the learning objective, which may require further investigation into the convergence and consistency. Additionally, the use of standard Gaussian codelength to encode the marginal distribution of the cause can introduce a bias toward more Gaussian causes, potentially posing a hindrance in intricate settings. In future work, we plan to adapt our approach to multivariate settings, explore alternative priors and likelihoods, and consider other marginal likelihood estimation techniques for neural networks, such as Laplace approximation, to enhance its applicability. Regarding the multivariate extension, we also provide a brief discussion in App. H on potential adaptations of the bivariate methods, including ours, to multivariate data. Impact Statement This paper introduces a novel method whose goal is to advance the field of causal discovery, with potential applications across various scientific disciplines. While our work potentially carries broad societal implications, we do not believe that there are any specific consequences that require particular emphasis here. Blei, D. M. and Jordan, M. I. Variational inference for Dirichlet process mixtures. Bayesian Analysis, 1(1):121 143, 2006. Blei, D. M., Kucukelbir, A., and Mc Auliffe, J. D. Variational inference: A review for statisticians. Journal of the American Statistical Association, 112(518):859 877, 2017. Blier, L. and Ollivier, Y. The description length of deep learning models. In Advances in Neural Information Processing Systems (Neur IPS), volume 31, 2018. Bl obaum, P., Janzing, D., Washio, T., Shimizu, S., and Sch olkopf, B. Cause-effect inference by comparing regression errors. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), volume 84 of Proceedings of Machine Learning Research, pp. 900 909, 2018. Bornschein, J., Chiappa, S., Malek, A., and Ke, R. N. Prequential MDL for causal structure learning with neural networks. In Proceedings of the Conference on Neural Informational Processing Systems (Neur IPS) WHY-21 Workshop, 2021. B uhlmann, P., Peters, J., and Ernest, J. CAM: Causal additive models, high-dimensional order search and penalized Identifying Causal Direction via Variational Bayesian Compression regression. The Annals of Statistics, 42(6):2526 2556, 2014. Chickering, D. M. Optimal structure identification with greedy search. Journal of Machine Learning Research, 3 (Nov):507 554, 2002. Coker, B., Bruinsma, W. P., Burt, D. R., Pan, W., and Doshi Velez, F. Wide mean-field Bayesian neural networks ignore the data. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), volume 151 of Proceedings of Machine Learning Research, pp. 5276 5333, 2022. Cundy, C., Grover, A., and Ermon, S. BCD Nets: Scalable variational approaches for Bayesian causal discovery. In Advances in Neural Information Processing Systems (Neur IPS), volume 34, pp. 7095 7110, 2021. Daniuˇsis, P., Janzing, D., Mooij, J., Zscheischler, J., Steudel, B., Zhang, K., and Sch olkopf, B. Inferring deterministic causal relations. In Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), 2010. Dawid, A. P. Present position and potential developments: Some personal views: Statistical theory: The prequential approach. Journal of the Royal Statistical Society. Series A (General), 147(2):278 292, 1984. Deleu, T., G ois, A., Emezue, C., Rankawat, M., Lacoste Julien, S., Bauer, S., and Bengio, Y. Bayesian structure learning with generative flow networks. In Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), volume 180 of Proceedings of Machine Learning Research, pp. 518 528, 2022. Dhir, A., Power, S., and van der Wilk, M. Bivariate causal discovery using Bayesian model selection. In Proceedings of the International Conference on Machine Learning (ICML), volume 235 of Proceedings of Machine Learning Research, pp. 10710 10735, 2024a. Dhir, A., Sedgwick, R., Kori, A., Glocker, B., and van der Wilk, M. Continuous Bayesian model selection for multivariate causal discovery. ar Xiv preprint ar Xiv:2411.10154, 2024b. Dhir, A., Ashman, M., Requeima, J., and van der Wilk, M. A meta-learning approach to Bayesian causal discovery. In Proceedings of the International Conference on Learning Representations (ICLR), 2025. Duong, B. and Nguyen, T. Bivariate causal discovery via conditional divergence. In Proceedings of the Conference on Causal Learning and Reasoning (CLea R), volume 177 of Proceedings of Machine Learning Research, pp. 236 252, 2022. Duong, B. and Nguyen, T. Heteroscedastic causal structure learning. In Proceedings of the European Conference on Artificial Intelligence (ECAI), volume 372 of Frontiers in Artificial Intelligence and Applications, pp. 598 605, 2023. Goudet, O., Kalainathan, D., Caillou, P., Guyon, I., Lopez Paz, D., and Sebag, M. Learning functional causal models with generative neural networks. Explainable and Interpretable Models in Computer Vision and Machine Learning, pp. 39 80, 2018. Gretton, A., Bousquet, O., Smola, A., and Sch olkopf, B. Measuring statistical dependence with Hilbert-Schmidt norms. In Proceedings of the International Conference on Algorithmic Learning Theory (ALT), pp. 63 77, 2005. Gr unwald, P. D. The minimum description length principle. Adaptive Computation and Machine Learning. The MIT Press, 2007. Guyon, I., Statnikov, A., and Batu, B. B. Cause effect pairs in machine learning. Springer, 2019. Hinton, G. E. and van Camp, D. Keeping the neural networks simple by minimizing the description length of the weights. In Proceedings of the Annual Conference on Computational Learning Theory (COLT), pp. 5 13, 1993. Honkela, A. and Valpola, H. Variational learning and bitsback coding: An information-theoretic view to Bayesian learning. IEEE Transactions on Neural Networks, 15(4): 800 810, 2004. Hornik, K., Stinchcombe, M., and White, H. Multilayer feedforward networks are universal approximators. Neural Networks, 2(5):359 366, 1989. Hoyer, P., Janzing, D., Mooij, J. M., Peters, J., and Sch olkopf, B. Nonlinear causal discovery with additive noise models. In Advances in Neural Information Processing Systems (NIPS), volume 21, 2008. Immer, A., Schultheiss, C., Vogt, J. E., Sch olkopf, B., B uhlmann, P., and Marx, A. On the identifiability and estimation of causal location-scale noise models. In Proceedings of the International Conference on Machine Learning (ICML), volume 202 of Proceedings of Machine Learning Research, pp. 14316 14332, 2023. Janzing, D. and Sch olkopf, B. Causal inference using the algorithmic Markov condition. IEEE Transactions on Information Theory, 56(10):5168 5194, 2010. Janzing, D., Mooij, J., Zhang, K., Lemeire, J., Zscheischler, J., Daniuˇsis, P., Steudel, B., and Sch olkopf, B. Information-geometric approach to inferring causal directions. Artificial Intelligence, 182 183:1 31, 2012. Identifying Causal Direction via Variational Bayesian Compression Kalainathan, D., Goudet, O., and Dutta, R. Causal discovery toolbox: Uncovering causal relationships in Python. Journal of Machine Learning Research, 21(37):1 5, 2020. Kaltenpoth, D. Don t Confound Yourself: Causality from Biased Data. Ph D thesis, Saarland University, 2024. Kaltenpoth, D. and Vreeken, J. Causal discovery with hidden confounders using the algorithmic Markov condition. In Proceedings of the Uncertainty in Artificial Intelligence (UAI), volume 216 of Proceedings of Machine Learning Research, pp. 1016 1026, 2023. Kingma, D. P. and Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), 2015. Kolmogorov, A. N. On tables of random numbers. Sankhy a: The Indian Journal of Statistics, Series A (1961 2002), 4 (25):369 376, 1963. Li, M. and Vit anyi, P. An Introduction to Kolmogorov Complexity and Its Applications. Texts in Computer Science. Springer, 4th edition, 2019. Lin, Y., Huang, Y., Liu, W., Deng, H., Ng, I., Zhang, K., Gong, M., Ma, Y.-A., and Huang, B. A skewness-based criterion for addressing heteroscedastic noise in causal discovery. In Proceedings of the International Conference on Learning Representations (ICLR), 2025. Lorch, L., Rothfuss, J., Sch olkopf, B., and Krause, A. Di BS: Differentiable Bayesian structure learning. In Advances in Neural Information Processing Systems (Neur IPS), volume 34, pp. 24111 24123, 2021. Lorch, L., Sussex, S., Rothfuss, J., Krause, A., and Sch olkopf, B. Amortized inference for causal structure learning. In Advances in Neural Information Processing Systems (Neur IPS), volume 35, pp. 13104 13118, 2022. Louizos, C., Ullrich, K., and Welling, M. Bayesian compression for deep learning. In Advances in Neural Information Processing Systems (Neur IPS), volume 30, 2017. Mac Kay, D. J. C. Comparison of approximate methods for handling hyperparameters. Neural Computation, 11(5): 1035 1068, 1999. Marx, A. and Vreeken, J. Telling cause from effect using MDL-based local and global regression. In Proceedings of the IEEE International Conference on Data Mining (ICDM), pp. 307 316, 2017. Marx, A. and Vreeken, J. Identifiability of cause and effect using regularized regression. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD), pp. 852 861, 2019a. Marx, A. and Vreeken, J. Telling cause from effect by local and global regression. Knowledge and Information Systems, 60(3):1277 1305, 2019b. Marx, A. and Vreeken, J. Formally justifying MDL-based inference of cause and effect. In Proceedings of the AAAI Workshop on Information-Theoretic Causal Inference and Discovery (ITCI 22), 2022. Mian, O. A., Marx, A., and Vreeken, J. Discovering fully oriented causal networks. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), volume 35, pp. 8975 8982, 2021. Mooij, J. M., Stegle, O., Janzing, D., Zhang, K., and Sch olkopf. Probabilistic latent variable models for distinguishing between cause and effect. In Advances in Neural Information Processing Systems (Neur IPS), volume 23, 2010. Mooij, J. M., Peters, J., Janzing, D., Zscheischler, J., and Sch olkopf, B. Distinguishing cause from effect using observational data: Methods and benchmarks. Journal of Machine Learning Research, 17(32):1 102, 2016. Neal, R. M. Priors for infinite networks. In Bayesian Learning for Neural Networks, volume 118 of Lecture Notes in Statistics, chapter 2, pp. 29 53. Springer, 1996. Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., De Vito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., and Chintala, S. Py Torch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems (Neur IPS), volume 32, 2019. Pearl, J. Causality. Cambridge University Press, 2009. Peters, J., Mooij, J. M., Janzing, D., and Sch olkopf, B. Causal discovery with continuous additive noise models. Journal of Machine Learning Research, 15(58):2009 2053, 2014. Peters, J., Janzing, D., and Sch olkopf, B. Elements of Causal Inference: Foundations and Learning Algorithms. Adaptive Computation and Machine Learning. The MIT press, 2017. Reisach, A., Seiler, C., and Weichwald, S. Beware of the simulated DAG! Causal discovery benchmarks may be easy to game. In Advances in Neural Information Processing Systems (Neur IPS), volume 34, pp. 27772 27784, 2021. Rissanen, J. Modeling by shortest data description. Automatica, 14(5):465 471, 1978. Identifying Causal Direction via Variational Bayesian Compression Shimizu, S., Hoyer, P. O., Hyv arinen, A., Kerminen, A., and Jordan, M. A linear non-Gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(72), 2006. Solomonoff, R. J. A formal theory of inductive inference. Part I. Information and Control, 7(1):1 22, 1964a. Solomonoff, R. J. A formal theory of inductive inference. Part II. Information and Control, 7(2):224 254, 1964b. Sønderby, C. K., Raiko, T., Maaløe, L., Sønderby, S. K., and Winther, O. Ladder variational autoencoders. In Advances in Neural Information Processing Systems (NIPS), volume 29, 2016. Tagasovska, N., Chavez-Demoulin, V., and Vatter, T. Distinguishing cause from effect using quantiles: Bivariate quantile causal discovery. In Proceedings of the International Conference on Machine Learning (ICML), volume 119 of Proceedings of Machine Learning Research, pp. 9311 9323, 2020. Titsias, M. and Lawrence, N. D. Bayesian Gaussian process latent variable model. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), volume 9 of Proceedings of Machine Learning Research, pp. 844 851, 2010. Tran, Q.-D., Nguyen, P., Duong, B., and Nguyen, T. Differentiable Bayesian structure learning with acyclicity assurance. In Proceedings of the IEEE International Conference on Data Mining (ICDM), pp. 598 607, 2023. Tran, Q.-D., Duong, B., Nguyen, P., and Nguyen, T. Robust estimation of causal heteroscedastic noise models. In Proceedings of the SIAM International Conference on Data Mining (SDM), pp. 788 796, 2024a. Tran, Q.-D., Nguyen, P., Duong, B., and Nguyen, T. Constraining acyclicity of differentiable Bayesian structure learning with topological ordering. Knowledge and Information Systems, 66(9):5605 5630, 2024b. Voita, E. and Titov, I. Information-theoretic probing with minimum description length. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 183 196, 2020. Wallace, C. S. Classification by minimum-message-length inference. In Advances in Computing and Information Proceedings of the International Conference on Computing and Information (ICCI), volume 468 of Lecture Notes in Computer Science, pp. 72 81, 1990. Wallace, C. S. and Freeman, P. R. Estimation and inference by compact coding. Journal of the Royal Statistical Society. Series B (Methodological), 49(3):240 265, 1987. Williams, C. Computing with infinite networks. In Advances in Neural Information Processing Systems (NIPS), volume 9, 1996. Zhang, K. and Hyv arinen, A. On the identifiability of the post-nonlinear causal model. In Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI), pp. 647 655, 2009. Identifying Causal Direction via Variational Bayesian Compression A. Kolmogorov Complexity, Algorithmic Mutual Information, & Minimum Description Length Kolmogorov Complexity In this section, we will introduce some related definitions and notations regarding the algorithmic complexity, which is also known as the Kolmogorov complexity. This complexity represents the length of the ultimate lossless compression of the data (Marx & Vreeken, 2019b). Definition A.1 (Kolmogorov Complexity, Kolmogorov, 1963; Li & Vit anyi, 2019). With a universal Turing machine U, for every program p {0, 1} that generates x and halts from an input y, the Kolmogorov complexity is the length of the shortest program. Formally, the Kolmogorov complexity is defined as K (x | y) := min p {length (p) | U ( y, p ) = x} . (20) When the input y is an empty string ϵ, we have K (x) = K (x | ϵ). In various causal discovery literature, it is necessary to evaluate the complexity of a distribution with a probability density function. As a consequence, we consider the Kolmogorov complexity of a function f (x). Definition A.2 (Kolmogorov Complexity of a Function, Li & Vit anyi, 2019). To compute the complexity of a function f (x), we need to find the shortest program p that generates f (x) from x with precision q as follows K (p (x)) := min p length (p) | |U ( x, q, p ) f (x)| < 1 Algorithmic Mutual Information From the Kolmogorov complexity, the algorithmic mutual information between two binary strings is also defined to quantify the amount of overlapping information. Definition A.3 (Algorithmic Mutual Information, Li & Vit anyi, 2019). The algorithmic mutual information between two binary strings x and y is IA (x : y) := K (y) K (y | x ) += K (x) + K (y) K (x, y) , (22) where x is the shortest description of the string x and += denotes equality up to an additive constant, which only depends on U and is independent of x and y. Minimum Description Length in Causal Discovery Literature In the context of causal discovery task, there are two major hindrances (Kaltenpoth & Vreeken, 2023) when estimating the Kolmogorov complexity K (p (X)) of a distribution p (X) from the data samples x: (1) there is no knowledge about the true underlying distribution p (X), and (2) the Kolmogorov complexity is incomputable. The former problem can be resolved by estimating the model through the joint complexity K (x, p (X)), which on expectation over p (X) will yield K (p (X)) + H (p (X)) up to an additive constant (Marx & Vreeken, 2022). The latter problem requires an approximating codelength L (x, p (X)) that mirrors K (x, p (X)), commonly selected according to the minimum description length (MDL, Gr unwald, 2007) principles. The MDL codelengths are computed by limiting the finding of the shortest program/model that generate the data and halt from a predefined set instead of the universal Solomonoff prior (Gr unwald, 2007; Li & Vit anyi, 2019; Solomonoff, 1964a;b). By employing a sufficiently broad class of models, designed to maintain independence from their conditioning variables, and using a large enough number of samples, the gap between the approximated codelength and the Kolmogorov complexity can be expected to decrease (Kaltenpoth, 2024; Marx & Vreeken, 2022). While approximation gaps may still exist in our method, they should not affect the identifiability result, which is assessed through marginal likelihoods and remains independent of the Kolmogorov complexity-based postulates in Sec. 3. B. Prior and Variational Distributions Let us consider a fully-connected layer with the weight matrix W = [wij] and the bias vector b = [bj]. The parameters W and b in each layer form the set of parameters θ in Sec. 4. Priors over Parameters As the parameters are continuous, we consider Gaussian priors with zero means and variances as hyperparameters. For simplicity, we utilize the standard Gaussian distribution for the biases. The priors of the weights and Identifying Causal Direction via Variational Bayesian Compression biases can be formulated as i,j p (wij) , p (wij) = N wij | 0, z2 ij , (23) j p (bj) , p (bj) = N bj | 0, σ2 b . (24) Hyperparameters Selection Let z = [zij, σb] denote the scale hyperparameters of the priors as in Eq. (23) and (24). The marginal likelihood can be acquired by marginalizing over z as log p y(1:N) | x(1:N) = log Z z p y(1:N) | x(1:N), z p (z) dz, (25) where p (z) is the prior over the hyperparameters z. Following previous work by Dhir et al. (2024a), the integral in the marginal likelihood above is approximated with Laplace approximation (LA, Mac Kay, 1999). In this approximation method, we assume a Gaussian distribution around the maximum a posteriori (MAP) solution and consider the marginal likelihood as the normalizing constant of this Gaussian distribution. The formulation of the normalizing constant computed around the MAP solution in LA is as follows log p y(1:N) | x(1:N) log p y(1:N) | x(1:N), ˆz p (ˆz) 1 2π Λz where ˆz := arg max z h log p y(1:N), z | x(1:N) i , and (27) Λz := 2 z log p y(1:N), z | x(1:N) ˆz . (28) We also follow Dhir et al. (2024a) by choosing a uniform hyperprior over the hyperparameters z and assuming a sufficient amount of samples so that the distribution concentrates around a single point. Given these conditions, we can simply approximate log p y(1:N) | x(1:N) log p y(1:N) | x(1:N), ˆz . From an implementation perspective, this means that we can regard the priors z as additional parameters to be optimized in conjunction with the following variational parameters. Variational Posteriors over Parameters To evaluate the posteriors of W and b, we employ mean-field variational inference (MF-VI, Blei et al., 2017) where the posterior of each weight and bias is assumed to be independent. The posterior distribution of each weight and bias is modeled as a Gaussian distribution, given by ij qϕ (wij) , qϕ (wij) = N wij | µW,ij, σ2 W,ij , (29) j qϕ (bj) , qϕ (bj) = N bj | µb,j, σ2 b,j . (30) Forward Pass With Gaussian Variational Posteriors The forward pass of a fully-connected Bayesian layer with the Gaussian variational posteriors is presented in Alg. 1. Hin RN A where N is the number of samples and A is the input dimension. Let MW = [µW,ij] and VW = σ2 W,ij represent the means and variances of the weight matrix W, µb = [µb,j] and σ2 b = h σ2 b,j i be the means and variances of the bias vector b, be the Hadamard (element-wise) product, and 2 and 1 2 respectively denote the Hadamard (element-wise) square and square root. Corresponding to this algorithm, the output matrix will include N samples of B nodes. Given that the cost of sampling one instance from the standard Gaussian distribution N (0, 1) is O (1), the computational complexity of forwarding through one layer using local reparametrization is O (NAB). C. Causal Identification via Bayesian Model Selection C.1. Bayesian Model Selection Criterion Let X and Y not be independent of each other, and assume there are no hidden confounders between these two random variables (i.e., the assumption of causal sufficiency). Dhir et al. (2024a) have introduced a Bayesian framework for inferring Identifying Causal Direction via Variational Bayesian Compression Algorithm 1 Forward pass of fully-connected Bayesian layer with Gaussian variational posteriors Require: input: Hin RN A; parameters: MW RA B, VW RA B + , µb RB, σ2 b RB + 1: Mout Hin MW + µb 2: Vout H 2 in VW + σ2 b 3: E N (0, 1) {dims: N B} 4: Hout Mout + V 1 2 out E 5: return Hout the causal direction between two possible choices MX Y and MY X from an observational dataset DN with N samples via the posterior p Mi | DN = p DN | Mi p (Mi) p (DN) , (31) where Mi M := {MX Y , MY X} and p DN = P Mi M p DN | Mi p (Mi). The likelihood p DN | Mi has a prior π p (π | Mi) and can be retrieved by marginalizing p DN | π, Mi over π. Hence, this likelihood is also referred as the marginal likelihood . As we do not possess any knowledge about each causal direction, the prior probabilities are set uniformly to p (Mi) = |M| 1 = 0.5. With this prior, the log-ratio between two posteriors of the causal directions are then computed to balance the evidence p (D) and achieve the following equations log p MX Y | DN p (MY X | DN) = log p DN | MX Y p (MX Y ) p (DN | MY X) p (MY X) (32) = log p DN | MX Y p (DN | MY X) . (as p (MX Y ) = p (MY X) = 0.5) (33) Eq. (33) shows that choosing a model Mi so that log p Mi | DN > log p M i | DN is equivalent to log p DN | Mi > log p DN | M i , where M i is the inverse direction of Mi; therefore, we can select the causal direction depending on the log-ratio of the marginal likelihoods as follows MX Y if log p DN | MX Y log p DN | MY X > 0, MY X if log p DN | MX Y log p DN | MY X < 0, indecisive if log p DN | MX Y log p DN | MY X = 0. (34) C.2. Causal Identifiability of Marginal Likelihoods From Eq. (34), Dhir et al. (2024a) study the causal identifiability based on the separability of the two marginal likelihoods of the data given the models. First, the distribution-equivalent causal models and Bayesian distribution-equivalent causal models are defined, which are the cases where the causal direction can not be determined. Definition C.1 (Distribution-Equivalence of Causal Models, restated from Def. 2.2 of Dhir et al., 2024a). Let two causal models MX Y and MY X respectively have m X, c Y |X CX CY |X and m Y , c X|Y CY CX|Y , where CX and CY denote the sets of marginal distributions of X and Y , and CY |X and CX|Y denote the sets of conditional distributions of Y given X and X given Y . If MX Y and MY X are distribution-equivalent, there exists a unique bijective map γ : CX CY |X CY CX|Y such that for every m X, c Y |X CX CY |X, there holds an equality of joint likelihoods m X (x) m Y |X (y | x) = m Y (y) c X|Y (x | y) , x, y, (35) where m Y , c X|Y = γ m X, c Y |X . If the two causal models are distribution-equivalent, the maximum likelihoods cannot distinguish the causal direction. Otherwise, there exists some dataset DN (N sufficiently large) such that the maximum likelihood can determine the cause direction. For Bayesian causal models, which include auxiliary prior distributions over the marginal distributions (CX and CY ) and the conditional distributions (CY |X and CX|Y ), we have a definition of Bayesian distribution-equivalent causal models as follows: Identifying Causal Direction via Variational Bayesian Compression Definition C.2 (Bayesian Distribution-Equivalence of Causal Models, restated from Def. 4.4 of Dhir et al., 2024a). Given two Bayesian causal models MX Y and MY X, they are Bayesian distribution-equivalent if for all N N and for all DN, the marginal likelihoods of the data given the models are equal, i.e., p DN | MX Y = p DN | MY X . If two possible causal models MX Y and MY X are Bayesian distribution-equivalent according to this definition, regardless of the observational dataset DN, the log-ratio of marginal likelihoods cannot be utilized as criterion for distinguishing the causal direction (third case in Eq. (34)). Bayesian distribution-equivalence is a more specific case of distribution-equivalence since it not only implies the equality of maximum likelihoods but also imposes the equality of marginal likelihoods, which is the expectations of the likelihoods over their prior distributions. Hence, if two causal models are not distribution-equivalent (i.e., identifiable via maximum likelihoods), they are not Bayesian distribution-equivalent (i.e., identifiable via marginal likelihoods). In cases of distribution-equivalent causal models, these models need a necessary condition of separable-compatibility for them to be Bayesian distribution-equivalence, which is also proposed by Dhir et al. (2024a) to determine the set of distributions C that can be utilized for modeling the marginal likelihoods. The definition of this condition is as follows: Definition C.3 (Separable-Compatibility of Bayesian Causal Models, restated from Def. 4.6 of Dhir et al., 2024a). Let two Bayesian causal models MX Y and MY X include m X, c Y |X CX CY |X and m Y , c X|Y CY CX|Y with their corresponding priors being πX Y m X, c Y |X = πX (m X) πY |X c Y |X and πY X m Y , c X|Y = πY (m Y ) πX|Y c X|Y , respectively. Let us denote γ as in Def. C.1, if the push-forward γ πX Y is separable with respect to CY CX|Y , i.e., πY X γ m X, m Y |X = πY (m Y ) πX|Y c X|Y , and γ 1 πY X is separable with respect to CX CY |X, then the two causal models are separable-compatible. This definition is similar to the principle of independent causal mechanisms where the joint density can only be factorized (separable) with respect to the causal direction and the anti-causal factorization do not satisfy separability. This means that even the prior of the anti-causal models can be represented as the product of two priors of distributions, these two priors are not independent of each other. As the separable-compatibility is the necessary condition of the Bayesian distributionequivalence, if the two causal models are not separable-compatible, they are not Bayesian distribution-equivalent, implying that they are identifiable via the marginal likelihoods. C.3. Analysis of Our Models (Proof of Prop. 4.2, Sec. 4.4) To validate the identifiability of the causal models of COMIC, we need to analyze their separable-compatibility. Let us recall the non-separable-compatibility of our proposed causal models in Prop. 4.2 as follows: Proposition 4.2 (Non-Separable-Compatibility of Our Causal Models). Let the two Bayesian causal models MX Y and MY X respectively be factorized into the marginal densities p (x | MX Y ) and p (y | MY X) and the conditional densities p (y | x, θY , MX Y ) and p (x | y, θX, MY X). The marginal densities are standard Gaussian N ( | 0, 1). The conditional densities are Gaussian likelihoods N | µ ( ; θ) , σ2 ( ; θ) , where µ ( ; θ) and σ ( ; θ) are parametrized by neural networks as described in Eq. (14), Sec. 4.3, and the parameters θ follow Gaussian priors specified in App. B. Under these conditions, the Bayesian causal models are not separable-compatible. The proof of this non-separable-compatibility is provided below. For ease of proof, neural networks with one hidden layer can be considered as Gaussian processes under Central Limit Theorem when the number of hidden units is large enough (Neal, 1996; Williams, 1996). Additive Noise Models (ANMs) First, we study the case of an ANM where the mean of y is modeled by a neural network f Y : R R with one hidden layer. With an input x, the hidden layer is computed as follows: h Y = h Y (x) = σ w h Y x + bh Y , (36) where wh Y = [wh Y ,ih] R1 Dh, wh Y ,ih is sampled from a Gaussian prior N 0, z2 h Y ,ih , bh Y = [bh Y ,ih] RDh, bh Y ,ih N 0, σ2 bh , and σ ( ) is an activation function such as the hyperbolic tangent (tanh). From the features h Y in the hidden layer, the output function f Y is then linearly represented as f Y (x) = w f Y h (x) + bf Y , (37) Identifying Causal Direction via Variational Bayesian Compression where wf Y = [wf Y ,ih] RDh 1, wf Y ,ih is also sampled from a Gaussian distribution N 0, z2 f Y ,ih , and bf Y N 0, σ2 bf Y . Let us denote that θf Y = (wf Y , bf Y ), θh Y = (wh Y , bh Y ), z2 f Y = h z2 f Y ,ih, σ2 f Y i , and z2 h Y = h z2 h Y ,ih, σ2 h Y i , the Gaussian process of f Y is GP (mf Y (x) , Kf Y (x, x )) where mf Y (x) = Eθf Y ,θh Y [f Y (x)] = 0, (38) Kf Y (x, x ) = Eθf Y ,θh Y [f Y (x) f Y (x )] = Eθf Y ,θh Y w f Y h Y + bf Y w f Y h Y + bf Y (h Y = h Y (x )) = Eθf Y ,θh Y w f Y h Y w f Y h Y + w f Y h Y bf Y + bf Y w f Y h Y + b2 f Y = Ewf Y ,θh Y w f Y h Y w f Y h Y + Ebf Y b2 f Y = Ewf Y ,θh Y ih wf Y ,ihh Y,ih ih wf Y ,ihh Y,ih + Ebf Y b2 f Y = Ewf Y ,θh Y jh wf Y ,ihwf Y ,jhh Y,ihh Y,jh + Ebf Y b2 f Y = Ewf Y ,θh Y ih w2 f Y ,ihh Y,ihh Y,ih + Ebf Y b2 f Y ih z2 f Y ,ihh Y,ihh Y,ih + Ebf Y b2 f Y = Eθh Y h Y diag z2 f Y h Y + σ2 bf Y h h Y (x) diag z2 f Y h Y (x ) i + σ2 bf Y . (39) With y = f Y (x) + εY and εY N 0, σ2 εY , the generative processes are f Y | x, z2 f Y , z2 h Y N (0, Kf Y (x, x)) , (40) y | f Y N f Y , σ2 εY , (41) = y | x, z2 f Y , z2 h Y N 0, Kf Y (x, x) + σ2 εY . (42) For the causal model MX Y , the marginal distribution p (x) and the conditional distribution p (y | x) with respect to the Gaussian process above are as follows p (x | MX Y ) = N (x | 0, 1) , (43) p y | x, z2 f Y , z2 h Y , MX Y = N y | 0, Kf Y (x, x) + σ2 εY , (44) where z2 f Y and z2 h Y are hyperparameters. The selection method for these hyperparameters is discussed in App. B. The corresponding distributions for the causal model MY X are p (y | MY X) = N (y | 0, 1) , (45) p x | y, z2 f X, z2 h X, MY X = N x | 0, Kf X (y, y) + σ2 εX . (46) To analyze the separable-compatible, we need to compare the causal factorization of MX Y with the anti-causal factorization of MY Y , which is formulated as p x | z2 f X, z2 h X, MY X = Z p x | y, z2 f X, z2 h X, MY X p (y | MY X) dy (47) Identifying Causal Direction via Variational Bayesian Compression p y | x, z2 f X, z2 h X, MY X = p x | y, z2 f X, z2 h X, MY X p (y | MY X) p x | z2 f X, z2 h X, MY X . (48) As Kf X (y, y) is non-linearly dependent on y, the marginal distribution of x | z2 f X, z2 h X, MY X is non-Gaussian, which makes it not belong to the same class of distribution as p (x | MX Y ). Moreover, because the Gaussian distribution is not a conjugate prior if the likelihood is a variance-parametrized Gaussian distribution, the posterior distribution y | x, z2 f X, z2 h X, MY X will not be Gaussian. Hence, these distributions are not in the same classes of distributions as those of MX Y . In addition, let us examine the variance x with respect to the anti-causal marginal Ep x|z2 f X ,z2 h X ,MY X x2 = Z N (y | 0, 1) Z x2N x | 0, Kf X (y, y) + σ2 εX dx dy (49) = Z N (y | 0, 1) Kf X (y, y) + σ2 εX dy. (50) Although this integral does not have a closed form, it is to be expected that the variance of x | z2 f X, z2 h X, MY X is dependent on z2 f X and z2 h X, which should also be the case for the variance of y | x, z2 f X, z2 h X, MY X. Hence, the distributions in the anti-causal factorization share identical priors, which make them not independent of each other and violate the principle of independent causal mechanisms (ICMs, Peters et al., 2017, Sec. 2.1). From these aspects, we can accept that the models MX Y and MY X are not separable-compatible. Location-Scale Noise Models (LSNMs) In a LSNM, not only the mean but also the scale of the noise is parametrized. As mentioned in Eq. (14), Sec. 4.3, we consider the case where y is generated from x via a neural network f Y : R R2 with two output nodes f Y,1 and f Y,2. The generating process of y can be presented as follows y | f Y N f Y,1, ζ2 (f Y,2) , (51) where ζ ( ) is a positive link function, such as exponential or softplus function. In our work, as f Y,1 and f Y,2 are two output nodes of the same network, the Gaussian processes of f Y,1 and f Y,2 from x can be described as f Y,i | x, z2 f Y,i, z2 h Y N 0, Kf Y,i (x, x) , (52) where Kf Y,i (x, x ) = Eθh Y h h Y (x) diag z2 f Y,i h Y (x ) i + σ2 bf Y,i and i {1, 2}. The conditional distributions p y | x, z2 f Y,1, z2 f Y,2, z2 h Y , MX Y and p x | y, z2 f X,1, z2 f X,2, z2 h X, MY X will be more complicated in this setting, however, we can expect similar separable-compatibility as in the location-only setting above. Since the property of separable-compatibility does not hold in our causal models, following the definition of separablecompatibility in Def. C.3 and Bayesian distribution-equivalence in Def. C.2, our causal models are not distribution-equivalent, which indicates that there exists a dataset where the causal direction is identifiable via marginal likelihoods. As a result, we propose Cor. 4.3, which is recalled as follows Corollary 4.3 (Causal Identifiability of Our Causal Models). Let p ( | MX Y ) and p ( | MY X) respectively be the marginal likelihoods of the data given the causal models MX Y and MY X. Because of the non-separable-compatibility in Prop. 4.2, there exists a dataset DN whose causal direction is identifiable via the difference between Bayesian indicator scores: Bayes DN := Bayes Y X DN Bayes X Y DN . (19) D. Implementations & Hyperparameters All the experiments are conducted on a workstation with an Intel Core i7 processor, 64 GB of memory, and 3 TB of storage, except for those involving GPLVM, which are executed on NVIDIA Tesla V100 GPUs. Our proposed method COMIC is implemented in Python with the Py Torch library (Paszke et al., 2019). The implementation of the fully-connected Bayesian layers with Gaussian priors is adapted from the publicly available code3 of Louizos 3Bayesian layers: https://github.com/Karen Ullrich/Tutorial_Bayesian Compression For DL Identifying Causal Direction via Variational Bayesian Compression et al. (2017). As mentioned in App. C, Bayesian neural networks are employed to model the mean and the standard deviation of each assumed direction, i.e., X Y and Y X. Each neural network includes one hidden layer with Dh = 50 nodes with the hyperbolic tangent (tanh) as the activation function and a fully-connected output layer. The natural exponential function is chosen as the positive link function for ensuring the positivity of the predicted standard deviation values. The neural networks of COMIC are optimized to minimize the variational Bayesian objective with the Adam optimizer (Kingma & Ba, 2015) and the cosine learning rate scheduler with a maximum learning rate of 10 2, a minimum learning rate of 10 6, and T = 2, 500 training epochs. To avoid the bad local optima of the variational objective, Louizos et al. (2017) suggest the adoption of the warm-up scheme from Sønderby et al. (2016), where the model complexity term KL (qϕ (w) || p W (w)) is weighted by a hyperparameter β annealed linearly from 0 to 1 for the first TWU training epochs. In our implementation, we choose TWU = 250. In addition, to avoid overfitting the hyperparameters of the priors, we pretrain the models and the hyperparameters with maximum a posteriori (MAP) estimation before the variational optimization for TWU = 2, 500 epochs, where the parameters and hyperparameters are optimized to minimize the negative joint log-likelihood of the data and the parameters. All samples of each pair are utilized in the training process, i.e., we set the batch size to be equal to the number of available samples of each dataset. D.2. Baselines For information theory-based methods including SLOPPY4 (Marx & Vreeken, 2019a), SLOPE5 (Marx & Vreeken, 2017; 2019b), and QCCD6 (Tagasovska et al., 2020), we utilize their original repositories in R. The rpy2 interface is employed to incorporate these methods into our source code. As a Python version of IGCI (Daniuˇsis et al., 2010; Janzing et al., 2012) is available in the Causal Discovery Toolkit (CDT7, Kalainathan et al., 2020), we employ this version for our experiments. With GPLVM8, we utilize the Python implementation from the repository of Dhir et al. (2024a). Following to the recommendation of Dhir et al. (2024a), we choose the closed form GPLVM for all datasets except for the T ubingen dataset where the stochastic GPLVM is employed. For regression-based methods, the Python source code of LOCI9 is obtainable from the repository of Immer et al. (2023). The R version of CAM10 (B uhlmann et al., 2014) can be accessed via the Comprehensive R Archive Network (CRAN). E. Descriptions of the Benchmarks In the upcoming descriptions, we denote X as the cause, Y as the effect, and EY as the noise term (EY X). AN, AN-s, LS, LS-s, & MN-U (Tagasovska et al., 2020) As their names are the abbreviations of their generating models, the AN and AN-s are simulated from additive noise models Y := f (X) + EY . Correspondingly, the LS and LS-s are generated from the location-scale (heteroscedastic) noise models Y := f (X) + g (X) EY , and the MN-U pairs are sampled with the multiplicative noise models Y := f (X) EY as the data generating processes. Injective sigmoid-type functions from B uhlmann et al. (2014) are chosen in contrast to Gaussian processes for the model functions f and g in the MN-U set and the ones with the -s suffix to produce more difficult settings. Except for the MN-U benchmark with uniform noises, other datasets are sampled with the Gaussian distribution for EY . Each dataset contains 100 pairs with 1, 000 samples for each pair. SIM, SIM-c, SIM-G, & SIM-ln (Mooij et al., 2016) In each dataset, similar to the five benchmarks mentioned above, there are also 100 pairs where each pair has the number of samples of 1, 000. The cause-and-effect samples are generated from a more complicated process as X := f X (EX) + MX and Y := f Y (X, EY ) + MY , where EX and EY are sampled from randomly generated distributions P (EX) and P (EY ), f X and f Y are drawn from Gaussian processes, and MX and MY are Gaussian measurement noises. The SIM dataset has this default configuration. In SIM-c, there is a one-dimensional confounder Z included as an additional input of both f X and f Y . The measurement noise levels are reduced in the SIM-ln pairs to create approximately deterministic relations. The cause X in each SIM-G pair follows a distribution that is close to 4SLOPPY: https://eda.rg.cispa.io/prj/sloppy/ 5SLOPE: https://eda.rg.cispa.io/prj/slope/ 6QCCD: https://github.com/tagas/QCCD 7CDT: https://github.com/Fen Tech Solutions/Causal Discovery Toolbox 8GPLVM: https://github.com/Anish144/causal_discovery_bayesian_model_selection 9LOCI: https://github.com/aleximmer/loci 10CAM: https://github.com/cran/CAM Identifying Causal Direction via Variational Bayesian Compression the Gaussian distribution and the corresponding effect Y imitates a nonlinear additive noise model with a Gaussian noise. CE-Multi, CE-Net, & CE-Cha (Goudet et al., 2018) CE-Multi encompasses 300 pairs generated from four families of noise models: pre-additive noise model Y := f (X + EY ), post-additive noise model Y := f (X) + EY , pre-multiplicative noise model Y := f (X EY ), and post-multiplicative noise model Y := f (X) EY , where f can be both linear and nonlinear functions. CE-Net also involves 300 pairs, whose causes are sampled from random distributions and analogous effects synthesized from a neural network with random weights. The CE-Cha set originates from the Cha Learn Cause-Effect Pairs Challenge (Guyon et al., 2019) where the selected 300 pairs only contain either X Y or Y X causal relations with X and Y being continuous variables. Each pair of these datasets includes 1, 500 samples. T ubingen Cause-Effect Pairs (Mooij et al., 2016) This real-world benchmark involves 108 cause-effect pairs collected from various domains, such as meteorology, biology, economy, engineering, medicine, etc. As most related approaches (B uhlmann et al., 2014; Immer et al., 2023; Marx & Vreeken, 2017; 2019a;b; Peters et al., 2014) and COMIC are designed for univariate numeric cause-effect pairs, we adopt existing experimental setups and consider 99 pairs that have one-dimensional continuous causes and effects. Every pair in this dataset is weighted to avoid biased results toward related and similar examples of causal relations. Accessing the Benchmarks All the benchmarks are available in the repository of LOCI (Immer et al., 2023)11. F. Detailed Experimental Results The detailed experimental results in Fig. 1 (Sec. 5) are listed in Tab. 1. 11See Fn. 9 Identifying Causal Direction via Variational Bayesian Compression Table 1. Detailed accuracy and bidirectional AUROC scores of our COMIC approach and baseline methods across all benchmarks. Higher accuracy and bidirectional AUROC (Bi-AUROC) are preferable. The highest scores are listed in bold, the second-best overall results are listed in italic. Our COMIC method is compared against SLOPPY (including the AIC and BIC variants, Marx & Vreeken, 2019a), SLOPER (Marx & Vreeken, 2019b) and SLOPE (Marx & Vreeken, 2017), QCCD (Tagasovska et al., 2020), IGCI (with uniform and Gaussian reference measures, Daniuˇsis et al., 2010), GPLVM (Dhir et al., 2024a), LOCI (Immer et al., 2023), and CAM (B uhlmann et al., 2014). Our COMIC method ranks second overall while maintaining decently consistent performance, as indicated by its lower standard deviations compared to most baseline methods. Method AN AN-s LS LS-s MN-U SIM SIM-c SIM-G SIM-ln CE-Multi CE-Net CE-Cha T ubingen Overall COMIC (Ours) 1.00 1.00 1.00 1.00 1.00 0.57 0.54 0.78 0.89 0.79 0.78 0.43 0.67 0.804 0.200 SLOPPY-AIC 0.75 0.31 0.71 0.12 0.04 0.54 0.64 0.52 0.57 0.90 0.69 0.58 0.71 0.544 0.249 SLOPPY-BIC 1.00 1.00 1.00 0.56 0.96 0.64 0.62 0.81 0.77 0.46 0.79 0.49 0.61 0.749 0.199 SLOPER 0.91 0.56 0.80 0.06 0.10 0.57 0.66 0.60 0.81 0.86 0.73 0.59 0.67 0.609 0.260 SLOPE 0.18 0.28 0.20 0.12 0.07 0.46 0.54 0.46 0.47 0.89 0.62 0.56 0.72 0.429 0.246 QCCD 1.00 0.85 1.00 0.99 0.99 0.64 0.72 0.66 0.85 0.52 0.82 0.54 0.77 0.796 0.171 IGCI-Uniform 0.20 0.35 0.46 0.34 0.11 0.37 0.45 0.53 0.51 0.92 0.56 0.55 0.68 0.464 0.207 IGCI-Gaussian 0.89 0.97 0.95 0.94 0.86 0.36 0.42 0.86 0.59 0.68 0.57 0.55 0.63 0.712 0.210 GPLVM 1.00 1.00 1.00 1.00 1.00 0.83 0.79 0.92 0.90 0.96 0.96 0.75 0.70 0.909 0.105 LOCI 1.00 1.00 1.00 1.00 1.00 0.49 0.50 0.78 0.79 0.72 0.77 0.43 0.56 0.773 0.219 CAM 1.00 1.00 0.98 0.52 0.86 0.56 0.62 0.82 0.87 0.35 0.78 0.53 0.52 0.725 0.217 Bidirectional AUROC COMIC (Ours) 1.00 1.00 1.00 1.00 1.00 0.58 0.59 0.85 0.95 0.91 0.84 0.47 0.75 0.843 0.186 SLOPPY-AIC 0.84 0.27 0.80 0.03 0.00 0.60 0.68 0.58 0.67 0.97 0.78 0.62 0.79 0.587 0.301 SLOPPY-BIC 1.00 1.00 1.00 0.58 1.00 0.74 0.74 0.90 0.90 0.59 0.88 0.54 0.59 0.805 0.182 SLOPER 0.97 0.60 0.87 0.01 0.03 0.67 0.71 0.68 0.93 0.96 0.84 0.62 0.79 0.668 0.313 SLOPE 0.09 0.23 0.12 0.03 0.01 0.48 0.57 0.45 0.46 0.97 0.68 0.59 0.81 0.422 0.307 QCCD 1.00 0.90 1.00 1.00 1.00 0.67 0.80 0.71 0.91 0.56 0.91 0.56 0.79 0.833 0.164 IGCI-Uniform 0.17 0.37 0.51 0.42 0.02 0.38 0.43 0.57 0.51 0.98 0.59 0.58 0.69 0.478 0.235 IGCI-Gaussian 0.98 1.00 0.99 0.99 0.94 0.32 0.38 0.95 0.66 0.78 0.57 0.56 0.64 0.750 0.244 GPLVM 1.00 1.00 1.00 1.00 1.00 0.92 0.91 0.99 0.97 0.98 0.99 0.82 0.80 0.951 0.069 LOCI 1.00 1.00 1.00 1.00 1.00 0.53 0.56 0.85 0.91 0.88 0.82 0.45 0.65 0.819 0.202 CAM 1.00 1.00 1.00 0.35 0.82 0.62 0.65 0.88 0.86 0.38 0.82 0.59 0.38 0.719 0.240 G. Ablation Studies Layer Width of Bayesian Neural Networks Coker et al. (2022) have indicated that mean-field variational inference (MF-VI) of wide Bayesian neural networks can cause the network to ignore the data, where the variational posteriors will collapse to the priors instead of the true posteriors. In this part, we examine the behavior of our models with different hidden layer widths. The result is depicted in Tab. 2. From these results, there are impacts on performance when the width of the hidden layer increases. When the width reaches 100, the reductions in accuracy scores become noticeable. Nonetheless, the Bi-AUROC scores still remain satisfactory. At 200 hidden nodes, we begin to experience severer declines in performance. However, it is important to note that such excessively large width is not recommended in the bivariate setting, where there is only one input dimension and up to two output dimensions. Thus, our approach will not be substantially affected by the ignorance of the data (Coker et al., 2022), if the width of the hidden layer stays within an appropriate range ( 100). Identifying Causal Direction via Variational Bayesian Compression Table 2. The effect of hidden layer width on the performance of COMIC. Higher accuracy and bidirectional AUROC (Bi-AUROC) are preferable. The impact of the width on the Bi-AUROC is tolerable up until 100 nodes. When the width reaches 200 a discouraged choice in the bivariate setting the decline in performance is severe. Width AN AN-s LS LS-s MN-U SIM SIM-c SIM-G SIM-ln CE-Multi CE-Net CE-Cha T ubingen 10 1.00 1.00 1.00 1.00 1.00 0.53 0.53 0.83 0.81 0.76 0.77 0.43 0.45 20 1.00 1.00 1.00 1.00 1.00 0.54 0.53 0.81 0.89 0.77 0.77 0.43 0.60 50 1.00 1.00 1.00 1.00 1.00 0.57 0.54 0.78 0.89 0.79 0.78 0.43 0.67 100 0.98 0.94 0.95 0.98 0.96 0.54 0.55 0.76 0.88 0.79 0.72 0.48 0.63 200 0.97 0.91 0.89 0.98 0.87 0.54 0.53 0.74 0.84 0.76 0.69 0.47 0.57 Bidirectional AUROC 10 1.00 1.00 1.00 1.00 1.00 0.58 0.58 0.90 0.92 0.90 0.83 0.46 0.53 20 1.00 1.00 1.00 1.00 1.00 0.58 0.59 0.89 0.96 0.90 0.83 0.46 0.68 50 1.00 1.00 1.00 1.00 1.00 0.58 0.59 0.85 0.95 0.91 0.84 0.47 0.75 100 1.00 0.99 0.99 1.00 0.99 0.62 0.60 0.84 0.95 0.93 0.80 0.49 0.70 200 0.99 0.97 0.97 0.99 0.96 0.62 0.57 0.81 0.93 0.92 0.76 0.48 0.69 Sparsity-Inducing Priors In addition to the MAP approach for selecting the hyperparameters in App. B, the sparsityinducing prior proposed by Louizos et al. (2017) can be used as an alternative hyperprior. In this prior, the scale hyperparameters follow the non-informative Jeffreys hyperprior p (z) |z| 1 rather than a uniform hyperprior. With this choice of prior, the posteriors of the hyperparameters are inferred using MF-VI with Gaussian distributions adopted as the variational posteriors (Louizos et al., 2017). The Gaussian posteriors of the hyperparameters enable a pruning mechanism in which the weights satisfying log σ2 z log µ2 z > t are pruned from the network. We compare the results between these two hyperpriors in Tab. 3. The results indicate that the uniform hyperprior is generally more effective in comparison to the sparsity-inducing Jeffreys hyperprior on most benchmarks, except for SIM-c. Moreover, pruning yields no significant differences in scores compared to the unpruned case of the Jeffreys hyperprior. Table 3. Performance of COMIC with uniform and Jeffreys hyperpriors (with and without the post-hoc pruning, Louizos et al., 2017). Higher accuracy and bidirectional AUROC (Bi-AUROC) are preferable. The uniform hyperpriors provide better performance on most datasets, except for SIM-c. Hyperprior AN AN-s LS LS-s MN-U SIM SIM-c SIM-G SIM-ln CE-Multi CE-Net CE-Cha T ubingen Uniform 1.00 1.00 1.00 1.00 1.00 0.57 0.54 0.78 0.89 0.79 0.78 0.43 0.67 Jeffreys w/o Pruning 1.00 1.00 1.00 1.00 1.00 0.54 0.62 0.66 0.85 0.84 0.70 0.47 0.58 Jeffreys w/ Pruning 1.00 1.00 1.00 1.00 1.00 0.55 0.62 0.67 0.85 0.84 0.70 0.47 0.58 Bidirectional AUROC Uniform 1.00 1.00 1.00 1.00 1.00 0.58 0.59 0.85 0.95 0.91 0.84 0.47 0.75 Jeffreys w/o Pruning 1.00 1.00 1.00 1.00 1.00 0.58 0.63 0.80 0.93 0.94 0.80 0.47 0.63 Jeffreys w/ Pruning 1.00 1.00 1.00 1.00 1.00 0.57 0.62 0.80 0.92 0.94 0.80 0.47 0.63 Choices for Encoding the Causes While the standard Gaussian distribution is a common choice in previous methods for encoding the marginal distributions of the causes, as discussed in Sec. 4.3, we also investigate a more advanced alternative for modeling the marginal distribution of the causes, which is the variational Bayesian Gaussian mixture model (VB-GMM, Blei & Jordan, 2006). The VB-GMM is a flexible model that can capture multimodal distributions, which may be beneficial in scenarios where the causes exhibit complex distributions. Despite that, the identifiability of the causal direction is more challenging to evaluate theoretically with the VB-GMM. As shown in Tab. 4, VB-GMM substantially improves prediction accuracy and Bi-AUROC over the standard Gaussian on certain synthetic benchmarks, such as SIM, SIM-c, and CE-Net. This improvement, however, is not uniform across datasets: on SIM-G and CE-Multi, performance drops notably, while on Identifying Causal Direction via Variational Bayesian Compression the real-world T ubingen dataset, VB-GMM increases accuracy but slightly reduces Bi-AUROC. Table 4. The effect of different cause encoding choices on the performance of COMIC, including the standard Gaussian and variational Bayesian Gaussian mixture models (VB-GMM) encoding methods. Higher accuracy and bidirectional AUROC (Bi-AUROC) are preferable. Using the more advanced VB-GMM method for encoding the causes yields mixed results, showing significant improvements in performance on some datasets, but reducing on some others, including the real-world T ubingen benchmark. Cause Encoding AN AN-s LS LS-s MN-U SIM SIM-c SIM-G SIM-ln CE-Multi CE-Net CE-Cha T ubingen Standard Gaussian 1.00 1.00 1.00 1.00 1.00 0.57 0.54 0.78 0.89 0.79 0.78 0.43 0.67 VB-GMM 1.00 1.00 1.00 1.00 1.00 0.77 0.71 0.57 0.90 0.74 0.86 0.48 0.73 Bidirectional AUROC Standard Gaussian 1.00 1.00 1.00 1.00 1.00 0.58 0.59 0.85 0.95 0.91 0.84 0.47 0.75 VB-GMM 1.00 1.00 1.00 1.00 1.00 0.80 0.79 0.67 0.95 0.81 0.96 0.47 0.73 H. From Bivariate to Multivariate Settings There are some probable approaches to adapting this work to multivariate settings, which we will explore in this section. Order-Based Causal Discovery The order-based approach is applied by FCM-based methods (Zhang & Hyv arinen, 2009; Peters et al., 2014; Duong & Nguyen, 2023; Lin et al., 2025), which includes choosing a criterion for discovering the topological order of the variables to construct a corresponding full directed acyclic graph (DAG) and then pruning the excessive edges from this full DAG. For example, the multivariate version of RESIT by Peters et al. (2014) utilizes the HSIC (Gretton et al., 2005) score between each variable with the remaining one as a sorting criterion, whose value will be highest in a sink node (the last node in a topological order). In HOST by Duong & Nguyen (2023), with the assumption of standard Gaussian noise in heteroscedastic noise models (HNMs), the normality of the residuals is evaluated with the Shapiro Wilk test after regression, which is expected to be the highest in sink nodes. Lin et al. (2025) present a more generalized criterion based on minimal skewness in sink nodes, which enables the identification of HNMs with symmetric noise distributions. As the joint distribution of a sink node Xj and its set of non-descendants NDG (Xj) in the causal graph G can be factorized as p (Xj, NDG (Xj)) = p (Xj | NDG (Xj)) p (NDG (Xj)) under the principle of independent causal mechanisms, the Kolmogorov complexity can be computed from the Postulate of Algorithmic Independence of Conditionals (Pos. 1, Sec. 3) as K (p (Xj, NDG,j)) += K (p (Xj | NDG,j)) + K (p (NDG,j)). We can expect the factorization of the codelengths corresponding to the sink node as the most succinct. As a consequence, we can choose a node yielding the shortest sum of codelengths as the sink node. Score-Based Causal Discovery MDL-based scores have also appeared in score-based causal discovery literature (Bornschein et al., 2021; Mian et al., 2021) as evaluation metrics for candidate causal graphs. GLOBE from Mian et al. (2021) is an extension of SLOPE (Marx & Vreeken, 2017), which employs the algorithm of greedy equivalence search (GES, Chickering, 2002) for searching in the graph space. Bornschein et al. (2021) also propose a prequential MDL-based score for ranking graphs via their conditional probability distributions of a causal graph. In these approaches, the graphs maximizing the criteria are chosen as the causal graphs. Following this learning scheme, an extension of GPLVM (Dhir et al., 2024a) have also been introduced by Dhir et al. (2024b), which adapt the posterior-based model selection criterion in App. C to multivariate formulation. In addition to these point-estimation approaches, a Bayesian causal discovery approach can also be utilized to infer the full posterior distribution over the causal graphs given the data p (G | D). This posterior is proportional to the product of the marginal likelihood p (D | G) = R p (D | G, θ) p (θ) dθ and a prior over the structures p (G). The posterior p (G | D) can either be estimated in an unsupervised manner, such as those proposed by Cundy et al. (2021); Lorch et al. (2021); Deleu et al. (2022); Tran et al. (2023; 2024b), or learned via supervised training, as demonstrated by Lorch et al. (2022); Dhir et al. (2025). Since the marginal likelihood p (D | G) is equivalent to the Bayesian codelength, which can be effectively Identifying Causal Direction via Variational Bayesian Compression evaluated via the variational Bayesian framework. Hence, our approach is naturally well-suited for this Bayesian approach to multivariate causal discovery.