# robust_model_selection_of_gaussian_graphical_models__90a2f933.pdf Published in Transactions on Machine Learning Research (04/2025) Robust Model Selection of Gaussian Graphical Models Abrar Zahin azahin@asu.edu School of Electrical, Computer & Energy Engineering Arizona State University Rajasekhar Anguluriú rajangul@umbc.edu Department of Computer Science & Electrical Engineering University of Maryland, Baltimore County Lalitha Sankar lsankar@asu.edu School of Electrical, Computer & Energy Engineering Arizona State University Oliver Kosut okosut@asu.edu School of Electrical, Computer & Energy Engineering Arizona State University Gautam Dasarathy gautamd@asu.edu School of Electrical, Computer & Energy Engineering Arizona State University Reviewed on Open Review: https: // openreview. net/ forum? id= AIby9MQXbu In Gaussian graphical model selection, noise-corrupted samples present significant challenges. It is known that even minimal amounts of noise can obscure the underlying structure, leading to fundamental identifiability issues. A recent line of work addressing this robust model selection problem narrows its focus to tree-structured graphical models. Even within this specific class of models, exact structure recovery is shown to be impossible. However, several algorithms have been developed that are known to provably recover the underlying tree-structure up to an (unavoidable) equivalence class. In this paper, we extend these results beyond tree-structured graphs. We first characterize the equivalence class up to which general graphs can be recovered in the presence of noise. Despite the inherent ambiguity (which we prove is unavoidable), the structure that can be recovered reveals local clustering information and global connectivity patterns in the underlying model. Such information is useful in a range of real-world problems, including power grids, social networks, protein-protein interactions, and neural structures. We then propose an algorithm which provably recovers the underlying graph up to the identified ambiguity. We further provide finite sample guarantees in the high-dimensional regime for our algorithm and validate our results through numerical simulations. 1 Introduction Probabilistic graphical models have emerged as a powerful and flexible formalism for expressing and leveraging relationships among entities in large interacting systems (Lauritzen, 1996). They have found application in úThe work was primarily done while R.A. was at Arizona State University. Published in Transactions on Machine Learning Research (04/2025) a range of areas including signal processing (Kim & Smaragdis, 2013; Ott & Stoop, 2006; Murphy et al., 2013), power systems (Anguluri et al., 2022; Deka et al., 2020; 2015), (phylo)genomics (Zuo et al., 2017; Dasarathy et al., 2014; 2022), and neuroscience (Bullmore & Bassett, 2011; Vinci et al., 2019). Gaussian graphical models are an important subclass of graphical models and are the main focus of this paper; our techniques do apply more broadly, as discussed in Section 7. c In several applications, we do not know the underlying graph structure, and the goal is to learn this from data a problem dubbed graphical model selection. This is important because the graph structure provides a succinct representation of the complex multivariate distribution and can reveal important relationships among the underlying variables. See, e.g., Drton & Maathuis (2017); Maathuis et al. (2018) and references therein for more on this problem. Here, we focus on a relatively new but important task where samples from the underlying distribution are corrupted by independent noise with unknown variances. This occurs in a wide variety of applications where sensor data or experimental measurements suffer from statistical uncertainty or measurement noise. In these situations, we refer to the task of graph structure learning as robust model selection. This problem was recently considered by Katiyar et al. (2019) and a line of follow-up work (Katiyar et al., 2020; Casanellas et al., 2024; Tandon et al., 2021) who show that unfortunately the conditional independence structure of the underlying distribution can be completely lost in general under such corruption; see Section 3.1 for more on this. The authors show that even in the often tractable case of tree-structured graphical models, one can only identify the structure up to an equivalent class. In fact, the assumption that the underlying uncorrupted graphical model has a tree structure is critical to the techniques of this line of work. As Casanellas et al. (2024) astutely observe, when the random vector associated with the true underlying graph is corrupted with independent but non-identical additive noise, the robust estimation problem reduces to a latent tree structure learning problem. We improve on the algorithmic and theoretical results of this line of work by considering the robust model selection problem for general graphs. Our main contributions are summarized below. We establish a fundamental identifiability result (c.f. Theorem 3.1) for general graphs in the robust Gaussian graphical model selection problem. This confirms that the identifiability problem is exacerbated if one considers more general graphs. More importantly, this also generalizes the identifiability results from earlier lines of work and identifies an equivalence class up to which one may hope to recover the underlying structure. We devise a novel algorithm, called No MAD (for Noisy Model selection based on Ancestor Discovery), that tackles the robust model selection problem for general graphs extending the results of (Katiyar et al., 2019; 2020; Casanellas et al., 2024; Tandon et al., 2021). Our algorithm is based on a novel ancestor discovery procedure (see Section 3.4) that we expect to be of independent interest. It is worth observing that the tree-based algorithms previously proposed fail, often catastrophically, when there are loops in the underlying graph. We show that No MAD provably recovers the underlying graph up to a small equivalence class (c.f. Theorem 4.2) and establish sample complexity results (c.f. Theorem 5.3) for partial structure recovery in the high-dimensional regime. We also show the efficacy of our algorithm through experiments on synthetic and realistic network structures. 2 Related Work Several lines of research have tackled the problem of robust estimation of high-dimensional graphical models under corruption. This includes graphical modeling with missing data, outliers, or bounded noise, see, for instance, Loh & Wainwright (2011); Chen et al. (2013); Wang & Lin (2014); Nguyen et al. (2022) and references therein. For the missing data problem, several other algorithms have been proposed for estimating mean values and covariance matrices from the incomplete dataset available to the learner Little & Rubin (2019); Schneider (2001); Lounici (2014). Zheng & Allen (2024) considered a variant of the missing value problem where instead of missing values, the measurements are irregular; that is, different vertex pairs have Published in Transactions on Machine Learning Research (04/2025) vastly different sample sizes. Vinci et al. (2019); Chang et al. (2023); Dasarathy (2019) explored the situations where one is only able to obtain samples from subsets of variables, possibly missing joint observations from several pairs. Sun & Li (2012) and Yang & Lozano (2015) proposed algorithms for handling the outliers. There is another line of work that treats this problem using the error-in-variables lens (see the books and papers Hwang (1986); Carroll et al. (1995); Iturria et al. (1999); Xu & You (2007) and references therein). For the problem of model selection from bounded noisy measurements, see Wang & Lin (2014); Öllerer & Croux (2015); Loh & Tan (2018); Chen et al. (2015). However, these papers do not consider the setting of unknown additive noise and the corresponding implications on the conditional independence structure. Recently, Nikolakakis et al. (2019) considered recovering foreststructured graphical models assuming that noise distribution across all vertices is identical. In contrast, our setting allows for unknown and non-identical noise. The robust model selection problem, as considered here, had not been adequately addressed even for the tree-structured graphical models until the recent work by Katiyar et al. (2019; 2020) who showed that the structure recovery in the presence of unknown noise is possible only up to an equivalence class. These studies also proposed algorithms to recover the correct equivalence class from noisy samples. Using information-theoretic methods, Tandon et al. (2021) improved the sample complexity result of Katiyar et al. (2020); Nikolakakis et al. (2019) and provided a more statistically robust algorithm for partial tree recovery. Finally, Zhang & Tan (2021) studied the structure recovery problem under noise when the nodes of the GGM are vector-valued random variables. However, these results are limited to tree-structured graphical models, as they inherently use the additive distance metric property to learn the full (or partial) structure. However, the additive distance metric does not hold for general structures (see more on Section 3.3). The results of this paper significantly extend this line of work, and are applicable to general graphs. 3 Preliminaries and Problem Statement Graph theory. Let G = (V, E) be an undirected graph on vertex set V (with cardinality p) and edge set E µ . For a vertex v œ V , let Nv , {u œ V : {u, v} œ E} be the neighborhood of the vertex v œ V and degree deg(v) be the size of Nv. A vertex v is said to be leaf if deg(v) = 1. A subgraph of G is any graph whose vertices and edges are subsets of those of G. For V Õ V the induced subgraph G(V Õ) has the vertex set V Õ and the edge set EÕ = {{u, v} œ E : u, v œ V Õ}. A path between the vertices u, v is a sequence of distinct vertices v1 = u, v2, . . . , vk = v such that {vi, vi+1} œ E, for 1 Æ i < k. We let Puv denote the set of all paths between u and v. If Puv is not empty, we say u and v are connected. The graph G is connected if every pair of vertices in G is connected. A set S V separates two disjoint subsets A, B V if any path from A to B contains a vertex in S. We denote this separation as A B | S. The resemblance of this notation to that of the conditional independence of random variables in the graphical model will be made clear later. Gaussian graphical models. Let X = (X1, X2, . . . , Xp) œ Rp be a zero-mean Gaussian random vector with a covariance matrix Σ œ Rp p. Compactly, X N(0, Σ), where 0 is the p-dimensional vector of all zeros. Let G = ([p], E) be a graph on the vertex set [p] , {1, 2, . . . , p} representing the coordinates of X. Let K , Σ 1 is called the precision matrix of X. The distribution of X is said to be a Gaussian graphical model (or equivalently, Markov) with respect to G if Kij = 0 for all {i, j} /œ E 1. In other words, for any {i, j} /œ E, Xi and Xj are conditionally independent given all the other coordinates of X (see Lauritzen (1996) for more details). In the sequel, we will use a generic set V to denote the vertex set of our graph with the understanding that every element in V is uniquely mapped to a coordinate of the corresponding random vector X. For a vertex v œ V , with a slight abuse of notation, we write Xv to denote the corresponding coordinate of X. Similarly, we write Σuv to mean the covariance between Xu and Xv. 3.1 The Robust Model Selection Problem In this paper, we consider a variant of the model selection problem, which we refer to as robust model selection. Formally, let X N(0, Σ) be a Gaussian graphical model with respect to an unknown graph G = (V, E). In the robust model selection problem, the goal is to estimate the edge set E (or equivalently the sparsity 1Hence, the sparsity pattern of K is represented by the edge set of G Published in Transactions on Machine Learning Research (04/2025) pattern of K = Σ 1) when one only has access to noisy samples of X. That is, we suppose we have access to the corrupted version Y of the underlying random vector X such that Y = X + Z, where the noise Z N(0, D) is independent of X and D is assumed to be diagonal with possibly distinct and even zero entries. In other words, the noise is assumed to be independent and heteroscedastic while potentially allowing for some coordinates of X to be observed uncorrupted. Observe that Y N(0, Σo), where Σo , Σ + D. Indeed, D is assumed to be unknown. Unfortunately, such corruption can completely obliterate the conditional independence structure of X. For instance, suppose that D = eje T j , where ej is a vector of zeros except in the jth entry where it is one. By the Sherman-Morrison identity (see e.g., Horn & Johnson, 2012), we have (Σo) 1 = K c Keje j K for some c Ø 0. The term Keje j K can be dense in general and, hence, can fully distort the sparsity of K (the conditional independence structure of X). In view of the above example, the robust model selection problem appears intractable. As outlined in Section 1, recent studies show that this problem is partially tractable for tree-structured graphs. Notably, as Casanellas et al. (2024) astutely observes, one can reduce the problem of robust structure estimation of trees to the problem of learning the structure of latent tree graphical models, which enjoy several efficient algorithms and rich theoretical results (see, e.g., Choi et al. (2011); Erdös et al. (1999); Semple & Steel (2003); Dasarathy et al. (2014)). To see this, suppose that X is Markov according to a graph G = (V = [p], E) that is tree-structured. Let the joint graph Gj denote the graph obtained by creating a copy of each node in G and linking the copies to their counterparts in G. Formally, we define Gj = (V j, Ej), where V j , V fi{1e, 2e, . . . , pe} and Ej , E fi{{i, ie} : i œ [p]}. In subsequent sections, for any vertex subset B 2V , we let Be denote the vertices associated with the noisy observations from B and call it the noisy counterpart of B. Notice that with this definition, if we associate the coordinates of Y to the newly added leaf vertices, the concatenated random vector [X; Y], obtained by stacking X on top of Y, is Markov according to Gj. Casanellas et al. (2024) then uses the fact that when given samples of Y, one can reconstruct a reduced latent tree representation of Gj, which in turn can be used to infer an equivalence class of trees that contains the true tree G. Indeed, the equivalence class thus obtained is the same one identified by Katiyar et al. (2019; 2020). We next state our general identifiability result after introducing some more graph-theoretic concepts. 3.2 An Identifiability Result A connected graph G is said to be biconnected if at least 2 vertices need to be removed to disconnect the graph. A subgraph H of G is said to be a biconnected component if it is a maximal biconnected subgraph of G. That is, H is not a strict subgraph of any other biconnected subgraph of G. The vertex set of such a biconnected component will be referred to as a block. A block is non-trivial if it has more than two vertices. For example, in Fig. 1a, the vertex set {1, 2, 3, 4} and B1 fi{6, 8} are non-trivial blocks where B1 is an arbitrary set of vertices such that the subgraph on B1 fi{6, 8} is a biconnected component, whereas, the set {10, 8} is a trivial block. In what follows, we will often be interested in the vertices of such non-trivial blocks and toward this we write Bnt to denote the set of all vertex sets of non-trivial blocks in G. From these definitions, it follows that trees (which are cycle free) do not have any non-trivial blocks. It also follows that two blocks can share at most one vertex; we refer to such shared vertices as cut vertices. In Fig. 1a, the vertices 4 and 10 are cut vertices. The vertices in Bnt which are not cut vertices are referred to as non-cut vertices. In Fig. 1a, the vertex 1 is a non-cut vertex. With these definitions, we now introduce a novel representation for a graph G that will be crucial to stating our results. This representation is a tree-structured graph Tast(G), which we call the articulated set tree of G, whose vertices correspond to (a) non-trivial blocks in G, and (b) vertices in G that are not a member of any non-trivial blocks. Vertices in this tree-structured representation are connected by edges if the corresponding sets in G either share a vertex or are connected by a single edge. The vertices in the original graph that are responsible for the edges in the representation are called articulation points2. We formally define this below. For example, for G in Fig. 1a, the articulated set tree is illustrated in Fig. 1c, where the sets {1, 2, 3, 4} and 2Articulation (points) vertices are cut vertices that separate non-trivial blocks from the rest of the graph. Published in Transactions on Machine Learning Research (04/2025) (c) Tast(G) Figure 1: (a) a true underlying graph where both B1 fi{6, 8} (B2 fi{7, 9}) are non-trivial blocks where B1 (B2) is an arbitrary set of vertices such that the subgraph on B1 fi{6, 8} (B2 fi{7, 9}) is a biconnected component, (b) joint graph Gj; noisy vertices associated with the non-trivial blocks containing B1 and B2, and some other vertices are not numbered to reduce the clutter, and (c) the articulated set tree Tast(G). {17, 18, 19} are associated to vertices in the articulated set tree representation, and 4, 6, 7, 14, and 17 are examples of articulation points. Definition 3.1 (Articulated Set Tree). For an undirected graph G = (V, E), the articulated set tree Tast(G) is a tuple (P, E, A) where (a) the set P = {B : B œ Bnt} fi{{v} : v œ V \ fiBœBnt B}; (b) an edge {P, P Õ} œ E if and only if (i) vertices v, vÕ œ V are such that v œ P, vÕ œ P Õ, and {v, vÕ} œ E or (ii) there exists a vertex v œ V such that v œ P flP Õ, and (c) the articulation function A : E æ V V returns the articulation points of each edge. Notice that the articulated set tree (AST), as the name suggests, is indeed a tree. Otherwise, by definition, the set of non-trivial blocks Bnt would be incorrect (we show this formally in Lemma B.1 in the Appendix). Readers familiar with graph theory may have observed that the AST representation is quite similar to the block-cut tree representation (see e.g., Harary, 1971; Biggs et al., 1986), but unlike a block-cut tree, the subgraph associated with any non-trivial block does not matter in the articulated set tree. We will now define the equivalence class of graphs up to which robust recovery is possible. Let L(G) denote the set of all leaves in G (i.e., all vertices of degree one). A subset R µ L(G) is said to be remote if no two elements of R share a common neighbor. Let R be the set of all remote subsets of L(G). For each R œ R, define a graph GR on V by exchanging each vertex in R with its (unique) neighbor. Definition 3.2 (Equivalence Relation, ). Two graphs G, H are said to be equivalent if and only if R œ R such that Tast(GR) = Tast(H). Symbolically, we write as H G. We let [G] denote the equivalence class of G with respect to . It is not hard to verify that Definition 3.2 is a valid equivalence relation. Furthermore, it can be readily checked that this notion of equivalence subsumes the ones defined for trees in Katiyar et al. (2019); Casanellas et al. (2024). Fig. 2 illustrates three graphs from the same equivalence class. Notice that G1 can be constructed from G by: (i) exchanging the labels between the leaf vertices {13, 17, 15} with their corresponding neighbors {10, 21, 14}; (ii) adding an edge {2, 4} inside a non-trivial block. Similarly, G2 can be constructed from G by: (i) exchanging the labels between the leaf vertices {11, 20} with their corresponding neighbors {10, 21}; (ii) removing an edge {1, 4} from a non-trivial block3. Therefore, for any two graphs in the equivalence class, the non-cut vertices of any non-trivial block remain unchanged, whereas, the edges in the non-trivial block can be arbitrarily changed; the labels of the leaves can be swapped with their neighbor. In the following we will show that our identifiability result also complements the equivalence class. Finally, notice that with the similar operation described above, given a Tast which equals to Tast(G), one can recover all the graphs in [G]. We now state our unidentifiability result, which establishes that the true covariance matrix of any graph in the equivalence class [G], under the noise model of Subsection 3.1, will result in the same observed covariance matrix Σo. Theorem 3.1 (Undentifiability). Fix a covariance matrix Σú whose conditional independence structure is given by the graph G. Suppose we are given a noisy covariance matrix Σo = Σú + D where D = 3Notice that the sets {13, 17, 15} and {11, 20} are remote according to Definition 3.2. Published in Transactions on Machine Learning Research (04/2025) Figure 2: An illustration of three graphs from the same equivalence class of G in Fig. 1a. For all three graphs, B1 fi{6, 8} (B2 fi{7, 9}) can have any induced subgraph as long as subgraphs on B1 fi{6, 8} (B2 fi{7, 9}) is a biconnected component. diag(D11, . . . , Dpp) Ø 0. Then, for any H œ [G] where H = G, there exists matrices ΣH, DH such that Σo = ΣH + DH, DH is a diagonal matrix with non-negative entries, and the sparsity pattern of (ΣH) 1 is described by H. This theorem is proved in Section D. Notice that this undentifiability result shows it is impossible to uniquely recover G as there are other confounding graphs whose noisy observations would be indistinguishable from those of G. However, this theorem does not rule out the possibility of recovering the equivalence class to which G belongs since all the confounding examples are confined to [G]. In Section 3.3, we devise an algorithm that precisely does this. Before we conclude this section, we introduce the notion of partial structure recovery and discuss how recovering the equivalence class still reveals useful information about the true graph. Partial Structure Recovery of G. Given the noise model described in Subsection 3.1 we know that any graph is only identifiable upto the equivalence relation in Definition 3.2. This is not only the best one can do (by Theorem 3.1), but also preserves useful partial structure of the graph. In particular, such a partial structure recovery is able to identify the (non-cut) constituents of the non-trivial blocks and the set of leaf vertices (and neighbors thereof). As the following examples illustrate, we conclude this subsection by arguing that even such partially recovered graphs are instrumental in several application domains. 1. Electrical distribution networks usually have radial (or globally tree-like) network structures. Recently, several parts of such networks have become increasingly locally interconnected due to the adoption of technologies like roof-top solar panels and battery power storage that enable more flexible power flow. Nonetheless, practitioners critically rely on (learning) the global structure for most operations and maintenance tasks, such as state estimation, power flow, and cybersecurity. 2. Neuronal networks. Network models are commonly used to describe structural and functional connectivity in the brain Sporns (2018). An important property of networks that model the brain is their modular structure; modules or communities correspond to clusters of nodes that are densely connected (and are presumably functionally related). Such modular structure has long been regarded as a hallmark of many complex systems; see Simon (2012) for more details. Module detection, that is, understanding which nodes belong to which modules can yield important insights into how networks function and also uncover a network s latent community structure. 3.3 The Robust Model Selection Algorithm We now present an algorithm that can recover the partial structure of a graph G for which only noisy samples are available based on the setup from Subsection 3.1. Before we describe our algorithm, we introduce a few more concepts that will play a key role. We start with a well-known fact about the factorization of pairwise correlations for a faithful4 Gaussian graphical model. First, recall that for two random variables Xi and Xj, the correlation coefficient is defined as flij , Σij/ 4The global Markov property for GGMs ensures that graph separation implies conditional independence. The reverse implication need not to hold. However, for a faithful GGM, the reverse implication does hold. Published in Transactions on Machine Learning Research (04/2025) Fact 1 (see e.g., Soh & Tatikonda (2014)). For a faithful Gaussian graphical model, Xi Xk|Xj if and only if flik = flij fljk. Figure 3: Graph with multiple minimal mutual separators. We now define information distances, which play a vital role in designing our algorithm. Definition 3.3 (Information distances). For (X1, . . . , Xp) N(0, Σ), the information distance between Xi and Xj is defined by dij , log |flij| Ø 0, where flij is the pairwise correlation coefficient between Xi and Xj. For tree-structured Gaussian graphical models, the strength of the correlation dictates the information (or graphical) distance between vertices i and j. Higher the strength, the smaller is the distance, and vice versa. In fact, the information distance defined this way is an additive metric on the vertices of the tree. Although the graphs we consider are not necessarily trees, we still refer to this quantity as a distance throughout the paper for convenience. We now define the notion of ancestors for a triplet of vertices using the notion of minimal mutual separator. Definition 3.4 (Minimal mutual separators, Star triplets, Ancestors). Fix a triplet of vertices U œ . A vertex set S V is called a mutual separator of U if S separates each pair i, j œ U; that is, every path fiœ Pij contains at least one element of S. The set S is called a minimal mutual separator of the triplet U if no proper subset of S is a mutual separator of U. We let Smin(U) denote the set of all minimal mutual separators of U. U is said to be a star triplet if |Smin(U)| = 1 and the separator in Smin(U) is a singleton. The unique vertex that mutually separates U is called the ancestor of U. For instance, for the graph in Fig. 3, the set {2, 4, 7, 8, 3, 5} is a mutual separator for the triple {1, 6, 9}. Further notice that minimal mutual separator set may not be unique for a triplet: here, the sets {2, 3, 7} and {4, 5, 8} are minimal mutual separators of the triple {1, 6, 9}. In Fig. 1a, for the triplet {1, 11, 12}, minimal mutual separator Smin ({1, 11, 12}) = {10}. Further, notice that an ancestor of U can be one of the elements of U. In Fig. 1a, the vertex {10} is the ancestor of triplet {4, 10, 11} , U. Notice that for triplet U, the distance between 10 and the ancestor is zero. For a star triplet {i, j, k} with ancestor r, it is clear that the following holds true based on the relationship between graph separation and conditional independence: Xi Xj | Xr, Xi Xk | Xr, and Xj Xk | Xr. As a consequence, from Fact 1 and Definition 3.3, the pairwise distances dij, dik, and djk satisfy the following equations: dij = dir + drj, dik = dir + drk, and djk = djr + drk. Some straightforward algebra results in the following identities that allows us to compute the distance between each vertex in {i, j, k} and the ancestor vertex r. In particular, for any ordering {x, y, z} of the set {i, j, k} notice that the following is true: dxr = 0.5 (dxy + dxz dyz) (1) For a triplet U = {i, j, k}, we will let d U 2(dij + dik djk). If U is a star triplet, then d U i reveals the distance between i and the ancestor of U. However, we do not restrict this definition to star triplets alone. When U is not a star triplet, d U i is some arbitrary (operationally non-significant) number; in fact, for non-star triplets this quantity may even be negative. Notice that all vertex triplets in a tree are star triplets. Hence, for a tree-structured graphical model, we can choose any arbitrary triplet and if we can find a vertex r for which d U i = dir, i œ U, then we can identify the ancestor of U. If such a vertex does not exist, we can deduce the existence of a latent ancestor. Therefore, iterating through all possible triplets, one can recover the true (latent) tree structure underlying the observed variables. In fact, several algorithms in the literature use similar techniques to learn trees (see e.g., Saitou & Nei (1987); Krishnamurthy & Singh (2012); Choi et al. (2011); Dasarathy et al. (2014). 3.4 The No MAD Algorithm In this section, we introduce our algorithm No MAD, for Noisy Model selection based on Ancestor Discovery, for robust model selection. We describe No MAD in the population setting (i.e., in the infinite sample limit) for clarity of presentation; the modifications required for the finite sample setting are discussed in Section 5. Published in Transactions on Machine Learning Research (04/2025) Figure 4: (a) The joint graph Gj, (b) the leaf clusters and internal clusters of Gj; Be 2) denote the set of noisy vertices associated with the vertices in B1 (B2); Also, recall that clusters are a set of vertices; grey vertices are identified but unlabeled articulation points associated with the clusters, (c) non-trivial blocks and trivial blocks along with the identified and labeled articulation points, and (d) the edges between different articulation points. In the population setting, No MAD takes as input the (exact) pairwise information distances dij, for all i, j in the observed vertex set V o, and returns an articulated set tree Top , (Pop, Aop, Eop) (see Definition 3.1). A high level overview of the algorithm is given in Algorithm 1. Its operation may be divided into two main steps: (a) learning Pop and Aop; and (b) learning Eop. These steps are summarized in the following. A formal algorithmic listing and a full description can be found in the Appendix. (a) Learning Pop and Aop. Inspired by the aforementioned ancestor based tree reconstruction algorithms, No MAD first identifies the ancestors5 in Gj, and learns the pairwise distances between them. Notice that finding the ancestors in Gj is challenging for the following reasons: (a) since Gj is not a tree, some vertex triplets are not star triplets (e.g., {1e, 3e, 11e} in Fig. 1b), and (b) a subset of vertices (which may include ancestors) in Gj are unobserved or latent. Hence, we can not guarantee the identification of a star triplet following the procedure for trees. No MAD instead uses a novel procedure that compares two triplets of vertices to identify the ancestors; we call this the TIA (Test Identical Ancestor) test which is defined as follows: Algorithm 1 No MAD 1: Input: Pairwise distances D = {dij}i,jœV o. 2: Output: Top , (Pop, Aop, Eop). 3: Identify Ancestors(Subroutine 3). Accepts: D; Returns: (I) A set Aobs (Ahid resp.) of observed (hidden resp.) ancestors, and the corresponding collection of vertex triplets Vobs (Vhid resp.), and (II) The set of pairwise distances {dij} for each pair i, j œ V o fiAhid 4: Learn Clusters (Subroutine 4). Accepts: Aobs, Ahid, and D; Returns: A collection of (I) leaf clusters L, and (II) internal clusters I. 5: Vertex Set-AST(Subroutine 6).Accepts: L and I; Returns: (I) the vertex set Pop (II) the articulation points Aop, and (III) a subset of the edge set Eop. 6: Edge Set-AST(Subroutine 8). Accepts: Pop and Aop; Returns: Eop. Definition 3.5 (TIA Test). The Test Identical Ancestor (TIA) test accepts a triplet pair U, W œ , and returns True if and only if for all x œ U, there exists at least one pair y, z œ W such that d U y = dxy and d U In words, in order for a pair of triplets U, W to share an ancestor in Gj, each vertex in one triplet (say, U) needs to be separated from at least a pair in W by the (shared) ancestor in Gj. We now describe the fist step of the No MAD in three following sub-steps: Identifying the ancestors in Gj. In the first sub-step, No MAD (a) uses the TIA test to create the set V = {V µ : each U œ V has the same ancestor }, (b) it then assigns each triplet collection V œ V to 5We say that a vertex r is an ancestor if there is a triple U such that r is an ancestor of U. Published in Transactions on Machine Learning Research (04/2025) either Vobs or Vhid; the former is the collection of vertex triples whose ancestor is observed and the latter has ancestors that are hidden, and (c) identifies the observed ancestors and enrolls them into a set of observed ancestors Aobs. Furthermore, for each collection Vi œ Vhid, No MAD introduces a hidden vertex, and enrolls it in Ahid such that, |Ahid| equals to the number of hidden ancestors in Gj. For example, consider the joint graph Gj in Fig. 4a. In Gj, the vertex {4} is the observed ancestor of the pair {1e, 4e, 10} and {3e, 4e, 8e}, and the vertex {8} is the hidden ancestor of the pair {3e, 8e, 9e}, {1e, 8e, 7e}. Complete pseudocode for this step appears in Subroutine 3 in Appendix. Extending the distance set {dij}i,jœV o. In the next sub-step, using pairwise distances {dij}i,jœV o, and Ahid, No MAD learns the following distances: (a) dij for all i, j œ Ahid, and (b) dij for all i, j œ AhidfiV o. For learning (a), notice from the last sub-step that each ai œ Ahid is assigned to a collection of triplets Vi œ Vhid. In order to compute the distance between two hidden ancestors (say ap, aq œ Ahid), the No MAD chooses two triplets Vp œ Vp and Vq œ Vq, and computes the set pq as follows: pq = dxy (d Vp x + d Vq y ) : x œ Vp, y œ Vq . Then, the most frequent element in pq, i.e., mode( pq), is declared as dpq. We show in Appendix that No MAD not only correctly learns the distance set {dij}i,jœAhid, but also learns dij for any i œ Ahid and j œ V o. A pseudocode for this step appears in Subroutine 3 in Appendix. Learning Pop and Aop. In the final sub-step, No MAD learns the clusters of vertices in V o \ Aobs using the separation test in Fact 1 which eventually lead to finding Pop and Aop. Specifically, No MAD learns (a) all the leaf clusters, each of which is a set of vertices that are separated from the rest of the graph by a single ancestor, and (b) all the internal clusters, each of which is a set of vertices that are separated from the rest of the graph by multiple ancestors. For example, in Fig. 4b, the set {17e, 18e, 19e, 20e, 21e} is a leaf cluster separated from the rest of the graph by the (hidden) ancestor a9. The set Be 2 is an internal cluster separated from the rest of the graph by the set of ancestors {a5, a6, a7}. Next, No MAD uses the clusters to learn Pop and Aop for Top by applying the TIA test on each cluster to identify the non-cut vertices and potential cut vertices in it. For example, for the leaf cluster {17e, 18e, 19e, 20e, 21e}, 18 and 19 are non-cut vertices, and a vertex from 17, 20, and 21 may be declared as a cut-vertex arbitrarily (see Fig. 4c). A pseudocode for this step appears as Subroutine 4 and Subroutine 6 in Appendix. (b) Learning Eop. In this step, No MAD learns the edge set Eop. Notice from Definition 3.1 that any two elements of Pop are connected with each other through their respective articulation points. Hence, in order to learn Eop, No MAD needs to learn the neighborhood of each articulation point in G. To this end, No MAD first learns this neighborhood for each articulation point in G using Fact 1. Then, in the next step, No MAD creates an edge between two elements of Pop if the articulation points from each element are neighbors in G (dotted lines in Fig. 4d). A pseudocode for this step appears as Subroutine 8 in Appendix. 4 Performance Analysis of No MAD in the Population Setting In this section, we show the correctness of No MAD in returning the equivalence class of a graph G while having access only to the noisy samples according to the problem setup in Section 3.1. We now make an assumption that will be crucial to show the correctness of No MAD. This is similar to the faithfulness assumption common in the graphical modeling literature Choi et al. (2011); Kalisch & Bühlman (2007); Uhler et al. (2013), and like the latter, it rules out spurious cancellations . To that end, let Vstar be the set of all star triplets in G (see Definition 3.4). Let Vsep be the set of triplets V such that one of the vertices in V separates the other two vertices. Assumption 4.1 (Ancestor faithfulness). Let U, W œ \ Vstar fiVsep. Then, (i) there are no vertices x œ U and a œ W that satisfy d U a = dxa, and (ii) there does not exist any vertex r œ V and x œ U for which the distance dxr satisfies relation in equation 1. Published in Transactions on Machine Learning Research (04/2025) Notice that Assumption 4.1 is only violated when there are explicit constraints on the corresponding covariance values which, like the faithfulness assumption, only happens on a set of measure zero. We next state the main result of our paper in the population settings. Theorem 4.2. Consider a covariance matrix Σú whose conditional independence structure is given by the graph G, and the model satisfies Assumption 4.1. Suppose that according to the problem setup in Section 3.1, we are given pairwise distance dij of a vertex pair (i, j) in the observed vertex set V o, that is, dij , log|flij| where flij , Σo jj. Then, given the pairwise distance set {dij}i,jœV o as inputs, No MAD outputs the equivalence class [G]. Proof Outline. In order to show that No MAD correctly learns the equivalence class, it suffices to show that it can correctly deduce the articulated set tree Top. Given this, and the equivalence relation from Definition 3.2, the entire equivalence class can be readily generated. Our strategy will be to show that No MAD learns Top correctly by showing that it learns (a) the vertex set Pop, (b) the articulation points Aop, and (c) the edge set Eop correctly. We will now establish (a) and (b). From the description of the algorithm in Section 3.3, it is clear that No MAD succeeds in finding the ancestors, which is the first step, provided the TIA tests succeed. Indeed, in the first stage of our proof, we establish Lemma B.7 in Appendix which shows that the TIA test passes with two triplets if and only if they share a common ancestor in Gj. Next, we show in Lemma B.9 and Claim 3 in Appendix that No MAD correctly learns the distances dij for all vertices i, j that either in the set of observed vertices V o or in the set of hidden ancestors Ahid. Proposition B.13 establishes the correctness of No MAD in learning the leaf clusters and internal clusters, and in learning Pop and Aop. The proof correctness of this step crucially depends on identifying the non-cut vertices in G from different clusters which is proved in Lemma B.12. Finally, we outline the correctness of No MAD in learning the edge set Eop. Recall that No MAD learns the neighbor articulation points of each articulation point. Proposition B.14 in Appendix shows that No MAD correctly achieves this task. Using the neighbors of different articulation points, No MAD correctly learns the edges between different elements in Pop. 5 Performance Analysis of No MAD in Finite Sample Setting In describing No MAD (cf. Section 3.4) and in the analysis in the population setting (cf. Section 4), we temporarily assumed that we have access to the actual distances dij for the sake of exposition. However, in practice, these distances need to be estimated from samples {Y1, . . . , Yn}. In what follows, we show that No MAD, with high probability, correctly outputs the equivalence class [G] even if we replace dij with the estimate dij = log ----, where Σoij is the (i, j)-th element in 1 i . We recall from Section 3.4 that the subroutines in No MAD depend on the TIA test, which relies on the distances dij. Thus, we establish the correctness of No MAD in the finite sample setting by showing that the empirical TIA (TIA with estimated distances) correctly identifies the ancestors in GJ with high probability. We begin with the following assumptions. Assumption 5.2. [ -Strong Faithfulness Assumption] For any vertex triplet i, j, k œ , if i j|k, then |dij dik djk| > . -Strong Faithfulness assumption is a standard assumption used in the literature (Kalisch & Bühlman (2007); Uhler et al. (2013)). 0-Strong-Faithfulness is just the usual Faithfulness assumption discussed in Section 3.3. This motivates our next assumption which strengthens our requirement on spurious cancellations involving ancestors. Assumption 5.3 (Strong Ancestor consistency). For any triplet pair U, W œ \ Vstar fiVsep and any vertex pair (x, a) œ U W, there exists a constant > 0, such that Published in Transactions on Machine Learning Research (04/2025) (a) Synthetic graph, Gsyn (b) IEEE 33 Bus Graph Figure 5: Synthetic graph and IEEE-33 Bus system considered for our simulation Assumption 5.3 is in direct analogy with Assumption 5.2. As we show in Lemma B.7, for any pair U, W œ \ Vstar fiVsep (i.e., any pair that would fail the TIA test), there exists at least one triplet {x, a, b} where x œ U and a, b œ W such that dxa d U a = 0 and dxb d U b = 0. This observation motivates us to replace the exact equality testing in the TIA test in Definition 3.5 with the following hypothesis test against zero: max Ó--- dxa d U --- dxb d U Æ . We set < Furthermore, in order to learn the distance between two hidden ancestors in Ahid, the mode test introduced in Subsection 3.4 needs to be replaced with a finite sample version; we call this the d-mode test and this is formalized in Appendix C. Finally, for any triplet (i, j, k) œ , in order to check whether i j|k, the test in Fact 1 needs to be replaced as follows: | dij dik djk| < d 6 . We now introduce two new notations to state our main result. Let flmin(p) = mini,jœ( p 2)|flij| and Ÿ(p) = log((16 + (flmin(p))2 2 d)/(16 (flmin(p))2 2 where d = min( 14, ), where is from Assumption 5.2. Theorem 5.3. Suppose the underlying graph G of a faithful GGM satisfies Assumptions 5.2-5.3. Fix any œ (0, 1]. Then, there exists a constant C > 0 such that if the number of samples n satisfies n > , then with probability at least 1 , No MAD accepting dij outputs the equivalence class [G]. Remark 5.1. Theorem 5.3 indicates that the sample complexity of the No MAD is dependent on the absolute minimum and maximum pairwise correlations flmin and flmax, the number of vertices p, and the magnitude of the quantity from Assumption 5.3. Specifically, in regimes of interest, we see that the sample complexity scales as a logarithm in the number of vertices p and inversely in fl2 min(p) and 2 d, thus allowing for robust model selection in the high-dimensional regime. Further note that our sample requirement can have an exponential dependence on p based on the degree of G. However, we expect that this dependency can be improved; see Section 7 for more on this avenue for future work. 6 Experiments We perform experiments on both a synthetic graph and an IEEE 33 bus system, which is a graphical representation of established IEEE-33 bus benchmark distribution system (Baran & Wu, 1989), to assess the validity of our theoretical results and to demonstrate the performance of No MAD. In particular, our experiments demonstrate how an unmodified graphical modeling algorithm (in this case GLASSO Friedman et al. (2008)) compares to No MAD. The synthetic graph Gsyn we consider and graph associated with the IEEE-33 bus system are given in Fig. 5a and Fig. 5b, respectively. As a first step, we need some new performance metrics to make a fair comparison in light of the unidentifiability of the underlying graph structure from noisy data. In the following we will introduce some sets, and show that if an algorithm can recover all these sets, then the output graph (from that algorithm) is in equivalence class. 1. Families. For a vertex i, define its family Fi as {v : deg(v) = 1 and {v, i} œ E(G)} fi{i}. Notice that the subgraph associated with each family is a tree. For synthetic graph Gsyn, the sets {1, 2, 3}, {8, 9, 10} are some of the families in Gsyn; For IEEE 33 Bus Graph {15, 16, 17, 18}, {30, 31, 32, 33} are some of the families. Let F = t Published in Transactions on Machine Learning Research (04/2025) (a) Performance for Gsyn (b) Performance for IEEE-33 Bus. Figure 6: Performance of GLASSO and No MAD for various degree of noises. Up to the values of 2.5 and 1.5 dollars for Gsyn and IEEE-33 Bus, respectively, GLASSO successfully reconstructed an equivalent graph. To investigate this influence, we conducted the experiment with a fixed sample size across 15 trials. In contrast, No MAD consistently demonstrated its ability to recover an equivalent graph under various noise influences. 2. Non-Cut Vertices. For graph G, let Bnon-cut be the set of all non-cut vertices in a non-trivial block B. Define Bnon-cut , t BœBNT Bnon-cut, where BNT is the set of all non-trivial blocks. Recall from Section 3 that in all the equivalent graphs the set of non-cut vertices remain unchanged. 3. Global Edges. Let K be the set of cut vertices who do not share an edge with a leaf vertex in G. For any vertex k œ K, let a family Fk œ F be such that there exists a vertex f œ Fk such that {k, f} œ E(G). Now, we will note two following condition for any neighbor i œ N(k): (a) if i œ K, then {i, k} œ E(G), and (b) otherwise, there exists a vertex j œ Fk such that {j, k} œ E(G). If condition (a) and condition (b) are met, global edge associated with i are recovered correctly . In Lemma B.15, we demonstrate that if an algorithm learns these sets correctly, and the conditions associated with the global edges are met, then an equivalent graph can be recovered. Experimental Setup. We now describe our experimental setup. We generated precision matrices associated with Gsyn, and for IEEE-33 bus system, we added some loops. We then inverted the corresponding precision matrices to obtain the respective covariance matrices in the population setting. Next, we added a diagonal matrix of random positive values to the population covariance matrices at each entry to generate the corresponding noisy covariance matrices. We will now compare the performance of 0-1 loss of No MAD to GLASSO under the influence of noise. In order to compare, our goal is to compare No MAD with the best GLASSO. Our protocol for selecting the best GLASSO is as follows: for a fixed maximum allowable diagonals, we selected the regularization parameter for which the output graph given the noise covariance matrix to GLASSO is in equivalence class. Then, for that fixed regularization parameter we report the performance of GLASSO for varying numbers of (increased) maximum allowable diagonals. Notice that the maximum allowed values of the diagonal elements of the matrix D (which is being added to the generated covariance matrix) contain information about the potential influence of the noise in the setup. In order to study this influence we ran the experiment for a fixed sample size over 15 trials. In Fig. 6, we observe that up to the value of 2.5 and 1.5 for Gsyn and IEEE-33 Bus, respectively, GLASSO was able to recover an equivalent graph. On the contrary, No MAD was able to show a consistent performance in recovering an equivalent graph with various influence of noises. 7 Conclusion and Future Directions Conclusion. In this paper, we consider model selection of Gaussian graphical models when the observations are corrupted by independent but non-identically distributed noise with unknown statistics. We first show that this ambiguity is unavoidable. Finally, we devise a novel algorithm, referred to as No MAD, which learns Published in Transactions on Machine Learning Research (04/2025) structure up to a small equivalence class. Future Directions. This paper opens up several exciting avenues for future research. First, our novel ancestor testing method can be used to identify the ancestors for other graphical models (beyond Gaussians) where information distance satisfies the factorization property in Fact 1; e.g., Discrete graphical models, where the random vector X takes values in the product space X p, where |X| = k. For any i, j œ [p], let Υij œ Rk k denote the tabular representation of the marginal distribution of the pair (Xi, Xj) and Υii denote a diagonal matrix with the marginal distribution of Xi on the diagonal. Then, it is known that the following quantity: dij = det( ij) Ô det( ii jj), can be taken to be the information distance. We refer the reader to Semple & Steel (2003) for more on this, including a proof of an equivalent version of Fact 1. Note that the Ising model is a special case, and our results naturally extend to them. Second, it is well known that flmin could scale exponentially in the diameter of the graph; this could imply that the sample complexity will scale polynomially in the number of vertices p even for balanced binary trees, and as bad as exponential for more unbalanced graphs. Now, notice that in Subroutine 3, No MAD identifies all the star triplets for any ancestor in Gj. Hence, this identification procedure in quite computationally expensive. As we reason this in theoretical section of the appendix that this computation is required in order to learn the pairwise distances dij for each pair (i, j) such that i œ V o and j œ Ahid, where V o and Ahid is the set of observed vertices, and hidden ancestors, respectively. A promising future research work is to develop a TIA test which can obtain {dij}i,jœV ofiAhid without iterating all the triplets in . Furthermore, the Subroutine 4 can be redesigned to a computationally efficient one by learning the clusters in divide and conquer manner. Another promising avenue for future research work is to obtain an upper bound on the diagonal entries Dii for which the underlying graph G is identifiable. Finally, future research can be done to understand to understand the behavior of the hyperparameters and using the stability selection method (see Meinshausen & Bühlmann, 2010) in finite sample settings. Acknowledgments The work was supported in part by the National Science Foundation (NSF) under the grants EPCN-2246658, OAC-1934766, and CCF-2048223. It was also supported by the National Institutes of Health (NIH) under the grant 1R01GM140468-01. Rajasekhar Anguluri, Gautam Dasarathy, Oliver Kosut, and Lalitha Sankar. Grid topology identification with hidden nodes via structured norm minimization. IEEE Control Systems Letters, 6:1244 1249, 2022. Mesut E Baran and Felix F Wu. Network reconfiguration in distribution systems for loss reduction and load balancing. IEEE Transactions on Power delivery, 4(2):1401 1407, 1989. Norman Biggs, E Keith Lloyd, and Robin J Wilson. Graph Theory. Oxford University Press, 1986. Edward T Bullmore and Danielle S Bassett. Brain graphs: graphical models of the human brain connectome. Annual review of clinical psychology, 7:113 140, 2011. Raymond J Carroll, David Ruppert, and Leonard A Stefanski. Measurement error in nonlinear models, volume 105. CRC press, 1995. Marta Casanellas, Marina Garrote-López, and Piotr Zwiernik. Identifiability in robust estimation of tree structured models. Bernoulli, 30(1):1 21, 2024. Andersen Chang, Lili Zheng, Gautam Dasarathy, and Genevera I Allen. Nonparanormal graph quilting with applications to calcium imaging. Stat, 12(1):e623, 2023. Mengjie Chen, Chao Gao, and Zhao Ren. Robust covariance matrix estimation via matrix depth. ar Xiv preprint ar Xiv:1506.00691, 2015. Published in Transactions on Machine Learning Research (04/2025) Yudong Chen, Constantine Caramanis, and Shie Mannor. Robust sparse regression under adversarial corruption. In International Conference on Machine Learning, pp. 774 782. PMLR, 2013. Myung Jin Choi, Vincent YF Tan, Animashree Anandkumar, and Alan S Willsky. Learning latent tree graphical models. J. of Machine Learning Research, 12:1771 1812, 2011. Gautam Dasarathy. Gaussian graphical model selection from size constrained measurements. In 2019 IEEE International Symposium on Information Theory (ISIT), pp. 1302 1306. IEEE, 2019. Gautam Dasarathy, Robert Nowak, and Sebastien Roch. Data requirement for phylogenetic inference from multiple loci: a new distance method. IEEE/ACM transactions on computational biology and bioinformatics, 12(2):422 432, 2014. Gautam Dasarathy, Elchanan Mossel, Robert Nowak, and Sebastien Roch. A stochastic farris transform for genetic data under the multispecies coalescent with applications to data requirements. Journal of Mathematical Biology, 84(5):1 37, 2022. Deepjyoti Deka, Ross Baldick, and Sriram Vishwanath. One breaker is enough: Hidden topology attacks on power grids. In 2015 IEEE Power & Energy Society General Meeting, pp. 1 5. IEEE, 2015. Deepjyoti Deka, Saurav Talukdar, Michael Chertkov, and Murti V Salapaka. Graphical models in meshed distribution grids: Topology estimation, change detection & limitations. IEEE Transactions on Smart Grid, 11(5):4299 4310, 2020. Mathias Drton and Marloes H Maathuis. Structure learning in graphical modeling. Annual Review of Statistics and Its Application, 4:365 393, 2017. Péter L Erdös, Michael A Steel, LászlóA Székely, and Tandy J Warnow. A few logs suffice to build (almost) all trees: Part ii. Theoretical Computer Science, 221(1-2):77 118, 1999. Jerome Friedman, Trevor Hastie, and Robert Tibshirani. Sparse inverse covariance estimation with the graphical lasso. Biostatistics, 9(3):432 441, 2008. F. Harary. Graph Theory. Addison Wesley series in mathematics. Addison-Wesley, 1971. Roger A Horn and Charles R Johnson. Matrix analysis. Cambridge university press, 2012. Jiunn T Hwang. Multiplicative errors-in-variables models with applications to recent data released by the us department of energy. Journal of the American Statistical Association, 81(395):680 688, 1986. Stephen J Iturria, Raymond J Carroll, and David Firth. Polynomial regression and estimating functions in the presence of multiplicative measurement error. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 61(3):547 561, 1999. Markus Kalisch and Peter Bühlman. Estimating high-dimensional directed acyclic graphs with the pc- algorithm. Journal of Machine Learning Research, 8(3), 2007. Ashish Katiyar, Jessica Hoffmann, and Constantine Caramanis. Robust estimation of tree structured gaussian graphical models. In International Conference on Machine Learning, pp. 3292 3300. PMLR, 2019. Ashish Katiyar, Vatsal Shah, and Constantine Caramanis. Robust estimation of tree structured ising models. ar Xiv preprint ar Xiv:2006.05601, 2020. Minje Kim and Paris Smaragdis. Single channel source separation using smooth nonnegative matrix factorization with markov random fields. In 2013 IEEE International Workshop on Machine Learning for Signal Processing (MLSP), pp. 1 6. IEEE, 2013. Akshay Krishnamurthy and Aarti Singh. Robust multi-source network tomography using selective probes. In 2012 Proceedings IEEE INFOCOM, pp. 1629 1637. IEEE, 2012. Steffen L Lauritzen. Graphical models, volume 17. Clarendon Press, 1996. Published in Transactions on Machine Learning Research (04/2025) Roderick JA Little and Donald B Rubin. Statistical analysis with missing data. John Wiley & Sons, 2019. Po-Ling Loh and Xin Lu Tan. High-dimensional robust precision matrix estimation: Cellwise corruption under epsilon-contamination. Electronic Journal of Statistics, 12(1):1429 1467, 2018. Po-Ling Loh and Martin J Wainwright. High-dimensional regression with noisy and missing data: Provable guarantees with non-convexity. Advances in Neural Information Processing Systems, 24, 2011. Karim Lounici. High-dimensional covariance matrix estimation with missing observations. Bernoulli, 20(3): 1029 1058, 2014. Marloes Maathuis, Mathias Drton, Steffen Lauritzen, and Martin Wainwright. Handbook of graphical models. CRC Press, 2018. Nicolai Meinshausen and Peter Bühlmann. Stability selection. Journal of the Royal Statistical Society Series B: Statistical Methodology, 72(4):417 473, 2010. Kevin Murphy, Yair Weiss, and Michael I Jordan. Loopy belief propagation for approximate inference: An empirical study. ar Xiv preprint ar Xiv:1301.6725, 2013. Viet Anh Nguyen, Daniel Kuhn, and Peyman Mohajerin Esfahani. Distributionally robust inverse covariance estimation: The wasserstein shrinkage estimator. Operations Research, 70(1):490 515, 2022. Konstantinos E Nikolakakis, Dionysios S Kalogerias, and Anand D Sarwate. Learning tree structures from noisy data. In The 22nd International Conference on Artificial Intelligence and Statistics, pp. 1771 1782. PMLR, 2019. Viktoria Öllerer and Christophe Croux. Robust high-dimensional precision matrix estimation. In Modern nonparametric, robust and multivariate methods, pp. 325 350. Springer, 2015. Thomas Ott and Ruedi Stoop. The neurodynamics of belief propagation on binary markov random fields. Advances in neural information processing systems, 19, 2006. Naruya Saitou and Masatoshi Nei. The neighbor-joining method: a new method for reconstructing phylogenetic trees. Molecular biology and evolution, 4(4):406 425, 1987. Tapio Schneider. Analysis of incomplete climate data: Estimation of mean values and covariance matrices and imputation of missing values. Journal of climate, 14(5):853 871, 2001. Charles Semple and Mike Steel. Phylogenetics, volume 24. Oxford University Press on Demand, 2003. Herbert A Simon. The architecture of complexity. In The Roots of Logistics, pp. 335 361. Springer, 2012. De Wen Soh and Sekhar C Tatikonda. Testing unfaithful gaussian graphical models. Advances in Neural Information Processing Systems, 27:2681 2689, 2014. Olaf Sporns. Graph theory methods: applications in brain networks. Dialogues in clinical neuroscience, 2018. Hokeun Sun and Hongzhe Li. Robust gaussian graphical modeling via l1 penalization. Biometrics, 68(4): 1197 1206, 2012. Anshoo Tandon, Aldric Han, and Vincent Tan. Sga: A robust algorithm for partial recovery of tree-structured graphical models with noisy samples. In International Conference on Machine Learning, pp. 10107 10117. PMLR, 2021. Caroline Uhler, Garvesh Raskutti, Peter Bühlmann, and Bin Yu. Geometry of the faithfulness assumption in causal inference. The Annals of Statistics, pp. 436 463, 2013. Giuseppe Vinci, Gautam Dasarathy, and Genevera I Allen. Graph quilting: graphical model selection from partially observed covariances. ar Xiv preprint ar Xiv:1912.05573, 2019. Published in Transactions on Machine Learning Research (04/2025) Jun-Kun Wang and Shou-de Lin. Robust inverse covariance estimation under noisy measurements. In International Conference on Machine Learning, pp. 928 936. PMLR, 2014. Qinfeng Xu and Jinhong You. Covariate selection for linear errors-in-variables regression models. Communi- cations in Statistics Theory and Methods, 36(2):375 386, 2007. Eunho Yang and Aurélie C Lozano. Robust gaussian graphical modeling with the trimmed graphical lasso. Advances in Neural Information Processing Systems, 28, 2015. Fengzhuo Zhang and Vincent Tan. Robustifying algorithms of learning latent trees with vector variables. Advances in Neural Information Processing Systems, 34, 2021. Lili Zheng and Genevera I Allen. Graphical model inference with erosely measured data. Journal of the American Statistical Association, 119(547):2282 2293, 2024. Yiming Zuo, Yi Cui, Guoqiang Yu, Ruijiang Li, and Habtom W Ressom. Incorporating prior biological knowledge for network-based differential gene expression analysis using differentially weighted graphical lasso. BMC bioinformatics, 18(1):1 14, 2017.