# active_causal_structure_learning_with_advice__5b85a03c.pdf Active causal structure learning with advice Davin Choo 1 Themis Gouleakis 1 Arnab Bhattacharyya 1 We introduce the problem of active causal structure learning with advice. In the typical wellstudied setting, the learning algorithm is given the essential graph for the observational distribution and is asked to recover the underlying causal directed acyclic graph (DAG) G while minimizing the number of interventions made. In our setting, we are additionally given side information about G as advice, e.g. a DAG G purported to be G . We ask whether the learning algorithm can benefit from the advice when it is close to being correct, while still having worst-case guarantees even when the advice is arbitrarily bad. Our work is in the same space as the growing body of research on algorithms with predictions. When the advice is a DAG G, we design an adaptive search algorithm to recover G whose intervention cost is at most O(max{1, log ψ}) times the cost for verifying G ; here, ψ is a distance measure between G and G that is upper bounded by the number of variables n, and is exactly 0 when G = G . Our approximation factor matches the state-of-the-art for the advice-less setting. 1. Introduction A causal directed acyclic graph on a set V of n variables is a Bayesian network in which the edges model direct causal effects. A causal DAG can be used to infer not only the observational distribution of V but also the result of any intervention on any subset of variables V V . In this work, we restrict ourselves to the causally sufficient setting where there are no latent confounders, no selection bias, and no missingness in data. The goal of causal structure learning is to recover the underlying DAG from data. This is an important problem with applications in multiple fields including philosophy, 1School of Computing, National University of Singapore. Correspondence to: Davin Choo . Proceedings of the 40 th International Conference on Machine Learning, Honolulu, Hawaii, USA. PMLR 202, 2023. Copyright 2023 by the author(s). medicine, biology, genetics, and econometrics (Reichenbach, 1956; Hoover, 1990; King et al., 2004; Woodward, 2005; Rubin & Waterman, 2006; Eberhardt & Scheines, 2007; Sverchkov & Craven, 2017; Rotmensch et al., 2017; Pingault et al., 2018). Unfortunately, in general, it is known that observational data can only recover the causal DAG up to an equivalence class (Pearl, 2009; Spirtes et al., 2000). Hence, if one wants to avoid making parametric assumptions about the causal mechanisms, the only recourse is to obtain experimental data from interventions (Eberhardt et al., 2005; 2006; Eberhardt, 2010). Such considerations motivate the problem of interventional design where the task is to find a set of interventions of optimal cost which is sufficient to recover the causal DAG. There has been a series of recent works studying this problem (He & Geng, 2008; Hu et al., 2014; Shanmugam et al., 2015; Kocaoglu et al., 2017; Lindgren et al., 2018; Greenewald et al., 2019; Squires et al., 2020; Choo et al., 2022; Choo & Shiragur, 2023) under various assumptions. In particular, assuming causal sufficiency, (Choo et al., 2022) gave an adaptive algorithm that actively generates a sequence of interventions of bounded size, so that the total number of interventions is at most O(log n) times the optimal. Typically though, in most applications of causal structure learning, there are domain experts and practitioners who can provide additional advice about the causal relations. Indeed, there has been a long line of work studying how to incorporate expert advice into the causal graph discovery process; e.g. see (Meek, 1995a; Scheines et al., 1998; De Campos & Ji, 2011; Flores et al., 2011; Li & Beek, 2018; Andrews et al., 2020; Fang & He, 2020). In this work, we study in a principled way how using purported expert advice can lead to improved algorithms for interventional design. Before discussing our specific contributions, let us ground the above discussion with a concrete problem of practical importance. In modern virtualized infrastructure, it is increasingly common for applications to be modularized into a large number of interdependent microservices. These microservices communicate with each other in ways that depend on the application code and on the triggering userflow. Crucially, the communication graph between microservices is often unknown to the platform provider as the application code may be private and belong to different entities. Active causal structure learning with advice However, knowing the graph is useful for various critical platform-level tasks, such as fault localization (Zhou et al., 2019), active probing (Tan et al., 2019), testing (Jha et al., 2019), and taint analysis (Clause et al., 2007). Recently, (Wang et al., 2023) and (Ikram et al., 2022) suggested viewing the microservices communication graph as a sparse causal DAG. In particular, (Wang et al., 2023) show that arbitrary interventions can be implemented as fault injections in a staging environment, so that a causal structure learning algorithm can be deployed to generate a sequence of interventions sufficient to learn the underlying communication graph. In such a setting, it is natural to assume that the platform provider already has an approximate guess about the graph, e.g. the graph discovered in a previous run of the algorithm or the graph suggested by public metadata tagging microservice code. The research program we put forth is to design causal structure learning algorithms that can take advantage of such potentially imperfect advice1. 1.1. Our contributions In this work, we study adaptive intervention design for recovering non-parametric causal graphs with expert advice. Specifically, our contributions are as follows. Problem Formulation. Our work connects the causal structure learning problem with the burgeoning research area of algorithms with predictions or learningaugmented algorithms (Mitzenmacher & Vassilvitskii, 2022) where the goal is to design algorithms that bypass worst-case behavior by taking advantage of (possibly erroneous) advice or predictions about the problem instance. Most work in this area has been restricted to online algorithms, data structure design, or optimization, as described later in Section 2.5. However, as we motivated above, expert advice is highly relevant for causal discovery, and to the best of our knowlege, ours is the first attempt to formally address the issue of imperfect advice in this context. Adaptive Search Algorithm. We consider the setting where the advice is a DAG G purported to be the orientations of all the edges in the graph. We define a distance measure which is always bounded by n, the number of variables, and equals 0 when G = G . For any integer k 1, we propose an adaptive algorithm to generate a sequence of interventions of size at most k that recovers the true DAG G , such that the total number of interventions is O(log ψ(G, G ) log k) times the optimal number of interventions of size k. Thus, 1Note however that the system in (Wang et al., 2023) is not causally sufficient due to confounding user behavior and (Ikram et al., 2022) does not actively perform interventions. So, the algorithm proposed in this work cannot be used directly for the microservices graph learning problem. our approximation factor is never worse than the factor for the advice-less setting in (Choo et al., 2022). Our search algorithm also runs in polynomial time. Verification Cost Approximation. For a given upper bound k 1, a verifying intervention set for a DAG G is a set of interventions of size at most k that, together with knowledge of the Markov equivalence class of G , determines the orientations of all edges in G . The minimum size of a verifying intervention set for G , denoted νk(G ), is clearly a lower bound for the number of interventions required to learn G (regardless of the advice graph G). One of our key technical results is a structural result about ν1. We prove that for any two DAGs G and G within the same Markov equivalence class, we always have ν1(G) 2 ν1(G ) and that this is tight in the worst case. Beyond an improved structural understanding of minimum verifying intervention sets, which we believe is of independent interest, this enables us to blindly trust the information provided by imperfect advice to some extent. Similar to prior works (e.g. (Squires et al., 2020; Choo et al., 2022; Choo & Shiragur, 2023)), we assume causal sufficiency and faithfulness while using ideal interventions. Under these assumptions, running standard causal discovery algorithms (e.g. PC (Spirtes et al., 2000), GES (Chickering, 2002)) will always successfully recover the correct essential graph from data. We also assume that the given expert advice is consistent with observational essential graph. See Appendix A for a discussion about our assumptions. 1.2. Paper organization In Section 2, we intersperse preliminary notions with related work. Our main results are presented in Section 3 with the high-level technical ideas and intuition given in Section 4. Section 5 provides some empirical validation. See the appendices for full proofs, source code, and experimental details. 2. Preliminaries and Related Work Basic notions about graphs and causal models are defined in Appendix B. To be very brief, if G = (V, E) is a graph on |V | = n nodes/vertices where V (G), E(G), and A(G) E(G) denote nodes, edges, and arcs of G respectively, we write u v to denote that two nodes u, v V are connected in G, and write u v or u v when specifying a certain direction. The skeleton skel(G) refers to the underlying graph where all edges are made undirected. A v-structure in G refers to a collection of three distinct vertices u, v, w V such that u v w and u w. Let G = (V, E) be fully unoriented. For vertices u, v V , subset of vertices V V and integer r 0, we define dist G(u, v) as the Active causal structure learning with advice shortest path length between u and v, and N r G(V ) = {v V : minu V dist G(u, v) r} V as the set of vertices that are r-hops away from V in G. A directed acyclic graph (DAG) is a fully oriented graph without directed cycles. For any DAG G, we denote its Markov equivalence class (MEC) by [G] and essential graph by E(G). DAGs in the same MEC have the same skeleton and the essential graph is a partially directed graph such that an arc u v is directed if u v in every DAG in MEC [G], and an edge u v is undirected if there exists two DAGs G1, G2 [G] such that u v in G1 and v u in G2. It is known that two graphs are Markov equivalent if and only if they have the same skeleton and v-structures (Verma & Pearl, 1990; Andersson et al., 1997) and the essential graph E(G) can be computed from G by orienting v-structures in skel(G) and applying Meek rules (see Appendix D). In a DAG G, an edge u v is a covered edge if Pa(u) = Pa(v) \ {u}. We use C(G) E(G) to denote the set of covered edges of G. 2.1. Ideal interventions An intervention S V is an experiment where all variables s S is forcefully set to some value, independent of the underlying causal structure. An intervention is atomic if |S| = 1 and bounded size if |S| k for some k 1; observational data is a special case where S = . The effect of interventions is formally captured by Pearl s docalculus (Pearl, 2009). We call any I 2V a intervention set: an intervention set is a set of interventions where each intervention corresponds to a subset of variables. An ideal intervention on S V in G induces an interventional graph GS where all incoming arcs to vertices v S are removed (Eberhardt et al., 2005). It is known that intervening on S allows us to infer the edge orientation of any edge cut by S and V \ S (Eberhardt, 2007; Hyttinen et al., 2013; Hu et al., 2014; Shanmugam et al., 2015; Kocaoglu et al., 2017). We now give a definition and result for graph separators. Definition 2.1 (α-separator and α-clique separator, Definition 19 from (Choo et al., 2022)). Let A, B, C be a partition of the vertices V of a graph G = (V, E). We say that C is an α-separator if no edge joins a vertex in A with a vertex in B and |A|, |B| α |V |. We call C is an α-clique separator if it is an α-separator and a clique. Theorem 2.2 ((Gilbert et al., 1984), instantiated for unweighted graphs). Let G = (V, E) be a chordal graph with |V | 2 and p vertices in its largest clique. There exists a 1/2-clique-separator C involving at most p 1 vertices. The clique C can be computed in O(|E|) time. For ideal interventions, an I-essential graph EI(G) of G is the essential graph representing the Markov equivalence class of graphs whose interventional graphs for each intervention is Markov equivalent to GS for any intervention S I. There are several known properties about I-essential graph properties (Hauser & B uhlmann, 2012; 2014): Every I-essential graph is a chain graph2 with chordal3 chain components. This includes the case of I = . Orientations in one chain component do not affect orientations in other components. In other words, to fully orient any essential graph E(G ), it is necessary and sufficient to orient every chain component in E(G ). For any intervention set I 2V , we write R(G, I) = A(EI(G)) E to mean the set of oriented arcs in the Iessential graph of a DAG G. For cleaner notation, we write R(G, I) for single interventions I = {I} for some I V , and R(G, v) for single atomic interventions I = {{v}} for some v V . For any interventional set I 2V , define GI = G[E \ R(G, I)] as the fully directed subgraph DAG induced by the unoriented arcs in EI(G), where G is the graph obtained after removing all the oriented arcs in the observational essential graph due to v-structures. See Figure 1 for an example. In the notation of R( , ), the following result justifies studying verification and adaptive search via ideal interventions only on DAGs without v-structures, i.e. moral DAGs (Definition 2.4): since R(G, I) = R(G , I) R(G, ), any oriented arcs in the observational graph can be removed before performing any interventions as the optimality of the solution is unaffected.4 Theorem 2.3 ((Choo & Shiragur, 2023)). For any DAG G = (V, E) and intervention sets A, B 2V , = R(GA, B) R(GB, A) (R(G, A) R(G, B)) Definition 2.4 (Moral DAG). A DAG G is called a moral DAG if it has no v-structures. So, E(G) = skel(G). 2.2. Verifying sets A verifying set I for a DAG G [G ] is an intervention set that fully orients G from E(G ), possibly with repeated applications of Meek rules (see Appendix D), i.e. EI(G ) = G . Furthermore, if I is a verifying set for G , then so is I S for any additional intervention S V . While there may be multiple verifying sets in general, we are often interested in finding one with a minimum size. Definition 2.5 (Minimum size verifying set). An intervention set I 2V is called a verifying set for a DAG G if EI(G ) = G . I is a minimum size verifying set if EI (G ) = G for any |I | < |I|. 2A partially directed graph is a chain graph if it does not contain any partially directed cycles where all directed arcs point in the same direction along the cycle. 3A chordal graph is a graph where every cycle of length at least 4 has an edge that is not part of the cycle but connects two vertices of the cycle; see (Blair & Peyton, 1993) for an introduction. 4The notation A B denotes disjoint union of sets A and B. Active causal structure learning with advice For bounded size interventions, the minimum verification number νk(G) denotes the size of the minimum size verifying set for any DAG G [G ]; we write ν1(G) for atomic interventions. That is, any revealed arc directions when performing interventions on E(G ) respects G. (Choo et al., 2022) tells us that it is necessary and sufficient to intervene on a minimum vertex cover of the covered edges C(G) in order to verify a DAG G, and that ν1(G) is efficiently computable given G since C(G) induces a forest. Theorem 2.6 ((Choo et al., 2022)). Fix an essential graph E(G ) and G [G ]. An atomic intervention set I is a minimal sized verifying set for G if and only if I is a minimum vertex cover of covered edges C(G) of G. A minimal sized atomic verifying set can be computed in polynomial time since the edge-induced subgraph on C(G) is a forest. For any DAG G, we use V(G) 2V to denote the set of all atomic verifying sets for G. That is, each atomic intervention set in V(G) is a minimum vertex cover of C(G). 2.3. Adaptive search using ideal interventions Adaptive search algorithms have been studied in earnest (He & Geng, 2008; Hauser & B uhlmann, 2014; Shanmugam et al., 2015; Squires et al., 2020; Choo et al., 2022; Choo & Shiragur, 2023) as they can use significantly less interventions than non-adaptive counterparts.5 Most recently, (Choo et al., 2022) gave an efficient algorithm for computing adaptive interventions with provable approximation guarantees on general graphs. Theorem 2.7 ((Choo et al., 2022)). Fix an unknown underlying DAG G . Given an essential graph E(G ) and intervention set bound k 1, there is a deterministic polynomial time algorithm that computes an intervention set I adaptively such that EI(G ) = G , and |I| has size 1. O(log(n) ν1(G )) when k = 1 2. O(log(n) log(k) νk(G )) when k > 1. Meanwhile, in the context of local causal graph discovery where one is interested in only learning a subset of causal relationships, the Subset Search algorithm of (Choo & Shiragur, 2023) incurs a multiplicative overhead that scales logarithmically with the number of relevant nodes when orienting edges within a node-induced subgraph. Definition 2.8 (Relevant nodes). Fix a DAG G = (V, E) and arbitrary subset V V . For any intervention set I 2V and resulting interventional essential graph EI(G ), we define the relevant nodes ρ(I, V ) V as the set of nodes within V that is adjacent to some unoriented arc within the node-induced subgraph EI(G )[V ]. 5If the essential graph E(G ) is a path of n nodes, then nonadaptive algorithms need Ω(n) atomic interventions to recover G while O(log n) atomic interventions suffices for adaptive search. Figure 1. (I) Ground truth DAG G ; (II) Observational essential graph E(G ) where C E D is a v-structure and Meek rules orient arcs D F and E F; (III) G = G[E \ R(G, )] where oriented arcs in E(G ) are removed from G ; (IV) MPDAG G [G ] incorporating the following partial order advice (S1 = {B}, S2 = {A, D}, S3 = {C, E, F}), which can be converted to required arcs B A and B D. Observe that A C is oriented by Meek R1 via B A C, the arc A D is still unoriented, the arc B A disagrees with G , and there are two possible DAGs consistent with the resulting MPDAG. For an example of relevant nodes, see Figure 1: For the subset V = {A, C, D, E, F} in (II), only {A, C, D} are relevant since incident edges to E and F are all oriented. Theorem 2.9 ((Choo & Shiragur, 2023)). Fix an unknown underlying DAG G . Given an interventional essential graph EI(G ), node-induced subgraph H with relevant nodes ρ(I, V (H)) and intervention set bound k 1, there is a deterministic polynomial time algorithm that computes an intervention set I adaptively such that EI I (G )[V (H)] = G [V (H)], and |I | has size 1. O(log(|ρ(I, V (H))|) ν1(G )) when k = 1 2. O(log(|ρ(I, V (H))|) log(k) νk(G )) when k > 1. Note that k = 1 refers to the setting of atomic interventions and we always have 0 |ρ(I, V (H))| n. 2.4. Expert advice in causal graph discovery There are three main types of information that a domain expert may provide (e.g. see the references given in Section 1): (I) Required parental arcs: X Y (II) Forbidden parental arcs: X Y (III) Partial order or tiered knowledge: A partition of the n variables into 1 t n sets S1, . . . , St such that variables in Si cannot come after Sj, for all i < j. In the context of orienting unoriented X Y edges in an essential graph, it suffices to consider only information of type (I): X Y implies Y X, and a partial order can be converted to a collection of required parental arcs.6 6For every edge X Y with X Si and Y Sj, enforce the required parental arc X Y if and only if i < j. Active causal structure learning with advice Maximally oriented partially directed acyclic graphs (MPDAGs), a refinement of essential graphs under additional causal information, are often used to model such expert advice and there has been a recent growing interest in understanding them better (Perkovic et al., 2017; Perkovic, 2020; Guo & Perkovic, 2021). MPDAGs are obtained by orienting additional arc directions in the essential graph due to background knowledge, and then applying Meek rules. See Figure 1 for an example. 2.5. Other related work Causal Structure Learning Algorithms for causal structure learning can be grouped into three broad categories, constraint-based, score-based, and Bayesian. Previous works on the first two approaches are described in Appendix C. In Bayesian methods, a prior distribution is assumed on the space of all structures, and the posterior is updated as more data come in. Heckerman (1995) was one of the first works on learning from interventional data in this context, which spurred a series of papers (e.g. Heckerman et al. (1995); Cooper & Yoo (1999); Friedman & Koller (2000); Heckerman et al. (2006)). Research on active experimental design for causal structure learning with Bayesian updates was initiated by Tong & Koller (2000; 2001) and Murphy (2001). Masegosa & Moral (2013) considered a combination of Bayesian and constraint-based approaches. Cho et al. (2016) and Agrawal et al. (2019) have used active learning and Bayesian updates to help recover biological networks. While possibly imperfect expert advice may be used to guide the prior in the Bayesian approach, the works mentioned above do not provide rigorous guarantees about the number of interventions performed or about optimality, and so they are not directly comparable to our results here. Algorithms with predictions Learning-augmented algorithms have received significant attention since the seminal work of Lykouris & Vassilvitskii (2021), where they investigated the online caching problem with predictions. Based on that model, Purohit et al. (2018) proposed algorithms for the ski-rental problem as well as non-clairvoyant scheduling. Subsequently, Gollapudi & Panigrahi (2019), Wang et al. (2020), and Angelopoulos et al. (2020) improved the initial results for the ski-rental problem. Several works, including (Rohatgi, 2020; Antoniadis et al., 2020a; Wei, 2020), improved the initial results regarding the caching problem. Scheduling problems with machine-learned advice have been extensively studied in the literature (Lattanzi et al., 2020; Bamas et al., 2020a; Antoniadis et al., 2022). There are also results for augmenting classical data structures with predictions (e.g. indexing (Kraska et al., 2018) and Bloom filters (Mitzenmacher, 2018)), online selection and matching problems (Antoniadis et al., 2020b; D utting et al., 2021), online TSP (Bernardini et al., 2022; Gouleakis et al., 2023), and a more general framework of online primal-dual algorithms (Bamas et al., 2020b). In the above line of work, the extent to which the predictions are helpful in the design of the corresponding online algorithms, is quantified by the following two properties. The algorithm is called (i) α-consistent if it is αcompetitive with no prediction error and (ii) β-robust if it is β-competitive with any prediction error. In the language of learning augmented algorithms or algorithms with predictions, our causal graph discovery algorithm is 1-consistent and O(log n)-robust when competing against the verification number ν1(G ), the minimum number of interventions necessary needed to recover G . Note that even with arbitrarily bad advice, our algorithm uses asymptotically the same number of interventions incurred by the best-known advice-free adaptive search algorithm (Choo et al., 2022). Our exposition here focuses on interpreting and contextualizing our main results while deferring technicalities to Section 4. We first focus on the setting where the advice is a fully oriented DAG e G [G ] within the Markov equivalence class [G ] of the true underlying causal graph G , and explain in Appendix E how to handle the case of partial advice. Full proofs are provided in the appendix. 3.1. Structural property of verification numbers We begin by stating a structural result about verification numbers of DAGs within the same Markov equivalence class (MEC) that motivates the definition of a metric between DAGs in the same MEC our algorithmic guarantees (Theorem 3.5) are based upon. Theorem 3.1. For any DAG G with MEC [G ], we have that max G [G ] ν1(G) 2 min G [G ] ν1(G). Theorem 3.1 is the first known result relating the minimum and maximum verification numbers of DAGs given a fixed MEC. The next result tells us that the ratio of two is tight. Lemma 3.2 (Tightness of Theorem 3.1). There exist DAGs G1 and G2 from the same MEC with ν1(G1) = 2 ν1(G2). Theorem 3.1 tells us that we can blindly intervene on any minimum verifying set e V V( e G) of any given advice DAG e G while incurring only at most a constant factor of 2 more interventions than the minimum verification number ν(G ) of the unknown ground truth DAG G . 3.2. Adaptive search with imperfect DAG advice Recall the definition of r-hop from Section 2. To define the quality of the advice DAG e G, we first define the notion of min-hop-coverage which measures how far a given Active causal structure learning with advice verifying set of e G is from the set of covered edges of G . Definition 3.3 (Min-hop-coverage). Fix a DAG G with MEC [G ] and consider any DAG e G [G ]. For any minimum verifying set e V V( e G), we define the minhop-coverage h(G , e V ) {0, 1, 2, . . . , n} as the minimum number of hops such that both endpoints of covered edges C(G ) of G belong in N h(G ,e V ) skel(E(G ))(e V ). Using min-hop-coverage, we now define a quality measure ψ(G , e G) for DAG e G [G ] as an advice for DAG G . Definition 3.4 (Quality measure). Fix a DAG G with MEC [G ] and consider any DAG e G [G ]. We define ψ(G , e G) as follows: ψ(G , e G) = max e V V( e G) ρ e V , N h(G ,e V ) skel(E(G ))(e V ) By definition, ψ(G , G ) = 0 and max G [G ] ψ(G , G) n. In words, ψ(G , e G) only counts the relevant nodes within the min-hop-coverage neighborhood after intervening on the worst possible verifying set e V of e G. We define ψ via the worst set because any search algorithm cannot evaluate h(G , e V ), since G is unknown, and can only consider an arbitrary e V V( e G). See Figure 2 for an example. Our main result is that it is possible to design an algorithm that leverages an advice DAG e G [G ] and performs interventions to fully recover an unknown underlying DAG G , whose performance depends on the advice quality ψ(G , e G). Our search algorithm only knows E(G ) and e G [G ] but knows neither ψ(G , e G) nor ν(G ). Theorem 3.5. Fix an essential graph E(G ) with an unknown underlying ground truth DAG G . Given an advice graph e G [G ] and intervention set bound k 1, there exists a deterministic polynomial time algorithm (Algorithm 1) that computes an intervention set I adaptively such that EI(G ) = G , and |I| has size 1. O(max{1, log ψ(G , e G)} ν1(G )) when k = 1 2. O(max{1, log ψ(G , e G)} log k νk(G )) when k > 1. Consider first the setting of k = 1. Observe that when the advice is perfect (i.e. e G = G ), we use O(ν(G )) interventions, i.e. a constant multiplicative factor of the minimum number of interventions necessary. Meanwhile, even with low quality advice, we still use O(log n ν(G )) interventions, asymptotically matching the best known guarantees for adaptive search without advice. To the best of our knowledge, Theorem 3.5 is the first known result that principally employs imperfect expert advice with provable guarantees in the context of causal graph discovery via interventions. Consider now the setting of bounded size interventions where k > 1. The reason why we can obtain such a result is precisely because of our algorithmic design: we deliberately zn . . . z2 z1 zn . . . z2 z1 Figure 2. Consider the moral DAGs G and e G [G ] on n + 5 nodes, where dashed arcs represent the covered edges in each DAG. A minimum sized verifying set e V = {a, e, z2} V( e G) of e G is given by the boxed vertices on the right. As N 1 skel(G )(e V ) = {a, b, c, d, e, z1, z2, z3} includes both endpoints of all covered edges of G , we see that h(G , e V ) = 1. Intervening on e V = {a, e, z2} in G orients the arcs b a c, c e d, and z3 z2 z1 respectively which then triggers Meek R1 to orient c b via e c b and to orient z4 z3 via e c . . . z4 z3 (after a few invocations of R1), so {a, b, e, z1, z2, z3} will not be relevant nodes in E e V (G ). Meanwhile, the edge c d remains unoriented in E e V (G ), so ρ(e V , N 1 skel(G )(e V )) = |{c, d}| = 2. One can check that ψ(G , e G) = 2 while n could be arbitrarily large. On the other hand, observe that ψ is not symmetric: in the hypothetical situation where we use G as an advice for e G, the min-hop-coverage has to extend along the chain z1 . . . zn to reach {z1, z2}, so h(G , V ) n and ψ( e G, G ) n since the entire chain remains unoriented with respect to any V V(G ). designed an algorithm that invokes Subset Search as a black-box subroutine. Thus, the bounded size guarantees of Subset Search given by Theorem 2.9 carries over to our setting with a slight modification of the analysis. 4. Techniques Here, we discuss the high-level technical ideas and intuition behind how we obtain our adaptive search algorithm with imperfect DAG advice. See the appendix for full proofs; in particular, see Appendix F for an overview of Theorem 3.1. For brevity, we write ψ to mean ψ(G , e G) and drop the subscript skel(E(G )) of r-hop neighborhoods in this section. We also focus our discussion to the atomic interventions. Our adaptive search algorithm (Algorithm 1) uses Subset Search as a subroutine. We begin by observing that Subset Search(E(G ), A) fully orients E(G ) into G if the covered edges of G lie within the node-induced subgraph induced by A. Lemma 4.1. Fix a DAG G = (V, E) and let V V be any subset of vertices. Suppose IV V is the set of nodes intervened by Subset Search(E(G ), V ). If Active causal structure learning with advice C(G ) E(G [V ]), then EIV (G ) = G . Motivated by Lemma 4.1, we design Algorithm 1 to repeatedly invoke Subset Search on node-induced subgraphs N r(e V ), starting from an arbitrary verifying set e V V( e G) and for increasing values of r. For i N {0}, let us denote r(i) N {0} as the value of r in the i-th invocation of Subset Search, where we insist that r(0) = 0 and r(j) > r(j 1) for any j N. Note that r = 0 simply implies that we intervene on the verifying set e V , which only incurs O(ν1(G )) interventions due to Theorem 3.1. Then, we can appeal to Lemma 4.1 to conclude that E(G ) is completely oriented into G in the t-th invocation if r(t) h(G , e V ). While the high-level subroutine invocation idea seems simple, one needs to invoke Subset Search at suitably chosen intervals in order to achieve our theoretical guarantees we promise in Theorem 3.5. We now explain how to do so in three successive attempts while explaining the algorithmic decisions behind each modification introduced. As a reminder, we do not know G and thus do not know h(G , e V ) for any verifying set e V V( e G) of e G [G ]. NAIVE ATTEMPT: INVOKE FOR r = 0, 1, 2, 3, . . . The most straightforward attempt would be to invoke Subset Search repeatedly each time we increase r by 1 until the graph is fully oriented in the worst case, t = h(G , e V ). However, this may cause us to incur way too many interventions. Suppose there are ni relevant nodes in the i-th invocation. Using Theorem 2.9, one can only argue that the overall number interventions incurred is O(Pt i=0 log ni ν(G )). However, P i log ni could be significantly larger than log(P i ni) in general, e.g. log 2 + . . . + log 2 = (n/2) log 2 log n. In fact, if G was a path on n vertices v1 v2 . . . vn and e G [G ] misleads us with v1 v2 . . . vn, then this approach incurs Ω(n) interventions in total. TWEAK 1: ONLY INVOKE PERIODICALLY Since Theorem 2.9 provides us a logarithmic factor in the analysis, we could instead consider only invoking Subset Search after the number of nodes in the subgraph increases by a polynomial factor. For example, if we invoked Subset Search with ni previously, then we will wait until the number of relevant nodes surpasses n2 i before invoking Subset Search again, where we define n0 2 for simplicity. Since log ni 2 log ni 1, we can see via an inductive argument that the number of interventions used in the final invocation will dominate the total number of interventions used so far: nt 2 log nt 1 log nt 1 + 2 log nt 2 . . . Pt 1 i=0 log ni. Since ni n for any i, we can already prove that O(log n ν1(G )) inter- Figure 3. Consider the ground truth DAG G with unique minimum verifying set {v2} and an advice DAG e G [G ] with chosen minimum verifying set e V = {v1}. So, h(G , e V ) = 1 and ideally we want to argue that our algorithm uses a constant number of interventions. Without tweak 2 and n0 = 2, an algorithm that increases hop radius until the number of relevant nodes is squared will not invoke Subset Search until r = 3 because ρ(e V , N 1) = 1 < n2 0 and ρ(e V , N 2) = 2 < n2 0. However, ρ(e V , N 3) = n 1 and we can only conclude that the algorithm uses O(log n) interventions by invoking Subset Search on a subgraph on n 1 nodes. ventions suffice, matching the advice-free bound of Theorem 2.7. However, this approach and analysis does not take into account the quality of e G and is insufficient to relate nt with the advice measure ψ. TWEAK 2: ALSO INVOKE ONE ROUND BEFORE Suppose the final invocation of Subset Search is on r(t)- hop neighborhood while incurring O(log nt ν1(G )) interventions. This means that C(G ) lies within N r(t)(e V ) but not within N r(t 1)(e V ). That is, N r(t 1)(e V ) N h(G ,e V )(e V ) N r(t)(e V ). While this tells us that nt 1 |ρ(e V , N r(t 1)(e V ))| < |ρ(e V , N h(G ,e V )(e V ))| = ψ, what we want is to conclude that nt O(ψ). Unfortunately, even when ψ = r(t 1) + 1, it could be the case that |ρ(e V , N h(G ,e V )(e V ))| |N r(t)(e V )| as the number of relevant nodes could blow up within a single hop (see Figure 3). To control this potential blow up in the analysis, we can introduce the following technical fix: whenever we want to invoke Subset Search on r(i), first invoke Subset Search on r(i) 1 and terminate earlier if the graph is already fully oriented into G . PUTTING TOGETHER Algorithm 1 presents our full algorithm where the inequality ρ(Ii, N r skel(E(G ))(e V )) n2 i corresponds to the first tweak while the terms Ci and C i correspond to the second tweak. In Appendix H, we explain why our algorithm (Algorithm 1) is simply the classic binary search with prediction 7 when the given essential graph E(G ) is an undirected path. So, 7e.g. see https://en.wikipedia.org/wiki/ Learning_augmented_algorithm#Binary_search Active causal structure learning with advice another way to view our result is a generalization that works on essential graphs of arbitrary moral DAGs. Algorithm 1 Adaptive search algorithm with advice. 1: Input: Essential graph E(G ), advice DAG e G [G ], intervention size k N 2: Output: An intervention set I such that each intervention involves at most k nodes and EI(G ) = G . 3: Let us write SS to mean Subset Search. 4: Let e V V( e G) be any atomic verifying set of e G. 5: if k = 1 then 6: Define I0 = e V as an atomic intervention set. 7: else 8: Define k = min{k, |e V |/2}, a = |e V |/k 2, and ℓ= loga |C| . Compute labelling scheme on e V with (|e V |, k, a) via Lemma 4.3 and define I0 = {Sx,y}x [ℓ],y [a], where Sx,y e V is the subset of vertices whose xth letter in the label is y. 9: end if 10: Intervene on I0 and initialize r 0, i 0, n0 2. 11: while EIi(G ) still has undirected edges do 12: if ρ(Ii, N r skel(E(G ))(e V )) n2 i then 13: Increment i i + 1 and record r(i) r. 14: Update ni ρ(Ii, N r skel(E(G ))(e V )) 15: Ci SS(EIi(G ), N r 1 skel(E(G ))(e V ), k) 16: if EIi 1 Ci(G ) still has undirected edges then 17: C i SS(EIi 1 Ci(G ), N r skel(E(G ))(e V ), k) 18: Update Ii Ii 1 Ci C i. 19: else 20: Update Ii Ii 1 Ci. 21: end if 22: end if 23: Increment r r + 1. 24: end while 25: return Ii For bounded size interventions, we rely on the following known results. Theorem 4.2 (Theorem 12 of (Choo et al., 2022)). Fix an essential graph E(G ) and G [G ]. If ν1(G) = ℓ, then νk(G) ℓ k and there exists a polynomial time algo. to compute a bounded size intervention set I of size |I| ℓ k + 1. Lemma 4.3 (Lemma 1 of (Shanmugam et al., 2015)). Let (n, k, a) be parameters where k n/2. There exists a polynomial time labeling scheme that produces distinct ℓ length labels for all elements in [n] using letters from the integer alphabet {0} [a] where ℓ= loga n . Further, in every digit (or position), any integer letter is used at most n/a times. This labelling scheme is a separating system: for any i, j [n], there exists some digit d [ℓ] where the labels of i and j differ. Theorem 4.2 enables us to easily relate ν1(G) with νk(G) while Lemma 4.3 provides an efficient labelling scheme to partition a set of n nodes into a set S = {S1, S2, . . .} of bounded size sets, each Si involving at most k nodes. By invoking Lemma 4.3 with a n /k where n is related to ν1(G), we see that |S| n k log k. As νk(G) ν1(G)/k, this is precisely why the bounded intervention guarantees in Theorem 2.7, Theorem 2.9 and Theorem 3.5 have an additional multiplicative log k factor. 5. Empirical validation While our main contributions are theoretical, we also performed some experiments to empirically validate that our algorithm is practical, outperforms the advice-free baseline when the advice quality is good, and still being at most a constant factor worse when the advice is poor. Motivated by Theorem 2.3, we experimented on synthetic moral DAGs from Wien obst et al. (2021b): For each undirected chordal graph, we use the uniform sampling algorithm of Wien obst et al. (2021b) to uniformly sample 1000 moral DAGs e G1, . . . , e G1000 and randomly choose one of them as G . Then, we give {(E(G ), e Gi)}i [1000] as input to Algorithm 1. Figure 4 shows one of the experimental plots; more detailed experimental setup and results are given in Appendix I. On the X-axis, we plot ψ(G , e V ) = ρ e V , N h(G ,e V ) skel(E(G ))(e V ) , which is a lower bound and proxy8 for ψ(G , e G). On the Y-axis, we aggregate advice DAGs based on their quality measure and also show (in dashed lines) the empirical distribution of quality measures of all DAGs within the Markov equivalence class. As expected from our theoretical analyses, we see that the number of interventions by our advice search starts from ν1(G ), is lower than advice-free search of (Choo et al., 2022) when ψ(G , e V ) is low, and gradually increases as the advice quality degrades. Nonetheless, the number of interventions used is always theoretically bounded below O(ψ(G , e V ) ν1(G )); we do not plot ψ(G , e V ) ν1(G ) since plotting it yields a squashed graph as the empirical counts are significantly smaller. In this specific graph instance, Figure 4 suggests that our advice search outperforms its advice-free counterpart when given an advice DAG e G that is better than 40% of all possible DAGs consistent with the observational essential graph E(G ). 8We do not know if there is an efficient way to compute ψ(G , e G) besides the naive (possibly exponential time) enumeration over all possible minimum verifying sets. Active causal structure learning with advice Figure 4. Experimental plot for one of the synthetic graphs G , with respect to 1000 |[G ]| 1.4 106 uniformly sampled advice DAGs e G from the MEC [G ]. The solid lines indicate the number of atomic interventions used while the dotted lines indicate the empirical cumulative probability density of e G. The true cumulative probability density lies within the shaded area with probability at least 0.99 (see Appendix I for details). 6. Conclusion and discussion In this work, we gave the first result that utilizes imperfect advice in the context of causal discovery. We do so in a way that the performance (i.e. the number of interventions in our case) does not degrade significantly even when the advice is inaccurate, which is consistent with the objectives of learning-augmented algorithms. Specifically, we show a smooth bound that matches the number of interventions needed for verification of the causal relationships in a graph when the advice is completely accurate and also depends logarithmically on the distance of the advice to the ground truth. This ensures robustness to bad advice, the number of interventions needed is asymptotically the same as in the case where no advice is available. Our results do rely on the widely-used assumptions of sufficiency and faithfulness as well as access to ideal iterventions; see Appendix A for a more detailed discussion. Since wrong causal conclusions may be drawn when these assumptions are violated by the data, thus it is of great interest to remove/weaken these assumptions while maintaining strong theoretical guarantees in future work. 6.1. Interesting future directions to explore Partial advice In Appendix E, we explain why having a DAG e G as advice may not always be possible and explain how to extend our results to the setting of partial advice by considering the worst case DAG consistent with the given partial advice A. The question is whether one can design and analyze a better algorithm than a trivial max e G A. For example, maybe one could pick e G = argmin G A max H [G ] ψ(H, G)? The motivation is as follows: If [G ] is a disc in R2 and ψ is the Euclidean distance, then e G should be the point within A that is closest to the center of the disc. Note that we can only optimize with respect to max H [G ] because we do not actually know G . It remains to be seen if such an object can be efficiently computed and whether it gives a better bound than max e G A. Incorporating expert confidence The notion of confidence level and correctness of an advice are orthogonal issues an expert can be confidently wrong. In this work, we focused on the case where the expert is fully confident but may be providing imperfect advice. It is an interesting problem to investigate how to principally handle both issues simultaneously; for example, what if the advice is not a DAG e G [G ] in the essential graph but a distribution over all DAGs in [G ]? Bayesian ideas may apply here. Better analysis? Empirically, we see that the log factor is a rather loose upper bound both for blind search and advice search. Can there be a tighter analysis? (Choo et al., 2022) tells us that Ω(log n ν1(G )) is unavoidable when E(G ) is a path on n vertices with ν1(G ) = 1 but this is a special class of graphs. What if ν1(G ) > 1? Can we give tighter bounds in other graph parameters? Furthermore, in some preliminary testing, we observed that implementing tweak 2 or ignoring it yield similar empirical performance and we wonder if there is a tighter analysis without tweak 2 that has similar guarantees. Acknowledgements This research/project is supported by the National Research Foundation, Singapore under its AI Singapore Programme (AISG Award No: AISG-Ph D/2021-08-013). TG and AB are supported by the National Research Foundation Fellowship for AI (Award NRF-NRFFAI-0002), an Amazon Research Award, and a Google South & Southeast Asia Research Award. Part of this work was done while the authors were visiting the Simons Institute for the Theory of Computing. We would like to thank Kirankumar Shiragur and Joy Qiping Yang for valuable feedback and discussions. Agrawal, R., Squires, C., Yang, K., Shanmugam, K., and Uhler, C. ABCD-Strategy: Budgeted Experimental Design for Targeted Causal Structure Discovery. In International Conference on Artificial Intelligence and Statistics, pp. 3400 3409. PMLR, 2019. Andersen, H. When to expect violations of causal faithful- Active causal structure learning with advice ness and why it matters. Philosophy of Science, 80(5): 672 683, 2013. Andersson, S. A., Madigan, D., and Perlman, M. D. A characterization of Markov equivalence classes for acyclic digraphs. The Annals of Statistics, 25(2):505 541, 1997. Andrews, B., Spirtes, P., and Cooper, G. F. On the Completeness of Causal Discovery in the Presence of Latent Confounding with Tiered Background Knowledge. In International Conference on Artificial Intelligence and Statistics, pp. 4002 4011. PMLR, 2020. Angelopoulos, S., D urr, C., Jin, S., Kamali, S., and Renault, M. Online Computation with Untrusted Advice. In 11th Innovations in Theoretical Computer Science Conference (ITCS 2020). Schloss Dagstuhl-Leibniz-Zentrum f ur Informatik, 2020. Antoniadis, A., Coester, C., Elias, M., Polak, A., and Simon, B. Online Metric Algorithms with Untrusted Predictions. In International Conference on Machine Learning, pp. 345 355. PMLR, 2020a. Antoniadis, A., Gouleakis, T., Kleer, P., and Kolev, P. Secretary and Online Matching Problems with Machine Learned Advice. Advances in Neural Information Processing Systems, 33:7933 7944, 2020b. Antoniadis, A., Jabbarzade, P., and Shahkarami, G. A Novel Prediction Setup for Online Speed-Scaling. In 18th Scandinavian Symposium and Workshops on Algorithm Theory (SWAT 2022). Schloss Dagstuhl-Leibniz-Zentrum f ur Informatik, 2022. Bamas, E., Maggiori, A., Rohwedder, L., and Svensson, O. Learning Augmented Energy Minimization via Speed Scaling. Advances in Neural Information Processing Systems, 33:15350 15359, 2020a. Bamas, E., Maggiori, A., and Svensson, O. The Primal-Dual method for Learning Augmented Algorithms. Advances in Neural Information Processing Systems, 33:20083 20094, 2020b. Bernardini, G., Lindermayr, A., Marchetti-Spaccamela, A., Megow, N., Stougie, L., and Sweering, M. A Universal Error Measure for Input Predictions Applied to Online Graph Problems. In Advances in Neural Information Processing Systems, 2022. Blair, J. R. S. and Peyton, B. W. An introduction to chordal graphs and clique trees. In Graph theory and sparse matrix computation, pp. 1 29. Springer, 1993. Canonne, C. L. A short note on learning discrete distributions. ar Xiv preprint ar Xiv:2002.11457, 2020. Chickering, D. M. A Transformational Characterization of Equivalent Bayesian Network Structures. In Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence, UAI 95, pp. 87 98, San Francisco, CA, USA, 1995. Morgan Kaufmann Publishers Inc. ISBN 1558603859. Chickering, D. M. Optimal Structure Identification with Greedy Search. Journal of Machine Learning Research, 3:507 554, 2002. Cho, H., Berger, B., and Peng, J. Reconstructing Causal Biological Networks through Active Learning. PLo S ONE, 11(3):e0150611, 2016. Choo, D. and Shiragur, K. Subset verification and search algorithms for causal DAGs. In International Conference on Artificial Intelligence and Statistics, 2023. Choo, D., Shiragur, K., and Bhattacharyya, A. Verification and search algorithms for causal DAGs. Advances in Neural Information Processing Systems, 35, 2022. Clause, J., Li, W., and Orso, A. Dytan: A Generic Dynamic Taint Analysis Framework. In Proceedings of the 2007 international symposium on Software testing and analysis, pp. 196 206, 2007. Colombo, D., Maathuis, M. H., Kalisch, M., and Richardson, T. S. Learning high-dimensional directed acyclic graphs with latent and selection variables. The Annals of Statistics, pp. 294 321, 2012. Cooper, G. F. and Yoo, C. Causal Discovery from a Mixture of Experimental and Observational Data. In Proceedings of the Fifteenth conference on Uncertainty in artificial intelligence, pp. 116 125, 1999. De Campos, C. P. and Ji, Q. Efficient Structure Learning of Bayesian Networks using Constraints. The Journal of Machine Learning Research, 12:663 689, 2011. D utting, P., Lattanzi, S., Paes Leme, R., and Vassilvitskii, S. Secretaries with Advice. In Proceedings of the 22nd ACM Conference on Economics and Computation, pp. 409 429, 2021. Eberhardt, F. Causation and Intervention. Unpublished doctoral dissertation, Carnegie Mellon University, pp. 93, 2007. Eberhardt, F. Causal Discovery as a Game. In Causality: Objectives and Assessment, pp. 87 96. PMLR, 2010. Eberhardt, F. and Scheines, R. Interventions and Causal Inference. Philosophy of science, 74(5):981 995, 2007. Active causal structure learning with advice Eberhardt, F., Glymour, C., and Scheines, R. On the number of experiments sufficient and in the worst case necessary to identify all causal relations among N variables. In Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence, pp. 178 184, 2005. Eberhardt, F., Glymour, C., and Scheines, R. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables. In Innovations in machine learning, pp. 97 112. Springer, 2006. Fang, Z. and He, Y. IDA with Background Knowledge. In Conference on Uncertainty in Artificial Intelligence, pp. 270 279. PMLR, 2020. Flores, M. J., Nicholson, A. E., Brunskill, A., Korb, K. B., and Mascaro, S. Incorporating expert knowledge when learning Bayesian network structure: A medical case study. Artificial intelligence in medicine, 53(3):181 204, 2011. Friedman, N. and Koller, D. Being Bayesian about Network Structure. In Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence, pp. 201 210, 2000. Gilbert, J. R., Rose, D. J., and Edenbrandt, A. A separator theorem for chordal graphs. SIAM Journal on Algebraic Discrete Methods, 5(3):306 313, 1984. Gollapudi, S. and Panigrahi, D. Online Algorithms for Rentor-Buy with Expert Advice. In International Conference on Machine Learning, pp. 2319 2327. PMLR, 2019. Gouleakis, T., Lakis, K., and Shahkarami, G. Learning Augmented Algorithms for Online TSP on the Line. In 37th AAAI Conference on Artificial Intelligence. AAAI, 2023. Greenewald, K., Katz, D., Shanmugam, K., Magliacane, S., Kocaoglu, M., Boix-Adser a, E., and Bresler, G. Sample Efficient Active Learning of Causal Trees. Advances in Neural Information Processing Systems, 32, 2019. Guo, R. and Perkovic, E. Minimal Enumeration of All Possible Total Effects in a Markov Equivalence Class. In International Conference on Artificial Intelligence and Statistics, pp. 2395 2403. PMLR, 2021. Hauser, A. and B uhlmann, P. Characterization and Greedy Learning of Interventional Markov Equivalence Classes of Directed Acyclic Graphs. The Journal of Machine Learning Research, 13(1):2409 2464, 2012. Hauser, A. and B uhlmann, P. Two Optimal Strategies for Active Learning of Causal Models from Interventions. International Journal of Approximate Reasoning, 55(4): 926 939, 2014. He, Y.-B. and Geng, Z. Active Learning of Causal Networks with Intervention Experiments and Optimal Designs. Journal of Machine Learning Research, 9:2523 2547, 2008. Heckerman, D. A Bayesian Approach to Learning Causal Networks. In Proceedings of the Eleventh conference on Uncertainty in artificial intelligence, pp. 285 295, 1995. Heckerman, D., Geiger, D., and Chickering, D. M. Learning Bayesian Networks: The Combination of Knowledge and Statistical Data . Machine learning, 20:197 243, 1995. Heckerman, D., Meek, C., and Cooper, G. A Bayesian Approach to Causal Discovery. Innovations in Machine Learning, pp. 1 28, 2006. Hoover, K. D. The logic of causal inference: Econometrics and the Conditional Analysis of Causation. Economics & Philosophy, 6(2):207 234, 1990. Hu, H., Li, Z., and Vetta, A. Randomized Experimental Design for Causal Graph Discovery. Advances in Neural Information Processing Systems, 27, 2014. Hyttinen, A., Eberhardt, F., and Hoyer, P. O. Experiment Selection for Causal Discovery. Journal of Machine Learning Research, 14:3041 3071, 2013. Ikram, M. A., Chakraborty, S., Mitra, S., Saini, S., Bagchi, S., and Kocaoglu, M. Root Cause Analysis of Failures in Microservices through Causal Discovery. In Advances in Neural Information Processing Systems, 2022. Jha, S., Banerjee, S., Tsai, T., Hari, S. K., Sullivan, M. B., Kalbarczyk, Z. T., Keckler, S. W., and Iyer, R. K. MLbased Fault Injection for Autonomous Vehicles: A Case for Bayesian Fault Injection. In 2019 49th annual IEEE/IFIP international conference on dependable systems and networks (DSN), pp. 112 124. IEEE, 2019. King, R. D., Whelan, K. E., Jones, F. M., Reiser, P. G. K., Bryant, C. H., Muggleton, S. H., Kell, D. B., and Oliver, S. G. Functional genomic hypothesis generation and experimentation by a robot scientist. Nature, 427(6971): 247 252, 2004. Kocaoglu, M., Dimakis, A., and Vishwanath, S. Cost Optimal Learning of Causal Graphs. In International Conference on Machine Learning, pp. 1875 1884. PMLR, 2017. Kraska, T., Beutel, A., Chi, E. H., Dean, J., and Polyzotis, N. The Case for Learned Index Structures. In Proceedings of the 2018 international conference on management of data, pp. 489 504, 2018. Active causal structure learning with advice Lattanzi, S., Lavastida, T., Moseley, B., and Vassilvitskii, S. Online Scheduling via Learned Weights. In Proceedings of the Fourteenth Annual ACM-SIAM Symposium on Discrete Algorithms, pp. 1859 1877. SIAM, 2020. Li, A. and Beek, P. Bayesian Network Structure Learning with Side Constraints. In International conference on probabilistic graphical models, pp. 225 236. PMLR, 2018. Lindgren, E. M., Kocaoglu, M., Dimakis, A. G., and Vishwanath, S. Experimental Design for Cost-Aware Learning of Causal Graphs. Advances in Neural Information Processing Systems, 31, 2018. Lykouris, T. and Vassilvitskii, S. Competitive Caching with Machine Learned Advice. Journal of the ACM (JACM), 68(4):1 25, 2021. Masegosa, A. R. and Moral, S. An interactive approach for Bayesian network learning using domain/expert knowledge. International Journal of Approximate Reasoning, 54(8):1168 1181, 2013. Meek, C. Causal Inference and Causal Explanation with Background Knowledge. In Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence, UAI 95, pp. 403 410, San Francisco, CA, USA, 1995a. Morgan Kaufmann Publishers Inc. ISBN 1558603859. Meek, C. Strong completeness and faithfulness in Bayesian networks. In Proceedings of the Eleventh conference on Uncertainty in artificial intelligence, pp. 411 418, 1995b. Mitzenmacher, M. A Model for Learned Bloom Filters, and Optimizing by Sandwiching. Advances in Neural Information Processing Systems, 31, 2018. Mitzenmacher, M. and Vassilvitskii, S. Algorithms with Predictions. Communications of the ACM, 65(7):33 35, 2022. Murphy, K. P. Active Learning of Causal Bayes Net Structure. Technical report, UC Berkeley, 2001. Pearl, J. Causality: Models, Reasoning and Inference. Cambridge University Press, USA, 2nd edition, 2009. ISBN 052189560X. Perkovic, E. Identifying causal effects in maximally oriented partially directed acyclic graphs. In Conference on Uncertainty in Artificial Intelligence, pp. 530 539. PMLR, 2020. Perkovic, E., Kalisch, M., and Maathuis, M. H. Interpreting and using CPDAGs with background knowledge. In Proceedings of the 2017 Conference on Uncertainty in Artificial Intelligence (UAI2017), pp. ID 120. AUAI Press, 2017. Pingault, J.-B., O reilly, P. F., Schoeler, T., Ploubidis, G. B., Rijsdijk, F., and Dudbridge, F. Using genetic data to strengthen causal inference in observational research. Nature Reviews Genetics, 19(9):566 580, 2018. Purohit, M., Svitkina, Z., and Kumar, R. Improving Online Algorithms via ML Predictions. Advances in Neural Information Processing Systems, 31, 2018. Reichenbach, H. The Direction of Time, volume 65. University of California Press, 1956. Rohatgi, D. Near-Optimal Bounds for Online Caching with Machine Learned Advice. In Proceedings of the Fourteenth Annual ACM-SIAM Symposium on Discrete Algorithms, pp. 1834 1845. SIAM, 2020. Rotmensch, M., Halpern, Y., Tlimat, A., Horng, S., and Sontag, D. Learning a Health Knowledge Graph from Electronic Medical Records. Scientific reports, 7(1):1 11, 2017. Rubin, D. B. and Waterman, R. P. Estimating the Causal Effects of Marketing Interventions Using Propensity Score Methodology. Statistical Science, pp. 206 222, 2006. Scheines, R., Spirtes, P., Glymour, C., Meek, C., and Richardson, T. he TETAD Project: Constraint Based Aids to Causal Model Specification. Multivariate Behavioral Research, 33(1):65 117, 1998. Shanmugam, K., Kocaoglu, M., Dimakis, A. G., and Vishwanath, S. Learning Causal Graphs with Small Interventions. Advances in Neural Information Processing Systems, 28, 2015. Solus, L., Wang, Y., and Uhler, C. Consistency guarantees for greedy permutation-based causal inference algorithms. Biometrika, 108(4):795 814, 2021. Spirtes, P., Glymour, C. N., Scheines, R., and Heckerman, D. Causation, Prediction, and Search. MIT press, 2000. Squires, C., Magliacane, S., Greenewald, K., Katz, D., Kocaoglu, M., and Shanmugam, K. Active Structure Learning of Causal DAGs via Directed Clique Trees. Advances in Neural Information Processing Systems, 33:21500 21511, 2020. Sverchkov, Y. and Craven, M. A review of active learning approaches to experimental design for uncovering biological networks. PLo S computational biology, 13(6): e1005466, 2017. Tan, C., Jin, Z., Guo, C., Zhang, T., Wu, H., Deng, K., Bi, D., and Xiang, D. Net Bouncer: Active Device and Link Failure Localization in Data Center Networks. In Proceedings of the 16th USENIX Conference on Networked Systems Design and Implementation, pp. 599 613, 2019. Active causal structure learning with advice Tong, S. and Koller, D. Active Learning for Parameter Estimation in Bayesian Networks. Advances in Neural Information Processing Systems, 13, 2000. Tong, S. and Koller, D. Active Learning for Structure in Bayesian Networks. In International joint conference on artificial intelligence, volume 17, pp. 863 869. Citeseer, 2001. Uhler, C., Raskutti, G., B uhlmann, P., and Yu, B. Geometry of the faithfulness assumption in causal inference. The Annals of Statistics, pp. 436 463, 2013. Verma, T. and Pearl, J. Equivalence and Synthesis of Causal Models. In Proceedings of the Sixth Annual Conference on Uncertainty in Artificial Intelligence, UAI 90, pp. 255 270, USA, 1990. Elsevier Science Inc. ISBN 0444892648. Wang, Q., Aliaga, J. R., Jha, S., Shanmugam, K., Bagehorn, F., Yang, X., Filepp, R., Abe, N., and Shwartz, L. Fault Injection based Interventional Causal Learning for Distributed Applications. In Innovative Applications of Artificial Intelligence Conference, 2023. Wang, S., Li, J., and Wang, S. Online Algorithms for Multi-shop Ski Rental with Machine Learned Advice. Advances in Neural Information Processing Systems, 33: 8150 8160, 2020. Wang, Y., Solus, L., Yang, K., and Uhler, C. Permutationbased Causal Inference Algorithms with Interventions. Advances in Neural Information Processing Systems, 30, 2017. Wei, A. Better and Simpler Learning-Augmented Online Caching. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2020). Schloss Dagstuhl-Leibniz Zentrum f ur Informatik, 2020. Wien obst, M., Bannach, M., and Li skiewicz, M. Extendability of Causal Graphical Models: Algorithms and Computational Complexity. In Uncertainty in Artificial Intelligence, pp. 1248 1257. PMLR, 2021a. Wien obst, M., Bannach, M., and Li skiewicz, M. Polynomial-Time Algorithms for Counting and Sampling Markov Equivalent DAGs. In Proccedings of the 35th Conference on Artificial Intelligence, AAAI, 2021b. Woodward, J. Making Things Happen: A Theory of Causal Explanation. Oxford University Press, 2005. Zhang, J. and Spirtes, P. Strong Faithfulness and Uniform Consistency in Causal Inference . In Proceedings of the Nineteenth conference on Uncertainty in Artificial Intelligence, pp. 632 639, 2002. Zhou, X., Peng, X., Xie, T., Sun, J., Ji, C., Liu, D., Xiang, Q., and He, C. Latent error prediction and fault localization for microservice applications by learning from system trace logs. In Proceedings of the 2019 27th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, pp. 683 694, 2019. Active causal structure learning with advice A. Remark about assumptions Under causal sufficiency, there are no hidden confounders (i.e. unobserved common causes to the observed variables). While causal sufficiency may not always hold, it is still a reasonable assumption to make in certain applications such as studying gene regulatory networks (e.g. see (Wang et al., 2017)). Faithfulness assumes that independencies that occur in the data do not occur due to cancellations in the functional relationships, but rather due to the causal graph structure. It is known (Meek, 1995b; Spirtes et al., 2000) that, under many natural parameterizations and settings, the set of unfaithful parameters for any given causal DAG has zero Lebesgue measure (i.e. faithfulness holds; see also Section 3.2 of (Zhang & Spirtes, 2002) for a discussion about faithfulness). However, one should be aware that the faithfulness assumption may be violated in reality (Andersen, 2013; Uhler et al., 2013), especially in the presence of sampling errors in the finite sample regime. Ideal interventions assume hard interventions (forcefully setting a variable value) and the ability to obtain as many interventional samples as desired, ensuring that we always recover the directions of all edges cut by interventions. Without this assumption, we may fail to correctly infer some arc directions and our algorithms will only succeed with some success probability. Our assumption that the given expert advice is consistent with observational essential graph is purely for simplicity and can be removed by deciding which part of the given advice to discard so that the remaining advice is consistent. However, we feel that deciding which part of the inconsistent advice to discard will unnecessarily complicate our algorithmic contributions without providing any useful insights, and thus we made such an assumption. B. Additional Preliminaries For any set A, we denote its powerset by 2A. We write {1, . . . , n} as [n] and hide absolute constant multiplicative factors in n using standard asymptotic notations O( ), Ω( ), and Θ( ). The indicator function 1predicate is 1 if the predicate is true and 0 otherwise. Throughout, we use G to denote the (unknown) ground truth DAG, its Markov equivalence class by [G ] and the corresponding essential graph by E(G ). We write A B and A \ B to represent the disjoint union and set difference of two sets A and B respectively. B.1. Graph basics We consider partially oriented graphs without parallel edges. Let G = (V, E) be a graph on |V | = n nodes/vertices where V (G), E(G), and A(G) E(G) denote nodes, edges, and arcs of G respectively. The graph G is said to be fully oriented if A(G) = E(G), fully unoriented if A(G) = , and partially oriented otherwise. For any subset V V and E E, we use G[V ] and G[E ] to denote the node-induced and edge-induced subgraphs respectively. We write u v to denote that two nodes u, v V are connected in G, and write u v or u v when specifying a certain direction. The skeleton skel(G) refers to the underlying graph where all edges are made undirected. A v-structure in G refers to a collection of three distinct vertices u, v, w V such that u v w and u w. A directed cycle refers to a sequence of k 3 vertices where v1 v2 . . . vk v1. An acyclic completion / consistent extension of a partially oriented graph refers to an assignment of edge directions to the unoriented edges E(G) \ A(G) such that the resulting fully oriented graph has no directed cycles. Suppose G = (V, E) is fully unoriented. For vertices u, v V , subset of vertices V V and integer r 0, define dist G(u, v) as the shortest path length between u and v, dist G(V , v) = minu V dist G(u, v), and N r G(V ) = {v V : dist G(v, V ) r} V as the set of vertices that are r-hops away from V , i.e. r-hop neighbors of V . We omit the subscript G when it is clear from context. Suppose G = (V, E) is fully oriented. For any vertex v V , we write Pa(v), Anc(v), Des(v) to denote the parents, ancestors and descendants of v respectively and we write Des[v] = Des(v) {v} and Anc[v] = Anc(v) {v} to include v itself. We define Ch(v) Des(v) as the set of direct children of v, that is, for any w Ch(v) there does not exists z V \ {v, w} such that z Des(v) Anc(w). Note that, Ch(v) {w V : v w} Des(v). Active causal structure learning with advice B.2. Causal graph basics A directed acyclic graph (DAG) is a fully oriented graph without directed cycles. By representing random variables by nodes, DAGs are commonly used as graphical causal models (Pearl, 2009), where the joint probability density f factorizes according to the Markov property: f(v1, . . . , vn) = Qn i=1 f(vi | pa(v)), where pa(v) denotes the values taken by v s parents. One can associate a (not necessarily unique) valid permutation / topological ordering π : V [n] to any (partially directed) DAG such that oriented arcs (u, v) satisfy π(u) < π(v) and unoriented arcs {u, v} can be oriented as u v without forming directed cycles when π(u) < π(v). For any DAG G, we denote its Markov equivalence class (MEC) by [G] and essential graph by E(G). DAGs in the same MEC have the same skeleton and the essential graph is a partially directed graph such that an arc u v is directed if u v in every DAG in MEC [G], and an edge u v is undirected if there exists two DAGs G1, G2 [G] such that u v in G1 and v u in G2. It is known that two graphs are Markov equivalent if and only if they have the same skeleton and v-structures (Verma & Pearl, 1990; Andersson et al., 1997). In fact, the essential graph E(G) can be computed from G by orienting v-structures in the skeleton skel(G) and applying Meek rules (see Appendix D). An edge u v is a covered edge (Chickering, 1995) if Pa(u) = Pa(v) \ {u}. We use C(G) E(G) to denote the set of covered edges of G. The following is a well-known result relating covered edges and MECs. Lemma B.1 ((Chickering, 1995)). If G and G belong in the same MEC if and only if there exists a sequence of covered edge reversals to transform between them. C. Additional Related Works on Causal Structure Learning Constraint-based algorithms, such as ours, use information about conditional independence relations to identify the underlying structure. From purely observational data, the PC (Spirtes et al., 2000), FCI (Spirtes et al., 2000) and RFCI algorithms (Colombo et al., 2012) have been shown to consistently recover the essential graph, assuming causal sufficiency, faithfulness, and i.i.d. samples. The problem of recovering the DAG using constraints from interventional data was first studied by Eberhardt et al. (2006; 2005); Eberhardt (2007). Many recent works (Hu et al., 2014; Shanmugam et al., 2015; Kocaoglu et al., 2017; Lindgren et al., 2018; Greenewald et al., 2019; Squires et al., 2020; Choo et al., 2022; Choo & Shiragur, 2023) have followed up on these themes. Score-based methods maximize a particular score function over the space of graphs. For observational data, the GES algorithm (Chickering, 2002) uses the BIC to iteratively add edges. Extending the GES, Hauser & B uhlmann (2012) proposed the GIES algorithm that uses passive interventional data to orient more edges. Hybrid methods, like Solus et al. (2021) for observational and Wang et al. (2017) for interventional data, use elements of both approaches. D. Meek rules Meek rules are a set of 4 edge orientation rules that are sound and complete with respect to any given set of arcs that has a consistent DAG extension (Meek, 1995a). Given any edge orientation information, one can always repeatedly apply Meek rules till a unique fixed point (where no further rules trigger) to maximize the number of oriented arcs. Definition D.1 (The four Meek rules (Meek, 1995a), see Figure 5 for an illustration). R1 Edge {a, b} E(G) \ A(G) is oriented as a b if c V such that c a and c b. R2 Edge {a, b} E(G) \ A(G) is oriented as a b if c V such that a c b. R3 Edge {a, b} E(G) \ A(G) is oriented as a b if c, d V such that d a c, d b c, and c d. R4 Edge {a, b} E(G) \ A(G) is oriented as a b if c, d V such that d a c, d c b, and b d. There exists an algorithm (Algorithm 2 of (Wien obst et al., 2021a)) that runs in O(d |E(G)|) time and computes the closure under Meek rules, where d is the degeneracy of the graph skeleton9. 9A d-degenerate graph is an undirected graph in which every subgraph has a vertex of degree at most d. Note that the degeneracy of a graph is typically smaller than the maximum degree of the graph. Active causal structure learning with advice Figure 5. An illustration of the four Meek rules E. Imperfect partial advice via MPDAGs In the previous sections, we discuss advice that occurs in the form of a DAG e G [G ]. However, this may be too much to ask for in certain situations. For example: The Markov equivalence class may be too large for an expert to traverse through and propose an advice DAG. The expert only has opinions about a subset of a very large causal graph involving millions of nodes / edges. As discussed in Section 2.4, we can formulate such partial advice as MPDAGs. Given a MPDAG as expert advice, a natural attempt would be to sample a DAG e G from it to use the full advice. Unfortunately, it is #P-complete even to count the number of DAGs consistent with a given MPDAG in general (Wien obst et al., 2021b) and we are unaware of any efficient way to sample uniformly at random from it. Instead, we propose to pick an arbitrary DAG e G as advice within the given MPDAG: pick any unoriented edge, orient arbitrarily, apply Meek rules, repeat until fully oriented. The following result follows naturally by maximizing over all possible DAGs consistent with the given partial advice. Theorem E.1. Fix an essential graph E(G ) with an unknown underlying ground truth DAG G . Given a set A of DAGs consistent with the given partial advice and intervention set bound k 1, there exists a deterministic polynomial time algorithm that computes an intervention set I adaptively such that EI(G ) = G , and |I| has size 1. O(max{1, log max e G A ψ(G , e G)} ν1(G )) 2. O(max{1, log max e G A ψ(G , e G)} log k νk(G )) when k = 1 and k > 1 respectively. F. Technical Overview for Theorem 3.1 As discussed in Section 2, it suffices to prove Theorem 3.1 with respect to moral DAGs. Our strategy for proving Theorem 3.1 is to consider two arbitrary DAGs Gs (source) and Gt (target) in the same equivalence class and transform a verifying set for Gs into a verifying set for Gt using Lemma B.1 (see Algorithm 2 for the explicit algorithm10). Instead of proving Theorem 3.1 by analyzing the exact sequence of covered edges produced by Algorithm 211 when transforming between the DAGs Gmin = argmin G [G ]ν1(G) and Gmax = argmax G [G ]ν1(G), we will prove something more general. Observe that taking both endpoints of any maximal matching of covered edges is a valid verifying set that is at most twice the size of the minimum verifying set. This is because maximal matching is a 2-approximation to the minimum vertex cover. Motivated by this observation, our proof for Theorem 3.1 uses the following transformation argument (Lemma F.3): for two DAGs G and G that differ only on the arc direction of a single covered edge x y, we show that given a conditional-root-greedy (CRG) maximal matching12 on the covered edges of G, we can obtain another CRG maximal matching of the same size on the covered edges of G , after reversing x y and transforming G to G . So, starting from Gs, we compute a CRG maximal matching, then we apply the transformation argument above on the sequence of covered edges given by Algorithm 2 until we get a CRG maximal matching of Gt of the same size. Thus, we can conclude that the minimum vertex cover sizes of Gs and Gt differ by a factor of at most two. This argument holds for any pair of DAGs (Gs, Gt) from the same MEC. 10Lemma 2 of (Chickering, 1995) guarantees that x y is a covered edge of the current Gs whenever step 9 is executed. 11The correctness of Algorithm 2 is given in (Chickering, 1995) where the key idea is to show that x y found in this manner is a covered edge. This is proven in Lemma 2 of (Chickering, 1995). 12A special type of maximal matching (see Definition F.1). Active causal structure learning with advice Algorithm 2 (Chickering, 1995): Transforms between two DAGs within the same MEC via covered edge reversals 1: Input: Two DAGs Gs = (V, Es) and Gt = (V, Et) 2: Output: A sequence seq of covered edge reversals that transforms Gs to Gt 3: seq 4: while Gs = Gt do 5: Fix an arbitrary valid ordering π for Gs. 6: Let A A(Gs) \ A(Gt) be the set of differing arcs. 7: Let y argminy V : Pa A(y) = {π(y)}. 8: Let x argmaxz Pa A(y){π(z)}. 9: Add x y to seq. 10: Update Gs by replacing x y with y x. 11: end while 12: return seq We now define what is a conditional-root-greedy (CRG) maximal matching. As the set of covered edges C(G) of any DAG G induces a forest (see Theorem 2.6), we define the CRG maximal matching using a particular greedy process on the tree structure of C(G). The CRG maximal matching is unique with respect to a fixed valid ordering π of G and subset S. We will later consider CRG maximal matchings with S = A(Gs) A(Gt), where the arc set S remains unchanged throughout the entire transformation process. Definition F.1 (Conditional-root-greedy (CRG) maximal matching). Given a DAG G = (V, E) with a valid ordering πG and a subset of edges S E, we define the conditional-root-greedy (CRG) maximal matching MG,πG,S as the unique maximal matching on C(G) computed via Algorithm 3: greedily choose arcs x y where the x has no incoming arcs by minimizing πG(y), conditioned on favoring arcs outside of S. Algorithm 3 Conditional-root-greedy maximal matching 1: Input: A DAG G = (V, E), a valid ordering πG, a subset of edges S E 2: Output: A CRG maximal matching MG,πG,S 3: Initialize MG,πG,S and C C(G) 4: while C = do 5: x argminz {u V | u v C}{πG(z)} 6: y argminz V : x z C{πG(z) + n2 1x z S} 7: Add the arc x y to MG,πG,S 8: Remove all arcs with x or y as endpoints from C 9: end while 10: return MG,πG,S To prove the transformation argument (Lemma F.3), we need to first understand how the status of covered edges evolve when we perform a single edge reversal. The following lemma may be of independent interest beyond this work. Lemma F.2 (Covered edge status changes due to covered edge reversal). Let G be a moral DAG with MEC [G ] and consider any DAG G [G ]. Suppose G = (V, E) has a covered edge x y C(G) E and we reverse x y to y x to obtain a new DAG G [G ]. Then, all of the following statements hold: 1. y x C(G ). Note that this is the covered edge that was reversed. 2. If an edge e does not involve x or y, then e C(G) if and only if e C(G ). 3. If x Ch G(a) for some a V \ {x, y}, then a x C(G) if and only if a y C(G ). 4. If b Ch G(y) and x b E(G) for some b V \ {x, y}, then y b C(G) if and only if x b C(G ). Using Lemma F.2, we derive our transformation argument. Active causal structure learning with advice Lemma F.3. Consider two moral DAGs G1 and G2 from the same MEC such that they differ only in one covered edge direction: x y E(G1) and y x E(G2). Let vertex a be the direct parent of x in G1, if it exists. Let S E be a subset such that a x S and x y, y x S (if a does not exist, ignore condition a x S). Suppose πG1 is an ordering for G1 such that y = argminz : x z C(G1){πG1(z) + n2 1x z S} and denote MG1,πG1,S as the corresponding CRG maximal matching for C(G1). Then, there exists an explicit modification of πG1 to πG2, and MG1,πG1,S to a CRG maximal matching MG2,πG2,S for C(G2) such that |MG1,πG1,S| = |MG2,πG2,S|. To be precise, given πG1, we will define πG2 in our proofs as follows: πG1(x) if v = y πG1(u) if v = x πG1(y) if v = u πG1(v) else As discussed earlier, Theorem 3.1 follows by picking Gs = argmax G [G ]ν1(G) and Gt = argmin G [G ]ν1(G), applying Algorithm 2 to find a transformation sequence of covered edge reversals between them, and repeatedly applying Lemma F.3 with the conditioning set S = A(Gs) A(Gt) to conclude that Gs and Gt have the same sized CRG maximal matchings, and thus implying that min G [G ] ν1(G) = ν1(Gs) 2 ν1(Gt) = 2 argmax G [G ]ν1(G). Note that we keep the conditioning set S unchanged throughout the entire transformation process from Gs to Gt. For an illustrated example of conditional-root-greedy (CRG) maximal matchings and how we update the permutation ordering, see Figure 6 and Figure 7. Figure 6. Consider the following simple setup of two DAGs G1 and G2 which agree on all arc directions except for x y in G1 and y x in G2. Dashed arcs represent the covered edges in each DAG. The numbers below each vertex indicate the πG1 and πG2 orderings respectively. In G1, u = argminz Ch G1 (x){πG1(z)}. Observe that Equation (1) modifies the ordering only for {x, y, u} (in blue) while keeping the ordering of all other vertices fixed. Suppose S = A(G1) A(G2) = {a b, a x, a y, a u, x b, x u, y b}. With respect to πG1 and S, The conditional-root-greedy maximal matchings (see Algorithm 3) are MG1,πG1 ,S = {a x, y b} and MG2,πG2 ,S = {a y, x b}. G. Deferred proofs G.1. Preliminaries Our proofs rely on some existing results which we first state and explain below. Lemma G.1 (Lemma 27 of (Choo et al., 2022)). Fix an essential graph E(G ) and G [G ]. If I 2V is a verifying set, then I separates all unoriented covered edge u v of G. Lemma G.2 (Lemma 28 of (Choo et al., 2022)). Fix an essential graph E(G ) and G [G ]. If I 2V is an intervention set that separates every unoriented covered edge u v of G, then I is a verifying set. Lemma G.1 tells us that we have to intervene on one of the endpoints of any covered edge in order to orient it while Lemma G.2 tells us that doing so for all covered edges suffices to orient the entire causal DAG. Active causal structure learning with advice Figure 7. Consider the following simple setup of two DAGs G3 and G4 which agree on all arc directions except for x y in G3 and y x in G4. Dashed arcs represent the covered edges in each DAG. The numbers below each vertex indicate the πG3 and πG4 orderings respectively. Observe that C(G3) = {x u, x y, y b}. If we define S = A(G3) A(G4) = {x b, x u, y b}, we see that the conditional-root-greedy maximal matchings (see Algorithm 3) are MG3,πG3 ,S = {x y} and MG4,πG4 ,S = {y x}. Note that Algorithm 3 does not choose x u C(G1) despite π(u) < π(y) because x u S, so π(y) < π(u) + n2. G.2. Verification numbers of DAGs within same MEC are bounded by a factor of two We use the following simple lemma in our proof of Lemma F.2. Lemma G.3. For any covered edge x y in a DAG G = (V, E), we have y Ch G(x). Furthermore, each vertex only appears as an endpoint in the collection of covered edges C(G) at most once. Proof. For the first statement, suppose, for a contradiction, that y Ch(x). Then, there exists some z V \ {x, y} such that z Des(x) Anc(y). Fix an arbitrary ordering π for G and let z = argmaxz Des(x) Anc(y){π(z)}. Then, we see that z y while z x since z Des(x). So, x y cannot be a covered edge. Contradiction. For the second statement, suppose, for a contradiction, that there are two covered edges u x, v x C(G) that ends with x. Since u x C(G), we must have v u. Since v x C(G), we must have u v. We cannot have both u v and v u simultaneously. Contradiction. Lemma F.2 (Covered edge status changes due to covered edge reversal). Let G be a moral DAG with MEC [G ] and consider any DAG G [G ]. Suppose G = (V, E) has a covered edge x y C(G) E and we reverse x y to y x to obtain a new DAG G [G ]. Then, all of the following statements hold: 1. y x C(G ). Note that this is the covered edge that was reversed. 2. If an edge e does not involve x or y, then e C(G) if and only if e C(G ). 3. If x Ch G(a) for some a V \ {x, y}, then a x C(G) if and only if a y C(G ). 4. If b Ch G(y) and x b E(G) for some b V \ {x, y}, then y b C(G) if and only if x b C(G ). Proof. The only parental relationships that changed when we reversing x y to y x are Pa G (y) = Pa G(y) \ {x} and Pa G (x) = Pa G(x) {y}. For any other vertex u V \ {x, y}, we have Pa G (u) = Pa G(u). The first two points have the same proof: as parental relationships of both endpoints are unchanged, the covered edge status is unchanged. 3. Since x y C(G), we have a y E(G). We prove both directions separately. Suppose a x C(G). Then, Pa G(a) = Pa G(x) \ {a}. Since x y C(G), then Pa G(x) = Pa G(y) \ {x}. So, we have Pa G (a) = Pa G(a) = Pa G(x) \ {a} = Pa G(y) \ {x, a} = Pa G (y) \ {a}. Thus, a y C(G ). Suppose a x C(G). Then, one of the two cases must occur: (a) There exists some vertex u such that u a and u x in G. Since x y is a covered edge, u x implies u y in G. Therefore, a y C(G ) due to u a. (b) There exists some vertex v such that v x and v a in G. There are two possibilities for v a: v a or v a. If v a, then v x a is a v-structure. If v a, then x Ch(a) since we have a v x. Both possibilities lead to contradictions. The first case implies a y C(G ) while the second case cannot happen. Active causal structure learning with advice 4. We prove both directions separately. Suppose y b C(G). Then, Pa G(y) = Pa G(b) \ {y}. Since x y C(G), then Pa G(x) = Pa G(y) \ {x}. So, we have Pa G (b) \ {x} = Pa G(b) \ {x} = Pa G(y) {y} \ {x} = Pa G(x) {y} = Pa G (x). Thus, x b C(G ). Suppose y b C(G). Then, one of the two cases must occur: There exists some vertex u y and u b. Since x y is a covered edge, u y implies u x. Therefore, x b C(G ) due to u b. There exists some vertex v b and v y. There are two possibilities for v y: v y or v y. If v y, then v b y is a v-structure. If v y, then b Ch(y) since we have y v b. Both possibilities lead to contradictions. The first case implies x b C(G ) while the second case cannot happen. Lemma F.3. Consider two moral DAGs G1 and G2 from the same MEC such that they differ only in one covered edge direction: x y E(G1) and y x E(G2). Let vertex a be the direct parent of x in G1, if it exists. Let S E be a subset such that a x S and x y, y x S (if a does not exist, ignore condition a x S). Suppose πG1 is an ordering for G1 such that y = argminz : x z C(G1){πG1(z) + n2 1x z S} and denote MG1,πG1,S as the corresponding CRG maximal matching for C(G1). Then, there exists an explicit modification of πG1 to πG2, and MG1,πG1,S to a CRG maximal matching MG2,πG2,S for C(G2) such that |MG1,πG1,S| = |MG2,πG2,S|. Proof. Define u = argminz Ch G1(x){πG1(z)} as the lowest ordered child of x. Note that Algorithm 3 chooses x y instead of x u by definition of y. This implies that x u S whenever u = y. Let us define πG2 as follows: πG1(x) if v = y πG1(u) if v = x πG1(y) if v = u πG1(v) else Clearly, πG1(x) < πG1(y) and πG2(x) > πG2(y). Meanwhile, for any other two adjacent vertices v and v , observe that πG1(v) < πG1(v ) πG2(v) < πG2(v ) so πG2 agrees with the arc orientations of πG1 except for x y. See Figure 6 for an illustrated example. Define vertex b as follows: b = argminz V : z Des(x) and y z C(G1){πG1(z) + n2 1x z S} If vertex b exists, then we know that b Ch G1(y) and x b C(G2) by Lemma G.3 and Lemma F.2. By minimality of b, Definition F.1 will choose y b if picking a covered edge starting with y for MG1,πG1,S. So, we can equivalently define vertex b as follows: b = argminz V : z Des(y) and x z C(G2){πG2(z) + n2 1x z S} By choice of πG2, Definition F.1 will choose x b if picking a covered edge starting with x for MG2,πG2,S. We will now construct a same-sized maximal matching MG2,πG2,S from MG1,πG1,S (Step 1), argue that it is maximal matching of C(G2) (Step 2), and that it is indeed a conditional-root-greedy matching for C(G2) with respect to πG2 and S (Step 3). There are three cases that cover all possibilities: Case 1 Vertex a exists, a x MG1,πG1,S, and vertex b exists. Case 2 Vertex a exists, a x MG1,πG1,S, and vertex b does not exist. Case 3 a x MG1,πG1,S. This could be due to vertex a not existing, or a x C(G1), or MG1,πG1,S containing a covered edge ending at a so a x was removed from consideration. Active causal structure learning with advice Step 1: Construction of MG2,πG2,S such that |MG2,πG2,S| = |MG1,πG1,S|. By Lemma F.2, covered edge statuses of edges whose endpoints do not involve x or y will remain unchanged. By definition of y, we know that Definition F.1 will choose x y if picking a covered edge starting with x for MG1,πG1,S. Since a x MG1,πG1 in cases 1 and 2, we know that there is no arcs of the form x in MG1,πG1,S. Since there is no arc of the form x in MG1,πG1,S in case 3, we know that x y MG1,πG1,S. Case 1 Define MG2,πG2,S = MG1,πG1,S {a y, x b} \ {a x, y b}. Case 2 Define MG2,πG2,S = MG1,πG1,S {a y} \ {a x}. Case 3 Define MG2,πG2,S = MG1,πG1,S {y x} \ {x y}. By construction, we see that |MG2,πG2,S| = |MG1,πG1,S|. Step 2: MG2,πG2,S is a maximal matching of the covered edge C(G2) of G2. To prove that MG2,πG2,S is a maximal matching of C(G2), we argue in three steps: 2(i) Edges of MG2,πG2,S belong to C(G2). 2(ii) MG2,πG2,S is a matching of C(G2). 2(iii) MG2,πG2,S is maximal matching of C(G2). Step 2(i): Edges of MG2,πG2,S belong to C(G2). By Lemma F.2, covered edge statuses of edges whose endpoints do not involve x or y will remain unchanged. Since MG1,πG1,S is a matching, it has at most one edge e involving endpoint x and at most one edge e involving endpoint y (e could be e). Case 1 Since b exists, the edges in MG1,πG1,S with endpoints involving {x, y} are a x and y b. By Lemma F.2, we know that a y, x b C(G2). Case 2 Since b does not exist, the only edge in MG1,πG1,S with endpoints involving {x, y} is a x. By Lemma F.2, we know that a y C(G2). Case 3 Since a x MG1,πG1,S, we have x y MG1,πG1,S by minimality of y. In all cases, we see that MG2,πG2,S C(G2). Step 2(ii): MG2,πG2,S is a matching of C(G2). It suffices to argue that there are no two edges in MG2,πG2,S sharing an endpoint. Since MG1,πG1,S is a matching, this can only happen via newly added endpoints in MG2,πG2,S. Case 1 The endpoints of newly added edges are exactly the endpoints of removed edges. Case 2 Since we removed a x and added a y, it suffices to check that there are no edges in MG1,πG1,S involving y. This is true since b does not exist in Case 2. Case 3 The endpoints of newly added edges are exactly the endpoints of removed edges. Therefore, we conclude that MG2,πG2,S is a matching of C(G2). Active causal structure learning with advice Step 2(iii): MG2,πG2,S is a maximal matching of C(G2). For any u v C(G2), we show that there is some edge in MG2,πG2,S with at least one of u or v is an endpoint. By Lemma F.2, covered edge statuses of edges whose endpoints do not involve x or y will remain unchanged, so it suffices to consider |{u, v} {x, y}| 1. We check the following 3 scenarios corresponding to |{u, v} {x, y}| 1 below: (i) y {u, v}. The endpoints of MG2,πG2 always contains y. (ii) y {u, v} and x v C(G2), for some v V \ {x, y}. Since x v C(G2) and y x in G2, it must be the case that y v in G2. Since G1 and G2 agrees on all arcs except x y, we have that y v in G1 as well. Since x v C(G2), we know that v Ch G2(x) via Lemma G.3. So, we have y v C(G1) via Lemma F.2. Since the set {v : y v C(G1)} is non-empty, vertex b exists. In both cases 1 and 3, the endpoints of MG2,πG2 includes x. (iii) y {u, v} and u x C(G2), for some u V \ {x, y}. By Lemma G.3, we know that x Ch G2(u). Meanwhile, since y x C(G2), we must have u y in G2. However, this implies that x Ch G2(u) since u y x exists. This is a contradiction, so this situation cannot happen. As the above argument holds for any u v C(G2), we see that MG2,πG2 is maximal matching for C(G2). Step 3: MG2,πG2,S is a conditional-root-greedy maximal matching. We now compare the execution of Algorithm 3 on (πG1, S) and (πG2, S). Note that S remains unchanged. We know the following: Since πG2(y) = πG1(x) and a x S, if a exists and a x is chosen by Algorithm 3 on (πG1, S), then it means that there are no a v arc in C(G1) such that a v S. So, a y will be chosen by Algorithm 3 on (πG2, S) if a exists. Since πG2(y) = πG1(x), x is chosen as a root by Algorithm 3 on (πG1, S) if and only if y is chosen as a root by Algorithm 3 on (πG2, S). By definition of b, if it exists, then y b MG1,πG1,S x b MG2,πG2,S. By the definition of πG2, we see that Algorithm 3 makes the same decisions when choosing arcs rooted on V \ {a, x, y, b}. Therefore, MG2,πG2,S is indeed a conditional-root-greedy maximal matching for C(G2) with respect to πG2 and S. Theorem 3.1. For any DAG G with MEC [G ], we have that max G [G ] ν1(G) 2 min G [G ] ν1(G). Proof. Consider any two DAGs Gs, Gt [G ]. To transform Gs = (V, Es) to Gt = (V, Et), Algorithm 2 flips covered edges one by one such that |Es \ Et| decreases in a monotonic manner. We will repeatedly apply Lemma F.3 with S = A(Gs) A(Gt) on the sequence of covered edge reversals produced by Algorithm 2. Let πGs be an arbitrary ordering for Gs and we compute an initial conditional-root-greedy maximal matching for C(Gs) with respect to some ordering πGs and conditioning set S. To see why Lemma F.3 applies at each step for reversing a covered edge from x y to y x, we need to ensure the following: 1. If x has a parent vertex a (i.e. x Ch G1(a)), then a x S. If a x S, then then a x is a covered edge that should be flipped to transform from Gs to Gt. However, this means that Algorithm 2 would pick a x to reverse instead of picking x y to reverse. Contradiction. Active causal structure learning with advice 2. x y, y x S. This is satisfied by the definition of S = Es Et since reversing x y to y x implies that neither of them are in S. 3. y = argminz : x z C(G1){πG1(z) + n2 1x z S}. Since x y S, this is equivalent to checking if y = argminz : x z C(G1){πG1(z)}. This is satisfied by line 7 of Algorithm 2. 4. MG1,πG1,S is a conditional-root-greedy maximal matching for C(G1) with respect to some ordering πG1 and conditioning set S. This is satisfied since we always maintain a conditional-root-greedy maximal matching and S is unchanged throughout. By applying Lemma F.3 with S = A(Gs) A(Gt) repeatedly on the sequence of covered edge reversals produced by Algorithm 2, we see that there exists a conditional-root-greedy maximal matching MGs,πGs for C(Gs) and a conditionalroot-greedy maximal matching MGt,πGt for C(Gt) such that |MGs,πGs | = |MGt,πGt|. The claim follows since maximal matching is a 2-approximation to minimum vertex cover, and the verification number ν(G) of any DAG G is the size of the minimum vertex cover of its covered edges C(G). Lemma 3.2 (Tightness of Theorem 3.1). There exist DAGs G1 and G2 from the same MEC with ν1(G1) = 2 ν1(G2). Proof. See Figure 8. Figure 8. The ratio of 2 in Theorem 3.1 is tight: G1 and G2 belong in the same MEC with ν(G1) = 2 and ν(G2) = 1. The dashed arcs represent the covered edges and the boxed vertices represent a minimum vertex cover of the covered edges. G.3. Adaptive search with imperfect advice Lemma 4.1. Fix a DAG G = (V, E) and let V V be any subset of vertices. Suppose IV V is the set of nodes intervened by Subset Search(E(G ), V ). If C(G ) E(G [V ]), then EIV (G ) = G . Proof. By Theorem 2.9, Subset Search fully orients edges within the node-induced subgraph induced by V , i.e. Subset Search will perform atomic interventions on IV V resulting in EIV (G )[V ] = G [V ]. Since C(G ) E(G [V ]) and all covered edges C(G ) were oriented, then according to Lemma G.1, it must be the case that V IV for some minimum vertex cover V of C(G ), so we see that R(G , V ) R(G , IV ). By Lemma G.2, we have R(G , V ) = A(G ) and so Subset Search(E(G ), V ) fully orients E(G ). We will now prove our main result (Theorem 3.5) which shows that the number of interventions needed is a function of the quality of the given advice DAG. Let us first recall how we defined the quality of a given advice and restate our algorithm. Active causal structure learning with advice Definition 3.4 (Quality measure). Fix a DAG G with MEC [G ] and consider any DAG e G [G ]. We define ψ(G , e G) as follows: ψ(G , e G) = max e V V( e G) ρ e V , N h(G ,e V ) skel(E(G ))(e V ) Algorithm 1 Adaptive search algorithm with advice. 1: Input: Essential graph E(G ), advice DAG e G [G ], intervention size k N 2: Output: An intervention set I such that each intervention involves at most k nodes and EI(G ) = G . 3: Let e V V( e G) be any atomic verifying set of e G. 4: if k = 1 then 5: Define I0 = e V as an atomic intervention set. 6: else 7: Define k = min{k, |e V |/2}, a = |e V |/k 2, and ℓ= loga |C| . Compute labelling scheme on e V with (|e V |, k, a) via Lemma 4.3 and define I0 = {Sx,y}x [ℓ],y [a], where Sx,y e V is the subset of vertices whose xth letter in the label is y. 8: end if 9: Intervene on I0 and initialize r 0, i 0, n0 2. 10: while EIi(G ) still has undirected edges do 11: if ρ(Ii, N r skel(E(G ))(e V )) n2 i then 12: Increment i i + 1 and record r(i) r. 13: Update ni ρ(Ii, N r skel(E(G ))(e V )) 14: Ci Subset Search(EIi(G ), N r 1 skel(E(G ))(e V ), k) 15: if EIi 1 Ci(G ) still has undirected edges then 16: C i Subset Search(EIi 1 Ci(G ), N r skel(E(G ))(e V ), k) 17: Update Ii Ii 1 Ci C i. 18: else 19: Update Ii Ii 1 Ci. 20: end if 21: end if 22: Increment r r + 1. 23: end while 24: return Ii Theorem 3.5. Fix an essential graph E(G ) with an unknown underlying ground truth DAG G . Given an advice graph e G [G ] and intervention set bound k 1, there exists a deterministic polynomial time algorithm (Algorithm 1) that computes an intervention set I adaptively such that EI(G ) = G , and |I| has size 1. O(max{1, log ψ(G , e G)} ν1(G )) when k = 1 2. O(max{1, log ψ(G , e G)} log k νk(G )) when k > 1. Proof. Consider Algorithm 1. Observe that n0 = 2 ensures that n2 0 > n0. In this proof, we will drop the subscript skel(E(G )) when we discuss the r-hop neighbors N r skel(E(G ))( ). We first prove the case where k = 1 then explain how to tweak the proof for the case of k > 1. If Algorithm 1 terminates when i = 0, then I = I0 = e V and Theorem 3.1 tells us that |I| O(ν1(G )). Now, suppose Algorithm 1 terminates with i = t, for some final round t > 0. As Algorithm 1 uses an arbitrary verifying set of e G in step 3, we will argue that O(max{1, log |N h(G ,e V )(e V )|} ν(G )) interventions are used in the while-loop, for any arbitrary chosen e V V( e G). The theorem then follows by taking a maximization over all possibilities in V( e G). In Line 12, r(i) records the hop value such that ρ(Ii, N r(i)(e V )) n2 i , for any 0 i < t. By construction of the algorithm, we know the following: Active causal structure learning with advice 1. For any 0 < i t, ni = ρ(Ii, N r(i)(e V )) n2 i 1 > ρ(Ii, N r(i) 1(e V )) (2) because r(i) 1 did not trigger Algorithm 1 to record r(i). 2. By Theorem 2.9 and Equation (2), for any 1 i t, |Ci| O(log ρ(Ii, N r(i) 1(e V )) ν1(G )) O(log ni 1 ν1(G )) |C i| O(log ρ(Ii, N r(i)(e V )) ν1(G )) O(log ni ν1(G )) (3) Note that the bound for |C i| is an over-estimation (but this is okay for our analytical purposes) since some nodes previously counted for ρ(Ii, N r(i)(e V )) may no longer be relevant in EIi Ci(G ) after intervening on Ci. 3. Since ni 1 ni for any 0 < i t, we know that nj n1/2t j t for any 0 j t. So, for any 0 t t, we have i=0 log(ni) i=0 log n1/2t i 2t i 2 log(nt ) (4) 4. By definition of t, h(G , e V ), and Lemma 4.1, r(t 1) < h(G , e V ) r(t) (5) and N r(t 1)(e V ) N h(G ,e V )(e V ) N r(t)(e V ) (6) Combining Equation (2), Equation (3), and Equation (4), we get i=1 (|Ci| + |C i|) O i=1 log ni 1 + log ni i=1 log ni ν1(G ) O (log nt 1 ν1(G )) (7) To relate |It| with |N h(G ,e V )(e V )|, we consider two scenarios depending on whether the essential graph was fully oriented after intervening on Ct or C t. Scenario 1: Fully oriented after intervening on Ct, i.e. EIt 1 Ct(G ) = G . Then, It = Ct It 1 = Ct (Ct 1 C t 1) It 2 = . . . = Ct i=1 (Ci C i) e V In this case, h(G , e V ) = r(t) 1. By definition, nt 1 |N r(t 1)(e V )| and we have nt 1 |N r(t 1)(e V )| < |N h(G ,e V )(e V )| (8) since N r(t 1)(e V ) N h(G ,e V )(e V ). So, |It| |e V | = |Ct| + i=1 (|Ci| + |C i|) O (log nt 1 ν1(G )) + O (log nt 1 ν1(G )) By Equation (3) and Equation (7) O log |N h(G ,e V )(e V )| ν1(G ) Equation (8) Scenario 2: Fully oriented after intervening on C t, i.e. EIt 1 Ct C t(G ) = G . Then, It = Ct C t It 1 = . . . = Ct C t i=1 (Ci C i) e V Active causal structure learning with advice In this case, h(G , e V ) = r(t) and N h(G ,e V )(e V ) = N r(t)(e V ). So, nt |N r(t)(e V )| = |N h(G ,e V )(e V )| (9) |It| |e V | = |Ct| + |C t| + i=1 (|Ci| + |C i|) O ((log nt 1 + nt) ν1(G )) + O (log nt 1 ν1(G )) By Equation (3) and Equation (7) O log |N h(G ,e V )(e V )| ν1(G ) Equation (9) Since |e V | O(ν1(G )), we can conclude |It| O ν(G ) + log |N h(G ,e V )(e V )| ν1(G ) O max n 1, log |N h(G ,e V )(e V )| o ν1(G ) in either scenario, as desired. The theorem then follows by taking a maximization over all e V V( e G). Adapting the proof for k > 1 By Theorem 4.2, νk(G ) ν1(G )/k . So, |I0| O(log k νk(G )) via Lemma 4.3. The rest of the proof follows the same structure except that we use the bounded size guarantee of Theorem 2.9, which incurs an additional multiplicative log k factor. Polynomial running time By construction, the Algorithm 1 is deterministic. Furthermore, Algorithm 1 runs in polynomial time because: Hop information and relevant nodes can be computed in polynomial time via breadth first search and maintaining suitable neighborhood information. It is known that performing Meek rules to obtain essential graphs takes polynomial time ((Wien obst et al., 2021a)). Algorithm 1 makes at most two calls to Subset Search whenever the number of relevant nodes is squared. Each Subset Search call is known to run in polynomial time (Theorem 2.9). Since this happens each time the number of relevant nodes is squared, this can happen at most O(log n) times. Theorem E.1. Fix an essential graph E(G ) with an unknown underlying ground truth DAG G . Given a set A of DAGs consistent with the given partial advice and intervention set bound k 1, there exists a deterministic polynomial time algorithm that computes an intervention set I adaptively such that EI(G ) = G , and |I| has size 1. O(max{1, log max e G A ψ(G , e G)} ν1(G )) 2. O(max{1, log max e G A ψ(G , e G)} log k νk(G )) when k = 1 and k > 1 respectively. Proof. Apply Theorem 3.5 while taking a maximization over all possible advice DAGs e G consistent with the given partial advice. H. Path essential graph In this section, we explain why our algorithm (Algorithm 1) is simply the classic binary search with prediction 13 when the given essential graph E(G ) is an undirected path on n vertices. So, another way to view our result is a generalization that works on essential graphs of arbitrary moral DAGs. 13e.g. see https://en.wikipedia.org/wiki/Learning_augmented_algorithm#Binary_search Active causal structure learning with advice When the given essential graph is a path E(G ) on n vertices, we know that there are n possible DAGs in the Markov equivalence class where each DAG corresponds to choosing a single root node and having all edges pointing away from it. Observe that a verifying set of any DAG is then simply the root node as the set of of covered edges in any rooted tree are precisely the edges incident to the root. Therefore, given any e G [G ], we se that h(G , e V ) measures the number of hops between the root of the advice DAG e G and the root of the true DAG G . Furthermore, by Meek rule R1, whenever we intervene on a vertex u on the path, we will fully orient the half of the path that points away from the root while the subpath between u and the root remains unoriented (except the edge directly incident to u). So, one can see that Algorithm 1 is actually mimicking exponential search from the root of e G towards the root of G . Then, once the root of G lies within the r-hop neighborhood H, Subset Search uses O(log |V (H)|) interventions, which matches the number of queries required by binary search within a fixed interval over |V (H)| nodes. I. Experiments In this section, we provide more details about our experiments. All experiments were run on a laptop with Apple M1 Pro chip and 16GB of memory. Source code implementation and experimental scripts are available at https://github.com/cxjdavin/ active-causal-structure-learning-with-advice. I.1. Experimental setup For experiments, we evaluated our advice algorithm on the synthetic graph instances of (Wien obst et al., 2021b)14 on graph instances of sizes n = {16, 32, 64}. For each undirected chordal graph instance, we do the following: 1. Set m = 1000 as the number of advice DAGs that we will sample. 2. Use the uniform sampling algorithm of (Wien obst et al., 2021b) to uniformly sample m advice DAGs e G1, . . . , e Gm. 3. Randomly select G from one of e G1, . . . , e Gm. 4. For each e G { e G1, . . . , e Gm}, Compute a minimum verifying set e V of e G. Define and compute ψ(G , e V ) = ρ e V , N h(G ,e V ) skel(E(G ))(e V ) . Compute a verifying set using (E(G ), e G) as input to Algorithm 1. 5. Aggregate the sizes of the verifying sets used based on ψ(G , e V ) and compute the mean and standard deviations. 6. Compare against verification number ν1(G ) and the number of interventions used by the fully adaptive search (without advice, which we denote as blind search in the plots) of (Choo et al., 2022). 7. Compute the empirical distribution of the quality measure amongst the m advice DAGs, then use standard sample complexity arguments for estimating distributions up to ε error in TV distance to compute a confidence interval for which the true cumulative probability density of all DAGs within the MEC lies within15. To be precise, it is known that for a discrete distribution P on k elements, when there are m max{k/ε2, (2/ε2) ln(2/δ)} uniform samples, the probability that the TV distance between the true distribution P and the empirical distribution P is less than ε is at least 1 δ. Since the upper bound on the domain size of quality measure is the number of nodes n, by setting m = 1000 and δ = 0.01, we can compute ε = max{ p (2/m) ln(2/δ)} and conclude that the probability that the true cumulative probability density of all DAGs within the MEC lies within ε distance (clipped to be between 0 and 1) of the empirical distribution is at least 99%. 14See Appendix E of (Wien obst et al., 2021b) for details about each class of synthetic graphs. Instances are available at https: //github.com/mwien/Clique Picking/tree/master/aaai_experiments 15For example, see Theorem 1 of (Canonne, 2020). Active causal structure learning with advice I.2. Experimental remarks The uniform sampling code of (Wien obst et al., 2021b) is written in Julia and it uses a non-trivial amount of memory, which may make it unsuitable for running on a shared server with memory constraints. Note that ψ(G , e V ) ψ(G , e G) = maxe V V( e G) ρ e V , N h(G ,e V ) skel(E(G ))(e V ) . We use ψ(G , e V ) as a proxy for ψ(G , e G) because we do not know if there is an efficient way to compute the latter besides the naive (possibly exponential time) enumeration over all possible minimum verifying sets. We also experimented with an unsafe variant of Algorithm 1 where we ignore the second tweak of intervening one round before. In our synthetic experiments, both variants use a similar number of interventions. We do not plot the theoretical upper bounds O(log ψ(G , e V ) ν1(G )) or O(log n ν1(G )) because these values are a significantly higher than the other curves and result in squashed (and less interesting/interpretable) plots. Even when ψ(G , e V ) = 0, there could be cases where (Choo et al., 2022) uses more interventions than ν1(G ). For example, consider Figure 8 with G = G2 and e G = G1. After intervening on e V = {b, c}, the entire graph will be oriented so the ψ(G , e V ) = 0 while ν1(G ) = 1 < 2 = |e V |. Fortunately, Theorem 3.1 guarantees that |e V | 2 ν1(G ). Note that the error bar may appear lower than the verification number even though all intervention sizes are at least as large as the verification number. For instance, if ν1(G ) = 6 and we used (6, 6, 7) interventions on three different e G s, each with ψ(G , e V ) = 0. In this case, the mean is 6.3333 . . . while the standard deviation is 0.4714 . . ., so the error bar will display an interval of [5.86 . . . , 6.80 . . .] whose lower interval is below ν1(G ) = 6. I.3. All experimental plots For details about the synthetic graph classes, see Appendix E of (Wien obst et al., 2021b). Each experimental plot is for one of the synthetic graphs G , with respect to 1000 |[G ]| uniformly sampled advice DAGs e G from the MEC [G ]. The solid lines indicate the number of atomic interventions used while the dotted lines indicate the empirical cumulative probability density of e G. The true cumulative probability density lies within the shaded area with probability at least 0.99. Figure 9. Subtree-logn synthetic graphs Active causal structure learning with advice Figure 10. Subtree-2logn synthetic graphs Figure 11. Subtree-sqrtn synthetic graphs Figure 12. Interval synthetic graphs Figure 13. peo-2 synthetic graphs Active causal structure learning with advice Figure 14. peo-4 synthetic graphs Figure 15. Thickening-3 synthetic graphs Figure 16. Thickening-logn synthetic graphs Figure 17. Thickening-sqrtn synthetic graphs