# the_parameterized_complexity_of_network_microaggregation__8b96a01c.pdf The Parameterized Complexity of Network Microaggregation V aclav Blaˇzej1, Robert Ganian2, Duˇsan Knop1, Jan Pokorn y1, ˇSimon Schierreich1, Kirill Simonov3 1Faculty of Information Technology, Czech Technical University in Prague, Prague, Czechia 2Algorithms and Complexity Group, Technische Universit at Wien, Vienna, Austria 3Hasso Plattner Institute, University of Potsdam, Postdam, Germany blazeva1@fit.cvut.cz, rganian@gmail.com, {dusan.knop,pokorj54,schiesim}@fit.cvut.cz, kirill.simonov@hpi.de Microaggregation is a classical statistical disclosure control technique which requires the input data to be partitioned into clusters while adhering to specified size constraints. We provide novel exact algorithms and lower bounds for the task of microaggregating a given network while considering both unrestricted and connected clusterings, and analyze these from the perspective of the parameterized complexity paradigm. Altogether, our results assemble a complete complexity-theoretic picture for the network microaggregation problem with respect to the most natural parameterizations of the problem, including input-specified parameters capturing the size and homogeneity of the clusters as well as the treewidth and vertex cover number of the network. Introduction In view of the rising importance of user anonymity and privacy, it is of paramount importance to ensure that release of data about relations in social or other networks does not lead to the disclosure of individual information, while also preserving the informational content. Such tasks lie at the heart of statistical disclosure control, and are typically tackled by creating synthetic data from the available real-world network (Das et al. 2022; Belhajjame et al. 2020; Kim, Venkatesha, and Panda 2022). One of the most classical approaches used to create such data is microaggregation (Domingo-Ferrer 2009; Yan et al. 2022), which typically aggregates available data into small homogeneous clusters and releases the centers of these clusters instead of the original data points. While it is always desirable to have a lower bound for the size of these clusters in order to achieve anonymity, depending on the context in which microaggregation is used it may either be useful to allow for clusters of variable size (so-called dataoriented microaggregation) or to require all the clusters to have roughly the same size (referred to as fixed-size microaggregation) (Domingo-Ferrer and Mateo-Sanz 2002; Sol e, Munt es-Mulero, and Nin 2012). The advantage of the former is that one can achieve more homogeneous clusters, while the latter ensures that each synthetic data point represents roughly the same amount of original data points. Indeed, without an upper bound on the cluster size, it may eas- Copyright 2023, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. ily happen that two data points in the microaggregated data represent sets of highly disproportionate sizes. While microaggregation has obvious connections to clustering, the addition of size control restrictions is not compatible with most clustering approaches which typically aim at maximizing the sizes of coherent clusters. By now, there is an extensive body of research studying typical network clustering models from an empirical (Rattigan, Maier, and Jensen 2007; Mukherjee, Sarkar, and Lin 2017) as well as complexity-theoretical perspective (Orecchia and Zhu 2014; Micha and Shah 2020). However, research on microaggregation of networks has so far focused on empirical aspects (Sun et al. 2012; Iftikhar, Wang, and Lin 2020) or (to a much smaller extent) on approximation specifically in the Euclidean space (Domingo-Ferrer, Seb e, and Solanas 2008; Domingo-Ferrer and Seb e 2006), while very little was known about the precise boundaries of tractability for optimal network microaggregation. In fact, we are aware of only a single paper that touched on this topic to date (Thaeter and Reischuk 2021), albeit the related topic of lower bounded clustering has been explored in two other works (Abu-Khzam et al. 2018; Casel 2019). While formal definitions are provided in the Preliminaries later, for the upcoming discussions it will be useful to already concretize the NETWORK MICROAGGREGATION problem (hereinafter NMA). On the input, we receive a network (modeled as an edge-weighted graph G), a lower bound ℓ, a distance bound d, and an upper bound u1. The task is then to partition the vertices of G into clusters such that each cluster (1) has size between ℓand u, and (2) admits a center (vertex) whose distance to every vertex in the cluster is at most d. We remark that, in view of the aim of anonymizing G, the center need not be part of the cluster and that a center can be reused for multiple clusters. Moreover, in connection to previous research on network clustering (Deligkas et al. 2022; Macgregor and Sun 2021; van Bevern et al. 2015), in some settings it is sensible to add a third requirement which avoids the creation of completely disconnected clusters: (3) each cluster must be connected in G. We denote this variant CONNECTED NETWORK MI- 1Setting u to a value slightly above or equal to ℓmodels fixedsize microaggregation, while setting u to the size of the input data set captures data-oriented microaggregation. The Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI-23) Figure 1: The figure depicts a network G which forms a double-tree; all edges have length 1. Setting ℓ= 4 and d = 2 might yield a single cluster over V (G) with a center in r. Setting additionally an upper bound u = 5 blocks this possibility, and can in particular obtain the highlighted clustering (with centers marked by squares), which is more succinct. CROAGGREGATION (hereinafter CNMA). The definition in the previous paragraph is close to that considered in previous works (Abu-Khzam et al. 2018), but it also generalizes these via the inclusion of an upper bound u on the size of the cluster. Note that in the standard case of NETWORK MICROAGGREGATION a lower bound ℓimmediately implies an upper bound of 2ℓ 1 since a cluster of size at least 2ℓcan be arbitrarily split in two while not increasing the distance to the center. However, in the case of CONNECTED NETWORK MICROAGGREGATION this operation is not necessarily possible, as the cluster might lose its connectivity. Therefore having only a lower bound might yield arbitrarily large clusters in this case, leading to an extremely non-informative representation of the data (see Figure 1 for an example). Moreover, even in the case of NETWORK MICROAGGREGATION, specifying an upper bound allows for a finer control of the cluster sizes, as in the fixed-size microaggregation setting. Contribution. We initiate the comprehensive study of the complexity of both NMA and CNMA. While both problems are easily seen to be NP-complete in their full generality, in many scenarios of interest the instances we deal with have specific structural properties and the natural question that arises is whether one can exploit such properties to obtain exact algorithms with good runtime guarantees. To provide a comprehensive answer to this question, we turn to the parameterized complexity paradigm (Downey and Fellows 2013; Cygan et al. 2015) that has by now been successfully applied to a large number of problems from numerous areas of computer science, including in many settings relevant to artificial intelligence research (Kronegger et al. 2014; Ganian et al. 2020; Dvoˇr ak et al. 2021). Essentially, in the parameterized refinement of classical complexity, one analyzes the performance of algorithms not only with respect to the input size n but also in view of one or several numerical parameters k which capture specific properties of the input. If the problem can be solved in time f(k) n O(1) for some computable function f, we say that it is fixed-parameter tractable (FPT). The associ- ated algorithm is then also called fixed-parameter, and this represents the most favorable outcome for an NP-hard problem. A weaker notion of tractability is captured by the class XP, which contains all problems that can be solved in polynomial time for every constant value of the parameter. Naturally, it may also happen that a problem remains NP-hard even for some constant parameter value, and similarly we may also use the established notion of W[1]-hardness to exclude fixed-parameter algorithms under a specific parameterization. When considering natural parameterizations of NMA and CNMA, two well-motivated parameters are d (which ensures that the aggregated clusters are coherent) and u (which ensures that the clusters are small). As our main contribution, we establish the complexity of both problems under all combinations of these two parameters along with structural parameterizations of the network. There, we consider treewidth (Robertson and Seymour 1984) as the most prominent and natural structural graph parameter, as well as the vertex cover number as a restriction of treewidth that has typically been used to achieve tractability for problems which are intractable w.r.t. treewidth. Our main results are summarized in Table 1; in the rest of this section, we provide an overview and context for the individual results. As an initial starting point, we observe that NMA as well as CNMA remain NP-hard even for fixed values of d and u, which means that to achieve tractability one needs to include at least some parameterization of the network (that is why Table 1 only considers combinations of d and u with network parameterizations; it is worth noting that the same considerations also apply to ℓ, since ℓ u). Next, when using the treewidth tw of the network (as our baseline structural parameter), we show that (a) NMA is XP parameterized by tw + d, (b) CNMA is XP parameterized by tw alone, and (c) CNMA is FPT parameterized by tw + d + u. We obtain these by designing two new algorithms one for each problem variant where results (b) and (c) follow as a corollary of the same algorithm for CNMA. These algorithms utilize the dynamic programming technique employed by virtually all treewidth-based algorithms, but are non-trivial and especially result (a) required the introduction of an advanced cluster-grouping step. We remark that none of these algorithms seem to be obtainable via known algorithmic metatheorems for treewidth such as Courcelle s Theorem (Courcelle 1990) or MSOL Partitioning (Rao 2007). Next, we turn to the question of whether one can obtain more favorable tractability results by exploiting the vertex cover number vc of the network. In this setting, we apply a range of techniques to establish the remaining tractability results depicted in Table 1. We begin with a trivial observation that CNMA is FPT parameterized by vc + u. To obtain fixed-parameter tractability w.r.t. the same parameterization for NMA, we make use of a new structural observation which guarantees that every sufficiently large YESinstance of the problem must admit a solution with homogeneous clusters, which in turn allows us to apply the classical kernelization technique (Cygan et al. 2015); for a similar approach see, e.g., (Mnich and Wiese 2015). To establish the fixed-parameter tractability of both problems tw vc tw vc NP-h W[1]-h (Thm. 13), XP (Thm. 9) W[1]-h, XP (Cor. 3) W[1]-h (Thm. 14), XP d W[1]-h, XP (Thm. 1) FPT (Thm. 8) W[1]-h (Thm. 10), XP FPT (Thm. 7) u NP-h (Thm. 12) FPT (Thm. 5) W[1]-h (Thm. 15), XP FPT (Obs. 4) d + u W[1]-h (Thm. 11), XP FPT FPT (Cor. 3) FPT Table 1: Our results the full complexity-theoretic landscape of (CONNECTED) NETWORK MICROAGGREGATION under all combinations of considered input-specified parameters (rows) and structural parameters (columns); NP-h means that the problem remains NP-hard even for a fixed value of the parameters. w.r.t. vc +d, we then use a different tool, notably exhaustive branching in combination with the construction of an ILP encoding with a bounded number of variables which can then be solved by Lenstra s celebrated result (Lenstra Jr. 1983) and its subsequent improvements (Kannan 1987; Frank and Tardos 1987). Interestingly, for our final result the XP-tractability of NMA when parameterized by vc we need to combine both the techniques mentioned above: in particular, we show that every sufficiently large YESinstance admits a solution which is well-structured (albeit in a different way than we required for fixed-parameter tractability), and then use this fact to design an ILP which captures the problem. Finally, we complement all of the aforementioned algorithms with lower bounds that justify the presence of every parameter in each of the algorithms; in other words, the classification arising from these algorithms is tight. First, we observe that the known intractability of EQUITABLE CONNECTED PARTITION parameterized by tw almost immediately yields the intractability of CNMA w.r.t. tw + d. Next, we establish W[1]-hardness of both problems parameterized by vc alone, as well as the W[1]-hardness of CNMA parameterized by tw + u using two reductions from the classical INDEPENDENT SET and MULTICOLORED CLIQUE problems; the main difficult here is that it was necessary to encode the structure of the input graph via distances in the network. For the last two open cases, notably NMA parameterized by tw + u and by tw + u + d, it is far from obvious how one could reduce from the usual problems used to establish intractability: indeed, the intractability of these variants stems from the possible presence of many incomplete clusters which need to be paired up with each other. Hence, to establish our final lower bounds, we use a more recent reduction technique and start from the MULTIDIMENSIONAL SUBSET SUM problem (Ganian, Klute, and Ordyniak 2021). Related Work. The general concept of k-anonymity, i.e., replacing a dataset with a similar one but where each entry is indistinguishable from at least k 1 others, was introduced by Sweeney (2002). Aggarwal et al. (2010) proposed the specific model of k-anonymity where the entries are represented by their respective centers in a clustering. In that work, the particular optimization problem is called r-GATHER; the difference to NETWORK MICROAGGREGATION is only that r-GATHER does not specify an upper bound on the size of the cluster, and the authors showed that r-GATHER is NP-complete, but admits a 2-approximation. In the Euclidean space with the sum-of-squares clustering objective, it is known that an analogue of r-GATHER admits a O(r3)-factor approximation (Domingo-Ferrer, Seb e, and Solanas 2008), and a 2-approximation for r = 2 (Domingo Ferrer and Seb e 2006). An overview of various versions of r-GATHER on networks was conducted (Abu-Khzam et al. 2018). The authors of that work considered several variants of interand intra-cluster distance measure, and showed a number of approximation algorithms and NP-hardness results. The NETWORK MICROAGGREGATION problem is related to the well-studied k-Center clustering problem, where clusters have no size restrictions, but instead the number of clusters k is predetermined. In particular, Feldmann and Marx (2020) studied parameterized algorithms and complexity of k-Center clustering on networks for various structural parameters of the network. A number of recent works studied parameterized algorithms for clustering with size constraints on the clusters, among others with k-Median as the optimization objective (Cohen-Addad and Li 2019; Bandyapadhyay, Fomin, and Simonov 2021). It is worth to note that all of the results on k-Center/k-Median mentioned above offer little insight in our setting, as they heavily exploit small number of clusters. Last but not least, we mention that there is a large body of work exploring the computational complexity of data clustering in various settings, e.g., when data is missing (Ganian et al. 2020, 2022). Preliminaries We use bold-face letters (e.g., x) for vectors and normal font for their entries (i.e., x2 is the second entry of vector x). For a positive integer n we denote by [n] the set of all positive integers up to (and including) n, that is, [n] = {1, 2, . . . , n}, and let [n]0 = [n] {0}. Network Microaggregation. Since the networks in our problems are typically modeled as graphs, we employ basic graph-theoretic terminology (Matouˇsek and Neˇsetˇril 2009; Diestel 2012). A graph G = (V, E) has a vertex set V and an edge set E. Most graphs in this paper are edge-weighted; we treat the edge-weight ω(e) of an edge e as its length. For two vertices u and v the distance dist G(u, v) is the length of a shortest path between u and v (and is if no such path exists). Our problem of interest is: NETWORK MICROAGGREGATION (NMA) Input: An undirected n-vertex graph G = (V, E), a lower bound ℓ N, an upper bound u N, a maximum allowed distance to a cluster center d N and a length function ω: E [d]. Question: Is there an integer m and a partition Π = (C1, . . . , Cm) of V together with a list of vertices C = (c1, . . . , cm) such that i [m]: ℓ |Ci| u, v Ci : dist G(v, ci) d? The CONNECTED NETWORK MICROAGGREGATION problem (CNMA) is defined analogously, but with the additional requirements that the subgraph induced by each cluster C Π is connected. We mention that while (C)NMA are formulated as decision problems for complexity-theoretic reasons, all our algorithms are constructive and can also output a microaggregation Π as a witness. The most crucial parts of the input, apart from the structure of the given network, are arguably those that determine the quality of the considered microaggregation. Clearly, this includes the parameter d as this one directly governs the quality of the microaggregated data (the closer the centers are to the original datapoints, the higher the quality of the data collected by the algorithm). Furthermore, by the parameter u we can directly affect the number of datapoints in the aggregated data. Therefore, in many settings it would be desirable to keep the value of both d and u rather low. In fact, the same applies to u ℓas this governs the proportionality of the collected data. We remark that both CONNECTED NETWORK MICROAGGREGATION and NETWORK MICROAGGREGATION remain NP-complete even for constant values of both u and d; indeed, for the former this follows from the known NP-hardness of the P3PARTITIONING graph problem (van Bevern et al. 2015; Kirkpatrick and Hell 1978; Hell and Kirkpatrick 1982) while for the latter it is, e.g., an immediate corollary of our stronger Theorem 11. Structural Parameters. Let G = (V, E) be a graph. A set M V is a vertex cover of G if M e = for every e E. The vertex cover number of G, denoted vc, is the minimum size of a vertex cover of G. A nice tree-decomposition T D of an undirected graph G = (V, E) is a pair (Tree = (W, F), {Xx | x W}), where Tree is a tree rooted at a node r W and a bag Xx V is a set of vertices of G associated with node x such that: For every vw E there is a node x for which v, w Xx. For every vertex v V , the set of nodes x satisfying v Xx forms a subtree of Tree. |Xx| = 0 for every leaf x of Tree and |Xr| = 0. There are only three kinds of non-leaf nodes in Tree: Introduce node: a node x with exactly one child y such that Xx = Xy {v} for some vertex v Xy. Forget node: a node x with exactly one child y such that Xx = Xy \ {v} for some vertex v Xy. Join node: a node x with two children y and z such that Xx = Xy = Xz. The width of a nice tree-decomposition (Tree, {Xx | x V (Tree)}) is maxx V (Tree)(|Xx| 1), and the treewidth of the graph G, denoted tw, is the minimum width of a nice tree-decomposition of G. Let x be a node of a nice tree decomposition. We denote by V x the set of all past vertices contained in the subtree rooted at x, while vertices outside of V x are said to be in the future. Formally, we have V x = S y predecessor of x Xy. It is worth noting that the structure of real-world data has been demonstrated to attain low treewidth in several settings (Maniu, Senellart, and Jog 2019). Moreover, it is well known (and easy to see) that vc tw. Algorithms for Tree-Like Networks In this section, we provide dynamic programming algorithms for NMA and CNMA that establish the tractable fragments of these problems when treewidth is used as a parameter. We remark that while the use of dynamic programming that relies on bags acting as separators in the graph is the golden standard for treewidth-based algorithms, the technical details here are far from standard. Among others, in order to obtain dynamic programming tables which are succinct enough for our purposes, we needed to identify a suitable notion of vertex types that capture their properties when used as a center and incorporate these into the tables. As the types used in this section will crucially depend on vertex distances, it will be useful to introduce some additional terminology to capture this. As we will not need to distinguish distances longer than d, let distd(v, w) = dist(v, w) if dist(v, w) d and otherwise. Our operations over distances are additive, assume the result is set to whenever the value exceeds d. For each node x in a tree-decomposition, we can now partition the vertices of G into center-types that are based on their membership to V x and their distances to Xx. Formally, let the center-type of a vertex v for v V (G) be tx,v = (distd(v, w1), distd(v, w2), . . . , distd(v, w|Xx|)), where wi are the vertices of Xx. A core ingredient for the proof is that if two vertices have the same type and are both in the past or both in the future, then they are indistinguishable from the viewpoint of the bag. We remark that the notion of center-types is related to the well-studied METRIC DIMENSION problem (Bonnet and Purohit 2021). Both of the presented algorithm in this section proceed by leaf-to-root dynamic programming along a nice tree-decomposition computed using well-known algorithms (Bodlaender 1996; Korhonen 2021). We begin by establishing the XP-tractability of NETWORK MICROAGGREGATION with respect to treewidth and d which, while still involving some technical challenges, is the easier of the two results in this section. Theorem 1. NETWORK MICROAGGREGATION can be solved in time n O((d+2)tw). Proof Sketch. The algorithm stores binary records at each node x in the following form: D[x, f x, px] {T, F}, where: f x : T x [n]0 is the total number of vertices in clusters in V x \ Xx that have a center with center-type T x that occurs in the future, px : T x [n]0 [n]0 is the lower and upper bound on the required number of vertices that are expected to be added to clusters that have a center with respective center-type in the past. Intuitively, the records hold aggregated information about clusters with centers of the same center-type. Vertices within clusters that have future centers with the same center-type are interchangeable so we just need their amount. Partially formed clusters that have centers in the past require some number of additional vertices to have the correct size. We can aggregate those clusters that have the same past centertype because it will not matter which one a new vertex satisfies; the total demand still decreases by one. Through the computation, we immediately resolve aggregated clusters with center-type t = |Xx| since no more vertices may be added to these. The semantics of the records are as follows: a record is true if and only if the subinstance induced on V x admits a partial solution satisfying the conditions given by f and p. Clearly, the number of records is upper-bounded by n O((d+2)tw). Moreover, the records can be trivially determined at each leaf node, and once we obtain the records for the root node r of the tree-decomposition, we can correctly answer Yes if and only if D[r, , ] = T since Xr = . Hence, to complete the proof it suffices to show that the records can be computed in a leaf-to-root fashion along the nodes of the tree-decomposition. Next, we show that for CONNECTED NETWORK MICROAGGREGATION one can also use treewidth to achieve tractability, albeit with an entirely different dynamic programming algorithm. Theorem 2. CONNECTED NETWORK MICROAGGREGATION can be solved in time O(n3) + d O(tw2)u O(tw)n. Corollary 3. CONNECTED NETWORK MICROAGGREGATION is in XP parameterized by tw only, and is fixedparameter tractable when parameterized by tw + u + d. Algorithms Using the Vertex Cover Number In this section, we investigate the complexity of both considered problems with respect to the vertex cover number. It is well known that the vertex cover can be computed by an efficient fixed-parameter algorithm, and hence throughout this section we assume that G already comes equipped with a minimum vertex cover M of size vc. The results of this section are obtained via the application of two distinct techniques, and to streamline the presentation we divide the section into two subsections accordingly. While the techniques differ, an idea that will be crucial in both subsections is that even though the number of vertices we need to deal with may be large, one can group them into equivalence classes based on a suitable notion of type . Interestingly, how these types need to be set up in order for the algorithms to work differs from scenario to scenario (and also differs from the previous section). Moreover, in some cases it will be necessary to define not only types for vertices, but also for clusters. Tractability via Kernelization Kernelization is a procedure that transforms the instance into an equivalent one whose size is bounded by the parameter value (where the reduced instance can then be solved, e.g., by brute force or any heuristic), and is the core technique used in this subsection. Kernelization is required to run in polynomial time and is typically achieved by using problemspecific reduction rules. For each such rule, one typically needs to establish that it is safe , meaning that it preserves Yes and No instances. We begin with a simple observation for CONNECTED NETWORK MICROAGGREGATION: every Yes-instance of CONNECTED NETWORK MICROAGGREGATION may only contain at most vc u vertices. Observation 4. CONNECTED NETWORK MICROAGGREGATION is fixed-parameter tractable parameterized by the vertex-cover number vc plus the upper bound u. Next, we turn our attention to NETWORK MICROAGGREGATION with the same parameterization. Before we introduce the reduction rule, we need a few definitions. We define the neighborhood-type of a vertex v for v V \ M, where M is a vertex cover of G. Let the neighborhoodtype tv of a vertex be the set of its neighbors, i.e., its open neighborhood. For a cluster C V \M we define its clustermultitype t(C) to be the multiset {tv | v C}, i.e., a multiset of neighborhood-types of its vertices. We stress that cluster-multitypes are defined only for those clusters that do not contain any vertices of M. We call a cluster C homogeneous if tv = tw for all v, w C, and heterogeneous otherwise. By proving that every Yes-instance also admits a nice solution where each heterogeneous cluster-multitype occurs only at most ℓmany times, we show: Reduction Rule 1. Suppose there are at least u (2vc u + ℓ u + vc) vertices of the same neighborhood-type. Then, remove ℓsuch vertices. The exhaustive application of Reduction Rule 1 then suffices to obtain a kernel of size at most 2O(vc u), and hence: Theorem 5. NETWORK MICROAGGREGATION is fixedparameter tractable when parameterized by vc + u. Algorithms Based on Branching and ILP Our next set of algorithms combines exhaustive branching with an encoding into an Integer Linear Program (ILP) where the number of variables is bounded by the parameters; such ILPs are known to be fixed-parameter tractable: Proposition 6 (Lenstra Jr. 1983; Kannan 1987; Frank and Tardos 1987). There is an algorithm that solves an input ILP instance I with p variables in time p O(p) |I|. We start by defining the types of vertices outside the vertex cover and the types of clusters, both of which are based on the distances to the vertices in M. Definition 1. Let M = {v1, v2, . . . vvc} be a vertex cover of the graph G = (V, E) and let v V \ M. The vertex-type tv {[d] { }}M of v is tv i = ω(v, vi) if {v, vi} E, otherwise. By T we denote the set of all vertex-types of vertices. Definition 2. Let C V be a cluster. The cluster-type t(C) T of C is defined as t(C) = {t T | v C : tv = t}. Moreover, let c T M. We call (t(C), c) the extended-cluster-type of cluster C with center c. We observe that there are |T| (d + 1)vc different vertex-types, plus additional vc vertices in the vertex cover that are dealt with separately. Moreover, the number of different cluster-types is at most 2vc +|T | = 2vc +(d+1)vc 2d O(vc). When also considering the centers, we get at most 2d O(vc) (vc +(d+1)vc) 2d O(vc) possible extended-clustertypes. Theorem 7. CONNECTED NETWORK MICROAGGREGATION is fixed-parameter tractable parameterized by the vertex-cover number vc and the maximum distance d. Proof Sketch. We begin by observing that the number of clusters can be upper bounded by vc. We then branch to determine the number m [vc] of clusters in the solution, their extended-cluster-types, and which vertex cover vertices they contain. In each branch, we perform a set of basic checks to prune out invalid choices (such as when the extendedcluster-types of clusters cannot correspond to a connected cluster). Then we construct an ILP with at most |T| vc variables that is a Yes-instance if and only if there is a solution that corresponds to the given branch. Next, we turn to NETWORK MICROAGGREGATION. The following algorithm is based on similar ideas as the previous Theorem, but there is an additional complication: the number of clusters is no longer upper bounded by vc. Hence, we use variables in the ILP formulation to capture how often each of the extended-cluster-types occurs. Theorem 8. NETWORK MICROAGGREGATION is fixedparameter tractable parameterized by vc + d. The final and the most complicated algorithm in this section combines both of the techniques: we first show that there is a solution with certain properties (as was done for Reduction Rule 1), and we then use this to design an ILP formulation as in the previous two algorithms. Theorem 9. NETWORK MICROAGGREGATION is in XP parameterized by the vertex-cover number vc. Proof Sketch. As the core ingredient in the proof, we show that every Yes-instance also admits a solution with at most 2O(vc) cluster centers not in M; the proof of this claim is non-trivial and involves a swapping argument on an auxiliary graph representation of the solution. Once we establish the claim, we use exhaustive branching to determine the exact centers which lie outside of M, and at that point we are at a stage where the existence of a solution can be checked by a non-trivial ILP with boundedly-many variables which also relies on a suitable partitioning of vertices into types. Lower Bounds This section contains proofs for all lower bounds (i.e., hardness results) required to complete Table 1. Before proceeding to the more interesting results, we first provide a straightforward reduction from EQUITABLE CONNECTED PARTITION a problem which is well-known to be W[1]-hard when parameterized by treewidth (Enciso et al. 2009). Theorem 10. CONNECTED NETWORK MICROAGGREGATION is W[1]-hard parameterized by tw, even if d = 1. Reductions from Multidimensional Subset Sum In this section, we provide lower-bounds for NETWORK MICROAGGREGATION when parameterized by treewidth by reducing from a problem called MULTIDIMENSIONAL SUBSET SUM (MSS for short), which is defined as follows The input of the MULTIDIMENSIONAL SUBSET SUM problem contains n dim-dimensional vectors a1, . . . , an and a target vector c = (c1, . . . , cdim). The question is whether there is a subset S [n] such that P i S ai = c. MSS is known to be W[1]-hard parameterized by the number of dimensions, even if all integers in the input are given in unary (Ganian, Klute, and Ordyniak 2021). The reason we use MSS here is that the core of the intractability in this case notably, the fact that one would need to store too much information about incomplete clusters when performing dynamic programming makes the use of direct reductions from classical W[1]-hard problems problematic. The following gadget is the centerpiece of our reductions. Definition 3. A choice gadget choice(k, ℓ, d) for ℓ> k + 3 is a graph consisting of k + ℓvertices constructed in such a way that there are only two possible clusters that may cover ℓ of the vertices these two possibilities correspond to the active and inactive state of the gadget (see Figure 2). Let {u V (G), X V (G)} denote the edge set {{u, x} | x X}. Formally, the choice gadget is defined as follows: choice(k, ℓ, d) = W A I {v1, v2, v3}, {{v1, W}, {v1, v2}, {v1, v3}, {v2, A}, {v3, I}} , where W = {w1, . . . , wℓ 3 k}, A = {a1, . . . , ak}, and I = {i1, . . . , ik}. Edges have lengths ω({v1, W}) = d 1, ω({v1, v2}) = ω({v1, v3}) = 1, and ω({v2, A}) = ω({v3, I}) = d. This gadget shall be connected to other gadgets only via active vertices A and inactive vertices I. Theorem 11. NETWORK MICROAGGREGATION is W[1]- hard parameterized by the treewidth tw even if ℓ, u, and d are fixed constants. Proof Sketch. Let I = (c, a1, a2, . . . , an) be an instance of the MSS problem. Assume, for the sake of the argument, that we would be able to use arbitrarily large clusters, and set ℓ= u 1. In this case, we could model each v2 v3 ℓ 3 k W d 1 1 1 Figure 2: An illustration of a choice gadget choice(k, ℓ, d). Figure 3: Example of a duplication gadget that duplicates a vertex s into 9 vertices S. vector ai as a choice gadget choice(2 |ai|, ℓ, d) where |ai| = P j [dim] ai j. The active state of the choice gadget for ai corresponds with i S in the MSS. Every inactive vertex in Ii of the choice gadget connects to a center of its auxiliary star with ℓ 1 leaves. This allows clusters that cover the auxiliary stars to optionally cover Ii. Hence, if a choice gadget is in the inactive state, then all of its vertices are covered with the cluster within the choice gadget and clusters that cover the auxiliary stars. Now, we split the active vertices Ai of each choice gadget into groups Ai j, and Ai j, so that |Ai j, | = |Ai j, | = cj. We first ensure that P i active Ai j, cj for every j [dim]. For every j [dim] we create a demand gadget Dj that is a tree where the root has cj children, each of which forms a star with it as the center and ℓ 2 leaves. Setting the edge lengths to be d 1 ensures that each such star must be covered by a separate cluster. We connect the root of Dj to Ai j, for every i [n] making it possible to add active vertices of Ai j, to the clusters that cover the tree. The gadget Dj requires that at least cj of the connected vertices are active. Because of particular lengths of the construction one may not exclude that the active vertices of Ai j, are also covered by clusters that do not cover the stars in the demand gadget. To guarantee equality, we need to ensure that P i active Ai j, cj for every j [dim] by extending the construction. This would complete the construction if we did not need to preserve a bound on the cluster sizes. To deal with this final obstacle, we replace the gadget by a different one depicted in Figure 3. The next reduction from MULTIDIMENSIONAL SUBSET SUM is also based on the idea of using choice and duplication gadgets to capture the selection of vectors in a solution. However, since the treewidth is bounded by a constant, we cannot use dim distinct demand gadgets to encode the coordinates of each vector; instead, the coordinates are encoded via different edge lengths of their connections. Theorem 12. NETWORK MICROAGGREGATION is NPhard even if tw and u are fixed constants. Closing the Final Gaps To finalize the complexity picture of NMA and CNMA, we provide two W[1]-hardness reductions. Both reductions are from a multicolored variant of some well-known combinatorial problem. In particular, we reduce from MULTICOLORED INDEPENDENT SET and MULTICOLORED CLIQUE, respectively. Theorem 13. NETWORK MICROAGGREGATION is W[1]- hard parameterized by the vertex-cover number vc. We note that the same construction as the one used above also works for the connected variant of the problem. Corollary 14. CONNECTED NETWORK MICROAGGREGATION is W[1]-hard parameterized by the vertex-cover number vc. Our final result establishes the W[1]-hardness of CNMA even when parameterized by treewidth and the upper bound on the cluster size, which finalizes the complexity picture of the studied problems. Theorem 15. CONNECTED NETWORK MICROAGGREGATION is W[1]-hard parameterized by tw + u. Conclusions Our algorithms and lower bounds shed light on the nonobvious complexity-theoretic behavior of the microaggregation problem and provide a comprehensive picture for both NMA and CNMA. It is important to note that as is typical for complexity-theoretic studies of problems central to AI research (Bentert et al. 2021; Gr uttemeier, Komusiewicz, and Morawietz 2021; Dvoˇr ak et al. 2021) the algorithms provided here are not likely to outperform existing heuristics in practically relevant settings. That being said, the techniques introduced here may in the future serve as inspiration for improvements in practice. Our study does not consider the lower bound ℓas a separate parameter since many of the considerations and results obtained when parameterizing by u immediately carry over to ℓas well; in particular, all algorithmic lower bounds parameterized by u can be directly carried over to ℓ. Moreover, since it is known that u can be assumed to be at most 2ℓin NMA, the two parameters are asymptotically equivalent for this problem. Last but not least, it may be interesting to investigate the complexity of the considered problems under less standard parameterizations, such as the cluster vertex deletion number. Acknowledgments The work was performed while VB, JP, DK, and ˇSS were visiting Algorithms and Complexity Group, TU Wien. The visit was financially supported by project 92p1 of the AKTION Austria - Czech Republic Program. VB, DK, JP, and ˇSS was supported by the Czech Science Foundation Grant No. 22-19557S. JP and ˇSS also acknowledges support from the Grant Agency of the Czech Technical University in Prague funded grant No. SGS23/205/OHK3/3T/18. RG acknowledges support from the Austrian Science Foundation (FWF, project Y1329) and Oe AD bilateral Czech-Austrian WTZ-funding Programme, projects No. CZ 05/2021 and 8J21AT02. KS acknowledges support by DFG Research Group ADYN under grant DFG 411362735. References Abu-Khzam, F. N.; Bazgan, C.; Casel, K.; and Fernau, H. 2018. Clustering with Lower-Bounded Sizes. Algorithmica, 80(9): 2517 2550. Aggarwal, G.; Panigrahy, R.; Feder, T.; Thomas, D.; Kenthapadi, K.; Khuller, S.; and Zhu, A. 2010. Achieving Anonymity via Clustering. ACM Transactions on Algorithms, 6(3): 49. Bandyapadhyay, S.; Fomin, F. V.; and Simonov, K. 2021. On Coresets for Fair Clustering in Metric and Euclidean Spaces and Their Applications. In Proceedings of the 48th International Colloquium on Automata, Languages, and Programming, ICALP 21, volume 198 of LIPIcs, 23:1 23:15. Schloss Dagstuhl - Leibniz-Zentrum f ur Informatik. Belhajjame, K.; Faci, N.; Maamar, Z.; Bur egio, V. A.; Soares, E.; and Barhamgi, M. 2020. On privacy-aware e Science workflows. Computing, 102(5): 1171 1185. Bentert, M.; Bredereck, R.; Gy orgyi, P.; Kaczmarczyk, A.; and Niedermeier, R. 2021. A Multivariate Complexity Analysis of the Material Consumption Scheduling Problem. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 21, 11755 11763. AAAI Press. van Bevern, R.; Feldmann, A. E.; Sorge, M.; and Such y, O. 2015. On the Parameterized Complexity of Computing Balanced Partitions in Graphs. Theory of Computing Systems, 57(1): 1 35. Bodlaender, H. L. 1996. A Linear-Time Algorithm for Finding Tree-Decompositions of Small Treewidth. SIAM Journal on Computing, 25(6): 1305 1317. Bonnet, E.; and Purohit, N. 2021. Metric Dimension Parameterized By Treewidth. Algorithmica, 83(8): 2606 2633. Casel, K. 2019. Resolving Conflicts for Lower-Bounded Clustering. In Proceedings of the 13th International Symposium on Parameterized and Exact Computation, IPEC 18, volume 115 of LIPIcs, 23:1 23:14. Dagstuhl, Germany: Schloss Dagstuhl Leibniz-Zentrum fuer Informatik. Cohen-Addad, V.; and Li, J. 2019. On the Fixed-Parameter Tractability of Capacitated Clustering. In Proceedings of the 46th International Colloquium on Automata, Languages, and Programming, ICALP 19, volume 132 of LIPIcs, 41:1 41:14. Schloss Dagstuhl - Leibniz-Zentrum f ur Informatik. Courcelle, B. 1990. The Monadic Second-Order Logic of Graphs. I. Recognizable Sets of Finite Graphs. Information and Computation, 85(1): 12 75. Cygan, M.; Fomin, F. V.; Kowalik, Ł.; Lokshtanov, D.; Marx, D.; Pilipczuk, M.; Pilipczuk, M.; and Saurabh, S. 2015. Parameterized Algorithms. Springer. ISBN 978-3319-21274-6. Das, H. P.; Tran, R.; Singh, J.; Yue, X.; Tison, G. H.; Sangiovanni-Vincentelli, A. L.; and Spanos, C. J. 2022. Conditional Synthetic Data Generation for Robust Machine Learning Applications with Limited Pandemic Data. In Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 22, 11792 11800. AAAI Press. Deligkas, A.; Eiben, E.; Ganian, R.; Hamm, T.; and Ordyniak, S. 2022. The Complexity of Envy-Free Graph Cutting. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 22, 237 243. ijcai.org. Diestel, R. 2012. Graph Theory, volume 173 of Graduate texts in mathematics. Springer, 4th edition. ISBN 978-3642-14278-9. Domingo-Ferrer, J. 2009. Microaggregation. In Encyclopedia of Database Systems, 1736 1737. Boston, MA: Springer. ISBN 978-0-387-39940-9. Domingo-Ferrer, J.; and Mateo-Sanz, J. M. 2002. Practical Data-Oriented Microaggregation for Statistical Disclosure Control. IEEE Transactions on Knowledge and Data Engineering, 14(1): 189 201. Domingo-Ferrer, J.; and Seb e, F. 2006. Optimal Multivariate 2-Microaggregation for Microdata Protection: A 2Approximation. In Proceedings of the International Conference on the Privacy in Statistical Databases, PSD 06, volume 4302 of LNCS, 129 138. Berlin, Heidelberg: Springer. Domingo-Ferrer, J.; Seb e, F.; and Solanas, A. 2008. A polynomial-time approximation to optimal multivariate microaggregation. Computers & Mathematics with Applications, 55(4): 714 732. Downey, R. G.; and Fellows, M. R. 2013. Fundamentals of Parameterized Complexity. Texts in Computer Science. Springer. ISBN 978-1-4471-5558-4. Dvoˇr ak, P.; Eiben, E.; Ganian, R.; Knop, D.; and Ordyniak, S. 2021. The complexity landscape of decompositional parameters for ILP: Programs with few global variables and constraints. Artificial Intelligence, 300: 103561. Enciso, R.; Fellows, M. R.; Guo, J.; Kanj, I.; Rosamond, F.; and Such y, O. 2009. What Makes Equitable Connected Partition Easy. In Proceedings of the 4th International Workshop on Parameterized and Exact Computation, IWPEC 09, volume 5917 of LNCS, 122 133. Berlin, Heidelberg: Springer. Feldmann, A. E.; and Marx, D. 2020. The Parameterized Hardness of the k-Center Problem in Transportation Networks. Algorithmica, 82(7): 1989 2005. Frank, A.; and Tardos, E. 1987. An application of simultaneous Diophantine approximation in combinatorial optimization. Combinatorica, 7(1): 49 65. Ganian, R.; Hamm, T.; Korchemna, V.; Okrasa, K.; and Simonov, K. 2022. The Complexity of k-Means Clustering when Little is Known. In Proceedings of the 39th International Conference on Machine Learning, ICML 22, volume 162 of Proceedings of Machine Learning Research, 6960 6987. PMLR. Ganian, R.; Kanj, I.; Ordyniak, S.; and Szeider, S. 2020. On the Parameterized Complexity of Clustering Incomplete Data into Subspaces of Small Rank. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 20, 3906 3913. AAAI Press. Ganian, R.; Klute, F.; and Ordyniak, S. 2021. On Structural Parameterizations of the Bounded-Degree Vertex Deletion Problem. Algorithmica, 83(1): 297 336. Gr uttemeier, N.; Komusiewicz, C.; and Morawietz, N. 2021. On the Parameterized Complexity of Polytree Learning. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 21, 4221 4227. ijcai.org. Hell, P.; and Kirkpatrick, D. G. 1982. Star factors and star packings. Technical Report 82-6, Department of Computing Science, Simon Fraser University, Burnaby, BC V5A1S6, Canada. Iftikhar, M.; Wang, Q.; and Lin, Y. 2020. d KMicroaggregation: Anonymizing Graphs with Differential Privacy Guarantees. In Proceedings of the 24th Pacific Asia Conference on Advances in Knowledge Discovery and Data Mining, PAKDD 20, volume 12085 of LNCS, 191 203. Springer. Kannan, R. 1987. Minkowski s Convex Body Theorem and Integer Programming. Mathematics of Operations Research, 12(3): 415 440. Kim, Y.; Venkatesha, Y.; and Panda, P. 2022. Private SNN: Privacy-Preserving Spiking Neural Networks. In Proceedings of the Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 22, 1192 1200. AAAI Press. Kirkpatrick, D. G.; and Hell, P. 1978. On the Completeness of a Generalized Matching Problem. In Proceedings of the 10th Annual ACM Symposium on Theory of Computing, STOC 78, 240 245. New York, NY, USA: ACM. Korhonen, T. 2021. A Single-Exponential Time 2Approximation Algorithm for Treewidth. In Proceedings of the 62nd IEEE Annual Symposium on Foundations of Computer Science, FOCS 21, 184 192. IEEE. Kronegger, M.; Lackner, M.; Pfandler, A.; and Pichler, R. 2014. A Parameterized Complexity Analysis of Generalized CP-Nets. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, AAAI 14, 1091 1097. AAAI Press. Lenstra Jr., H. W. 1983. Integer programming with a fixed number of variables. Mathematics of Operations Research, 8(4): 538 548. Macgregor, P.; and Sun, H. 2021. Local Algorithms for Finding Densely Connected Clusters. In Proceedings of the 38th International Conference on Machine Learning, ICML 21, volume 139 of Proceedings of Machine Learning Research, 7268 7278. PMLR. Maniu, S.; Senellart, P.; and Jog, S. 2019. An Experimental Study of the Treewidth of Real-World Graph Data. In Proceedings of the 22nd International Conference on Database Theory, ICDT 19, volume 127 of LIPIcs, 12:1 12:18. Schloss Dagstuhl - Leibniz-Zentrum f ur Informatik. Matouˇsek, J.; and Neˇsetˇril, J. 2009. Invitation to Discrete Mathematics. Oxford University Press, 2nd edition. ISBN 978-0-19-857042-4. Micha, E.; and Shah, N. 2020. Proportionally Fair Clustering Revisited. In Proceedings of the 47th International Colloquium on Automata, Languages, and Programming, ICALP 20, volume 168 of LIPIcs, 85:1 85:16. Schloss Dagstuhl - Leibniz-Zentrum f ur Informatik. Mnich, M.; and Wiese, A. 2015. Scheduling and fixedparameter tractability. Mathematical Programming, 154(12): 533 562. Mukherjee, S. S.; Sarkar, P.; and Lin, L. 2017. On clustering network-valued data. In Proceedings of the 31st Annual Conference on Neural Information Processing Systems, NIPS 17, 7071 7081. Orecchia, L.; and Zhu, Z. A. 2014. Flow-Based Algorithms for Local Graph Clustering. In Proceedings of the Twenty-Fifth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 14, 1267 1286. SIAM. Rao, M. 2007. MSOL partitioning problems on graphs of bounded treewidth and clique-width. Theoretical Computer Science, 377(1-3): 260 267. Rattigan, M. J.; Maier, M. E.; and Jensen, D. D. 2007. Graph clustering with network structure indices. In Proceedings of the Twenty-Fourth International Conference on Machine Learning, ICML 07, volume 227 of ACM International Conference Proceeding Series, 783 790. ACM. Robertson, N.; and Seymour, P. D. 1984. Graph minors. III. Planar tree-width. Journal of Combinatorial Theory, Series B, 36(1): 49 64. Sol e, M.; Munt es-Mulero, V.; and Nin, J. 2012. Efficient microaggregation techniques for large numerical data volumes. International Journal of Information Security, 11(4): 253 267. Sun, X.; Wang, H.; Li, J.; and Zhang, Y. 2012. An Approximate Microaggregation Approach for Microdata Protection. Expert Systems with Applications, 39(2): 2211 2219. Sweeney, L. 2002. K-Anonymity: A Model for Protecting Privacy. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, 10(5): 557 570. Thaeter, F.; and Reischuk, R. 2021. Hardness of kanonymous microaggregation. Discrete Applied Mathematics, 303: 149 158. Yan, Y.; Eyeleko, A. H.; Mahmood, A.; Li, J.; Dong, Z.; and Xu, F. 2022. Privacy preserving dynamic data release against synonymous linkage based on microaggregation. Scientific Reports, 12(1): 2352.