# scalable_laplacian_kmodes__d26dba63.pdf Scalable Laplacian K-modes Imtiaz Masud Ziko ÉTS Montreal Eric Granger ÉTS Montreal Ismail Ben Ayed ÉTS Montreal We advocate Laplacian K-modes for joint clustering and density mode finding, and propose a concave-convex relaxation of the problem, which yields a parallel algorithm that scales up to large datasets and high dimensions. We optimize a tight bound (auxiliary function) of our relaxation, which, at each iteration, amounts to computing an independent update for each cluster-assignment variable, with guaranteed convergence. Therefore, our bound optimizer can be trivially distributed for large-scale data sets. Furthermore, we show that the density modes can be obtained as byproducts of the assignment variables via simple maximum-value operations whose additional computational cost is linear in the number of data points. Our formulation does not need storing a full affinity matrix and computing its eigenvalue decomposition, neither does it perform expensive projection steps and Lagrangian-dual inner iterates for the simplex constraints of each point. Furthermore, unlike mean-shift, our density-mode estimation does not require inner-loop gradient-ascent iterates. It has a complexity independent of featurespace dimension, yields modes that are valid data points in the input set and is applicable to discrete domains as well as arbitrary kernels. We report comprehensive experiments over various data sets, which show that our algorithm yields very competitive performances in term of optimization quality (i.e., the value of the discrete-variable objective at convergence) and clustering accuracy. 1 Introduction We advocate Laplacian K-modes for joint clustering and density mode finding, and propose a concave-convex relaxation of the problem, which yields a parallel algorithm that scales up to large data sets and high dimensions. Introduced initially in the work of Wang and Carreira-Perpinán [33], the model solves the following constrained optimization problem for L clusters and data points X = {xp RD, p = 1, . . . , N}: l=1 zp,lk(xp, ml) + λ p,q k(xp, xq) zp zq 2 ) ml = arg max y X p zp,lk(xp, y) 1tzp = 1, zp {0, 1}L p (1) where, for each point p, zp = [zp,1, . . . , zp,L]t denotes a binary assignment vector, which is constrained to be within the L-dimensional simplex: zp,l = 1 if p belongs to cluster l and zp,l = 0 otherwise. Z is the N L matrix whose rows are given by the zp s. k(xp, xq) are pairwise affinities, which can be either learned or evaluated in an unsupervised way via a kernel function. Model (1) integrates several powerful and well-known ideas in clustering. First, it identifies density modes [18, 8], as in popular mean-shift. Prototype ml is a cluster mode and, therefore, a valid data Corresponding author email: imtiaz-masud.ziko.1@etsmtl.ca 32nd Conference on Neural Information Processing Systems (Neur IPS 2018), Montréal, Canada. point in the input set. This is important for manifold-structured, high-dimensional inputs such as images, where simple parametric prototypes such as the means, as in K-means, may not be good representatives of the data; see Fig. 1. Second, the pairwise term in E is the well-known graph Laplacian regularizer, which can be equivalently written as λtr(Zt LZ), with L the Laplacian matrix corresponding to affinity matrix K = [k(xp, xq)]. Laplacian regularization encourages nearby data points to have similar latent representations (e.g., assignments) and is widely used in spectral clustering [32, 26] as well as in semi-supervised and/or representation learning [2]. Therefore, the model can handle non-convex (or manifold-structured) clusters, unlike standard prototype-based clustering techniques such as K-means. Finally, the explicit cluster assignments yield straightforward out-of-sample extensions, unlike spectral clustering [3]. Optimization problem (1) is challenging due to the simplex/integer constraints and the non-linear/nondifferentiable dependence of modes ml on assignment variables. In fact, it is well known that optimizing the pairwise Laplacian term over discrete variables is NP-hard [30], and it is common to relax the integer constraint. For instance, [33] replaces the integer constraint with a probabilitysimplex constraint, which results in a convex relaxation of the Laplacian term. Unfortunately, such a direct convex relaxation requires solving for N L variables all together. Furthermore, it requires additional projections onto the L-dimensional simplex, with a quadratic complexity with respect to L. Therefore, as we will see in our experiments, the relaxation in [33] does not scale up for large-scale problems (i.e., when N and L are large). Spectral relaxation [32, 27] widely dominates optimization of the Laplacian term subject to balancing constraints in the context of graph clustering2. It can be expressed in the form of a generalized Rayleigh quotient, which yields an exact closed-form solution in terms of the L largest eigenvectors of the affinity matrix. It is well-known that spectral relaxation has high computational and memory load for large N as one has to store the N N affinity matrix and compute explicitly its eigenvalue decomposition, which has a complexity that is cubic with respect to N for a straightforward implementation and, to our knowledge, super-quadratic for fast implementations [30]. In fact, investigating the scalability of spectral relaxation for large-scale problems is an active research subject [26, 30, 31]. For instance, the studies in [26, 30] investigated deep learning approaches to spectral clustering, so as to ease the scalability issues for large data sets, and the authors of [31] examined the variational Nyström method for large-scale spectral problems, among many other efforts on the subject. In general, computational scalability is attracting significant research interest with the overwhelming widespread of interesting large-scale problems [11]. Such issues are being actively investigated even for the basic K-means algorithm [11, 22]. The K-modes term in (1) is closely related to kernel density based algorithms for mode estimation and clustering, for instance, the very popular mean-shift [8]. The value of ml globally optimizing this term for a given fixed cluster l is, clearly, the mode of the kernel density of feature points within the cluster [29]. Therefore, the K-mode term, as in [6, 25], can be viewed as an energy-based formulation of mean-shift algorithms with a fixed number of clusters [29]. Optimizing the K-modes over discrete variable is NP-hard [33], as is the case of other prototype-based models for clustering3. One way to tackle the problem is to alternate optimization over assignment variables and updates of the modes, with the latter performed as inner-loop mean-shift iterates, as in [6, 25]. Mean-shift moves an initial random feature point towards the closest mode via gradient ascent iterates, maximizing at convergence the density of feature points. While such a gradient-ascent approach has been very popular for lowdimensional distributions over continuous domains, e.g., image segmentation [8], its use is generally avoided in the context of high-dimensional feature spaces [7]. Mean-shift iterates compute expensive summations over feature points, with a complexity that depends on the dimension of the feature space. Furthermore, the method is not applicable to discrete domains [7] (as it requires gradient-ascent steps), and its convergence is guaranteed only when the kernels satisfy certain conditions; see [8]. Finally, the modes obtained at gradient-ascent convergence are not necessarily valid data points in the input set. We optimize a tight bound (auxiliary function) of our concave-convex relaxation for discrete problem (1). The bound is the sum of independent functions, each corresponding to a data point p. This yields a scalable algorithm for large N, which computes independent updates for assignment variables zp, while guaranteeing convergence to a minimum of the relaxation. Therefore, our bound optimizer can be trivially distributed for large-scale data sets. Furthermore, we show that the density modes can 2Note that spectral relaxation is not directly applicable to the objective in (1) because of the presence of the K-mode term. 3In fact, even the basic K-means problem is NP-hard. be obtained as byproducts of assignment variables zp via simple maximum-value operations whose additional computational cost is linear in N. Our formulation does not need storing a full affinity matrix and computing its eigenvalue decomposition, neither does it perform expensive projection steps and Lagrangian-dual inner iterates for the simplex constraints of each point. Furthermore, unlike mean-shift, our density-mode estimation does not require inner-loop gradient-ascent iterates. It has a complexity independent of feature-space dimension, yields modes that are valid data points in the input set and is applicable to discrete domains and arbitrary kernels. We report comprehensive experiments over various data sets, which show that our algorithm yields very competitive performances in term of optimization quality (i.e., the value of the discrete-variable objective at convergence)4 and clustering accuracy, while being scalable to large-scale and high-dimensional problems. 2 Concave-convex relaxation We propose the following concave-convex relaxation of the objective in (1): p=1 zt p log(zp) l=1 zp,lk(xp, ml) λ X p,q k(xp, xq)zt pzq where L denotes the L-dimensional probability simplex L = {y [0, 1]L | 1ty = 1}. It is easy to check that, at the vertices of the simplex, our relaxation in (2) is equivalent to the initial discrete objective in (1). Notice that, for binary assignment variables zp {0, 1}L, the first term in (2) vanishes and the last term is equivalent to Laplacian regularization, up to an additive constant: tr(Zt LZ) = X p zt pzpdp X p,q k(xp, xq)zt pzq = X p,q k(xp, xq)zt pzq, (3) where the last equality is valid only for binary (integer) variables and dp = P q k(xp, xq). When we replace the integer constraints zp {0, 1} by zp [0, 1], our relaxation becomes different from direct convex relaxations of the Laplacian [33], which optimizes tr(Zt LZ) subject to probabilistic simplex constraints. In fact, unlike tr(Zt LZ), which is a convex function5, our relaxation of the Laplacian term is concave for positive semi-definite (psd) kernels k. As we will see later, concavity yields a scalable (parallel) algorithm for large N, which computes independent updates for assignment variables zp. Our updates can be trivially distributed, and do not require storing a full N N affinity matrix. These are important computational and memory advantages over direct convex relaxations of the Laplacian [33], which require solving for N L variables all together as well as expensive simplex projections, and over common spectral relaxations [32], which require storing a full affinity matrix and computing its eigenvalue decomposition. Furthermore, the first term we introduced in (2) is a convex negative-entropy barrier function, which completely avoids expensive projection steps and Lagrangian-dual inner iterations for the simplex constraints of each point. First, the entropy barrier restricts the domain of each zp to non-negative values, which avoids extra dual variables for constraints zp 0. Second, the presence of such a barrier function yields closed-form updates for the dual variables of constraints 1tzp = 1. In fact, entropy-like barriers are commonly used in Bregman-proximal optimization [35], and have well-known computational and memory advantages when dealing with the challenging simplex constraints [35]. Surprisingly, to our knowledge, they are not common in the context of clustering. In machine learning, such entropic barriers appear frequently in the context of conditional random fields (CRFs) [14, 15], but are not motivated from optimization perspective; they result from standard probabilistic and mean-field approximations of CRFs [14]. 3 Bound optimization In this section, we derive an iterative bound optimization algorithm that computes independent (parallel) updates of assignment variables zp (z-updates) at each iteration, and provably converges to a minimum of relaxation (2). As we will see in our experiments, our bound optimizer yields 4We obtained consistently lower values of function E at convergence than the convex-relaxation proximal algorithm in [33]. 5For relaxed variables, tr(Zt LZ) is a convex function because the Laplacian is always positive semi-definite. consistently lower values of function E at convergence than the proximal algorithm in [33], while being highly scalable to large-scale and high-dimensional problems. We also show that the density modes can be obtained as byproducts of the z-updates via simple maximum-value operations whose additional computational cost is linear in N. Instead of minimizing directly our relaxation R, we iterate the minimization of an auxiliary function, i.e., an upper bound of R, which is tight at the current solution and easier to optimize. Definition 1 Ai(Z) is an auxiliary function of R(Z) at current solution Zi if it satisfies: R(Z) Ai(Z), Z (4a) R(Zi) = Ai(Zi) (4b) In (4), i denotes the iteration counter. In general, bound optimizers update the current solution Zi to the optimum of the auxiliary function: Zi+1 = arg min Z Ai(Z). This guarantees that the original objective function does not increase at each iteration: R(Zi+1) Ai(Zi+1) Ai(Zi) = R(Zi). Bound optimizers can be very effective as they transform difficult problems into easier ones [37]. Examples of well-known bound optimizers include the concave-convex procedure (CCCP) [36], expectation maximization (EM) algorithms and submodular-supermodular procedures (SSP) [21], among others. Furthermore, bound optimizers are not restricted to differentiable functions6, neither do they depend on optimization parameters such as step sizes. Proposition 1 Given current solution Zi = [zi p,l] at iteration i, and the corresponding modes mi l = arg maxy X P p zi p,lk(xp, y), we have the following auxiliary function (up to an additive constant) for the concave-convex relaxation in (2) and psd7 affinity matrix K: p=1 zt p(log(zp) ai p λbi p) (5) where ai p and bi p are the following L-dimensional vectors: ai p = [ai p,1, . . . , ai p,L]t, with ai p,l = k(xp, mi l) (6a) bi p = [bi p,1, . . . , bi p,L]t, with bi p,l = X q k(xp, xq)zi q,l (6b) Proof 1 See supplemental material. Notice that the bound in Eq. (5) is the sum of independent functions, each corresponding to a point p. Therefore, both the bound and simplex constraints zp L are separable over assignment variables zp. We can minimize the auxiliary function by minimizing independently each term in the sum over zp, subject to the simplex constraint, while guaranteeing convergence to a local minimum of (2): min zp L zt p(log(zp) ai p λbi p), p (7) Note that, for each p, negative entropy zt p log zp restricts zp to be non-negative, which removes the need for handling explicitly constraints zp 0. This term is convex and, therefore, the problem in (7) is convex: The objective is convex (sum of linear and convex functions) and constraint zp L is affine. Therefore, one can minimize this constrained convex problem for each p by solving the Karush-Kuhn-Tucker (KKT) conditions8. The KKT conditions yield a closed-form solution for both primal variables zp and the dual variables (Lagrange multipliers) corresponding to simplex constraints 6Our objective is not differentiable with respect to the modes as each of these is defined as the maximum of a function of the assignment variables. 7We can consider K to be psd without loss of generality. When K is not psd, we can use a diagonal shift for the affinity matrix, i.e., we replace K by K = K + δIN. Clearly, K is psd for sufficiently large δ. For integer variables, this change does not alter the structure of the minimum of discrete function E. 8Note that strong duality holds since the objectives are convex and the simplex constraints are affine. This means that the solutions of the (KKT) conditions minimize the auxiliary function. 1tzp = 1. Each closed-form update, which globally optimizes (7) and is within the simplex, is given by: zi+1 p = exp(ai p + λbi p) 1t exp(aip + λbip) p (8) Algorithm 1: SLK algorithm Input :X, Initial centers {m0 l }L l=1 Output :Z and {ml}L l=1 1 {ml}L l=1 {m0 l }L l=1 3 i 1 // iteration index 4 {mi l}L l=1 {ml}L l=1 // z-updates 5 foreach xp do 6 Compute ai p from (6a) 7 zi p = exp{ai p} 1t exp{aip} // Initialize 10 Compute zi+1 p using (6b) and (8) 12 until convergence 13 Z = [zi+1 p,l ] // Mode-updates 14 if SLK-MS then 15 update ml using (9) until converges 16 if SLK-BO then 17 ml arg max xp [zi+1 p,l ] 18 until convergence 19 return Z, {ml}L l=1 The pseudo-code for our Scalable Laplacian K-modes (SLK) method is provided in Algorithm 1. The complexity of each inner iteration in z-updates is O(NρL), with ρ the neighborhood size for the affinity matrix. Typically, we use sparse matrices (ρ << N). Note that the complexity becomes O(N 2L) in the case of dense matrices in which all the affinities are non-zero. However, the update of each zp can be done independently, which enables parallel implementations. Our SLK algorithm alternates the following two steps until convergence (i.e. until the modes {ml}L l=1 do not change): (i) z-updates: update cluster assignments using expression (8) with the modes fixed and (ii) Mode-updates: update the modes {ml}L l=1 with the assignment variable Z fixed; see the next section for further details on mode estimation. 3.1 Mode updates To update the modes, we have two options: modes via mean-shift or as byproducts of the z-updates. Modes via mean-shift: This amounts to updating each mode ml by running inner-loop mean-shift iterations until convergence, using the current assignment variables: p zp,lk(xp, mi l)xp P p zp,lk(xp, mi l) (9) Modes as byproducts of the z-updates: We also propose an efficient alternative to mean-shift. Observe the following: For each point p, bi p,l = P q k(xp, xq)zi q,l is proportional to the kernel density estimate Figure 1: Examples of mode images obtained with our SLK-BO, mean images and the corresponding 3-nearest-neighbor images within each cluster. We used Label Me (Alexnet) dataset. (KDE) of the distribution of features within current cluster l at point p. In fact, the KDE at a feature point y is: P q k(y, xq)zi q,l P Therefore, bi p,l Pi l (xp). As a result, for a given point p within the cluster, the higher bi p,l, the higher the KDE of the cluster at that point. Notice also that ai p,l = k(xp, mi l) measures a proximity between point xp and the mode obtained at the previous iteration. Therefore, given the current assignment zi p, the modes can be obtained as a proximal optimization, which seeks a high-density data point that does not deviate significantly from the mode obtained at the previous iteration: max y X [k(y, mi l) | {z } proximity p zp,lk(xp, y) | {z } density Now observe that the z-updates we obtained in Eq. (8) take the form of softmax functions. Therefore, they can be used as soft approximations of the hard max operation in Eq. (10): mi+1 l = xp, with p = arg max q [zq,l]i (11) This yields modes as byproducts of the z-updates, with a computational cost that is linear in N. We refer to the two different versions of our algorithm as SLK-MS, which updates the modes via mean-shift, and SLK-BO, which updates the modes as byproducts of the z-updates. (a) MNIST (small) (b) Label Me (Alexnet) Figure 2: Discrete-variable objective (1): Comparison of the objectives obtained at convergence for SLK-MS (ours) and LK [33]. The objectives at convergence are plotted versus different values of parameter λ. 4 Experiments We report comprehensive evaluations of the proposed algorithm9 as well as comparisons to the following related baseline methods: Laplacian K-modes (LK) [33], K-means, NCUT [27], K-modes [25, 5], Kernel K-means (KK-means) [9, 29] and Spectralnet [26]. Our algorithm is evaluated in terms of performance and optimization quality in various clustering datasets. Table 1: Datasets used in the experiments. Datasets Samples (N) Dimensions (D) Clusters (L) Imbalance MNIST (small) 2, 000 784 10 1 MNIST (code) 70, 000 10 10 1 MNIST 70, 000 784 10 1 MNIST (GAN) 70, 000 256 10 1 Shuttle 58, 000 9 7 4, 558 Label Me (Alexnet) 2, 688 4, 096 8 1 Label Me (GIST) 2, 688 44, 604 8 1 YTF 10, 036 9, 075 40 13 Reuters (code) 685, 071 10 4 5 4.1 Datasets and evaluation metrics We used image datasets, except Shuttle and Reuters. The overall summary of the datasets is given in Table 1. For each dataset, imbalance is defined as the ratio of the size of the biggest cluster to the size of the smallest one. We use three versions of MNIST [17]. MNIST contains all the 70, 000 images, whereas MNIST (small) includes only 2, 000 images by randomly sampling 200 images per class. We used small datasets in order to compare to LK [33], which does not scale up for large datasets. For MNIST (GAN), we train the GAN from [12] on 60, 000 training images and extract the 256-dimensional features from the discriminator network for the 70, 000 images. The publicly available autoencoder in [13] is used to extract 10-dimensional features as in [26] for MNIST (code) and Reuters (code). Label Me [23] consists of 2, 688 images divided into 8 categories. We used the pre-trained Alex Net [16] and extracted the 4096-dimensional features from the fully-connected layer. To show the performances on high-dimensional data, we extract 44604-dimensional GIST features [23] for the Label Me dataset. Youtube Faces (YTF) [34] consists of videos of faces with 40 different subjects. 9Code is available at: https://github.com/imtiazziko/SLK Table 2: Clustering results as NMI/ACC in the upper half and average elapsed time in seconds (s). (*) We report the results of Spectralnet with Euclidean-distance affinity for MNIST (code) and Reuters (code) from [26]. Algorithm MNIST MNIST (code) MNIST (GAN) Label Me (Alexnet) Label Me (GIST) YTF Shuttle Reuters K-means 0.53/0.55 0.66/0.74 0.68/0.75 0.81/0.90 0.57/0.69 0.77/0.58 0.22/0.41 0.48/0.73 K-modes 0.56/0.60 0.67/0.75 0.69/0.80 0.81/0.91 0.58/0.68 0.79/0.62 0.33/0.47 0.48/0.72 NCUT 0.74/0.61 0.84/0.81 0.77/0.67 0.81/0.91 0.58/0.61 0.74/0.54 0.47/0.46 - KK-means 0.53/0.55 0.67/0.80 0.69/0.68 0.81/0.90 0.57/0.63 0.71/0.50 0.26/0.40 - LK - - - 0.81/0.91 0.59/0.61 0.77/0.59 - - Spectralnet* - 0.81/0.80 - - - - - 0.46/0.65 SLK-MS 0.80/0.79 0.88/0.95 0.86/0.94 0.83/0.91 0.61/0.72 0.82/0.65 0.45/0.70 0.43/0.74 SLK-BO 0.77/0.80 0.89/0.95 0.86/0.94 0.83/0.91 0.61/0.72 0.80/0.64 0.51/0.71 0.43/0.74 K-means 119.9s 16.8s 51.6s 11.2s 132.1s 210.1s 1.8s 36.1s K-modes 90.2s 20.2s 20.3s 7.4s 12.4s 61.0s 0.5s 51.6s NCUT 26.4s 28.2s 9.3s 7.4s 10.4s 19.0s 27.4s - KK-means 2580.8s 1967.9s 2427.9s 4.6s 17.2s 40.2s 1177.6s - LK - - - 33.4s 180.9s 409.0s - - Spectralnet* - 3600.0s - - - - - 9000.0s SLK-MS 101.2s 82.4s 37.3s 4.7s 37.0s 83.3s 3.8s 12.5s SLK-BO 14.2s 23.1s 10.3s 1.8s 7.1s 12.4s 1.3s 53.1s To evaluate the clustering performance, we use two well adopted measures: Normalized Mutual Information (NMI) [28] and Clustering Accuracy (ACC) [26, 10]. The optimal mapping of clustering assignments to the true labels are determined using the Kuhn-Munkres algorithm [20]. 4.2 Implementation details We built k NN affinities as follows: k(xp, xq) = 1 if xq N ρ p and k(xp, xq) = 0 otherwise, where N ρ p is the set of the ρ nearest neighbors of data point xp. This yields a sparse affinity matrix, which is efficient in terms of memory and computations. In all of the datasets, we fixed ρ = 5. For the large datasets such as MNIST, Shuttle and Reuters, we used the Flann library [19] with the KD-tree algorithm, which finds approximate nearest neighbors. For the other smaller datasets, we used an efficient implementation of exact nearest-neighbor computations. We used the Euclidean distance for finding the nearest neighbors. We used the same sparse K for the pairwise-affinity algorithms we compared with, i.e., NCUT, KK-means, Laplacian K-modes. Furthermore, for each of these baseline methods, we evaluated the default setting of affinity construction with tuned σ, and report the best result found. Mode estimation is based on the Gaussian kernel k(x, y) = e ( x y) 2/2σ2), with σ2 estimated as: σ2 = 1 Nρ P xq N ρ p xp xq 2. Initial centers {m0 l }L l=1 are based on K-means++ seeds [1]. We choose the best initial seed and regularization parameter λ empirically based on the accuracy over a validation set (10% of the total data). The λ is determined from tuning in a small range from 1 to 4. In SLK-BO, we take the starting mode ml for each cluster from the initial assignments by simply following the mode definition in (1). In Algorithm 1, all assignment variables zp are updated in parallel. We run the publicly released codes for K-means [24], NCUT [27], Laplacian K-modes [4], Kernel K-means10 and Spectralnet [26]. 4.3 Clustering results Table 2 reports the clustering results, showing that, in most of the cases, our algorithms SLK-MS and SLK-BO yielded the best NMI and ACC values. For MNIST with the raw intensities as features, the proposed SLK achieved almost 80% NMI and ACC. With better learned features for MNIST (code) and MNIST (GAN), the accuracy (ACC) increases up to 95%. For the MNIST (code) and Reuters (code) datasets, we used the same features and Euclidean distance based affinity as Spectralnet [26], and obtained better NMI/ACC performances. The Shuttle dataset is quite imbalanced and, therefore, 10https://gist.github.com/mblondel/6230787 Table 3: Discrete-variable objectives at convergence for LK [33] and SLK-MS (ours). Datasets LK [33] SLK-MS (ours) MNIST (small) 273.25 67.09 Label Me (Alexnet) 1.513 84 103 1.807 77 103 Label Me (GIST) 1.954 90 103 2.024 10 103 YTF 1.000 32 104 1.000 35 104 all the baseline clustering methods fail to achieve high accuracy. Notice that, in regard to ACC for the Shuttle dataset, we outperformed all the methods by a large margin. One advantage of our SLK-BO over standard prototype-based models is that the modes are valid data points in the input set. This is important for manifold-structured, high-dimensional inputs such as images, where simple parametric prototypes such as the means, as in K-means, may not be good representatives of the data; see Fig. 1. 4.4 Comparison in terms of optimization quality To assess the optimization quality of our optimizer, we computed the values of discrete-variable objective E in model (1) at convergence for our concave-convex relaxation (SLK-MS) as well as for the convex relaxation in [33] (LK). We compare the discrete-variable objectives for different values of λ. For a fair comparison, we use the same initialization, σ, k(xp, xq), λ and mean-shift modes for both methods. As shown in the plots in Figure 2, our relaxation consistently obtained lower values of discrete-variable objective E at convergence than the convex relaxation in [33]. Also, Table 3 reports the discrete-variable objectives at convergence for LK [33] and SLK-MS (ours). These experiments suggest that our relaxation in Eq. (2) is tighter than the convex relaxation in [33]. In fact, Eq. (3) also suggests that our relaxation of the Laplacian term is tighter than a direct convex relaxation (the expression in the middle in Eq. (3)) as the variables in term P p dpzt pzp are not relaxed in our case. 4.5 Running Time The running times are given at the bottom half of Table 2. All the experiments (our methods and the baselines) were conducted on a machine with Xeon E5-2620 CPU and a Titan X Pascal GPU. We restrict the multiprocessing to at most 5 processes. We run each algorithm over 10 trials and report the average running time. For high-dimensional datasets, such as Label Me (GIST) and YTF, our method is much faster than the other methods we compared to. It is also interesting to see that, for high dimensions, SLK-BO is faster than SLK-MS which uses mean-shift for mode estimation. 5 Conclusion We presented Scalable Laplacian K-modes (SLK), a method for joint clustering and density mode estimation, which scales up to high-dimensional and large-scale problems. We formulated a concaveconvex relaxation of the discrete-variable objective, and solved the relaxation with an iterative bound optimization. Our solver results in independent updates for cluster-assignment variables, with guaranteed convergence, thereby enabling distributed implementations for large-scale data sets. Furthermore, we showed that the density modes can be estimated directly from the assignment variables using simple maximum-value operations, with an additional computational cost that is linear in the number of data points. Our solution removes the need for storing a full affinity matrix and computing its eigenvalue decomposition. Unlike the convex relaxation in [33], it does not require expensive projection steps and Lagrangian-dual inner iterates for the simplex constraints of each point. Furthermore, unlike mean-shift, our density-mode estimation does not require inner-loop gradient-ascent iterates. It has a complexity independent of feature-space dimension, yields modes that are valid data points in the input set and is applicable to discrete domains as well as arbitrary kernels. We showed competitive performances of the proposed solution in term of optimization quality and accuracy. It will be interesting to investigate joint feature learning and SLK clustering. [1] D. Arthur and S. Vassilvitskii. k-means++: The advantages of careful seeding. In ACM-SIAM symposium on Discrete algorithms, pages 1027 1035. Society for Industrial and Applied Mathematics, 2007. [2] M. Belkin, P. Niyogi, and V. Sindhwani. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of Machine Learning Research, 7:2399 2434, 2006. [3] Y. Bengio, J.-f. Paiement, P. Vincent, O. Delalleau, N. L. Roux, and M. Ouimet. Out-of-sample extensions for lle, isomap, mds, eigenmaps, and spectral clustering. In Neural Information Processing Systems (NIPS), pages 177 184, 2004. [4] M. Á. Carreira-Perpiñán. Gaussian mean-shift is an em algorithm. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(5):767 776, 2007. [5] M. Á. Carreira-Perpiñán. A review of mean-shift algorithms for clustering. ar Xiv preprint ar Xiv:1503.00687, 2015. [6] M. Á. Carreira-Perpiñán and W. Wang. The k-modes algorithm for clustering. ar Xiv preprint ar Xiv:1304.6478, 2013. [7] C. Chen, H. Liu, D. Metaxas, and T. Zhao. Mode estimation for high dimensional discrete tree graphical models. In Neural Information Processing Systems (NIPS), pages 1323 1331, 2014. [8] D. Comaniciu and P. Meer. Mean shift: A robust approach toward feature space analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 24(5):603 619, 2002. [9] I. S. Dhillon, Y. Guan, and B. Kulis. Kernel k-means: spectral clustering and normalized cuts. In International Conference on Knowledge Discovery and Data Mining (SIGKDD), pages 551 556, 2004. [10] K. Ghasedi Dizaji, A. Herandi, C. Deng, W. Cai, and H. Huang. Deep clustering via joint convolutional autoencoder embedding and relative entropy minimization. In International Conference on Computer Vision (ICCV), pages 5747 5756, 2017. [11] Y. Gong, M. Pawlowski, F. Yang, L. Brandy, L. Bourdev, and R. Fergus. Web scale photo hash clustering on a single machine. In Computer Vision and Pattern Recognition (CVPR), pages 19 27, 2015. [12] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In Neural Information Processing Systems (NIPS), pages 2672 2680, 2014. [13] Z. Jiang, Y. Zheng, H. Tan, B. Tang, and H. Zhou. Variational deep embedding: An unsupervised and generative approach to clustering. In International Joint Conference on Artificial Intelligence (IJCAI), pages 1965 1972, 2017. [14] P. Krähenbühl and V. Koltun. Efficient inference in fully connected crfs with gaussian edge potentials. In Neural Information Processing Systems (NIPS), pages 109 117, 2011. [15] P. Krähenbühl and V. Koltun. Parameter learning and convergent inference for dense random fields. In International Conference on Machine Learning (ICML), pages 513 521, 2013. [16] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In Neural Information Processing Systems (NIPS), pages 1097 1105, 2012. [17] Y. Le Cun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [18] J. Li, S. Ray, and B. G. Lindsay. A nonparametric statistical approach to clustering via mode identification. Journal of Machine Learning Research, 8:1687 1723, 2007. [19] M. Muja and D. G. Lowe. Scalable nearest neighbor algorithms for high dimensional data. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(11):2227 2240, 2014. [20] J. Munkres. Algorithms for the assignment and transportation problems. Journal of the society for industrial and applied mathematics, 5(1):32 38, 1957. [21] M. Narasimhan and J. Bilmes. A submodular-supermodular procedure with applications to discriminative structure learning. In Conference on Uncertainty in Artificial Intelligence (UAI), pages 404 412, 2005. [22] J. Newling and F. Fleuret. Nested mini-batch k-means. In Neural Information Processing Systems (NIPS), pages 1352 1360, 2016. [23] A. Oliva and A. Torralba. Modeling the shape of the scene: A holistic representation of the spatial envelope. International Journal of Computer Vision, 42(3):145 175, 2001. [24] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825 2830, 2011. [25] M. B. Salah, I. B. Ayed, J. Yuan, and H. Zhang. Convex-relaxed kernel mapping for image segmentation. IEEE Transactions on Image Processing, 23(3):1143 1153, 2014. [26] U. Shaham, K. Stanton, H. Li, R. Basri, B. Nadler, and Y. Kluger. Spectralnet: Spectral clustering using deep neural networks. In International Conference on Learning Representations (ICLR), 2018. [27] J. Shi and J. Malik. Normalized cuts and image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(8):888 905, 2000. [28] A. Strehl and J. Ghosh. Cluster ensembles a knowledge reuse framework for combining multiple partitions. Journal of Machine Learning Research, 3(12):583 617, 2002. [29] M. Tang, D. Marin, I. B. Ayed, and Y. Boykov. Kernel cuts: Kernel and spectral clustering meet regularization. International Journal of Computer Vision, In press:1 35, 2018. [30] F. Tian, B. Gao, Q. Cui, E. Chen, and T.-Y. Liu. Learning deep representations for graph clustering. In AAAI Conference on Artificial Intelligence, pages 1293 1299, 2014. [31] M. Vladymyrov and M. Carreira-Perpiñán. The variational nystrom method for large-scale spectral problems. In International Conference on Machine Learning (ICML), pages 211 220, 2016. [32] U. Von Luxburg. A tutorial on spectral clustering. Statistics and computing, 17(4):395 416, 2007. [33] W. Wang and M. A. Carreira-Perpinán. The laplacian k-modes algorithm for clustering. ar Xiv preprint ar Xiv:1406.3895, 2014. [34] L. Wolf, T. Hassner, and I. Maoz. Face recognition in unconstrained videos with matched background similarity. In Computer Vision and Pattern Recognition (CVPR), pages 529 534, 2011. [35] J. Yuan, K. Yin, Y. Bai, X. Feng, and X. Tai. Bregman-proximal augmented lagrangian approach to multiphase image segmentation. In Scale Space and Variational Methods in Computer Vision (SSVM), pages 524 534, 2017. [36] A. L. Yuille and A. Rangarajan. The concave-convex procedure (CCCP). In Neural Information Processing Systems (NIPS), pages 1033 1040, 2001. [37] Z. Zhang, J. T. Kwok, and D.-Y. Yeung. Surrogate maximization/minimization algorithms and extensions. Machine Learning, 69:1 33, 2007.