# adaptively_robust_and_sparse_kmeans_clustering__3b2f1722.pdf Published in Transactions on Machine Learning Research (11/2024) Adaptively Robust and Sparse K-means Clustering Hao Li lee1995hao@keio.jp Graduate School of Economics Keio University Shonosuke Sugasawa sugasawa@econ.keio.ac.jp Faculty of Economics Keio University Shota Katayama katayama@econ.keio.ac.jp Faculty of Economics Keio University Reviewed on Open Review: https: // openreview. net/ forum? id= Eh C84f T2y A While K-means is known to be a standard clustering algorithm, its performance may be compromised due to the presence of outliers and high-dimensional noisy variables. This paper proposes adaptively robust and sparse K-means clustering (ARSK) to address these practical limitations of the standard K-means algorithm. For robustness, we introduce a redundant error component for each observation, and this additional parameter is penalized using a group sparse penalty. To accommodate the impact of high-dimensional noisy variables, the objective function is modified by incorporating weights and implementing a penalty to control the sparsity of the weight vector. The tuning parameters to control the robustness and sparsity are selected by Gap statistics. Through simulation experiments and real data analysis, we demonstrate the proposed method s superiority to existing algorithms in identifying clusters without outliers and informative variables simultaneously. 1 Introduction Identifying clustering structures is recognized as an important task in pursuing potential heterogeneity in datasets. While there are a variety of clustering methods, K-means clustering (Forgy, 1965) is the most standard clustering algorithm and is widely used in many scientific applications. However, real-world datasets present significant difficulties, such as the presence of outliers and high-dimensional noisy variables. Regarding the outlier issues, there are some attempts to robustify the standard K-means such as trimmed K-means (Cuesta-Albertos et al., 1997) and robust K-means (Klochkov et al., 2021). On the other hand, high-dimensional noisy characteristics are typically addressed through clustering approaches that focus on sparsity for variable selection. In particular, Witten & Tibshirani (2010) proposed a lasso-type penalty to the variable weights incorporated in the clustering objective function. However, the efficacy of this algorithm might be lost if the dataset contains a significant number of outliers. While several methods address either of the two aspects (the existence of outliers and noisy variables), practically useful methods to address both aspects simultaneously are scarce. One exception is the method proposed by Kondo et al. (2016). However, this approach requires the assumption of a trimming level in advance. In contrast, Brodinová et al. (2019) introduces an approach beginning with eliminating outliers identified by the model. However, this methodology potentially leads to reduced sample size and may result in information loss. Moreover, the excessive variations in gradient magnitudes may destabilize when optimizing the variable weight process unless all outliers are eliminated in advance. In addition to clustering methods based on objective functions, probabilistic clustering using mixture models is also popular. It is typically done by fitting multivariate Gaussian mixtures. While several approaches Published in Transactions on Machine Learning Research (11/2024) have been proposed for stable estimation and clustering under existence of outliers (e.g. Coretto & Hennig, 2016; Fujisawa & Eguchi, 2006; Punzo & Mc Nicholas, 2016; Sugasawa & Kobayashi, 2022; Yang et al., 2012), such approach suffers from a huge number of parameters (especially for modelling covariance matrix) to be estimated when the dimension is large. Therefore, these methods would not be a reasonable solution when our primary focus is on clustering. To address the difficulty of a series of sparse trimming-cluster and probabilistic-cluster approaches, we propose a new approach based on regularization techniques for sparsity as well as robustness, called adaptively robust and sparse K-means (ARSK). We introduce an error component for each observation as an additional parameter to absorb the undesirable effects caused by outliers, making the clustering algorithm robust. We employ a group lasso penalty and a group smoothly clipped absolute deviation (SCAD) penalty (Wang et al., 2007) to estimate these parameters during clustering steps. Meanwhile, to reduce the data dimension, we also introduce a SCAD penalty (Fan & Li, 2001) to weights for each variable to eliminate noise variables irrelevant to clustering. Based on our analysis of real-world datasets, in the ultra-high dimensional setting (i.e., the dataset dimension exceeding 700), the result procured through the implementation of the lasso penalty proposed by Witten & Tibshirani (2010) frequently poor performance. Therefore, we have developed a methodology based on the SCAD penalty. The strong sparsity property of the SCAD penalty effectively improves the selection of relevant variables in the ultra-high dimensional scenario. We then develop an efficient computation algorithm to minimize the proposed objective function. While the proposed method has two tuning parameters controlling robustness (the number of outliers to be detected) and sparsity (the level of the sparsity of the variable weight), we adopt a modified version of the Gap statistics (Tibshirani et al., 2001) to determine the optimal values of them. Hence, the level of robustness and sparsity is adaptively determined in a fully data-dependent manner. The paper is organized as follows. Section 2 reviews the K-means algorithm and its extended model introduces the proposed method ARSK and presents its details and pseudocode. We also present the selection approach of tuning parameters and its pseudocode. while Section 3 presents the tuning parameter selection result. Both two algorithms are applied to various scenarios of artificial datasets and several real-world datasets. This paper ends in Section 4 with concluding remarks. R code implementing the proposed method is available at the Git Hub repository (https://github.com/lee1995hao/ARSK). 2 Robust and Sparse K-means Clustering 2.1 Conventional K-means algorithm Suppose that we observe a p-dimensional vector, Xi,: = (Xi1, . . . , Xip), for i = 1, . . . , n, and that we are interested in clustering Xi,:. Given the number of clusters, K, the conventional K-means algorithm is to find the optimal clustering by minimizing the following objective function: min c1,...,c K i,i ck Xi,: Xi ,: 2, (1) where Xi,: Xi ,: 2 is the squared value of L2-distance between two observed vectors. Here ck is a set of indices representing a cluster of observations and nk is the size of ck, such that K k=1ck = {1, . . . , n} and PK k=1 nk = n, where ck ck = ϕ for k = k . The objective function (1) can be easily optimized by iteratively updating cluster means and assignment (e.g. Lloyd, 1982). As demonstrated in Witten & Tibshirani (2010), the objective function (1) can be reformulated in terms of the between-cluster sum of squares, given by maxc1,..,ck Pp j=1 Qj(X:,j; Θ), where Qj(X:,j; Θ) = 1 i =1 (Xij Xi j)2 i,i ck (Xij Xi j)2, (2) Published in Transactions on Machine Learning Research (11/2024) with X:,j = (X1j, . . . , Xnj) and Θ being a collection of c1, . . . , c K. Witten & Tibshirani (2010) employed the modified version of the objective function (2), given by Pp j=1 ωj Qj(X:,j; Θ) with an appropriate constrained on ωj, for conducting variable selection and clustering simultaneously. 2.2 Robust and sparse K-means algorithm The objective function (2) is sensitive to outliers since the approach relies on the L2 distance between each observation. As a result, even a small quantity of deviating observations can affect the K-means algorithm. In order to robustify the K-means approach, we modify the function Qj(X:,j; Θ) as follows: QR j (X:,j; Θ, E) = 1 i =1 {(Xij Eij) (Xi j Ei j)}2 i,i ck {(Xij Eij) (Xi j Ei j)}2, where E is an n p error matrix, which is a collection of Eij and Eij is an additional location parameter for each Xij such that Eij will be non-zero values for outlying observations Xij to prevent an undesirable effects from outliers. While the number of elements of E is the same as the number of observations, it can be assumed that E is sparse in the sense that most elements of E are 0. That is, most elements are not outliers. To stably estimate E, we consider penalized estimation for Eij when optimizing the function (3). Such approaches are used in the robust fitting of regression models (e.g. She & Owen, 2011; Katayama & Fujisawa, 2017). Based on the robust objective function (3), we propose an objective function for robust and sparse K-means algorithm as follows: L(Θ, E, w; λ) = j=1 wj QR j (X:,j; Θ, E:,j) i=1 P1( Ei,: 2; λ1) j=1 (P2(wj; λ2) + 1 2w2 j), (4) where P1 and P2 are group penalty functions and penalty functions for sparse (shrinkage) estimation of Ei,: and wj, respectively, and λ = (λ1, λ2) is a set of tuning parameters. To find a unique solution for the weight coefficients wj, we add a quadratic term 1 2 Pp j=1 w2 j to the objective function. Then, the robust and sparse K-means clustering can be defined as bΘ such that (bΘ, b E, b w) = argmaxΘ,E,w HL(Θ, E, w; λ), j=1 w2 j = 1 In what follows, we employ two specific forms of group penalty functions P1( Ei,: 2; λ1): one type is the convex group penalty function, which is the group lasso penalty as λ1 Ei,: 2 (Yuan & Lin, 2006), and the other type is the nonconvex group penalty function, which is the group SCAD penalty (Huang et al., 2012) as P SCAD λ1 ( Ei,: 2) = λ1 Ei,: 2 if Ei,: 2 λ1; ( Ei,: 2 2 2aλ1 Ei,: 2 + λ2 1)/2(a 1) if λ1 < Ei,: 2 aλ1; (a + 1)λ2 1/2 if Ei,: 2 > aλ1. We also consider both convex and nonconvex penalties for the penalty term P2(wj; λ2): the convex standard lasso penalty P2(wj; λ2) = λ2|wj| and, the non-convex SCAD penalty expressed as P2(wj; λ2) = P SCAD λ2 (wj), where P SCAD λ2 (wj) define as: P SCAD λ2 (wj) = λ2 |wj| if |wj| λ2; (|wj|2 2aλ2 |wj| + λ2 2)/2(a 1) if λ2 < |wj| aλ2; (a + 1)λ2 2/2 if |wj| > aλ2. Published in Transactions on Machine Learning Research (11/2024) When applying the L1 or SCAD penalties, we seek the optimal weight vector of w = (w1, . . . , wp), which is normalized to have a unit L2-norm, i.e., q Pp j=1 w2 j = 1. This approach helps to promote a more balanced solution (Zou & Hastie, 2010). 2.3 Optimization algorithm The block-coordinated decent method can optimise the proposed objective function (4) for updating parameters. Regarding the update of the clustering assignment Θ, we note that the maximization of L(Θ, E, w; λ) with respect to Θ given E and w is equivalent to minimizing L(Θ|E, w; λ) = n (Xij Eij) (Xi j Ei j) o2 . The optimal solution to this equation is achieved by assigning each adjusted observation point to the sample mean of the adjusted observation with the smallest squared Euclidean distance (e.g. Lloyd, 1982), given by i ck wj n Xij Eij 1 i ck (Xij Eij) o2 , (5) which can be easily computed by using the standard K-means algorithm. Given clustering assignment Θ and variable weight w, the error matrix E can be updated as Xij E ij µ kj 2 + i=1 P1( Ei,: 2; λ1) where µ kj = n 1 k P i k(Xij E ij) and E ij is the optimized value of Eij from the previous iteration. As previously discussed, the variable E plays a crucial role in robustness. If an observation does not fall into any cluster, vector Ei,: elements will exhibit large values. The penalty P1( Ei,: 2; λ1) controls the robustness of the model. We also note that under λ1 , all Ei,: equal a p-dimensional zero vector. As a result, the objective function (6) will be equivalent to traditional K-means clustering. In the minimization problem (6), as previously mentioned, we employ the group lasso penalty and group SCAD penalty (Tibshirani, 1996; Antoniadis & Fan, 2001). The optimizers are obtained by applying the thresholding function corresponding to the group penalty to Xi,: µk,:. These two types of group penalty functions correspond to two distinct categories of thresholding functions: the group lasso penalty function corresponds to the multivariate soft-threshold operator, where Ei,: can be calculated through Ei,: = (Xi,: µk,:) max 0, 1 λ1 Xi,: µk,: 2 as proposed by Witten (2013), while the group SCAD penalty function corresponds to the multivariate SCAD-threshold operator introduced by Huang et al. (2012), which Ei,: can be calculated as: S(zi; λ1), if zi 2 2λ1, (a 2) 1(a 1)S(zi; (a 1) 1aλ1), if 2λ1 < zi 2 aλ1, zi, if zi 2 > aλ1. (8) where zi = Xi,: µk,: and S(zi; λ) represents multivariate soft-threshold operator. In this paper, we set a as equal to 3.7, as recommended in Fan & Li (2001). In addition to the previously mentioned multivariate soft-threshold operator and multivariate SCADthreshold operator, The application of the aforementioned methods can also be extended to the group Published in Transactions on Machine Learning Research (11/2024) minimax concave penalty (Huang & Zhang, 2010) associated with the multivariate minimax concave penaltythreshold operator. However, due to its similarity to the SCAD penalty, we exclusively employed the SCAD penalty. Moreover, we suggest avoiding applying the group hard penalty (Antoniadis, 1996), as optimizing it is inherently difficult. Furthermore, during our simulation process, we discovered that it is easy to converge to local minima. As discussed by Witten (2013) and She & Owen (2011), this formulation establishes a connection between the framework of robust M-estimation and the proposed model (6) with clustering centre µ kj for each k held fixed. As described above, the matrix of errors acquired from the algorithm is weighted based on the current weights. Therefore, it is necessary to restore the weighted error matrix E to its original unweighted state by dividing the current weight before we go to the next step to calculate the new weight for each variable (if the weight of a certain variable is zero, then E:,j will divide one instead of zero). Finally, given E and Θ, the minimization of L(Θ, E, w; λ) with respect to w is equivalent to j=1 wj QR j (X:,j; Θ, E:,j) P2(wj; λ2) + 1 where H = n wj R+ : q Pp j=1 w2 j = 1 o . We show in the Appendix that the final solution to (9) can be denoted as wj = S(QR j (X:,j; Θ, E:,j); λ2) q Pp j=1(S(QR j (X:,j; Θ, E:,j); λ2))2 . When applying the P2(wj; λ2) = λ2|wj| penalty in (9), S can be represented by the soft-thresholding operator as x λ2, if x > λ2 0, if |x| λ2 x + λ2, if x < λ2 When applying the P2(wj; λ2) = P SCAD λ2 (wj) penalty in (9), S can be represented by the SCAD-thresholding operator as sgn(x)(|x| λ2)+ if |x| 2λ2 {(a 1)x aλ2 sign(x)} /(a 2) if 2λ2 |x| < aλ2 x if aλ2 |x| The parameter λ2 has a crucial role in controlling the sparsity in the variable weight vector. A higher value of λ2 results in increased sparsity of the variable vector w. Specifically, a higher weight value wj indicates greater importance of the jth variable in the clustering process. Conversely, when wj is equal to zero, it signifies that the jth variable does not contribute to the clustering. In order to gain a deeper comprehension of the algorithmic process we have proposed, we summarize the aforementioned procedures in a pseudocode in Algorithm 1. Moreover, the computation complexity of T iteration of Algorithm 1 can be represented O(npt TK), assuming that the number of iterations in the 4th step is O(t). 2.4 Adaptation of the tuning parameters by the robust Gap statistics There are two tuning parameters in the proposed algorithm: λ1 and λ2. The tuning parameter λ1 is crucial for finding the error matrix and impacts the detection of outliers, and the tuning parameter λ2 plays a crucial role in filtering out variables contributing to the clustering process. We here provide the robust Gap statistics to determine these parameters. Published in Transactions on Machine Learning Research (11/2024) Algorithm 1 : Iterative algorithm for ARSK 1: Initialize the weight vector w as wj = 1/ p, set tolerance ε > 0 and r = 0. 2: Initialize the error matrix by setting E(0) ij of the 80% of data farthest from all data center points to Xij; the others set as 0. 3: repeat 4: Run the clustering algorithm on the weighted dataset X ij = w(r) j (Xij E(r) ij ) by max c1,...,ck X ij 1 |nk| i ck X ij 2 5: Calculate the new Eij by (7) or (8), which incorporates the cluster-specific means µk,: = |nk| 1 P 6: Until the (6) converges, we can obtain the error matrix E(r+1) ij and the cluster arrangement Θ(r+1). Keep those results for the next variable selection phase. 7: Restore the error matrix by setting: E(r+1) ij := E(r+1) ij / q 8: Arrange the cluster Θ(r+1) and compute the QR j (X:,j; Θ(r+1), E) for different variables by: QR j (X:,j; Θ(r+1), E) = X ij 1 |nk| where X ij = Xij E(r+1) ij . 9: Compute new variable weight by: w(r+1) j = S(QR j (X:,j; Θ, E:,j); λ2) q Pp j=1(S(QR j (X:,j; Θ, E:,j); λ2))2 10: r = r + 1 11: until convergence of the criterion (P j |w(r) j |) 1 P j |w(r+1) j w(r) j | < ε is met. 12: Output: w, E, Θ The Gap statistics were originally proposed in Tibshirani et al. (2001) for selecting the number of clusters in the K-means algorithm. Note that the Gap statistics is constructed as a function of the between-cluster sum of squares, D = Pp j=1 wj(n 1 Pn i=1 Pn i =1(Xij Xi j)2 PK k=1 n 1 k P i,i ck(Xij Xi j)2), for the original dataset. However, since D is sensitive to outliers, it is not suitable for selecting the tuning parameters under existence of outliers. Hence, we propose a robust version of Gap statistics by adding the error part to each observation when calculating the DR λ2,λ1. As we mentioned before, the error part can outweigh the influence of the outlier, so that we define Gapλ2,λ1 as Gapλ2,λ1 = log(DR λ2,λ1) 1 b=1 log(DR(b) λ2,λ1), DR λ2,λ1 = j=1 wj QR j (X:,j; Θ, E:,j), (10) and the optimal value of λ2 and λ1 corresponding to the largest value of Gapλ2,λ1. Note that DR λ2,λ1 is a weighted robust between-cluster sum of squares defined in (3), and DR(b) λ2,λ1 is a version with bth permuted datasets. Here, B denotes the number of permuted datasets generated by randomly selecting from the original dataset. Increasing the number of permuted datasets improves the accuracy of selecting the true parameters. The between-cluster sum of squares, taking account of the variable weight, is adopted in Witten & Tibshirani (2010), so our version used in (10) can be regarded as its robust extension. Published in Transactions on Machine Learning Research (11/2024) Since it may be computationally intensive to search the optimal value of (λ2, λ1) by a grid search method, we conduct an alternating optimization algorithm for tuning parameter search. Specifically, we first set a suitable value λ 1 for λ1 and compute the optimal value λ 2 of λ2 by maximizing Gapλ2,λ 1. Then, we obtain the optimal value λ 1 of λ1 by maximizing Gapλ 2,λ1. Using this search algorithm instead of the grid search method can save a significant computation time. We also offer a pseudocode for this procedure for easy understanding of this search algorithm given in Algorithm 2. Algorithm 2 : Selection of (λ2, λ1) by the robust Gap statistics 1: Set some value for λ 1 and maximize the following Gap statistics with respect to λ2: Gapλ2 = log(DR λ2,λ 1) 1 b=1 log(DR(b) λ2,λ 1) to obtain the optimal λ 2. 2: Fix λ2 = λ 2 and optimize the following Gap statistics with respect to λ1: Gapλ1 = log(DR λ 2,λ1) 1 b=1 log(DR(b) λ 2,λ1) to obtain the optimal λ 1. 3: Output: the optimal set of tuning parameters, (λ 2, λ 1) 3 Numerical Studies 3.1 Experimental setup In this section, we explore the ability of the proposed clustering method. We consider that each observation xi is generated independently from a multivariate normal distribution, given that the observation belongs to cluster k. Specifically, for an observation xi in cluster k, we have xi N(µk,:, Σp), where µk,: Rp denotes the mean vector for cluster k, and Σp Rp p represents the covariance matrix. Each element of µk,: is independently sampled from from either U( 6, 3) or U(3, 6). To simulate scenarios involving outliers, we introduce a contaminated error distribution with a multivariate normal mixture, represented as (1 π)N(µk,:, Σp) + πN(µk,: + bj, Σp) where j (1, . . . , p) and π [0, 1] represents the proportion of outliers for each cluster. In the simulation study, we consider two types of Σp. In the first scenario, the assumption is made that all variables are independent. In another scenario, we assume there is a correlation structure among variables. To introduce randomness to the appearance of outliers, bj is generated with a random number from one of two uniform distributions: U( 13, 7) or U(7, 13). In this setup, we assume that some explanatory variables are significantly associated with the response, while the remaining variables are redundant for variable selection. To mimic this situation, we randomly set q elements of µk,: to non-zero values while the remaining elements are set to zero. Consequently, a successful method should accurately identify the q positions where the weights wj corresponding to the true significant variables are non-zero while ensuring that the weights wj of the remaining variables are set to zero. In order to evaluate the approach s clustering accuracy capability, we employ the clustering error rate (CER) (Rand, 1971). The CER measures the extent to which the model s predicted partition ˆC for a set of n observations is consistent with the true labels C. The CER is defined as CER( ˆC, C) = n 2 1 ˆ C(i,i ) 1C(i,i ) , (11) where, 1C(i,i ) and 1 ˆ C(i,i ) indicate whether the ith and i th observations belong to the same cluster. Lower CER indicates better performance in clustering accuracy or outlier detection. Published in Transactions on Machine Learning Research (11/2024) Furthermore, in order to evaluate each method s proficiency in variable sparsity, we apply two criteria for variable selection. The variable true positive rate (TPR) indicates the approach s success in finding informative variables. The true negative rate (TNR) represents the successful identification of non-informative variables. If both criteria are closer to 1, it indicates that the model provides a more comprehensive explanation of the structure of the variables predicted in the study. 3.2 Simulation 1: selection of turning parameter In this subsection explores the new Gap statistics capability to directly search for the tuning parameters of robustness and variable vector sparsity using Algorithm 2. Initially, we considered 3 clusters, each containing 50 observations, number of variables as p = 50, with q = 5, and all variables are independent, i.e., Σp = Ip. A proper of hyperparameters λ1 and λ2 should make the model accurately identify the structure of the clustering data in different contamination levels for π in {0, 0.1, 0.2, 0.3}. The proposed tuning parameter search strategy generates both the robustness parameter λ1 and the variable selection tuning parameter λ2 using exponential decay. We consider the number of permuted datasets to be 25 (i.e., B = 25). We employed different thresholding function types, where the first thresholding function is used to control model sparsity, and the second thresholding function is used to control model robustness. The tuning parameter search process is repeated 30 times for 30 different datasets, and the resulting table is presented in Table 3.2. threshold contamination number of number of detected type level detected outliers informative variable soft-soft π = 0 0 (0) 4.300 (0.000) π = 0.1 15.56 (6.425) 4.533 (0.618) π = 0.2 28.73 (2.434) 4.200 (0.871) π = 0.3 33.10 (14.76) 13.60 (7.203) soft-SCAD π = 0 0 (0) 4.400 (0.000) π = 0.1 14.43 (0.495) 6.310 (1.854) π = 0.2 28.68 (1.349) 4.566 (0.495) π = 0.3 42.36 (2.287) 6.800 (2.946) SCAD-soft π = 0 0 (0) 4.545 (0.687) π = 0.1 14.30 (0.483) 5.400 (0.469) π = 0.2 27.11 (6.166) 5.941 (1.748) π = 0.3 38.63 (16.63) 9.272 (4.540) SCAD-SCAD π = 0 0.090 (0.333) 4.272 (0.881) π = 0.1 13.33 (1.258) 5.833 (3.125) π = 0.2 24.10 (5.065) 4.500 (2.068) π = 0.3 44.00 (4.2110) 7.187 (1.223) Table 1: Using the robust Gap statistic to select the optimal tuning parameters for the RSKC algorithm based on soft-soft-thresholding, soft-SCAD-thresholding, SCAD-soft-thresholding, and SCAD-SCAD-thresholding, we report the average and its standard error (in parentheses) of the number of detected outliers and the number of detected informative variables. Table 3.2 summarizes the results of the evaluation measures. Overall, a good tuning parameter λ1 should accurately identify the number of outliers under different contamination levels. As the contamination level π increases from 0 to 0.3, we observe that the four thresholding methods demonstrate varying outlier detection capabilities across different contamination levels. At π = 0.1, the detected number of outliers ranges from 13.33 to 15.56, with soft-soft-thresholding achieving the highest average of 15.56 (standard error: 6.425). When π increases to 0.2, the detected number of outliers falls between 24.10 and 28.73, with soft-soft and soft-SCAD-thresholding showing similar performance. This demonstrates that the Gap statistics can help the ARKC method determine λ1 to some extent. However, when the contamination level reaches 0.3, the soft-soft-thresholding ARSK method detects 33.10 outliers on average, which is lower than the SCAD-SCADthresholding ARSK method (44.00 outliers) and deviates from the true number of outliers. Regarding the selection of the tuning parameter λ2, a good tuning parameter should accurately identify the number of informative variables under different contamination levels. When the data is clean (π = 0), Published in Transactions on Machine Learning Research (11/2024) all four methods identify around 4 to 5 informative variables, which is close to 5. As the contamination level increases, the number of detected informative variables generally increases, with some fluctuations. At π = 0.3, the soft-soft-thresholding ARSK method detects 13.60 informative variables on average, which significantly deviates from the 5. In contrast, the SCAD-SCAD-thresholding ARSK method detects 7.187 informative variables, showing better performance in the presence of high contamination. Furthermore, errors in identifying informative variables can lead to significant biases in outlier detection. Considering these results, we can conclude that the proposed parameter selection algorithm based on the Gap statistic works relatively well under low to moderate contamination levels. However, its performance may degrade when the contamination level is high, especially for the soft-thresholding-based methods. The SCAD-thresholding-based methods generally show better robustness and accuracy in identifying outliers and informative variables under various contamination levels. 3.3 Simulation 2: comparison with other methods In this subsection, we present the findings of a comprehensive simulation study conducted to investigate the cluster data structure and properties by the ARSK algorithm based on soft-thresholding and SCADthresholding. The study extensively applied Gap statistics to estimate optimal parameter settings and compared with several benchmark approaches, including the original K-means (KC), PCA-K-means (PCA-KC) - a mixed approach that combines Principal Component Analysis (PCA) with the original K-means algorithm, as initially proposed by Ding & He (2004). It is noted that PCA is a widely recognized dimensionality reduction technique that transforms high-dimensional data into a lower-dimensional space. We also considered Trimmed K-means (TKM), which executes standard K-means clustering and subsequently eliminates a predetermined percentage of points exhibiting the greatest Euclidean distance from their respective cluster centroids. Furthermore, we employ Robust and Sparse K-means (RSKC), whose fundamental premise involves adapting the SK-means algorithm of Witten & Tibshirani (2010) in conjunction with Trimmed K-means. Lastly, we apply the Weighted Robust K-means (WRCK) method, which shares similarities with RSKC but distinguishes itself through the incorporation of observation-specific weights in its objective function. These weights are reflective of each observation s isolation level relative to its surrounding neighbourhood. Considering the mixture error model mentioned at the beginning, we also generated data for 3 clusters, each containing 50 observations. We set the number of variables as p = 50, with q = 5, and p = 500, with q = 50. Meanwhile, we consider π in {0, 0.1, 0.2} for specifying the distribution of outliers. As previously stated, we study two types of covariance matrices: one with Σp = Ip and the another generated according to the method proposed by Hirose et al. (2017) as ρt 1 ... ρt ... ... 1 ... ρt ρt 1 where the Q denotes a p p random rotation matrix satisfying QT = Q 1 and the ρt are randomly generated form an Uniform distribution U(0.1, 1). In order to evaluate the clustering accuracy and outlier detection capability of each approach, we employ the clustering error rate (CER) as described above. To assess the outlier detection performance, the outliers identified by each model are assigned to the (K + 1)-th cluster group. Table 3.3 presents the results of 100 simulations for the scenario where all variables are independent, i.e., Σp = Ip, while Table 3.3 summarizes the results of 100 simulations for the scenario where variables exhibit a certain correlation structure. In Tables 3.3 and 3.3, the results show that KC, PCA-KC, and TKM approaches, which lack robustness for high dimensions or outliers, are the worst performers, particularly when the dimensionality p and proportion of outliers π increase. For instance, in the independent case with p = 500 and q = 50, the CER of KC increases from 0.050 to 0.341 as π increases from 0 to 0.2. Similarly, the CER of PCA-KC and TKM Published in Transactions on Machine Learning Research (11/2024) p = 50 p = 500 q = 5 q = 50 π = 0 π = 0.1 π = 0.2 π = 0 π = 0.1 π = 0.2 KC 0.073 (0.119) 0.191 (0.124) 0.285 (0.107) 0.050 (0.105) 0.228 (0.154) 0.341 (0.145) PCA-KC 0.109 (0.113) 0.326 (0.167) 0.383 (0.132) 0.058 (0.113) 0.219(0.098) 0.300(0.123) TKM (α = 0.1) 0.140 (0.027) 0.093 (0.036) 0.220 (0.087) 0.128 (0.005) 0.098 (0.044) 0.302 (0.146) RSKC (α = 0.1) 0.119 (0.029) 0.008 (0.031) 0.169 (0.118) 0.105 (0.007) 0 (0) 0.123 (0.129) RSKC (α = 0.2) 0.208 (0.029) 0.127 (0.015) 0.005 (0.021) 0.187 (0.007) 0.123 (0.009) 0 (0) WRCK 0.129 (0.037) 0.208 (0.104) 0.136 (0.071) 0.118 (0.024) 0.084 (0.028) 0.046 (0.018) soft-soft-ARSK 0.010 (0.043) 0.014 (0.043) 0.032 (0.064) 0 (0) 0 (0) 0 (0) soft-SCAD-ARSK 0.010 (0.038) 0.013 (0.035) 0.026 (0.033) 0 (0) 0 (0) 0 (0) SCAD-soft-ARSK 0.021 (0.059) 0.016 (0.036) 0.031 (0.061) 0 (0) 0 (0) 0 (0) SCAD-SCAD-ARSK 0.017 (0.057) 0.023 (0.052) 0.019 (0.045) 0 (0) 0 (0) 0 (0) Table 2: When variables are independent, the average values of the CER and their standard errors (in parentheses) based on 100 Monte Carlo replications. p = 50 p = 500 q = 5 q = 50 π = 0 π = 0.1 π = 0.2 π = 0 π = 0.1 π = 0.2 KC 0.078(0.122) 0.199(0.123) 0.318(0.138) 0.037(0.096) 0.219(0.172) 0.306(0.136) PCA-KC 0.112(0.113) 0.284(0.169) 0.372(0.126) 0.113(0.131) 0.212(0.075) 0.278(0.046) TKM (α = 0.1) 0.083(0.055) 0.039(0.080) 0.240(0.104) 0.065(0.002) 0.139 (0.053) 0.136 (0.286) RSKC (α = 0.1) 0.082 (0.021) 0.005 (0.016) 0.235 (0.128) 0.082 (0.008) 0 (0) 0.071 (0.089) RSKC (α = 0.2) 0.194 (0.021) 0.123 (0.021) 0.008 (0.015) 0.162 (0.010) 0.122 (0.009) 0 (0) WRCK 0.119 (0.021) 0.203 (0.116) 0.146 (0.071) 0.126 (0.019) 0.102 (0.020) 0.022 (0.081) soft-soft-ARSK 0.009 (0.038) 0.014 (0.039) 0.037 (0.047) 0 (0) 0 (0) 0.000 (0.002) soft-SCAD-ARSK 0.005 (0.020) 0.015 (0.039) 0.032 (0.026) 0 (0) 0 (0) 0.001 (0.002) SCAD-soft-ARSK 0.007 (0.021) 0.009 (0.014) 0.047 (0.069) 0 (0) 0 (0) 0.000 (0.002) SCAD-SCAD-ARSK 0.005 (0.019) 0.004 (0.005) 0.029 (0.033) 0 (0) 0 (0) 0.002 (0.018) Table 3: When variables are correlated, the average values of the CER and their standard errors (in parentheses) based on 100 Monte Carlo replications. (α = 0.1) increase from 0.058 to 0.300 and 0.128 to 0.302, respectively, under the same scenario. Conversely, RSCK, WRCK, and ARSK perform exceptionally well in high-dimension and high-contamination scenarios. However, the effectiveness of the RSCK approach heavily depends on the choice of the parameter α. When α equals the true outlier proportion, RSCK performs impressively, with CER close to 0. For example, in the correlated case with p = 500 and q = 50, the CER of RSKC (α = 0.1) is 0 when π = 0.1. However, the approach loses effectiveness if α is incorrectly specified, leading to higher CER values. In the same scenario, the CER of RSKC (α = 0.2) is 0.122 when π = 0.1. The WRCK approach, based on the density clustering method, generally outperforms RSCK when outliers are present. Still, it may erroneously classify more normal data points as outliers when no contamination exists. In contrast, the ARSK approach, with different combinations of thresholding functions (e.g., soft-soft, soft-SCAD, SCAD-soft, and SCAD-SCAD), consistently maintains a low Clustering Error Rate (CER) across all simulated data scenarios. In summary, the simulation results demonstrate the effectiveness and robustness of the ARSK method in clustering high-dimensional datasets, particularly in the presence of outliers and variable correlations. Across diverse data scenarios, the ARSK variants consistently outperform traditional methods, including KC, PCAKC, TKM, RSKC, and WRCK, highlighting the distinct advantages of the ARSK approach. Both In Table 3.3 and Table 3.3, we observe that the CER of KC is worse than ARSK when there are no outliers in the dataset. Moreover, when π = 0.1, the CER of the TKM method is significantly higher than that of the RSCK method. We attribute this result to the presence of noisy variables, which contribute to the increased CER. Therefore, it is essential to identify these noisy variables to reduce their impact. Published in Transactions on Machine Learning Research (11/2024) Through 100 repetitions, the results of the TPR and TNR are presented in Tables 3.3 and 3.3. As the proportion of outliers increases from 0 to 0.2, both the TPR and TNR of most approaches demonstrate a decreasing trend across different scenarios, suggesting that the presence of outliers can negatively influence the variable selection performance. However, the TPR and TNR of each type of ARSK approach variant remain relatively high and stable across various data scenarios, demonstrating ARSK s superior variable selection capabilities in the presence of outlier contamination in datasets. For example, in the independent case with p = 500 and q = 50, the TPR of soft-soft-ARSK, soft-SCAD-ARSK, SCAD-soft-ARSK, and SCAD-SCAD-ARSK remain above 0.797, 0.810, 0.798, and 0.810, respectively, even when π = 0.2. Similarly, their TNR values stay above 0.977, 0.980, 0.978, and 0.981 in the same scenario. These results emphasize the superior ability of soft-ARSK and SCAD-ARSK methods to accurately identify informative and noninformative variables in the presence of outliers, making them valuable tools for handling contaminated clustering datasets. In contrast, during the simulation process, we observed that traditional trimmed approaches such as WRCK and RSKC often struggled to maintain their variable selection performance under high outlier proportions. In many cases, these methods incorrectly assigned non-zero weights to variables that originally had zero weights, leading to an increased number of false positives. This limitation makes it challenging for WRCK and RSKC to effectively detect informative variables in the presence of outliers, emphasizing the need for more effective methods like ARSK. p = 50 p = 500 q = 5 q = 50 π = 0 π = 0.1 π = 0.2 π = 0 π = 0.1 π = 0.2 TPR RSKC (α = 0.1) 0.924 (0.038) 0.890 (0.100) 0.710 (0.267) 0.845 (0.101) 0.908 (0.100) 0.780 (0.255) RSKC (α = 0.2) 0.654 (0.109) 0.728 (0.143) 0.880 (0.098) 0.690 (0.043) 0.993 (0.027) 0.917 (0.095) WRCK 0.888 (0.099) 0.664 (0.172) 0.834 (0.106) 0.734 (0.044) 0.978 (0.048) 0.841 (0.101) soft-soft-ARSK 0.922 (0.127) 0.842 (0.169) 0.810 (0.183) 0.961 (0.035) 0.853 (0.053) 0.797 (0.084) soft-SCAD-ARSK 0.944 (0.103) 0.910 (0.162) 0.800 (0.171) 0.932 (0.020) 0.836 (0.084) 0.818 (0.018) SCAD-soft-ARSK 0.962 (0.092) 0.906 (0.153) 0.766 (0.144) 0.964 (0.029) 0.969 (0.024) 0.798 (0.054) SCAD-SCAD-ARSK 0.970 (0.076) 0.802 (0.150) 0.810 (0.160) 0.969 (0.026) 0.820 (0.054) 0.810 (0.055) TNR RSKC (α = 0.1) 0.671 (0.427) 0.660 (0.475) 0.926 (0.102) 0.806 (0.377) 0.548 (0.490) 0.817 (0.235) RSKC (α = 0.2) 0.978 (0.142) 0.880 (0.326) 0.753 (0.430) 1 (0) 0.086 (0.259) 0.532 (0.482) WRCK 0.660 (0.476) 0.911 (0.199) 0.837 (0.309) 1 (0) 0.261 (0.416) 0.536 (0.478) soft-soft-ARSK 1 (0) 0.969 (0.048) 0.981 (0.042) 0.999 (0.001) 0.986 (0.012) 0.977 (0.009) soft-SCAD-ARSK 1 (0) 0.954 (0.040) 0.996 (0.008) 0.972 (0.001) 0.988 (0.014) 0.980 (0.009) SCAD-soft-ARSK 1 (0) 0.906 (0.153) 0.976 (0.021) 0.999 (0.001) 0.999 (0.001) 0.978 (0.008) SCAD-SCAD-ARSK 1 (0) 0.976 (0.021) 0.969 (0.017) 0.999 (0.000) 0.981 (0.010) 0.981 (0.010) Table 4: When variables are independent, the average of 100 tests of TPR and TNP for variable selections and its standard error(in parentheses) are presented for the various data scenarios 3.4 Applications to real data 3.4.1 UCI data application In this subsection, we apply ARSK with different thresholding configurations to real-life datasets and compare it with the benchmark methods previously mentioned in the simulation study. We consider the dataset all from the UCI Machine Learning Repository (Dua & Graff (2019)). The datasets evaluated include various applications, such as glass identification, breast cancer diagnosis, acoustic signal classification, spam base detection, mortality analysis, and Parkinson s disease detection. The real-life dataset comprises data with a wide range of dimensions, ranging from low to high. As in the previous analysis, we also applied Gap statistics to estimate the hyperparameter. Due to the different units of the continuous variables, we normalize all of continuous variables before conducting the study. Table 6 summarizes the result of the real-life dataset comparison. The reported values are CERs for each approach. In order to better illustrate the performance of that method, we have emphasized the first and second best CER values in bold when applied to a certain real-life dataset. While traditional methods such Published in Transactions on Machine Learning Research (11/2024) p = 50 p = 500 q = 5 q = 50 π = 0 π = 0.1 π = 0.2 π = 0 π = 0.1 π = 0.2 TPR RSKC (α = 0.1) 0.917 (0.013) 0.839 (0.125) 0.761 (0.236) 0.813 (0.125) 0.934 (0.092) 0.802 (0.294) RSKC (α = 0.2) 0.705 (0.132) 0.608 (0.260) 0.860 (0.015) 0.695 (0.232) 0.903 (0.057) 0.884 (0.072) WRCK 0.813 (0.126) 0.604 (0.212) 0.734 (0.126) 0.613 (0.213) 0.932 (0.074) 0.812 (0.126) soft-soft-ARSK 0.953 (0.086) 0.860 (0.175) 0.769 (0.042) 0.974 (0.031) 0.843 (0.071) 0.733 (0.095) soft-SCAD-ARSK 0.940 (0.105) 0.855 (0.181) 0.806 (0.177) 0.968 (0.022) 0.849 (0.071) 0.773 (0.109) SCAD-soft-ARSK 0.928 (0.096) 0.830 (0.177) 0.772 (0.169) 0.973 (0.029) 0.877 (0.051) 0.813 (0.063) SCAD-SCAD-ARSK 0.928 (0.104) 0.918 (0.120) 0.758 (0.186) 0.968 (0.036) 0.874 (0.053) 0.793 (0.058) TNR RSKC (α = 0.1) 0.680 (0.387) 0.730 (0.325) 0.864 (0.226) 0.776 (0.218) 0.632 (0.363) 0.743 (0.206) RSKC (α = 0.2) 0.837 (0.232) 0.935 (0.102) 0.794 (0.400) 1 (0) 0.136 (0.454) 0.542 (0.436) WRCK 0.697 (0.437) 0.954 (0.083) 0.803 (0.231) 1 (0) 0.476 (0.376) 0.644 (0.306) soft-soft-ARSK 1 (0) 0.961 (0.085) 0.985 (0.026) 0.998 (0.006) 0.977 (0.011) 0.984 (0.015) soft-SCAD-ARSK 1 (0) 0.876 (0.080) 0.997 (0.008) 0.999 (0.001) 0.978 (0.012) 0.991 (0.008) SCAD-soft-ARSK 1 (0) 0.978 (0.035) 0.992 (0.017) 0.983 (0.010) 0.977 (0.011) 0.983 (0.010) SCAD-SCAD-ARSK 1 (0) 0.968 (0.020) 0.994 (0.013) 0.997 (0.005) 0.986 (0.010) 0.981 (0.009) Table 5: When variables are correlated, the average of 100 tests of TPR and TNP for variable selections and its standard error(in parentheses) are presented for the various data scenarios RSKC ARSK (proposed) dataset K (n, p) KC PCA-KC TKM α = 0.1 α = 0.2 WRCK soft-soft soft-SCAD SCAD-soft SCAD-SCAD glass 7 (214, 9) 0.334 0.299 0.324 0.324 0.304 0.319 0.259 0.297 0.159 0.142 Breast Cancer 2 (699, 9) 0.080 0.080 0.107 0.071 0.081 0.104 0.065 0.057 0.079 0.084 Acoustic 4 (400, 50) 0.309 0.443 0.392 0.293 0.280 0.382 0.354 0.360 0.355 0.346 Spambase 2 (4601, 57) 0.476 0.481 0.310 0.406 0.563 0.493 0.434 0.432 0.364 0.413 mortality 3 (198, 90) 0.227 0.185 0.228 0.153 0.149 0.106 0.131 0.136 0.143 0.149 DARWIN 2 (174, 451) 0.391 0.400 0.447 0.402 0.434 0.465 0.385 0.398 0.454 0.401 Parkinson 2 (756, 754) 0.477 0.462 0.500 0.482 0.563 0.392 0.445 0.441 0.429 0.462 Table 6: Comparison of CERs of different algorithms for the real-world data. The smallest and second smallest CER values are shown in bold. as KC, PCA-KC, and TKM demonstrate effectiveness in certain contexts, they generally yield higher CERs across the datasets compared to the proposed ARSK algorithm and its variants. For instance, in the glass and Breast Cancer dataset, the CER values obtained by KC, PCA-KC, and TKM are considerably higher than those of the best-performing ARSK. In seven real-world datasets, our ARSK method ranked among the top two best methods in six of the datasets. Therefore, the proposed ARSK algorithm consistently achieves superior performance with lower CERs across most of the real datasets, particularly when using the multivariate soft-threshold operator to obtain the error matrix E. ARSK demonstrates competitive performance in this regard. Furthermore, Figure 3.4.1 summarizes the number of zero elements in the variable weight vector of each method, where ARKC adopted a combination of thresholding corresponding to the minimum CER. Figure 3.4.1 shows that when analyzing ultra-high-dimensional data, the variable weight vector predicted by ARSK exhibits strong sparsity. Published in Transactions on Machine Learning Research (11/2024) Figure 1: The number of zero elements in the variable weight vector of each method, with results ranging from low-dimension to high-dimension dataset This comparative analysis emphasizes the ARSK algorithm s success in significantly reducing classification error rates across diverse real-world datasets, showcasing the versatility and adaptability of our proposed clustering method. Its different configurations (e.g., soft-soft, soft-SCAD, SCAD-soft, and SCAD-SCAD) allow it to handle various data structures and sizes effectively. The consistent performance across different real-world datasets underscores the broad applicability of our approach in various fields. 3.4.2 Analyzing the group structure of recognition of handwritten digits The dataset was downloaded from Kaggle, a data science competition platform. It consists of digitized images of the numbers 0 to 9 handwritten by 13 subjects. The images were divided into 64 non-overlapping blocks of 4 4 bits. For each block, we observe the number of black pixels. Hence, each image is represented by p = 64 variables recording the count of pixels in each block, taking integer values between 0 and 16. Its row names identify the true digit in the image, and the column names identify the position of each block in the original bitmap. Figure 2: The images of handwritten digits data Figure 3: The heatmap of the variable weight vector obtained by SCAD-soft ARSK Our focus is to detect an entire variable weight and group structure, i.e., ten groups. Given the superior performance of SCAD-soft ARSK in subsection 3.4.1, we apply RSKC (α = 0.1), RSKC (α = 0.2), WRCK, and SCAD-soft ARSK to this dataset. Although the precise contribution of individual pixels to the clustering process remains indeterminate, visual inspection of the real images of handwritten digits data (0 and 9) as Figure 2 shows that the pixels situated in both the initial and terminal columns exhibit uniform white Published in Transactions on Machine Learning Research (11/2024) RSKC-0.2 RSKC-0.1 SCAD-soft ARSK WRCK CER 0.086 0.077 0.079 0.156 Table 7: Evaluation of the clustering performance on recognition of handwritten digits RSKC-0.2 RSKC-0.1 SCAD-soft ARSK WRCK CER 6 9 13 0 Table 8: The number of zero elements within the weight vector derived from each method colouration. Thus, these 16 pixels can be reasonably deemed redundant and inconsequential for the clustering. The optimal robustness and sparsity parameter for ARSK is also selected using the Gap statistic. The evaluation of the resulting clustering solution is presented in Table 3.4.2. We also examine the final variable weights obtained by the sparse k-means-based algorithms. Table 3.4.2 illustrates each method s number of zero elements within the weight vector derived. The analysis of the weight values reveals that WRCK is ineffective in achieving sparsity within the variable weight vector, as evidenced by the exclusive presence of nonzero elements throughout. Correspondingly, it also exhibits the highest CER. As shown in Table 3.4.2, the performance of RSKC and ARSK is similar, with CER values around 0.8. However, the RSKC (α = 0.2) produces weight vectors with only 6 zero elements, indicating a moderate sparsity level. In contrast, the proposed ARSK approach generates weight vectors with 13 zero elements that are close to 16. We also presented a heatmap, in Figure 3, visualization of the variable weight vector derived from the SCAD-soft ARSK algorithm as Figure 2, wherein the intensity of the chromatic scale correlates positively with the magnitude of the weights. We can conclude that these findings correspond with those shown in Figure 2. These results demonstrate that ARSK exhibits both sparsity and accuracy of feature selection capabilities. 4 Concluding remarks In this paper, we propose an Adaptively Robust and Sparse K-means Clustering algorithm designed to handle multiple tasks within a unified framework for a comprehensive analysis of contaminated high-dimensional datasets. Our approach demonstrates greater flexibility in handling outliers within real-world datasets. We also proposed a modified Gap statistic with an alternating optimization algorithm to search for tuning parameters. This approach significantly reduces the computational time required for parameter tuning compared to traditional methods. The experimental results demonstrate the significant capabilities of the proposed approach in revealing group structures in high-dimensional data. Our model not only performs well in detecting outliers but also identifies significant variables directly, leading to a more thorough understanding of complex clustering datasets. In our current study, we assume that the number of clusters is known and consistent with the traditional K-means method. Additionally, Chen & Witten (2023) proposed a test for detecting differences in means between each cluster estimated from K-means clustering. This technique can be applied to the modified dataset Xij Eij, where noise variables (i.e., wj = 0) are removed to help determine which clusters should be retained. Furthermore, we have consistently utilized a single penalty for outliers in the objective function (4). This implements a uniform approach to detecting outliers across various groups. However, given the complexity and diversity of the data, a more sophisticated strategy may be required. This might involve using distinct group penalty functions for different groups. Incorporating this approach into unsupervised learning presents challenges in selecting group penalty functions. These extensions are left to our future work. Acknowledgments This work is supported by the Japan Society of the Promotion of Science (JSPS KAKENHI) grant numbers, 20H00080 and 21H00699. Published in Transactions on Machine Learning Research (11/2024) Anestis Antoniadis. Smoothing noisy data with tapered coiflets series. Scandinavian Journal of Statistics, pp. 313 330, 1996. Anestis Antoniadis and Jianqing Fan. Regularization of wavelet approximations. Journal of the American Statistical Association, 96(455):939 967, 2001. Šárka Brodinová, Peter Filzmoser, Thomas Ortner, Christian Breiteneder, and Maia Rohm. Robust and sparse k-means clustering for high-dimensional data. Advances in Data Analysis and Classification, 13: 905 932, 2019. Yiqun T Chen and Daniela M Witten. Selective inference for k-means clustering. Journal of Machine Learning Research, 24(152):1 41, 2023. Pietro Coretto and Christian Hennig. Robust improper maximum likelihood: tuning, computation, and a comparison with other methods for robust gaussian clustering. Journal of the American Statistical Association, 111(516):1648 1659, 2016. Juan Antonio Cuesta-Albertos, Alfonso Gordaliza, and Carlos Matrán. Trimmed k-means: an attempt to robustify quantizers. The Annals of Statistics, 25(2):553 576, 1997. Chris Ding and Xiaofeng He. K-means clustering via principal component analysis. In Proceedings of the twenty-first international conference on Machine learning, pp. 29, 2004. Dheeru Dua and C Graff. Uci machine learning repository irvine. University of California, 2019. Jianqing Fan and Runze Li. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical Association, 96(456):1348 1360, 2001. Edward W Forgy. Cluster analysis of multivariate data: efficiency versus interpretability of classifications. biometrics, 21:768 769, 1965. Hironori Fujisawa and Shinto Eguchi. Robust estimation in the normal mixture model. Journal of Statistical Planning and Inference, 136(11):3989 4011, 2006. Kei Hirose, Hironori Fujisawa, and Jun Sese. Robust sparse gaussian graphical modeling. Journal of Multivariate Analysis, 161:172 190, 2017. Jian Huang, Patrick Breheny, and Shuangge Ma. A selective review of group selection in high-dimensional models. Statistical science: a review journal of the Institute of Mathematical Statistics, 27(4), 2012. Junzhou Huang and Tong Zhang. The benefit of group sparsity. 2010. Shota Katayama and Hironori Fujisawa. Sparse and robust linear regression: An optimization algorithm and its statistical properties. Statistica Sinica, pp. 1243 1264, 2017. Yegor Klochkov, Alexey Kroshnin, and Nikita Zhivotovskiy. Robust k-means clustering for distributions with two moments. The Annals of Statistics, 49(4):2206 2230, 2021. Yumi Kondo, Matias Salibian-Barrera, and Ruben Zamar. RSKC: An R package for a robust and sparse k-means clustering algorithm. Journal of Statistical Software, 72:1 26, 2016. Stuart Lloyd. Least squares quantization in PCM. IEEE transactions on information theory, 28(2):129 137, 1982. Antonio Punzo and Paul D Mc Nicholas. Parsimonious mixtures of multivariate contaminated normal distributions. Biometrical Journal, 58(6):1506 1537, 2016. William M Rand. Objective criteria for the evaluation of clustering methods. Journal of the American Statistical association, 66(336):846 850, 1971. Published in Transactions on Machine Learning Research (11/2024) Yiyuan She and Art B Owen. Outlier detection using nonconvex penalized regression. Journal of the American Statistical Association, 106(494):626 639, 2011. Shonosuke Sugasawa and Genya Kobayashi. Robust fitting of mixture models using weighted complete estimating equations. Computational Statistics & Data Analysis, 174:107526, 2022. Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1):267 288, 1996. Robert Tibshirani, Guenther Walther, and Trevor Hastie. Estimating the number of clusters in a data set via the gap statistic. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 63(2): 411 423, 2001. Lifeng Wang, Guang Chen, and Hongzhe Li. Group scad regression analysis for microarray time course gene expression data. Bioinformatics, 23(12):1486 1494, 2007. Daniela M Witten. Penalized unsupervised learning with outliers. Statistics and its Interface, 6(2):211, 2013. Daniela M Witten and Robert Tibshirani. A framework for feature selection in clustering. Journal of the American Statistical Association, 105(490):713 726, 2010. Miin-Shen Yang, Chien-Yo Lai, and Chih-Ying Lin. A robust EM clustering algorithm for Gaussian mixture models. Pattern Recognition, 45(11):3950 3961, 2012. Ming Yuan and Yi Lin. Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society Series B: Statistical Methodology, 68(1):49 67, 2006. Hui Zou and Trevor Hastie. Addendum: "regularization and variable selection via the elastic net [j. r. stat. soc. ser. b stat. methodol. 67 (2005), no. 2, 301 320; mr2137327]. journal of the royal statistical society, 67(5):768 768, 2010. A Appendix: Derivation of the solution to (9) Consider maximizing the following regularized objective function for estimating the weights wj. L(wj, λ2) = P2(wj; λ2) + 1 where Qj = QR j (X:,j; Θ, E:,j). Taking the derivative with respect to wj and setting it to zero, we obtain: Qj wj Γj = 0. When the L1 penalty P2(wj; λ2) is applied, Γj represents the subgradient of the L1 penalty with respect to wj. Solving the above equation yields Ssoft(x; λ2) = wj = Qj λ2, if Qj > λ2 0, if |Qj| λ2 Qj + λ2, if Qj < λ2. When the SCAD penalty P2(wj; λ2) is applied, Γj represents the subgradient of the SCAD penalty with respect to wj. Solving the above equation yields SSCAD(x; λ2) = wj = sgn(Qj)(|Qj| λ2)+ if |Qj| 2λ2 (a 2) 1 (a 1)Qj aλ2sign(Qj) if 2λ2 |Qj| < aλ2 Qj if aλ2 |Qj|. Projecting ˆ wj onto the parameter space q Pp j=1 wj = 1, the final solution can be rewritten as wj = S(Qj; λ2) q Pp j=1(S(Qj; λ2))2 .