# spectral_selfsupervised_feature_selection__10ec3205.pdf Published in Transactions on Machine Learning Research (12/2024) Spectral Self-supervised Feature Selection Daniel Segal segaldaniel23@gmail.com Hebrew University Ofir Lindenbaum ofirlin@gmail.com Bar Ilan University Ariel Jaffe ariel.jaffe@mail.huji.ac.il Hebrew University Reviewed on Open Review: https: // openreview. net/ forum? id= t0EJi Od9Lg Choosing a meaningful subset of features from high-dimensional observations in unsupervised settings can greatly enhance the accuracy of downstream analysis tasks, such as clustering or dimensionality reduction, and provide valuable insights into the sources of heterogeneity in a given dataset. In this paper, we propose a self-supervised graph-based approach for unsupervised feature selection. Our method s core involves computing robust pseudo-labels by applying simple processing steps to the graph Laplacian s eigenvectors. The subset of eigenvectors used for computing pseudo-labels is chosen based on a model stability criterion. We then measure the importance of each feature by training a surrogate model to predict the pseudo-labels from the observations. Our approach is shown to be robust to challenging scenarios, such as the presence of outliers and complex substructures. We demonstrate the effectiveness of our method through experiments on real-world datasets from multiple domains, with a particular emphasis on biological datasets. 1 Introduction Improvements in sampling technology enable scientists across many disciplines to acquire numerous variables from biological or physical systems. One of the critical challenges in real-world scientific data is the presence of noisy, information-poor, or nuisance features. While such features could be mildly harmful to supervised learning, they could dramatically affect the outcome of downstream analysis tasks (e.g., clustering or manifold learning) in unsupervised settings (Mahdavi et al., 2019). There is thus a growing need for unsupervised feature selection schemes that enhance latent signals of interest by removing nuisance variables and thus advance reliable data-driven scientific discovery. Unsupervised Feature Selection (UFS) methods are designed to identify a set of informative features that can improve the outcome of downstream analysis tasks such as clustering and manifold learning. With the lack of labels, however, selecting features becomes highly challenging since the downstream task cannot be used to drive the selection of features. As an alternative, most UFS methods use a label-free criterion that correlates with the downstream task. For instance, many UFS schemes rely on a reconstruction prior (Li et al., 2017) and seek a subset of features that can be used to reconstruct the entire dataset as accurately as possible. Several works use Autoencoders (AE) to learn a reduced representation of the data while applying a sparsification penalty to force the AE to remove redundant features. This idea was implemented with several types of sparsity-inducing regularizers, including ℓ2,1 based (Chandra and Sharma, 2015; Han et al., 2018), relaxed ℓ0 (Balın et al., 2019; Shaham et al., 2022; Svirsky and Lindenbaum) and more. One of the most commonly used criteria for UFS is feature smoothness. A common hypothesis is that the structure of interest, such as clusters or a manifold, can be captured using a graph representation(Ng et al., 2001). The smoothness of features is measured using the Laplacian Score (LS) (He et al., 2005), which is Published in Transactions on Machine Learning Research (12/2024) Figure 1: Illustration of SSFS. Panel (a) shows a t SNE scatter plot of noisy MNIST digits (3, 6, 8). Panel (b) presents the six leading eigenvectors computed based on the graph Laplacian of the data. Samples are ordered according to the identity of the digit. (c) We apply the k-medoids algorithm to compute pseudolabels y i . These are presented as colors overlayed on the eigenvectors. (d) We select the three eigenvectors whose pseudo-labels are the most stable with respect to several prediction models (see Section 3.2). (e) For each data feature we estimate its importance score for each of the selected eigenvectors (see Section 3.3). (f) We aggregate the feature scores across eigenvectors. equal to the Rayleigh quotient of a normalized feature vector and the graph Laplacian. A feature that is smooth with respect to the graph is considered to be associated with the primary underlying data structures. There are many other UFS methods that use a graph to select informative features Li et al. (2018); Roffo et al. (2017); Zhu et al. (2017; 2020); Xie et al. (2023). (Li et al., 2012) derived Nonnegative Discriminative Feature Selection (NDFS), which performs feature selection and spectral clustering simultaneously. Its extension Li and Tang (2015) adds a loss term to prevent the joint selection of correlated features. Embedded unsupervised feature selection schemes aim to cluster the data while simultaneously removing irrelevant features. Examples include Wang et al. (2015), which performs the selection directly on the clustering matrix, and Zhu and Yang (2018), which learns feature weights while maximizing the distance between clusters. In recent years, several works have derived self-supervised learning methods for feature selection. The key idea is to design a supervised type learning task with pseudo-labels that do not require human annotation. A seminal work based on this paradigm is Multi-Cluster Feature Selection (MCFS) (Cai et al., 2010). MCFS uses the eigenvectors of the graph Laplacian as pseudo-labels and learns the informative features by optimizing over an ℓ1 regularized least squares problem. More recently, Lee et al. (2021) used self-supervision with correlated random gates to enhance the performance of feature selection. In this work, we present a spectral self-supervised scheme for feature selection. Our main innovation is to properly select a subset of Laplacian eigenvectors by a multi-stage approach. Firstly, we generate robust discrete pseudo-labels from the eigenvectors and filter them based on a stability measure. Next, we fit flexible surrogate classification models on the selected eigenvectors and query the models for feature scores. Using these components, we can identify informative features that are effective for clustering on real-world datasets. 2 Preliminaries 2.1 Laplacian score and representation-based feature selection Generating a graph-based representation for a group of high-dimensional observations has become a common practice for unsupervised learning tasks. In manifold learning, methods such as ISOMAPS (Tenenbaum et al., 2000), LLE (Roweis and Saul, 2000), Laplacian eigenmaps (Belkin and Niyogi, 2003), and diffusion maps (Coifman and Lafon, 2006) compute a low-dimensional representation that is associated with the manifold s Published in Transactions on Machine Learning Research (12/2024) latent structure. In spectral clustering, a set of points is partitioned by applying the k-means algorithm to the leading Laplacian eigenvectors (Ng et al., 2001). In graph methods, each node vi corresponds to one of the observations xi Rp. The weight Wij between two nodes vi, vj is computed based on some kernel function K(xi, xj). For example, the popular Gaussian kernel is equal to, K(xi, xj) = exp xi xj 2 Where the parameter σ determines the bandwidth of the kernel function. Let D be a diagonal matrix with the degree of each node in the diagonal, such that Dii = P j Wij. The unnormalized graph Laplacian matrix is equal to L = D W . For any vector v Rn we have the following equality (Von Luxburg, 2007), vi vj 2Wij. (1) The quadratic form in equation 1 gives rise to a notion of graph smoothness. (Ricaud et al., 2019; Shuman et al., 2013). A vector is smooth with respect to a graph if it has similar values on pairs of nodes connected with an edge with a significant weight. This notion underlies the Laplacian score suggested as a measure for unsupervised feature selection (He et al., 2005). Let fm Rn denote the values of the m-th feature for all observations. The Laplacian score sm is equal to, sm = f T m Lfm = 1 fm,i fm,j 2Wij. (2) A low score indicates that a feature is smooth with respect to the computed graph and thus strongly associated with the latent structure of the high-dimensional data x1, . . . , xn. The notion of the Laplacian score has been the basis of several other feature selection methods as well (Lindenbaum et al., 2021; Shaham et al., 2022; Zhu et al., 2012). Let vi, λi denote the i-th smallest eigenvector and eigenvalue of the Laplacian L. A slightly different interpretation of equation 2 is that the score for each feature is equal to a weighted sum of its correlation with the eigenvectors, such that i=1 λi(f T mvi)2. A potential drawback of the Laplacian score is its dependence on all eigenvectors, which may reduce its stability in measuring a feature s importance to the data s main structures. To overcome this limitation, Zhao and Liu (2007) derived an alternative score based only on a feature s correlation to the leading Laplacian eigenvectors. A related, more sophisticated approach is Multi-Cluster Feature Selection (MCFS) (Cai et al., 2010), which computes the solutions to the generalized eigenvector problem Lv = λDv. The leading eigenvectors are then used as pseudo-labels for a regression task with l1 regularization. Specifically, MCFS applies Least Angle Regression (LARS) (Efron et al., 2004) to obtain, for each leading eigenvector vi, a sparse vector of coefficients βi Rp. The feature score is set as the maximum over the absolute values of the coefficients computed from the leading eigenvectors, sj = maxi |βi j|. The output of MCFS is the set of features with the highest score. In the next section, we derive Spectral Self-supervised Feature Selection (SSFS), which improves upon the MCFS algorithm in several critical aspects. 3 Spectral Self-supervised Feature Selection 3.1 Rationale As its title suggests, MCFS aims to uncover features that separate clusters in the data. Let us consider an ideal case where the observations are partitioned into k well-separated clusters, denoted A1, . . . , Ak, such that the weight matrix Wij = 0 if xi, xj are in separate clusters. Let ei denote an indicator vector for cluster Published in Transactions on Machine Learning Research (12/2024) (a) Prostate-GE eigenvectors. (b) TOX-171 eigenvectors. Figure 2: The first four Laplacian eigenvectors of two real datasets. Samples are sorted according to the class label and colored by the outcome of a one-dimensional k-medoids per eigenvector. The vertical bar indicates the separation between the classes. In Prostate-GE, v4 is the most informative to the class labels, and an outlier can be seen on the upper left in the third and fourth eigenvectors. In TOX-171, v3 is most informative to the class labels than v2. i such that |Ai| j Ai 0 otherwise, where |Ai| denotes the size of cluster Ai. In this scenario, the zero eigenvalue of the graph Laplacian has multiplicity k, and the corresponding eigenvectors are equal, up to a rotation matrix, to a matrix E Rn d whose columns are equal to e1, . . . , ek. In this ideal setting, the k leading eigenvectors are indeed suitable for use as pseudo-labels for the feature selection task. Thus, the MCFS algorithm should provide highly informative features in terms of cluster separation. In most applications, however, cluster separation is far from perfect, and the use of leading eigenvectors may be suboptimal. Here are some common scenarios: 1) High-dimensional datasets may contain substructures in top eigenvectors, while the main structure of interest appears deeper in the spectrum. For illustration, consider the MNIST dataset visualized via t-SNE in Figure 1a. The data contains images of 3, 6 and 8. Figure 1b shows the elements of the six leading eigenvectors of the graph Laplacian matrix, sorted by their corresponding digits. The leading eigenvector shows a clear gap between images of digit 6 and the rest of the data. However, there is no clear separation between digits 3 and 8. Indeed, the next eigenvector is not associated with such a separation. Applying feature selection with this eigenvector may produce spurious features irrelevant to separating the two digits. This scenario is prevalent in the real datasets used in the experimental section. For example, Figure 2a shows four eigenvectors of a graph computed from observations containing the genetic expression data from prostate cancer patients and controls (Singh et al., 2002). The leading two eigenvectors, however, are not associated with the patient-control separation. 2) The leading eigenvectors may be affected by outliers. For example, an eigenvector may indicate a small group of outliers separated from the rest of the data. This phenomenon can also be seen in the third and fourth vectors of the Prostate-GE example in Figure 2a. While the fourth eigenvector separates the categories, it is corrupted by outliers and, hence, unsuitable for use as pseudo-labels in a classical regression task, as it might highlight features associated with the outliers. 3) The relation between important features and the separation of clusters may be highly non-linear. In such cases, applying linear regression models to obtain feature scores may be too restrictive. Motivated by the above scenarios, we derive Spectral Self-supervised Feature Selection (SSFS). We explain our approach in detail in the following two sections. Published in Transactions on Machine Learning Research (12/2024) 3.2 Eigenvector processing and selection Generating binary labels. Given the Laplacian eigenvectors V = (v1, ..., vd), our goal is to generate pseudo-labels that are highly informative to the cluster separation in the data. To that end, for each eigenvector vi, we compute a binary label vector y i (pseudo-labels) by applying a one-dimensional k-medoids algorithm (Kaufman and Rousseeuw, 1990) to the elements of vi. In contrast to k-means, in k-medoids, the cluster centers are set to one of the input points, which makes the algorithm robust to outliers. In Figure 2, the eigenvectors are colored according to the output of the k-medoids. After binarization, the fourth eigenvector of the Prostate-GE dataset is highly indicative of the category. The feature selection is thus based on a classification rather than a regression task, which is more aligned with selecting features for clustering. In Section 5.2 we show the impact of the binarization step on multiple real-world datasets. Eigenvector selection. Selecting k eigenvectors according to their eigenvalues may be unstable in cases where the eigenvalues exhibit a small spectral gap. We derive a robust criterion for selecting informative eigenvectors that is based on the stability of a model learned for each vector. Formally, we consider a surrogate model h : Rp R, and a feature score function s(h) Rp, where p denotes the number of features. The feature scores are non-negative and their sum is normalized to one. For example, h can be the logistic regression model h(x) = σ(βT x). In that case, a natural score function is the absolute value of the coefficient vector β. For each eigenvector vi, we train a model hi on B (non-mutually exclusive) subsets of the input data X and the pseudo-labels y i . We then estimate the variance of the feature score function, for every feature m {1, ..., p}: d Var(sm(hi)) = 1 B 1 b=1 (sm(hi,b) sm(hi))2. This procedure is similar (though not identical) to the Delete-d Jackknife method for variance estimation (Shao and Wu, 1989). We keep, as pseudo-labels, the k binarized eigenvectors with the lowest sum of variance, ˆSi = Pp m=1 d Var(sm(hi)). We denote the set of selected eigenvectors by I. A pseudo-code for the pseudo-labels generation and eigenvector selection appears in Algorithm 1. 3.3 Feature selection For the feature selection step, we train k models, denoted {fi | i I}, to predict the selected binary pseudolabels based on the original data. Similarly to the eigenvector selection step, each model is associated with a feature score function s(fi). The features are then scored according to the following maximum criterion, score(m) = max i I sm(fi). Finally, the features are ranked by their scores, and the top-ranked features are selected for the subsequent analysis. The choice of model for this step can differ from that used in the eigenvector selection step, allowing for flexibility in the modeling approach (see Section 3.4 for details). Pseudo-code for SSFS appears in Algorithm 2. 3.4 Choice of Surrogate Models Our algorithm is compatible with any supervised model capable of providing feature importance scores. We combine the structural information from the graph Laplacian with the capabilities of various supervised models for unsupervised feature selection. Empirical evidence supports the use of more complex models such as Gradient-Boosted Decision Trees for various complex, real-world datasets (Mc Elfresh et al., 2023; Chen and Guestrin, 2016). These models are capable of capturing complex nonlinear relationships, which we leverage by training them on pseudo-labels derived from the Laplacian s eigenvectors. For example, for eigenvector selection, one can use a simple logistic regression model for fast training on the resampling procedure and a more complex gradient boosting model such as XGBoost (Chen and Guestrin, 2016) for the feature selection step. Published in Transactions on Machine Learning Research (12/2024) Algorithm 1 Pseudo-code for Eigenvector Selection and Pseudo-labels Generation Input: Dataset X Rn p (with n samples and p features), number of eigenvectors to select k, number of eigenvectors to compute d, surrogate models H = {hi | i [d]}, feature scoring function s : F Rp, number of resamples B 1: Initialize an empty list for the pseudo-labels Y and an empty list for the sums of features variance ˆS 2: Compute the significant d eigenvectors of the Laplacian of X: V = (v1, ..., vd) 3: for i = 1 to d do 4: Binarize the eigenvector vi using k-medoids to obtain y i , and append to Y 5: for b = 1 to B do 6: Subsample ((X)b, (y i )b) from (X, y i ) 7: Fit the model hi,b to ((X)b, (y i )b) 8: Compute the feature scores sm(hi,b) 9: end for 10: for m = 1 to p do 11: Estimate the variance of the m-th feature score: d Var(sm(hi)) = 1 B 1 b=1 (sm(hi,b) sm(hi))2 12: end for 13: ˆSi = Pp m=1 d Var(sm(hi)) 14: ˆS ˆS { ˆSi } 15: end for 16: Select the indices of the k smallest elements in ˆS and store in I 17: return Y , I Algorithm 2 Pseudo-code for Spectral Self-supervised Feature Selection (SSFS) Input: Dataset X Rn p (with n samples and p features) number of eigenvectors to select k, number of eigenvectors to compute d, surrogate eigenvector selection models H = {hi | i [d]}, surrogate feature selection models F = {fi | i [d]}, feature scoring function s : F Rp, number of resamples B, number of features to select ℓ. 1: Apply Algorithm 1 to obtain the pseudo-labels and the selected eigenvectors: Y , I = Eigenvector Selection(X, k, d, H, s, B) 2: for i in I do 3: Fit the model fi on (X, y i ) 4: Calculate the feature scores s(fi) 5: Normalize the feature scores such that their sum is one 6: end for 7: for m = 1 to p do 8: Compute the final score for the m-th feature: score(m) = max i I sm(fi) 10: return a list of ℓfeatures with the highest score. 4 The importance of a proper selection of eigenvectors: analysis of the product manifold model As described in Section 3.1, the principle of selecting the leading Laplacian eigenvectors as pseudo-labels is motivated by the case of highly separable clusters, where observations in different clusters have very low Published in Transactions on Machine Learning Research (12/2024) connectivity between them. In many cases, the separation between meaningful states (i.e., biological or medical conditions) may not be that clear. To illustrate this point, consider the MNIST example in Figure 1. While images with the digit 6 constitute a cluster, the separation between digits 8 and 3 are not clearly separated. Figure 3a shows a scatter plot of these digits, where each image is located according to its coordinates in the third and fourth eigenvectors. Even when considering the most relevant eigenvectors, there is no clear separation between the digits. Instead, the transition between 3 and 8 is smooth and depends on the properties of the digits. To provide insight into the importance of eigenvector selection, we analyze a product of manifold model. Our analysis is based on results from two research topics: (i) the convergence, under the manifold assumption, of the Laplacian eigenvectors to the eigenfunctions of the Laplace Beltrami operator associated with the manifold, and (ii) the properties of manifold products. We next provide a brief background on these two topics. 4.1 Convergence of the Laplacian eigenvectors In many applications, high dimensional observations are assumed to reside close to some manifold M with low intrinsic dimensionality, which we denote by d. Many papers in recent decades have analyzed the relation between the Laplacian eigenvectors and the manifold structure Von Luxburg et al. (2008); Singer and Wu (2017); García Trillos et al. (2020); Wormell and Reich (2021); Dunson et al. (2021); Calder and Trillos (2022). More formally, let vk denote the k-th eigenvector of the graph Laplacian, and let gk denote the k-th eigenfunction of the Laplace-Beltrami (LB) operator. We usually assume that gk is normalized such that Z M gk(x)2µ(x)d V (x) = 1, where µ is the distribution function over M. Under several assumptions and proper normalization of gk, we have vk n gk(X). where gk(X) is a vector of size n containing samples of the function gk at the n rows of the data matrix X. Let us consider a simple example. Suppose we have n points sampled uniformly at random over the interval [0, 1]. The LB operator over this interval is defined as the second derivative, whose eigenfunctions are the harmonic functions given by gk(x) = cos(kπx). Figure 3 shows the three Laplacian eigenvectors computed with n = 102, 103 and 3 103 points. As n , the vector vk converges to gk(X). Here, we use a convergence result from Cheng and Wu (2022), derived under the following assumptions: (i) The n observations are generated according to a uniform distribution over the manifold, such that µ(x) equals to a constant µ. (ii) Let λk denote the eigenvalue associated with the eigenfunction gk. To ensure the stability of the eigenvectors, we assume a spectral gap between the smallest K eigenvalues bounded away from 0 such that, K 1 min i=1 (λi+1 λi) > γ > 0. (iii) The graph weights are computed by a Gaussian kernel exp( xi xj 2/ϵn), with a bandwidth ϵn n 0+ that satisfies ϵd/2+2 n > Ck log n n for a constant CK. The following result from Cheng and Wu (2022) guarantees the convergence of the Laplacian eigenvectors to samples of the LB eigenfunction. Theorem 1 (Theorem 5.4 of Cheng and Wu (2022)) For n and under assumptions (i)-(iii), with probability larger than 1 4K2n 10 (2K+6)n 9, the k-th eigenvector vk of the unnormalized Laplacian satisfies vk αgk(X) 2 = O(ϵn) + O s where vk = 1 and |α| = o(1). We next address the eigenvector structure in case of a product-manifold model. Published in Transactions on Machine Learning Research (12/2024) 4.2 The product of manifold model In a product of two manifolds, denoted M = M1 M2, every point x M is associated with a pair of points x1, x2 where x1 M1 and x2 M2. We denote by π1(x), π2(x) the canonical projections of a point in M to its corresponding points x1, x2 in M1, M2, respectively. For example, a 2D rectangle is a product of two 1D manifolds, where π1(x) and π2(x) select, respectively, the first and second coordinates. We denote by g(1) i (x), g(2) i (x) the i-th eigenfunction of the LB operator of M1, M2, respectively, evaluated at a point x, and by λ(1) i , λ(2) i the corresponding eigenvalues. In a manifold product M1 M2, the eigenfunctions are equal to the pointwise product of the eigenfunctions of the LB operator of M1, M2, and the corresponding eigenvalues are equal to the sum of eigenvalues, such that gl,k(x) = g(1) l (π1(x)) g(2) k (π2(x)) λl,k = λ(1) l + λ(2) k . (4) For simplicity, we denote by vl,k the (l, k)-th eigenvector of the Laplacian matrix, as ordered by λl,k. An example of a product of 2 manifolds is illustrated in Figure 4b. The figure shows the leading eight eigenvectors of the graph Laplacian. The eigenvectors are indexed by the vector b = [l, k]. The full details of this example is provided in the next section. (a) Noisy MNIST: digits 3 and 8 0.0 0.2 0.4 0.6 0.8 1.0 X Leading eigenvector n = 100 n = 1000 n = 3000 cos( x) (b) Convergence of eigenvectors for graph on an interval Figure 3: Panel 3a shows a scatter plot of the noisy MNIST dataset, containing digits 3 and 8, where each image is located according to its coordinates in the third and fourth eigenvectors. Panel 3b shows the leading eigenvector of a graph computed over n points on a 1D interval and the leading eigenfunction cos(πx). 4.3 Considerations for eigenvector selection in a product-manifold model We analyze a setting where the p features can be partitioned into H sets according to their dependencies on a set of latent and independent random variables θ1, . . . , θH with some bounded support. A feature fm that depends on θh consists of samples from a smooth transformation θh Fm fm. We denote by X(h) the submatrix that contains the features associated with θh. The smoothness of the transformations implies that the rows of X(h) constitute random samples from a manifold of intrinsic dimension 1. Figure 4a shows a 3D scatter plot, where the axis are three such features with values generated by three polynomials of θ1. The figure is an illustration of a manifold with a single intrinsic dimension embedded in a 3D space. The independence of the latent variables θh implies that the observations xi Rp are samples from a product of H manifolds, each of dimensionality 1 Zhang et al. (2021); He et al. (2023). The canonical Published in Transactions on Machine Learning Research (12/2024) (a) Simulated data (b) Eigenvectors of the simulated data Figure 4: Panel (a) illustrates three features of a simulated dataset. Each feature is equal to a different polynomial of the same random latent variable θ1. Each point in the 3D scatter plot is located according to the values of the three features and colored by the value of θ1. Panel (b) shows the eigenvectors of the graph Laplacian matrix. Each point is located according to the value of (θ1, θ2) and colored by the value of its corresponding element in the eight leading eigenvectors. The eigenvectors are indexed by the vector b, whose elements bi determine the eigenvector order in the submanifold M(i). projection π(h)(x) selects the features associated with the latent variable θh. According to the eigenfunctions properties in equation 4, the eigenfunctions are equal to the product of H eigenfunctions of the submanifolds M(h), and can thus be indexed by a vector of size H, which we denote by b NH. h=1 g(h) bh π(h)(x) λb = h=1 λ(h) bh . Let e(h) denote an index vector with elements e(h) j = 1 if j = h and 0 otherwise. The first eigenfunctions g(h) 0 are equal to a constant for all submanifolds M(h). Thus, The eigenfunctions ge(h) are equal to ge(h)(x) = g(h) 1 π(h)(x) H Y j =h g(j) 0 π(j)(x) = Cg(h) 1 π(h)(x) , (5) where C is some constant. Importantly, the functions g(h) 1 and thus ge(h), depend only on the parameter θ(h). We define by E the family of vectors in NH that include the indicator vectors e(h) or their integer products (e.g. 2e(h), 3e(h) etc.). A similar derivation as in equation 5 shows that for every index vector b E, the eigenfunction gb depends on only one of the latent variable in θ1, . . . , θh. On the relevance of features for choosing eigenvectors as pseudo-labels. Our goal is to select a set of features that contains at least one (or more) features from each of the H partitions. Such a choice would ensure that the set contains information about all the H latent variables. Clearly, this imposes a requirement on the set of pseudo-label vectors: we would like at least one vector of pseudo-labels that is correlated with each latent variable. It is instructive to consider the asymptotic case where n and hence according to Theorem 1 and the properties of manifold products, the eigenvectors vb converge to gb(X). A proper choice of eigenvectors for pseudo-labels would be the set {ve(h)}H h=1, as each of these vectors converges to the samples g(h) 1 (X), and is thus associated with a different latent variable. However, there is no guarantee that these eigenvectors have the smallest eigenvalues. Published in Transactions on Machine Learning Research (12/2024) Consider for example the case for the data illustrated in Figure 4a. Panel (b) shows the leading eight eigenvectors of the graph Laplacian. The leading two eigenvectors are functions of θ1 and by choosing them we completely disregard θ2 with an obvious impact on the feature selection accuracy. A better choice for pseudo-labels would be the first and third eigenvectors, indexed by e1 and e2. Therefore, we need an improved criterion for selecting eigenvectors to serve as pseudolabels for the feature selection process. The following theorem, proven in Appendix A.1, implies that the feature vectors fi are relevant for developing such a criterion. Theorem 2 We assume that the samples are generated according to our specified latent variable model and that assumptions (i)-(iii) are satisfied. Let fi Rn be a normalized, zero mean feature vector associated with parameter θh. Then, f T i vb = O(ϵn) + O s The theorem is proved via the following two steps. The details of the proof are provided in the appendix. Step 1: We show that the inner product f T i gb(X) can be written as the inner product of two random vectors with independent elements. Thus, f T i gb(X) is of order O 1/ n by standard concentration inequalities. Step 2: Combine the convergence of vb to gb(X) with the concentration result of step 1. Theorem 2 implies that one can use the inner products to avoid selecting less informative eigenvectors that depend on more than one variable. Further guarantees, such as selection of a single vector from each variable, require additional assumptions on the feature values, which we do not make here. In Algorithm 1 we compute the normalized measure of stability for the feature scores {sm(hi)}p m=1 obtained by the model hi to predict the labels computed from the i-th eigenvector. When the model hi is linear (or generalized linear), the score is strongly related to the simple inner product of Theorem 2. In that case, Theorem 2 indicates that the inner product between uninformative eigenvectors and all features is close to zero. Thus, we expect the variance (after normalization) to be similar to the variance of random positive noise. The advantage of the stability measure over the simple linear product as an eigenvector selection criterion is that it allows for more flexibility in the choice of model. 4.4 Computational Complexity The computational complexity of our proposed method consists of several key steps. First, constructing the graph representation of the dataset using a Gaussian kernel requires computing pairwise distances between all data points, resulting in a complexity of O(n2p), where n is the number of samples and p is the number of features. Following this, the eigendecomposition of the graph has a complexity of O(n3). This can be reduced by randomized methods to O(n2d), since we only require the first d eigenvectors. Once eigenvectors are computed, we apply the k-medoids algorithm to cluster the eigenvectors and generate pseudo-labels, adding a further complexity of O(dkn2), where k is the number of clusters. After creating pseudo-labels, we use surrogate models such as logistic regression or gradient-boosted decision trees to evaluate the features. Logistic regression has a complexity of O(nd), while gradient-boosted trees, being more complex, scale with O(Tn log n), where T is the number of trees. We also use a resampling procedure to estimate the variance of feature importance scores. Each resampling involves training a surrogate model, resulting in an overall complexity of O(BTn log n) for gradient-boosted trees or O(Bnd) for logistic regression. As a result, the overall computational complexity of SSFS is mainly influenced by the graph construction and training the surrogate mode, resulting in a complexity of O(n2d + Bnd). Published in Transactions on Machine Learning Research (12/2024) 5 Experiments 5.1 Evaluation on real world datasets Data and experiment description. We applied SSFS to eight real-world datasets from various domains. Table 5 in Appendix F.2 gives the number of features, samples, and classes in each dataset. All datasets are available online 1. We compare the performance of our approach to the following alternatives: (i) standard Laplacian score (LS) (He et al., 2005), (ii) Multi-Cluster Feature Selection (MCFS) (Cai et al., 2010), (iii) Nonnegative Discriminative Feature Selection (NDFS), (Li et al., 2012), (iv) Unsupervised Discriminative Feature Selection (UDFS) (Yang et al., 2011), (v) Laplacian Score-regularized Concrete Autoencoder (LS-CAE) (Shaham et al., 2022), (vi) Unsupervised Feature Selection Based on Iterative Similarity Graph Factorization and Clustering by Modularity (KNMFS) (Oliveira et al., 2022) and (vii) a naive baseline, where random selection is applied with a different seed for each number of selected features. For evaluation, we adopt a criterion that is similar to, but not identical to, the one used in prior studies (Li et al., 2012; Wang et al., 2015). We select the top 2, 5, 10, 20, 30, 40, 50, 100, 150, 200, 250, and 300 features as scored by each method. Then, we apply k-means 20 times on the selected features and report the average clustering accuracy (along with the standard deviation), computed by (Cai et al., 2011): ACC = max π 1 N i=1 δ(π(ci), li), where ci and li are the assigned cluster and true label of the i-th data point, respectively, δ(x, y) is the delta function which equals one if x = y and zero otherwise, and π represents a permutation of the cluster labels, optimized via the Kuhn-Munkres algorithm (Munkres, 1957). Unlike the evaluation approach taken by Wang et al. (2015); Li et al. (2012), which entailed a grid search over hyper-parameters to report the optimum results for each method, our analysis employed the default hyperparameters as specified by the respective implementations, including SSFS. This approach aims for a fair comparison to avoid favoring methods that are more sensitive to hyper-parameter adjustments. In addition, it acknowledges the practical constraints in unsupervised settings where hyper-parameter tuning is typically infeasible. Such differences in the approach to hyper-parameter selection could account for discrepancies between the results reported in previous studies and those in our study. See Appendix F.1 for additional details. Table 1: Average clustering accuracy on benchmark datasets along with the standard deviation. The number of selected features yielding the best clustering performance is shown in parentheses, the best result for each dataset highlighted in bold. Dataset Random LS MCFS NDFS UDFS KNMFS LS-CAE SSFS COIL20 65.1 2.1(250) 61.9 2.4(300) 67.4 3.3(300) 63.4 2.6(200) 61.9 3.5(300) 68.1 2.0(300) 64.2 3.1(30) 67.1 2.8(300) GISETTE 70.2 0.1(150) 70.0 0.0(250) 70.7 0.0(5) 58.3 1.9(100) 69.1 0.1(50) 54.9 0.0(40) 70.8 0.0(200) 69.7 0.0(150) Yale 47.8 3.5(250) 43.9 3.2(300) 44.4 2.9(300) 43.5 2.5(250) 43.8 2.3(50) 47.2 4.3(300) 46.2 1.6(10) 50.3 2.3(100) TOX-171 44.2 1.8(250) 51.3 1.0(5) 44.5 0.5(5) 47.3 0.1(150) 40.2 3.8(250) 48.1 3.5(20) 50.1 5.3(200) 59.4 2.5(100) ALLAML 73.2 1.7(300) 72.2 0.0(200) 75.0 0.0(150) 76.6 0.7(2) 66.4 1.3(50) 59.9 9.2(150) 63.9 0.0(2) 75.4 3.2(100) Prostate-GE 63.0 0.7(30) 58.8 0.0(2) 61.8 0.0(100) 58.8 0.0(2) 63.6 0.3(50) 62.7 0.0(50) 63.7 0.0(40) 75.9 0.5(10) ORL 58.9 1.8(300) 51.6 1.7(300) 57.0 2.8(300) 59.1 2.5(300) 57.3 2.4(300) 63.2 2.0(150) 61.0 2.0(300) 61.0 2.2(200) ISOLET 59.5 1.8(300) 48.9 2.0(300) 50.7 1.5(300) 63.1 2.4(200) 44.6 1.7(300) 52.7 2.3(300) 63.0 2.6(300) 59.9 1.4(100) Mean rank 4.12 5.88 4.62 4.94 6.44 4.38 3.31 2.31 Median rank 4.0 6.5 5.5 5.5 6.5 4.5 2.75 2.25 Table 1 shows, for each method, the highest average accuracy and the number of features for which it was achieved similarly to (Li et al., 2012; Wang et al., 2015). Figure 5 presents a comparative analysis of clustering accuracy across various datasets and methods, considering the full spectrum of selected features. This comparison aims to account for the inherent variance in each method, addressing a limitation where the criterion of the maximum accuracy over the number of selected features might inadvertently favor methods exhibiting higher variance. 1https://jundongl.github.io/scikit-feature/datasets.html Published in Transactions on Machine Learning Research (12/2024) Figure 5: Clustering accuracy vs. the number of selected features on eight real-world datasets. For SSFS, we use the following surrogate models: (i) The eigenvector selection model hi is set to Logistic Regression with ℓ2 regularization. We use scikit-learn s (Pedregosa et al., 2011) implementation with a default regularization value of C = 1.0. Feature scores are equal to the absolute value of the model s coefficients. (ii) The feature selection model fi is set to XGBoost classifier with Gain feature importance. We use the popular implementation by DMLC (Chen and Guestrin, 2016). Note that we employ the default hyper-parameters for all surrogate models as provided in their widely used implementations. However, it s worth noting that one can undoubtedly leverage domain knowledge to select surrogate models and hyperparameters better suited to the specific domain. For each dataset, SSFS selects k from d = 2k eigenvectors, where k is the number of distinct classes in the data. Results. SSFS has been ranked as the best method in three out of eight datasets. It has shown a significant advantage over competing methods, especially in the Yale, TOX-171, and Prostate-GE datasets. As discussed in Section 3.1, the Prostate-GE dataset has several outliers, and the fourth eigenvector plays a vital role in providing information about the class labels compared to the earlier eigenvectors. SSFS can effectively deal with such challenging scenarios, and this might explain its superior performance. Although our method is not ranked first in the other five datasets, it has produced results comparable to the leading method. 5.2 Ablation study We demonstrate the importance of three SSFS components: (i) eigenvector selection, (ii) self-supervision with nonlinear surrogate models, and (iii) binarization of the Laplacian eigenvectors along with classifiers instead of regressors as surrogate models. The ablation study is performed on a synthetic dataset described in Section 5.2.1, and the eight real datasets used for evaluation in Section 5.1. 5.2.1 Synthetic data We generate a synthetic dataset as follows: the first five features are generated from two isotropic Gaussian blobs; these blobs define the clusters of interest. Additional 45 nuisance features are generated according to a multivariate Gaussian distribution, with zero mean and a block-structured covariance matrix Σ, such that each block contains 15 features. The covariance elements Σi,j are equal to 0.5 if i, j are in the same block and 0.01 otherwise. We generated a total of 500 samples; see Appendix E.1 for further details. In Figure 6a, you can see a scatter plot of the first five features, and in Figure 6b, you can see a visualization of the covariance matrix. Our goal is to identify the features that can distinguish between the two groups. Published in Transactions on Machine Learning Research (12/2024) (a) First 5 features (b) Covariance matrix (c) Eigenvectors Figure 6: Visualizations of the synthetic data: Panel (a): scatter plot of the first five features corresponding to the Gaussian blobs, colored by the real label. Panel (b): the covariance matrix of the dataset. Panel (c): the top-4 eigenvectors, samples are sorted by the label and are partitioned by the vertical bar, colored according to the output of k-medoids. Table 2: Ablation study: average clustering accuracy on benchmark datasets, the number of selected features is shown in parenthesis for the best clustering accuracy over the feature range. Dataset no selection no XGBoost no selection, regression regression SSFS COIL20 65.0 (150) 62.1 (150) 70.5 (100) 69.0 (300) 67.1 (300) GISETTE 72.5 (10) 64.9 (300) 64.6 (5) 64.6 (5) 69.7 (150) Yale 48.6 (50) 42.7 (250) 49.8 (200) 47.4 (250) 50.3 (100) TOX-171 50.9 (2) 45.6 (20) 45.0 (5) 45.5 (50) 59.4 (100) ALLAML 75.4 (100) 66.7 (50) 71.1 (300) 71.1 (300) 75.4 (100) Prostate-GE 59.8 (30) 69.6 (30) 61.8 (150) 61.8 (150) 75.9 (10) ORL 60.0 (300) 56.8 (300) 58.5 (300) 58.5 (200) 61.1 (200) ISOLET 57.0 (150) 57.1 (300) 61.3 (300) 58.7 (300) 59.9 (100) Mean rank 2.94 4.0 3.0 3.5 1.56 Median rank 2.5 4.5 3.5 3.5 1.25 As Figure 6a demonstrates, the two clusters are linearly separated by three distinct features. Furthermore, examining Figure 6c reveals that while the fourth eigenvector distinctly separates the clusters, the higherranked eigenvectors do not exhibit this behavior. This pattern arises due to the correlated noise, significantly influencing the graph structure. The evaluation of this dataset is performed by calculating the true positive rate (TPR) with respect to the top-selected features and the discriminative features sampled from the two Gaussian blobs. The performance on the real-world datasets is measured similarly to Section 5.1. Table 3: Synthetic data results: Top-3 selected features (sorted in descending order by rank), along with their TPR (relative to the first five features). Method Top-3 Features TPR SSFS 2, 9, 19 0.3 (no XGBoost) 4, 3, 2 1.0 (no selection) 43, 30, 49 0.0 (regression) 15, 17, 14 0.0 MCFS 47, 7, 43 0.0 5.2.2 Results Eigenvector Selection. We compare to a variation of SSFS termed SSFS (no selection), where we don t filter the eigenvectors. We train the surrogate feature selector model on the leading k eigenvectors, with k set to the number of distinct classes in the data. Figure 7b shows that our eigenvector selection scheme provides an advantage in seven out of eight datasets. Similarly to Sec. 5.1, filtering the eigenvectors is especially advantageous on the Prostate-GE dataset, as our method successfully selects the most discriminative eigenvectors (see Figure 2a ). On the synthetic dataset, the selection procedure provides a large advantage, Published in Transactions on Machine Learning Research (12/2024) as seen in Table 3. Figure 6c illustrates that the fourth eigenvector is the informative one with respect to the Gaussian blobs. Indeed, the fourth eigenvector and the third eigenvector are selected by the selection procedure. This eigenvector yields better features than MCFS and SSFS (no selection), which rely on the top two eigenvectors. Classification and regression. We compare the following regression variants of SSFS , which use the original continuous eigenvectors as pseudo-labels (without binarization): (i) SSFS (regression): uses ridge regression for eigenvector selection and XGBoost regression for the feature selection as surrogate models. (ii) SSFS (no selection, regression): uses the top k eigenvectors without binarization and XGBoost regression. Figure 7a and Table 2 show that SSFS performs best on six of the eight real-world datasets. Interestingly, when using continuous regression as a surrogate model, the selection procedure does not seem to provide an advantage compared to no selection. (a) Classification and regression (b) Selection, and no XGBoost (logistic regression) Figure 7: Ablation study results on the real-world datasets. The best clustering accuracy over the number of selected features is shown for each method. Complex nonlinear models as surrogate models. We compare SSFS to a variant of our method denoted SSFS (no XGBoost), which employs a logistic regression instead of XGBoost as the surrogate feature selector model. Figure 7b shows that XGBoost provides an advantage compared to the linear model on real-world datasets. On the synthetic dataset, the linear variant provides better coverage for the top-3 features that separate the Gaussian blobs, compared to XGBoost (see Table 3 and Figure 6a). That is not surprising since, in this example, the cluster separation is linear in each informative feature. We note, however, that the top-ranked feature by SSFS with XGBoost is a discriminative feature for the clusters in the data (see Figure 6a); therefore, its selection can still be considered successful in the case of a single feature selection. 6 Discussion and future work We proposed a simple procedure for filtering eigenvectors of the graph Laplacian and demonstrated that its application could have a significant impact on the outcome of the feature selection process. The selection is based on the stability of a classification model in predicting binary pseudo-labels. However, additional criteria, such as the accuracy of a specific model or the overlap of the chosen features for different eigenvectors, may provide information on the suitability of a specific vector for a feature selection task. We also illustrated the utility of expressive models, typically used for supervised learning, in unsupervised feature selection. Another direction for further research is using self-supervised approaches for group feature selection (GFS) for single modality (Sristi et al., 2022) or multi-modal data (Yang et al., 2023; Yoffe et al., 2024). In contrast to standard feature selection where the output is sparse, GFS aims to uncover groups of features with joint effects on the data. Learning models based on different eigenvectors may provide information about group effects with potential applications such as detecting brain networks in Neuroscience and gene pathways in genetics. Published in Transactions on Machine Learning Research (12/2024) Broader Impact Statement This work involves methods applicable to high-dimensional data, including sensitive domains like healthcare. While our experiments used publicly available datasets, minimizing immediate privacy concerns, practitioners applying this method to sensitive data should ensure compliance with data protection regulations and adopt robust privacy-preserving measures. There is also a risk of misapplication in critical areas such as medical diagnosis or criminal justice, where biased or incorrect feature selection could lead to harmful decisions. To mitigate this, we recommend integrating bias detection and mitigation techniques into the training pipeline to enhance fairness and reliability. Careful consideration of these ethical and societal implications is essential for the responsible use of this method. Muhammed Fatih Balın, Abubakar Abid, and James Zou. Concrete autoencoders: Differentiable feature selection and reconstruction. In International conference on machine learning, pages 444 453. PMLR, 2019. Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. Neural computation, 15(6):1373 1396, 2003. Deng Cai, Chiyuan Zhang, and Xiaofei He. Unsupervised feature selection for multi-cluster data. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 333 342, 2010. Deng Cai, Xiaofei He, and Jiawei Han. Locally consistent concept factorization for document clustering. IEEE Transactions on Knowledge and Data Engineering, 23(6):902 913, 2011. Jeff Calder and Nicolas Garcia Trillos. Improved spectral convergence rates for graph laplacians on ε-graphs and k-nn graphs. Applied and Computational Harmonic Analysis, 60:123 175, 2022. B Chandra and Rajesh K Sharma. Exploring autoencoders for unsupervised feature selection. In 2015 International Joint Conference on Neural Networks (IJCNN), pages 1 6. IEEE, 2015. Tianqi Chen and Carlos Guestrin. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pages 785 794, 2016. Xiuyuan Cheng and Nan Wu. Eigen-convergence of gaussian kernelized graph laplacian by manifold heat interpolation. Applied and Computational Harmonic Analysis, 61:132 190, 2022. Ronald R Coifman and Stéphane Lafon. Diffusion maps. Applied and computational harmonic analysis, 21 (1):5 30, 2006. David B Dunson, Hau-Tieng Wu, and Nan Wu. Spectral convergence of graph Laplacian and heat kernel reconstruction in l from random samples. Applied and Computational Harmonic Analysis, 55:282 336, 2021. Bradley Efron, Trevor Hastie, Iain Johnstone, and Robert Tibshirani. Least angle regression. The Annals of Statistics, 32(2):407 499, 2004. Nicolás García Trillos, Moritz Gerlach, Matthias Hein, and Dejan Slepčev. Error estimates for spectral convergence of the graph laplacian on random geometric graphs toward the Laplace-Beltrami operator. Foundations of Computational Mathematics, 20(4):827 887, 2020. Kai Han, Yunhe Wang, Chao Zhang, Chao Li, and Chao Xu. Autoencoder inspired unsupervised feature selection. In 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 2941 2945. IEEE, 2018. Published in Transactions on Machine Learning Research (12/2024) Jesse He, Tristan Brugère, and Gal Mishne. Product manifold learning with independent coordinate selection. In Topological, Algebraic and Geometric Learning Workshops 2023, pages 267 277. PMLR, 2023. Xiaofei He, Deng Cai, and Partha Niyogi. Laplacian score for feature selection. In Proceedings of the 18th International Conference on Neural Information Processing Systems, NIPS 05, page 507 514, Cambridge, MA, USA, 2005. MIT Press. Leonard Kaufman and Peter J. Rousseeuw. Finding Groups in Data: An Introduction to Cluster Analysis. John Wiley, 1990. ISBN 978-0-47031680-1. Changhee Lee, Fergus Imrie, and Mihaela van der Schaar. Self-supervision enhanced feature selection with correlated gates. In International Conference on Learning Representations, 2021. Jundong Li, Jiliang Tang, and Huan Liu. Reconstruction-based unsupervised feature selection: An embedded approach. In IJCAI, pages 2159 2165, 2017. Xuelong Li, Han Zhang, Rui Zhang, Yun Liu, and Feiping Nie. Generalized uncorrelated regression with adaptive graph for unsupervised feature selection. IEEE transactions on neural networks and learning systems, 30(5):1587 1595, 2018. Zechao Li and Jinhui Tang. Unsupervised feature selection via nonnegative spectral analysis and redundancy control. IEEE Transactions on Image Processing, 24(12):5343 5355, 2015. Zechao Li, Yi Yang, Jing Liu, Xiaofang Zhou, and Hanqing Lu. Unsupervised feature selection using nonnegative spectral analysis. In Proceedings of the AAAI conference on artificial intelligence, volume 26, pages 1026 1032, 2012. Ofir Lindenbaum, Uri Shaham, Erez Peterfreund, Jonathan Svirsky, Nicolas Casey, and Yuval Kluger. Differentiable unsupervised feature selection based on a gated laplacian. Advances in Neural Information Processing Systems, 34:1530 1542, 2021. Kaveh Mahdavi, Jesus Labarta, and Judit Gimenez. Unsupervised feature selection for noisy data. In Advanced Data Mining and Applications: 15th International Conference, ADMA 2019, Dalian, China, November 21 23, 2019, Proceedings 15, pages 79 94. Springer, 2019. Duncan Mc Elfresh, Sujay Khandagale, Jonathan Valverde, Ganesh Ramakrishnan, Micah Goldblum, Colin White, et al. When do neural nets outperform boosted trees on tabular data? ar Xiv preprint ar Xiv:2305.02997, 2023. Marina Meilă. Comparing clusterings by the variation of information. In Learning Theory and Kernel Machines: 16th Annual Conference on Learning Theory and 7th Kernel Workshop, COLT/Kernel 2003, Washington, DC, USA, August 24-27, 2003. Proceedings, pages 173 187. Springer, 2003. James R. Munkres. Algorithms for the assignment and transportation problems. Journal of The Society for Industrial and Applied Mathematics, 10:196 210, 1957. URL https://api.semanticscholar.org/ Corpus ID:15996572. Andrew Ng, Michael Jordan, and Yair Weiss. On spectral clustering: Analysis and an algorithm. Advances in neural information processing systems, 14, 2001. Marcos de S Oliveira, Sergio R de M Queiroz, and Francisco de AT de Carvalho. Unsupervised feature selection method based on iterative similarity graph factorization and clustering by modularity. Expert Systems with Applications, 208:118092, 2022. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825 2830, 2011. Published in Transactions on Machine Learning Research (12/2024) Benjamin Ricaud, Pierre Borgnat, Nicolas Tremblay, Paulo Gonçalves, and Pierre Vandergheynst. Fourier could be a data scientist: From graph fourier transform to signal processing on graphs. Comptes Rendus Physique, 20(5):474 488, 2019. Giorgio Roffo, Simone Melzi, Umberto Castellani, and Alessandro Vinciarelli. Infinite latent feature selection: A probabilistic latent graph-based ranking approach. In Proceedings of the IEEE international conference on computer vision, pages 1398 1406, 2017. Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linear embedding. science, 290(5500):2323 2326, 2000. Uri Shaham, Ofir Lindenbaum, Jonathan Svirsky, and Yuval Kluger. Deep unsupervised feature selection by discarding nuisance and correlated features. Neural Networks, 152:34 43, 2022. Jun Shao and CF Jeff Wu. A general theory for jackknife variance estimation. The annals of Statistics, pages 1176 1197, 1989. David I Shuman, Sunil K Narang, Pascal Frossard, Antonio Ortega, and Pierre Vandergheynst. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE signal processing magazine, 30(3):83 98, 2013. Amit Singer and Hau-Tieng Wu. Spectral convergence of the connection laplacian from random samples. Information and Inference: A Journal of the IMA, 6(1):58 123, 2017. Dinesh Singh, Phillip Febbo, Kenneth Ross, Donald Jackson, Judith Manola, Christine Ladd, Pablo Tamayo, Andrew Renshaw, Anthony D Amico, Jerome Richie, Eric Lander, Massimo Loda, Philip Kantoff, Todd Golub, and William Sellers. Gene expression correlates of clinical prostate cancer behavior. Cancer cell, 1:203 9, 04 2002. doi: 10.1016/S1535-6108(02)00030-2. Ram Dyuthi Sristi, Gal Mishne, and Ariel Jaffe. Disc: Differential spectral clustering of features. Advances in Neural Information Processing Systems, 35:26269 26282, 2022. Jonathan Svirsky and Ofir Lindenbaum. Interpretable deep clustering for tabular data. In Forty-first International Conference on Machine Learning. Joshua B Tenenbaum, Vin de Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction. science, 290(5500):2319 2323, 2000. Roman Vershynin. High-dimensional probability. University of California, Irvine, 2020. Ulrike Von Luxburg. A tutorial on spectral clustering. Statistics and computing, 17:395 416, 2007. Ulrike Von Luxburg, Mikhail Belkin, and Olivier Bousquet. Consistency of spectral clustering. The Annals of Statistics, pages 555 586, 2008. Suhang Wang, Jiliang Tang, and Huan Liu. Embedded unsupervised feature selection. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, 2015. Caroline L Wormell and Sebastian Reich. Spectral convergence of diffusion maps: Improved error bounds and an alternative normalization. SIAM Journal on Numerical Analysis, 59(3):1687 1734, 2021. Xijiong Xie, Zhiwen Cao, and Feixiang Sun. Joint learning of graph and latent representation for unsupervised feature selection. Applied Intelligence, pages 1 14, 2023. Junchen Yang, Ofir Lindenbaum, Yuval Kluger, and Ariel Jaffe. Multi-modal differentiable unsupervised feature selection. In Uncertainty in Artificial Intelligence, pages 2400 2410. PMLR, 2023. Yi Yang, Heng Tao Shen, Zhigang Ma, Zi Huang, and Xiaofang Zhou. L2,1-norm regularized discriminative feature selection for unsupervised learning. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence - Volume Volume Two, IJCAI 11, page 1589 1594. AAAI Press, 2011. ISBN 9781577355144. Published in Transactions on Machine Learning Research (12/2024) Shira Yoffe, Amit Moscovich, and Ariel Jaffe. Spectral extraction of unique latent variables. ar Xiv preprint ar Xiv:2402.18741, 2024. Sharon Zhang, Amit Moscovich, and Amit Singer. Product manifold learning. In International Conference on Artificial Intelligence and Statistics, pages 3241 3249. PMLR, 2021. Zheng Zhao and Huan Liu. Spectral feature selection for supervised and unsupervised learning. In Proceedings of the 24th international conference on Machine learning, pages 1151 1157, 2007. Linling Zhu, Linsong Miao, and Daoqiang Zhang. Iterative laplacian score for feature selection. In Chinese Conference on Pattern Recognition, 2012. Qi-Hai Zhu and Yu-Bin Yang. Discriminative embedded unsupervised feature selection. Pattern Recognition Letters, 112:219 225, 2018. Xiaofeng Zhu, Shichao Zhang, Rongyao Hu, Yonghua Zhu, et al. Local and global structure preservation for robust unsupervised spectral feature selection. IEEE Transactions on Knowledge and Data Engineering, 30(3):517 529, 2017. Xiaofeng Zhu, Shichao Zhang, Yonghua Zhu, Pengfei Zhu, and Yue Gao. Unsupervised spectral feature selection with dynamic hyper-graph learning. IEEE Transactions on Knowledge and Data Engineering, 34(6):3016 3028, 2020. Published in Transactions on Machine Learning Research (12/2024) A Product of manifold perspective. A.1 Proof of Theorem 1. As mentioned in the main text, the theorem is proven with the following two main steps: Step 1: Prove that the inner product f T i gb(X) is of order O 1/ n for all eigenfunctions gb(X) indexed by a vector b / E. Step 2: Combine the result of step 1 with the convergence guarantees in Theorem 1 to bound the inner product f T i vb. Step 1: According to our model, feature i is equal to a smooth transformation of a single latent variables. Assume w.l.o.g that the single variable is θ1 such that fi = Fi(θ1). By the product of manifold assumption, the eigenfunction gb is equal to h=1 gbh(π(h)(x)) = gb1(π(1)(x)) h=2 gbh(π(h)(x)). Let denote the Hadamard product. We can write the inner product f T i gb(X) as, f T i gb(X) = fi gb1(π(1)(X)) T gb2(π(2)(X)) , . . . , gb H(π(H)(X)). (6) The vectors fi and gb1(π(1)(X)) both depend on θ1 only. The vectors {gbh(π(h)(X)}H h=2 depend, respectively, on θ2, . . . , θH. We set a(θ1) = fi gb1(π(1)(X)) d(θ2, . . . , θH) = gb2(π(2)(X)) , . . . , gb H(π(H)(X)). The elements of the random vectors a(θ1) and d(θ2, . . . , θH) are statistically independent. In addition, we have that fi = 1 and gh(π(h)(X)) = 1 + o(1) (h), see for example (Cheng and Wu, 2022, Lemma 3.4). This implies that both a(θ1) and d(θ2, . . . , θH) are bounded by 1 + o(1). The inner product between two independent random vectors with unit norm and iid elements is of order O(1/ n), (see for example (Vershynin, 2020, Remark 3.2.5)). Thus, |f T i gb(X)| = |a(θ1)T d(θ2, . . . , θH)| = O(1/ n). Step 2: By the triangle inequality, |f T i vb| = |f T i (vb gb(X) + gb(X))| |f T i (vb gb(X))| + |f T i gb(X)|. (7) The first term on the right-hand side of equation 7 can be bounded by the Cauchy-Schwartz inequality and Theorem 1 via: |f T i (vb gb(X))| f T i vb gb(X) = O(ϵn) + O s The second term is bounded by step 1. Since the term in equation 8 dominates O(1/ n) for any ϵn, this concludes the proof. B Stability analysis We perform stability analysis using the Variation of Information (VI) criterion introduced by Meilă (2003), a measure of the distance between two clustering assignments. This criterion can be used to measure the Published in Transactions on Machine Learning Research (12/2024) (a) TOX 171 dataset (b) Yale dataset Figure 8: Histograms of VI value distributions for SSFS-selected features (orange) and all features (blue) on (a) TOX 171 and (b) Yale datasets. Lower VI values indicate higher stability in clustering results. stability of clustering results, which is valuable for assessing feature selection quality. For noisy features, we expect unstable clustering outputs with high variability between runs, while informative features should yield more stable results with less variability. Our analysis involves applying k-means clustering to the selected features 500 times with random initializations, calculating the VI between all pairs of clustering outputs, and analyzing the distribution of VI values. Lower average VI and tighter distribution indicate higher stability, suggesting more informative features. We compared the stability of clusters obtained using the original dataset versus those obtained using features selected by our Stable Spectral Feature Selection (SSFS) method. Table 4 presents results for the TOX171 and Yale datasets, demonstrating clear improvement in stability when using SSFS. For both datasets, SSFS-selected features yield lower mean VI values, indicating more consistent clustering outcomes. Figure 8 visualizes the distributions of VI values for both datasets, showing the shift towards lower VI values when using SSFS-selected features compared to using all features. Table 4: Stability Analysis Results: Mean Variation of Information (VI) Standard Deviation Dataset SSFS (top 100 features) All features TOX171 1.64 0.39 2.01 0.36 Yale 2.12 0.27 2.25 0.27 C Interpretability Interpretability is a fundamental aspect of unsupervised feature selection, particularly in domains where understanding the meaning and relevance of selected features is crucial. In the context of our proposed SSFS algorithm, interpretability can be enhanced in several ways. Feature ranking, by examining the weights or importance scores assigned to features during the selection process, can provide insights into which features are most important across the dataset. Visualization techniques, as demonstrated in Figure 9, can reveal patterns and distributions that highlight the discriminative capability of selected features across different classes or clusters. Domain-specific analysis is another effective approach. In biological datasets, for instance, selected features can be mapped back to specific genes or biomarkers, allowing domain experts to validate the biological relevance of the selected features and potentially uncover new insights about the underlying biological processes. These interpretability techniques enhance the utility of unsupervised feature selection algorithms like SSFS by providing insights into underlying relationships and feature relevance, which are essential for real-world applications across various domains. Published in Transactions on Machine Learning Research (12/2024) Figure 9 visualizes the top twenty features selected by SSFS on the COIL20 dataset. Each column corresponds to a sample, and each row represents a selected feature. The columns are ordered by the class label. The figure highlights the discriminative effect the selected features have over five classes in this dataset. Figure 9: Visualization of the top 20 selected features by SSFS across five classes in the COIL20 dataset. Samples are grouped by class, separated by white vertical lines, with class labels shown at the top. The color intensity indicates the relative feature values. This visualization allows for the comparison of feature patterns and distributions across different classes, highlighting potential discriminative features and classspecific characteristics. D Additional experiments on synthetic data We conducted further experiments to test our method s robustness to noise, using two synthetic datasets with 100 samples per trial over 10 trials. Both datasets contain five meaningful features from two isotropic Gaussian blobs but differ in their nuisance features. The first dataset s nuisance features are sampled from a multivariate Gaussian distribution, similar to Section 5.2.1, but with varying numbers of nuisance features. The second uses a copula of correlated nuisance features formed of marginals with a Laplace distribution. We measure performance using recall, comparing the top five ranked features to the five known meaningful ones. Results in Figure 10 show SSFS with logistic regression as a linear classifier (denoted by linear_clf) outperforming other methods across both experiments. For Gaussian nuisance features (Figure 10a), SSFS with a linear classifier maintained consistently high performance up to 60 features, while other methods exhibited a more pronounced decline as the number of nuisance features grew. In the case of Laplace-distributed nuisance features (Figure 10b), the advantage of SSFS with a linear classifier was even more evident, with its performance decreasing more slowly as the number of nuisance features increased. Overall, these experiments demonstrate that SSFS is a robust method for identifying relevant features, performing well for various noise distributions and maintaining high recall even with increasing noise levels. E Ablation study additional details E.1 Synthetic data generation For the synthetic data, we generated 500 samples, where we used the make_blobs function from scikit-learn to generate the first five features, with arguments cluster_std=1, centers=2. Published in Transactions on Machine Learning Research (12/2024) (a) Gaussian nuisance features (b) Laplace nuisance features Figure 10: Performance comparison of feature selection methods under increasing noise levels. The x-axis shows the number of nuisance features added to the give meaningful ones, while the y-axis represents the average recall (proportion of the top five ranked features matching the five meaningful features). Each point is the average over 10 trials with 100 samples each. SSFS with a linear classifier (logistic regression) consistently outperforms MCFS and the other SSFS variants. Figure 11: Ablation study: Clustering accuracy on real-world datasets E.2 Detailed Experimental Results In this section, we provide more detailed results of the ablation study. Figure 11 contains comparative analysis in terms of the performance for the whole selected feature range, F Experiments on real world datasets additional details F.1 Additional Experimental Details F.2 Datasets Table 5 provides information about the real-world datasets used in the experiments. For all datasets, the features are z-score normalized to have zero mean and unit variance. Published in Transactions on Machine Learning Research (12/2024) Table 5: Real-world datasets description. Dataset Samples Dim Classes Domain COIL20 1440 1024 20 Image ORL 400 1024 40 Image Yale 165 1024 15 Bio ALLAML 72 7129 2 Bio Prostate-GE 102 5966 2 Bio TOX 171 171 5748 4 Bio Isolet 1560 617 26 Speech GISETTE 7000 5000 2 Image F.3 Hyperparameters Hyperparameter selection in unsupervised feature selection presents unique challenges and may be infeasible due to the absence of labeled data for validation. In our methods, we have two types of hyperparameters: (i) For computing the spectral representation (kernel bandwidth and the number of eigenvectors) and (ii) parameters relating to the surrogate model, which predicts the pseudo-labels (i.e., regularization parameter for Logistic Regression). Note that the second type of parameters is used in a prediction task; they can be determined similarly to hyperparameters in supervised settings, for example, through cross-validation. In our experiments, for the surrogate models, we leverage the default hyperparameters from established machine learning packages such as XGBoost and scikit-learn s logistic regression, as these are generally robust for a wide range of supervised tasks. However, surrogate models and their hyperparameters can also be chosen by performing hyperparameter tuning with cross-validation based on pseudo-labels derived from the eigenvectors. The selection of surrogate models is flexible and can be additionally guided by domain knowledge. Practitioners can also compare results across various models and analyze their selected features. The first type of hyperparameters is inherent to the method, for example, the number of considered eigenvectors, which should increase with the number of classes and noise level in the data. For instance, in scenarios with more classes or where noise dominates the spectrum, a larger range of eigenvectors should be considered. While we used 500 resamples in our experiments, this number can be adjusted for larger datasets to balance computational cost and performance. Although traditional hyperparameter tuning may not be feasible in most unsupervised scenarios, practitioners working with domain-specific data could potentially fine-tune parameters using cross-validation on similar, labeled datasets from the same domain (such as images). For SSFS, we use the same hyperparameters, as follows: Number of eigenvectors to select k is set to the distinct number of classes in the specific dataset, they are selected from a total of d = 2k eigenvectors. Size of each subsample is 95% of the original dataset. 500 resamples are performed in every dataset. For the affinity matrix, we used a Gaussian kernel with an adaptive scale σiσj such that σi is the distance to the k = 2 neighbor of xi. The Laplacian we used was the symmetric normalized Laplacian. In the ablation study, for regression, we use scikit-learn ridge regression (for eigenvector selection) and DMLC XGBoost regressor (for the final feature scoring) with their default hyperparameters. For all of the baseline methods, we used the default hyperparameters. So, for all methods, including SSFS, the hyperparameters are fixed for all datasets (excluding parameters that correspond to the number of features to select and the number of clusters). Published in Transactions on Machine Learning Research (12/2024) For LS, MCFS, UDFS, and NDFS, we used an implementation from the scikit-feature library 2 and inputted the same similarity matrices as SSFS for the methods which accepted such an argument. We fixed a bug in MCFS implementation to choose by the max of the absolute value of the coefficients instead of the max of the coefficients (this improved MCFS performance). For LS-CAE, we used an implementation from 3. For KNMFS, we used an implementation from 4. 2https://github.com/jundongl/scikit-feature 3https://github.com/jsvir/lscae 4https://github.com/marcosd3souza/KNMFS