# structure_regularized_unsupervised_discriminant_feature_analysis__d6a30d1e.pdf Structure Regularized Unsupervised Discriminant Feature Analysis Mingyu Fan College of Maths & Info. Science, Wenzhou University, Wenzhou 325035, China fanmingyu@wzu.edu.cn Xiaojun Chang Centre for Artificial Intelligence University of Technology Sydney, Sydney, NSW 2007, Australia cxj273@gmail.com Dacheng Tao Centre for Artificial Intelligence University of Technology Sydney, Sydney, NSW 2007, Australia dacheng.tao@uts.edu.au Feature selection is an important technique in machine learning research. An effective and robust feature selection method is desired to simultaneously identify the informative features and eliminate the noisy ones of data. In this paper, we consider the unsupervised feature selection problem which is particularly difficult as there is not any class labels that would guide the search for relevant features. To solve this, we propose a novel algorithmic framework which performs unsupervised feature selection. Firstly, the proposed framework implements structure learning, where the data structures (including intrinsic distribution structure and the data segment) are found via a combination of the alternative optimization and clustering. Then, both the intrinsic data structure and data segmentation are formulated as regularization terms for discriminant feature selection. The results of the feature selection also affect the structure learning step in the following iterations. By leveraging the interactions between structure learning and feature selection, we are able to capture more accurate structure of data and select more informative features. Clustering and classification experiments on real world image data sets demonstrate the effectiveness of our method. Introduction Real world applications usually involve big data with high dimensionality, such as in computer vision (Collins, Liu, and Leordeanu 2005), bioinformatics (Saeys, Inza, and Larranaga 2007), and data mining (Liu et al. 2010). High dimensionality generally poses great challenges, including the curse of dimensionality , huge computation and storage cost, to conventional machine learning algorithms. In order to address this issue, feature selection is proposed to select a subset of features from the feature pool of high dimensional data for a compact and informative representation (Guyon and Elisseeff 2003). After the implementation of feature selection, conventional machine learning algorithms can be applied on data represented by only the selected relevant features instead of all the features. According to the availability of class labels of data, feature selection algorithms can be roughly classified into three groups, i.e., supervised feature selection (Song et al. 2007), semi-supervised feature selection (Xu et al. 2010), and un- Copyright c 2017, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. supervised feature selection (He, Cai, and Niyogi 2006). Supervised feature selection evaluates features by computing a feature s correlation with the class labels. Representative supervised feature selection methods include the Fisher score method (Duda, Hart, and Stork 2000), robust ℓ21 regression method (Nie et al. 2010) and the generalized Fisher score method (Gu, Li, and Han 2011). By exploiting the information of class labels, supervised feature selection is usually able to identify the discriminative and effective features for recognition and classification (Tao et al. 2016). On the other hand, with insufficient class labels, unsupervised and semisupervised methods have to consider the capability of the features in preserving or revealing of the underlying structure of data (He, Cai, and Niyogi 2006). A frequently used criterion is to select the features which best preserve intrinsic data structure. Recent research has witnessed several important data structures that should be preserved by features, where these data structures include, but not limit to, the sparse global structure (Du and Shen 2015) and the local manifold structure (Han et al. 2015). Generally speaking, because of insufficient class labels, it is more difficult for unsupervised and semi-supervised feature selection to find the discriminative and informative features. In practice, data structures for unsupervised feature selection are usually captured in the form of weighted graphs, such as the sample pair-wise similarity graph and the sparse graph (Du and Shen 2015) and the locally linear reconstruction graph (Hou et al. 2014). In graph based feature selection methods, the constructed graph is fixed in the following procedures and the performance of feature selection is largely determined by the quality of graph. Ideally, the quality of constructed graphs should be improved using the selected informative features instead of all the features that contain noisy and irrelevant ones. It is reasonable to alternatively optimize the data structure characterization using the selected features and then identify the selected feature set using the refined graph. Each sub-task and be iteratively boosted by using the result of the other one. Motivated by this, the main contributions of our work are: Data structure characterization is learned in two forms, which are referred as the soft structure and hard structure. The soft structure is defined by pair-wise similarities between data points and the hard structure is learned by data segmentation. Soft data structure is used to evaluate Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) the ability of features in preserving geometry of data and hard structure is used to extract the unsupervised discriminant information of data. The results of feature subset selection and data structure learning are optimized alternatively. In this way, each sub-task (structure learning and feature selection) and can boost the result of the other in the proposed feature selection framework. Both the soft and hard data structure can be naturally formulated as regularization terms in the regressional feature selection framework. And the derived regression algorithm can be efficiently optimized with convergence guarantee. The Proposed Framework Let X = [x1, , x N] be the given data matrix, where xi RD (1 i N) denotes the i-th data sample. Feature selection aims to evaluate the importance of all features of X, i.e, the row vectors of X. For an arbitrary matrix A Rm n, its ℓ1,2-norm is defined as A 1,2 = n j=1 Ai 2 = n j=1 m i=1 a2 ij. where Ai denotes the i-th column vector of matrix A. For simplicity, we assume that the elements of the D-th row of data matrix X are all 1s and thus the bias term in linear regression can be integrated in matrix A. Data Structure Learning A large number of unsupervised feature selection algorithms have been proposed based on the analysis of the data structures, such as the Maximum Variance (Max Var), manifold structure (Zhao et al. 2009; Du et al. 2013). Inspired by the recent development of compressive sensing, a popular approach to learn the affinity matrix of data is based on the self-expressiveness model. The basic assumption is that data points lie on a union of subspaces. Each data point can be expressed in term of a linear combination of other data points. The general problem can be formulated as below: min Z,E Z κ + λE E ω, (1) subj. to X = XZ + E, diag(Z) = 0, where matrix Z consists of self-expressive coefficients and E denotes the matrix of data noise, κ and ω are two properly chosen norms, λE > 0 is a tradeoff parameter. Many successful methods have been proposed based on different choices of the norms for coefficient Z and noise E. For example, in Sparse Subspace Clustering (SSC) (Elhamifar and Vidal 2013), ℓ1 norm is used for both κ and ω as a convex surrogate over ℓ0 norm to promote sparseness in coefficient matrix Z and handle the noises E. In Low-Rank Representation (LRR) (Liu et al. 2013), the nuclear norm is adopted for κ as a convex surrogate of the matrix rank function and ℓ1,2 norm is used for ω to handle noise or outlying entries E. Besides, a number of variants of (1) for data structure learning have been proposed for various applications in machine learning and pattern recognition. Based on the motivation that nearby data points should have large similarity and far away data points should have small similarity, (Nie, Wang, and Huang 2014) proposes to compute the similarities between pair wise data points by solving the following problem: min S=(sij) i,j xi xj 2sij + μs2 ij , (2) subj. to N j=1 sij = 1, sij 0, and sii = 0. where μ is the regularization parameter which is used to avoid trivial solution and add a prior of uniform distribution. It can be found that nearby data samples have large similarity sij. With such desirable property, the estimated similarity matrix S can be considered as an effective local data structure characterization. The self-expressive model (1) preserves global and sparse reconstruction data structure while the adaptive neighbor model (2) is based on the local similarity of data and focus on local data structure. Once the coding Z (or similarity matrix S) has been found, the segmentation of data can be obtained by applying Spectral Clustering (SC) (Ng, Jordan, and Weiss 2001) on the induced affinity matrix W = |Z| + |ZT | (or W = |S| + |ST |). The clustering result is assumed to be given as {t1, , t N}, where ti {1, , C} is the assigned cluster label of xi with C denotes the number of clusters. In this paper, the induced affinity matrix W is referred to as the soft data structure because it describes the pair-wise similarity using nonnegative real value, meanwhile, the result of data segmentation is called as the hard data structure characterization as it provides label attribute of the data points. Discriminant Feature Analysis Linear Discriminant Analysis (LDA) (Fukunaga 1990) is a popular supervised feature extraction method. It seek directions on which data points from the same classes are close and data points from different classes are far away from each other. Given the class labels of data points, the objective function of LDA is as follows A = arg min A Tr ASw AT Tr (ASb AT ) , (3) where Tr( ) indicates the matrix trace operator, A RD D is the desired projective matrix and Sw = C c=1 nc i=1(x(c) i x(c))(x(c) i x(c))T , and Sb = C c=1 nc( x(c) x)( x(c) x)T , are within-class scatter matrix and between-class scatter matrix respectively, with nc indicates the number of samples in the c-th class, x(c) i be the i-th sample in the c-th class, x(c) is the mean of the samples in the c-th class, x denotes the mean of all the samples. Define St = N i=1(xi x)(xi x)T as the total scatter matrix, then we have St = Sw + Sb. The objective function of LDA in (3) is equivalent to A = arg min A Tr ASt AT Tr (ASb AT ) = arg max A Tr ASb AT Tr (ASt AT ) . (4) The solution A is given by the eigenvectors of the top eigenvalues of the generalized eigen-problem Sbα = λStα, where λ is a eigenvalue and α denotes the corresponding eigenvector. Because of its simplicity and effectiveness, LDA has been widely used in machine learning research. Unsupervised Discriminative Feature Selection Framework In unsupervised scenarios, the label of data samples are unknown. As discussed above, one can implement data structure learning (Section ) to find the clustering labels of data samples, i.e., the hard data structure characterization. Then, the with assigned clustering labels, we convert LDA into a feature selection algorithm via adopting ℓ1,2-norm as a regularizer: A = arg min A Tr ASw AT Tr (ASb AT ) + γA A 1,2, (5) where the regularization term A 1,2 ensures that A is sparse in columns, making it suitable for feature selection. The balancing parameter γA controls the tradeoff between the two terms. Furthermore, we hope the result of feature selection can affect back on the data structure learning process. When Axi and Axj are close after feature selection, the similarity wij between samples xi and xj should be large. The objective to minimize the disagreement between the projected data matrix AX and the data similarity matrix W can be quantified as: 2 Axi Axj 2 = min A,W W Θ 1, (6) where indicates the hadamard product, W = (wij) and Θ = 1 2 Axi Axj 2 . Because of the fact that W = |Z| + |ZT | (or W = |S| + |ST |), the problem (6) is essentially equivalent to i,j=1 |zij| 1 2 Axi Axj 2 = min A,Z Z Θ 1, i,j=1 |sij| 1 2 Axi Axj 2 = min A,S S Θ 1. Due to the success of SSC method in subspace clustering, here we adopt the ℓ1-norm for both κ and ω in (1) for structure learning. Combining our discriminative feature selection term (5) and the disagreement term (6), the unified optimization framework for the Sr-UDFS algorithm is proposed as follows: min Z,E,A { Z 1 + λE E 1 + λZ Θ Z 1 Tr (ASb AT ) + γA A 1,2}, (7) Subj. to: X = XZ + E, diag(Z) = 0. As can be seen that when A is fixed, our method learns the data structure with the consideration of the refined data features (the third term). When the coding matrix Z is fixed, both the soft data structure and hard data structure are transformed into regularizers for the problem of feature selection. In our method, both the adverse effect of data noise E and noisy features are largely alleviated. The two sub tasks, data structure learning and feature selection, boosts each other within the unified learning framework. Optimization In this subsection, an efficient solution to the optimization problem in (7) is proposed based on solving the following subproblems iteratively: 1. Given A, optimize Z and E through solving a weighted sparse coding problem. 2. Implement SC on weight matrix W = |Z| + |ZT | to obtain the clustering labels of data. 3. Find the optimal A given Z and E. Given the projection matrix A (initialized as the identity matrix I), we solve for matrix Z and E through optimizing the following structured sparse problem: min Z,E { Z 1 + λE E 1 + λZ Θ Z 1} (8) Subj. to: X = XZ + E, diag(Z) = 0. To implement Alternating Direction Method of Multipliers (ADMM) method, an augmented matrix Q should be introduced, and the problem (8) is equivalent to: min Z,E { Z 1 + λE E 1 + λZ Θ Z 1} Subj. to: X = XQ + E, Q = Z diag(Z). The augmented Lagrangian function is given by: L (Z, Q, E, Y1, Y2) = Z 1 + λE E 1 + λZ Θ Z 1 + Y1, X XQ E + Y2, Q Z + diag(Z) 2 X XQ E 2 F + Q Z + diag(Z) 2 F , (9) where Y (1), Y (2) are matrices of Lagrange multipliers, and μ > 0 is a adaptive parameter. The iterative scheme of ADMM method for (8) can be presented as (at the t + 1th iteration) where ρ > 1 is a given parameter. For the Z-subproblem in (9), we solve the following problem Z(t+1) = arg min Z 11T + λZΘ Z 1+ 2 Q(t) Z + diag(Z) + Y (t) 2 μ 2 F where 1 indicates the vector with entries are all 1s. The closed-form solution for Z can be given as Z(t+1) = Z(t+1) diag( Z(t+1)), (10) where Z(t+1) ij = S 1 μ(t) (1+λZΘij) U (t) ij with U (t) = Q(t)+ Y (t) 2 μ . Here, Sτ ( ) is the element-wise shrinkage thresholding operator. It should be noted that instead of softthresholding all entries of matrix U (t) with a constant value as in SSC (Elhamifar and Vidal 2013), the proposed method thresholds the entries of U t with different values. To optimize Q in (9), taking derivative of the objective function with respect to Q and the solution is given by Q(t+1) = XT X + I 1 XT (X E(t) + Y (t) 1 μ(t) ) +Z(t+1) diag(Z(t+1)) . (11) Algorithm 1 ADMM for solving problem (8) Input: Data matrix X, projective mapping A Output: Sparse coding matrix Z and the noise matrix E 1: while not converged do 2: Update Z(t+1) as in (10); 3: Update Q(t+1) as in (11); 4: Update E(t+1) as in (12); 5: Update Y (t+1) 1 , Y (t+1) 2 and μ(t+1) as follows Y (t+1) 1 = Y (t) 1 + μ(t)(X XQ(t+1) E(t+1)), Y (t+1) 2 = Y (t) 2 + μ(t)(Q(t+1) Z(t+1) + diag(Z(t+1))), μ(t+1) = ρ μ(t) 6: If not converged, set t t + 1. 7: end while When other variables are fixed, the subproblem to find E is: E(t+1) = arg min E λE E 1 + μ 2 X XQ(t+1) E + Y (t) 1 μ(t) 2 F . The closed-form solution for E can be given as E(t+1) = S λE V (t) , (12) where V (t) = X XQ(t+1) + Y (t) 1 μ(t) . After the convergence of ADMM method (9), the selfexpressive sparse representation matrix Z is obtained. The next step is to infer the segmentation of data into different clusters. To address this problem, one can directly compute the similarity matrix as W = |Z| + |ZT | and then apply SC method to W. The segmentation of data can be obtained as T = {t1, , t N}, where ti {1, , C} is the assigned cluster label of xi. Solving Structure regularized Unsupervised Discriminant Feature Selection. Given the labels T , the problem (7) reduces to the following problem: min A Tr ASw AT Tr (ASb AT ) + λZ i,j wij Axi Axj 2 + γA A 1,2, (13) where the scatter matrices Sw and Sb are calculated based on the labels T . Because matrix A exists in the numerator, denominator and the summed terms, it is difficult to directly solve (13). In this paper, we resort to the spectral regression method (Cai, He, and Han 2008) which can transform the intricate problem (13) to an equivalent regression form and make it easier and more efficient to solve. Let X = [x1 x, , x N x] be the centered data matrix. The between-class scatter matrix Sb can be rewritten as c=1 nc x(c)( x(c))T = X W XT , where Wij = 1 nc if xi and xj are of the same class c and 0 otherwise. The following Theorem can be obtained Theorem 0.1. Let Y R(C 1) N be a matrix of which each row vector is an eigenvector of the eigen-problem Algorithm 2 The algorithm for solving problem (13) Input: Data matrix X, Sparse coding matrix Z Output: Converged matrix A 1: Implement SC on W = |Z| + |ZT | to obtain data labels T = {t1, , t N}; 2: Compute the regression target Y ; 3: while not converged do 4: Compute the diagonal matrix D(t+1) A as (16); 5: Update A(t+1) as (17); 6: If not converged, set t t + 1. 7: end while Wy = λy. If there exist a matrix A R(C 1) D such that A X = Y , then each row vector of A is an eigenvector of the generalized eigen-problem X W XT α = λ X XT α (i.e., eigen-problem for LDA) with the same eigenvalue λ. Proof. With A X = Y and Wy = λy, we have the following equation X W XT α = X Wy = Xλy = λ X XT α, where α is the transpose of a row vector of matrix A and y is the transpose of a row vector of Y . Theorem 0.1 indicates that under mild condition, the LDA problem (4) is essentially equivalent to the regression problem: A = arg min A A X Y 2 F , where the row vectors in Y are eigenvectors of eigen-problem Wy = λy. One advantage of LDA is that one need not to really solve the eigenproblem to obtain the eigenvectors Y . The C + 1 eigenvectors of W can be directly given as {1} {vc}C c=1 {0, 1}N, with the j-th entry of vc is 1 if and only if xj is in class c. Subsequently, we can get the C 1 useful orthogonal eigenvectors {yc}C 1 c=1 by implementing the Gram Schmidt orthogonalization algorithm on {1} {vc}C c=1. As is shown in (Cai, He, and Han 2008), the C 1 orthogonal eigenvectors are sufficient to represent a C class problem. Based on the above discussions, the problem (13) is equivalent to the following problem: min A A X Y 2 F + λZTr AXLXT AT + γA A 1,2, (14) where L = D W is the graph Laplacian matrix, D is a diagonal matrix with diagonal elements Dii = N j=1 wij, i = 1, , N. Motivated by the recent progress on ℓ1,2 norm minimization, the problem (14) can be efficiently solved by an iterative re-weighted approach which solve the following problems (at the (t + 1)-th iteration): A(t+1) = arg min A A X Y 2 F + λZTr AXLXT AT +γATr A(D(t) A ) 1AT , (15) ii = A(t+1) i 2. (16) The solution to (15) can be given as A(t+1) = Y XD(t) A X XT D(t) A + λZXLXT D(t) A + γAI 1 (17) We can show that the objective function of problem (14) is nonincreasing under the updating rules of A and D in Algorithm 2. Theorem 0.2. The Algorithm monotonically decrease the objective function of the problem (14) in each iteration, and converge to the global optimum of the problem. Proof. The proof follows the work (Nie et al. 2010) and can be found in the supplement material. Discussions In this section, we discuss the relationships between the proposed method and several algorithms, including TRACK (Wang, Nie, and Huang 2014), CGSSL (Li et al. 2014), and DFS (Tao et al. 2016). TRACK proposed an unsupervised feature selection by integrating Fisher criterion and clustering as below min AAT =I,G Ind Tr ASt AT + γA A 1,2 where G is the {0, 1} cluster indicator matrix. In TRACK, G is computed by the k-means method and A is given by the eigenvectors of the smallest eigenvalues of the matrix Sw Tr ASw AT Tr ASt AT St + γATr ASt AT DA, (19) where DA is a diagonal matrix whose i-th diagonal entry DA(ii) = 1 2 Ai 2 . As can be seen, TRACK implement clustering directly on transformed data matrix AX, meanwhile, Sr-UDFS implements SC on refined similarity matrix W. DFS proposed a supervised feature selection based on the Fisher criterion, which compute the projection matrix A by solving the generalized eigen-problem: (γADA Sb)α = λStα. (20) Compared with both TRACK and DFS, our Sr-UDFS efficiently transforms objective of Fisher criterion into a regression model and avoid solving the eigen-problem (19) or the generalized eigen-problem (20). CGSSL is an one-stop unsupervised feature selection method. The objective function of CGSSL can be presented as min A,P,Q,Y Tr(Y LY T ) + α Y AX 2 F + β A 1,2 +γ A PQ 2 F subj. to. Y Y T = I, Y 0, QQT = I, where L is the graph Laplacian matrix, α, β and γ are given parameters. As can be seen, different from Sr-UDFS, the entries of target Y are nonnegative reals instead of {0, 1} values. Besides, the data structure learned in CGSSL is based only on the local manifold assumption, which cannot utilize the refined data features to improve the quality of data structure learning. Table 1: Statistics of the data sets Data sets # of samples # of Dimension # of Classes Coil-20 1440 1024 20 Yale B 2414 1024 38 USPS 9298 256 10 CMUPIE 11554 1024 68 Experiments In this section, we evaluate the proposed Sr-UDFS to data clustering and classification on benchmark image datasets. Besides, state-of-the-art unsupervised feature selection methods are compared under various experimental settings. Datasets Description The experiments are conducted on publicly available image data sets: the Coil-20 data set (Coil-20)1, the USPS handwritten digits data set (USPS)2, the Yale-B Extended (Yale B) and the CMUPIE face data sets3. The statistics of the data sets are summarized in Table 1. Experiment Setup The validate the effectiveness of the proposed Sr-UDFS, we compare it with several state-of-the-art unsupervised feature selection methods, which includes Lap Score (He, Cai, and Niyogi 2006), SPFS (Zhao et al. 2013), UDFS (Yang et al. 2011), MCFS (Cai, Zhang, and He 2010), RUFS (Qian and Zhai 2013), and JELSR (Hou et al. 2014). Also, one baseline (all features) for data clustering and classification is also compared. There are some parameters to be set for the comparing methods. For methods require neighborhood sizes for data structure learning, the neighborhood size is searched in {4, 6, 8, 10}. For UDFS, RUFS, and JELSR, the regularization parameters are searched in the range {10 5, 10 4, , 101, 102}. The parameter γA is searched in the range {0.01, 0.05, 0.1, 0.5, 1}. To make the experimental results reproducible, λZ for Sr-UDFS is set as 0.1 respectively throughout the experiments. The experimental results of the parameters sensitivity for Sr-UDFS on Coil-20 data set is shown in Fig. 2. Given a data set, we randomly select p percents from every class in data X to formulate the training set Xtrain, and the left data are used as the test data Xtest. The results when p = 30 is presented here and the results when p = 10 is provided in the supplement material. The feature selection methods are implemented on training data to rank all the features. In our data clustering experiments, the test data is clustered by k-means method with some selected features. In our classification experiments, we assume the class labels of training data is known and implement the nearest neighbor classifier on the test data with a few of selected features. Both clustering and classification accuracies are computed 1http://www.cs.columbia.edu/CAVE/software/softlib/coil20.php 2http://www.escience.cn/people/fpnie/ 3http://www.cad.zju.edu.cn/home/dengcai/Data/Face Data.html 0 200 400 600 800 1000 1200 Number of Selected Features Accuracy (%) 1NN with p = 30 1-NN Sr-UDFS Lap Score SPFS UDFS MCFS RUFS JELSR 0 200 400 600 800 1000 1200 Number of Selected Features Accuracy (%) 1NN with p = 30 1-NN Sr-UDFS Lap Score SPFS UDFS MCFS RUFS JELSR 0 50 100 150 200 250 300 Number of Selected Features Accuracy (%) 1NN with p = 30 1-NN Sr-UDFS Lap Score SPFS UDFS MCFS RUFS JELSR 0 200 400 600 800 1000 1200 Number of Selected Features Accuracy (%) 1NN with p = 30 1-NN Sr-UDFS Lap Score SPFS UDFS MCFS RUFS JELSR (a) (b) (c) (d) Figure 1: 1-NN classification results of the comparing feature selection methods with 30% percents training data the on (a) Coil-20, (b) Yale B, (c) USPS and (d) CMUPIE data sets. (1-NN in legend means the 1-NN classifier on all features) High resolution figures can be found in supplement material Table 2: Clustering results when 30% of data are used as the training data. The result is in bold when one feature selection method outperforms the comparing feature selection methods. Clustering Acc p = 30 All Fea Lap Score SPFS UDFS MCFS RUFS JELSR Sr-UDFS Coil-20 48.72(0.17) 52.81(7.16) 50.59(4.47) 48.87(4.98) 50.76(4.61) 47.67(4.73) 48.27(7.88) 57.20(5.61) Yale B 8.44(0.21) 8.61(0.42) 12.72(2.65) 8.43(0.62) 9.83(1.62) 10.74(1.68) 7.94(0.50) 16.71(4.59) USPS 65.56(0.06) 54.68(7.55) 62.91(4.15) 46.88(11.93) 54.28(5.32) 59.15(7.60) 59.46(5.95) 65.41(4.21) CMUPIE 9.77(0.16) 7.91(0.18) 11.42(0.60) 8.87(0.24) 9.08(0.48) 9.64(0.45) 8.95(0.30) 11.91(1.13) Clustering NMI p = 30 All Fea Lap Score SPFS UDFS MCFS RUFS JELSR Sr-UDFS Coil-20 67.31(0.23) 64.18(6.32) 64.07(2.49) 62.84(4.05) 64.63(5.13) 62.93(4.61) 61.04(7.16) 71.46(3.08) Yale B 11.61(0.43) 11.80(0.82) 22.33(6.46) 12.52(1.75) 15.81(3.70) 17.39(4.13) 10.20(0.91) 29.30(6.39) USPS 62.79(0.05) 53.33(7.70) 60.19(2.81) 42.11(13.69) 53.08(7.05) 54.16(8.85) 56.70(5.79) 62.28(3.52) CMUPIE 21.57(0.77) 19.42(0.45) 25.07(1.67) 21.35(0.56) 21.25(1.56) 21.32(0.91) 19.05(1.32) 26.73(2.36) 1NN Accuracy kmeans Accuracy Figure 2: Performance variation of Sr-UDFS on Coil-20 with respect to different values of the parameters (p = 30%, number of features = 200) only on the test data. For each setting, the experiments are repeated 10 times, and both the mean of results and standard deviation are reported. Performance Evaluation For classification experiments, we choose the 1-Nearest Neighbor (1-NN) classifier because it is parameter free and the results will be easily reproducible. The results are measured by classification accuracy. Because the optimal number of features are unknown, we compare the algorithms with different percentages of training subset and varying number of selected features. The classification results are shown in Fig. 1. As can be seen, most of the time Sr-UDFS outperforms the comparing methods on the data sets. On Coil-20, Yale B and CMUPIE data sets, Sr-UDFS outperforms the 1-NN with all features. And on USPS data set, Sr-UDFS can achieve comparable performance as 1-NN with fewer selected features. These results indicate that the proposed method can effectively remove redundant and noisy features of data. With the selected features, we evaluate the performance of clustering by two common evaluation metrics, Accuracy (Acc) and Normalized Mutual Information (NMI). The range of the number of the selected feature for Coil-20, Yale B, and CMUPIE is {15, 25, 35, 45, 50, 100, 150, 200, 250, 300} and the range of selected features for USPS is {15, 25, 35, 45, 50, 100, 150, 200, 250}. We finally report the averaged results and standard deviation over the range of selected features. The clustering results in terms of Acc and NMI are reported in Table 2. Compared with clustering using all features, except the case on USPS data set when p = 30, the proposed method not only largely reduce the number of selected features, but also improve the clustering performance. Compared with other unsupervised feature selection methods, the Sr-UDFS produces better performance in most of the cases. And the performance are very close when some other methods outperform the proposed method. In this paper, we proposed a novel unsupervised feature selection method which can simultaneously perform data structure learning and feature selection. In our method, two data structures, soft structure and hard structure, are learned via a combination of the alternative optimization and clustering. Both the two types of data structures are formulated as regularization terms for our discriminant feature selection. An efficient algorithm for the proposed algorithm is proposed. The connections between our method with other counterparts are discussed. Experiments on benchmark image data sets have been presented to demonstrate the superior performance of our method. Acknowledgments This work was partially supported by the National Natural Science Foundation (NNSF) of China under Grants 61473212, 61203241, 61472285, and the Natural Science Foundation of Zhejiang Province under Grant LY15F030011, partially supported by the Australian Research Council (ARC) projects FT-130101457, DP140102164, and LE-140100061. Cai, D.; He, X.; and Han, J. 2008. Srda: An efficient algorithm for large-scale discriminant analysis. IEEE Transactions on Knowledge and Data Engineering 20(1):1 12. Cai, D.; Zhang, C.; and He, X. 2010. Unsupervised feature selection for multi-cluster data. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 333 342. Collins, R. T.; Liu, Y.; and Leordeanu, M. 2005. Online selection of discriminative tracking features. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(10):1631 1643. Du, L., and Shen, Y.-D. 2015. Unsupervised feature selection with adaptive structure learning. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 209 218. Du, L.; Shen, Z.; Li, X.; Zhou, P.; and Shen, Y. D. 2013. Local and global discriminative learning for unsupervised feature selection. In 2013 IEEE 13th International Conference on Data Mining, 131 140. Duda, R. O.; Hart, P. E.; and Stork, D. G. 2000. Pattern Classification (2Nd Edition). Wiley-Interscience. Elhamifar, E., and Vidal, R. 2013. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence 35(11):2765 2781. Fukunaga, K. 1990. Introduction to Statistical Pattern Recognition (2Nd Ed.). San Diego, CA, USA: Academic Press Professional, Inc. Gu, Q.; Li, Z.; and Han, J. 2011. Generalized fisher score for feature selection. In Barcelona, S., ed., In Proc. of the 27th Conference on Uncertainty in Artificial Intelligence, 266 273. Guyon, I., and Elisseeff, A. 2003. An introduction to variable and feature selection. Journal of Machine Learning Research 3:1157 1182. Han, Y.; Yang, Y.; Yan, Y.; Ma, Z.; Sebe, N.; and Zhou, X. 2015. Semisupervised feature selection via spline regression for video semantic recognition. IEEE Transactions on Neural Networks and Learning Systems 26(2):252 264. He, X.; Cai, D.; and Niyogi, P. 2006. Laplacian score for feature selection. In Weiss, Y.; Sch olkopf, B.; and Platt, J. C., eds., Advances in Neural Information Processing Systems 18, 507 514. MIT Press. Hou, C.; Nie, F.; Li, X.; Yi, D.; and Wu, Y. 2014. Joint embedding learning and sparse regression: A framework for unsupervised feature selection. IEEE Transactions on Cybernetics 44(6):793 804. Li, Z.; Liu, J.; Yang, Y.; Zhou, X.; and Lu, H. 2014. Clusteringguided sparse structural learning for unsupervised feature selection. IEEE Transactions on Knowledge and Data Engineering 26(9):2138 2150. Liu, H.; Motoda, H.; Setiono, R.; and Zhao, Z. 2010. Feature selection: An ever evolving frontier in data mining. In JMLR Workshop and Conference Proceedings, volume 10 of Feature Selection in Data Mining, 4 13. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; and Ma, Y. 2013. Robust recovery of subspace structures by low-rank representation. IEEE Transactions on Pattern Analysis and Machine Intelligence 35(1):171 184. Ng, A. Y.; Jordan, M. I.; and Weiss, Y. 2001. On spectral clustering: Analysis and an algorithm. In ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS, 849 856. MIT Press. Nie, F.; Huang, H.; Cai, X.; and Ding, C. H. 2010. Efficient and robust feature selection via joint ℓ2,1-norms minimization. In Advances in Neural Information Processing Systems 23, 1813 1821. Nie, F.; Wang, X.; and Huang, H. 2014. Clustering and projected clustering with adaptive neighbors. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 14, 977 986. New York, NY, USA: ACM. Qian, M., and Zhai, C. 2013. Robust unsupervised feature selection. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, 1621 1627. AAAI Press. Saeys, Y.; Inza, I.; and Larranaga, P. 2007. A review of feature selection techniques in bioinformatics. Bioinformatics 23(19):2507 2517. Song, L.; Smola, A.; Gretton, A.; Borgwardt, K. M.; and Bedo, J. 2007. Supervised feature selection via dependence estimation. In Proceedings of the 24th International Conference on Machine Learning, 823 830. Tao, H.; Hou, C.; Nie, F.; Jiao, Y.; and Yi, D. 2016. Effective discriminative feature selection with nontrivial solution. IEEE Transactions on Neural Networks and Learning Systems 27(4):796 808. Wang, D.; Nie, F.; and Huang, H. 2014. Unsupervised Feature Selection via Unified Trace Ratio Formulation and K-means Clustering (TRACK). Berlin, Heidelberg: Springer Berlin Heidelberg. 306 321. Xu, Z.; King, I.; Lyu, M. R. T.; and Jin, R. 2010. Discriminative semi-supervised feature selection via manifold regularization. IEEE Transactions on Neural Networks 21(7):1033 1047. Yang, Y.; Shen, H. T.; Ma, Z.; Huang, Z.; and Zhou, X. 2011. L2,1norm regularized discriminative feature selection for unsupervised learning. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence - Volume Volume Two, 1589 1594. AAAI Press. Zhao, B.; Kwok, J.; Wang, F.; and Zhang, C. 2009. Unsupervised maximum margin feature selection with manifold regularization. In Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on, 888 895. Zhao, Z.; Wang, L.; Liu, H.; and Ye, J. 2013. On similarity preserving feature selection. IEEE Transactions on Knowledge and Data Engineering 25(3):619 632.