# sepne_bringing_separability_to_network_embedding__b24d7d7f.pdf The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) Sep NE: Bringing Separability to Network Embedding Ziyao Li,1 Liang Zhang,1 Guojie Song2 1Yuanpei College, Peking University, China 2Key Laboratory of Machine Perception, Ministry of Education, Peking University, China {leeeezy, zl515, gjsong}@pku.edu.cn Many successful methods have been proposed for learning low dimensional representations on large-scale networks, while almost all existing methods are designed in inseparable processes, learning embeddings for entire networks even when only a small proportion of nodes are of interest. This leads to great inconvenience, especially on super-large or dynamic networks, where these methods become almost impossible to implement. In this paper, we formalize the problem of separated matrix factorization, based on which we elaborate a novel objective function that preserves both local and global information. We further propose Sep NE, a simple and flexible network embedding algorithm which independently learns representations for different subsets of nodes in separated processes. By implementing separability, our algorithm reduces the redundant efforts to embed irrelevant nodes, yielding scalability to super-large networks, automatic implementation in distributed learning and further adaptations. We demonstrate the effectiveness of this approach on several real-world networks with different scales and subjects. With comparable accuracy, our approach significantly outperforms state-of-the-art baselines in running times on large networks. 1 Introduction Learning low dimensional representations of network data, or network embedding (NE), is a challenging task on large networks, of which the scales can reach billion-level and is growing rapidly. For example, the number of monthlyactive users of Facebook reaches 2.23 billion and increases 11% yearly. 1 At the same time, although sizes of networks may infinitely grow as data accumulates, it is often the case that only small proportions of nodes are of interest in downstream applications. This is the starting point of this paper: can we respectively learn representations for different subsets of nodes-very small compared to the collectivity-while preserving information of the entire network? If so, we can obtain good representations for the requested nodes without the redundant efforts to embed irrelevant ones. This work was supported by the National Natural Science Foundation of China (Grant No. 61876006 and No. 61572041). Corresponding author. Email: gjsong@pku.edu.cn Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 1https://investor.fb.com/investor-news/ Figure 1: Inseparable and separable NE processes. Efficiency is a major aspect of contemporary NE studies, and various methods that are applicable to large-scale networks have been proposed (Perozzi, Al-Rfou, and Skiena 2014; Tang et al. 2015; Grover and Leskovec 2016). Almost all of these methods embed entire networks with inseparable processes, in which the representation of one node depends on represented outcomes of every other node. A globallydefined optimum can be achieved under this framework, while it also causes great inconvenience: the maximum network size such methods can handle is eventually limited. For example, it takes LINE (2015), one of the fastest algorithms, several hours to embed a million-level network. Thousands of hours may be spent to achieve equivalent performance on billion-level networks. Another type of methods learns models in an inductive manner and conduct inferences over unseen data (Hamilton, Ying, and Leskovec 2017). These models have convenient inference processes, while they cannot variate over time and rely on large training data and time to achieve good performance. The efficiency problem over super-scale networks is impossible to solve directly, as the running time of algorithms inevitably grows proportionally to problem scales. Therefore, we bring up a new perspective of solving efficiency problems: separability. The separability of an algorithm indicates an ability of being conducted by different workers without exchanging information and merging outputs. In plain words, a separable algorithm divides the original problem into different self-standing sub-problems and separately solves each, and the solution to the sub-problems are directly usable answers instead of intermediate results. As networks are naturally composed of nodes and their relationships, an instinctive way to design separable NE algorithms is to partition the entire node set into small subsets and to separably embed each set. The solutions to the sub-problems yield direct meanings as the representations of the corresponding set of nodes. In this paper, we implement separability in NE problems under matrix-factorization-based framework. We formalize the problem of separated matrix factorization (SMF) and elaborate a novel loss function that preserves matrix information on local, global and landmark level. We then propose Sep NE (SEParated Network Embedding), a separable NE method based on SMF. Figure 1 illustrates the major difference between Sep NE and existing methods. Sep NE first partitions the node set into small subsets. A special landmark set is then established as references for partitioned subsets to implement comparability, that representations of different sets lie in the same linear space. After the landmarks are embedded, representations of different subsets are derived from the objective function defined in SMF. Separability in NE problems yields several specific advantages. Firstly, separability makes it available to embed only the requested nodes and thus reduces the vain efforts in embedding irrelevant ones; in addition, the optimization complexity of Sep NE is relevant only to the number of requested nodes instead of the entire network scale, leading to scalability to super-large networks. Secondly, even if entire networks are on request, Sep NE shows higher speed than state-of-the-art algorithms due to its simplicity, while yielding comparable accuracy. Thirdly, separability leads to automatic implementations in multi-agent systems and further adaptations in dynamic networks. We evaluated Sep NE with real-world networks of different sizes and topics. With equivalent or better accuracy, Sep NE is at least 50% faster than current NE methods including Deep Walk (2014) and LINE. A potentially more important contribution of this paper is the generalization of SMF. Maintaining competent performance, SMF reduces the complexity of MF problems from cubic to almost linear. This leads to intriguing further applications in the massive collectivity of MF-based algorithms. 2 Separated Matrix Factorization 2.1 Preliminaries Given a matrix M, matrix factorization (MF) aims to find two matrices W and C that both satisfy given constraints and minimize the residuals of reconstructing M with M = W T C. Denoted in formulas, we have min W,C M W T C . (1) W, C have lower ranks than M. In the embedding task of an n-node network, M is of size (n n), in which each entry indicates a proximity between the two corresponding nodes. The proximities can be defined in various metrics, such as edge weights between nodes. Columns in W are desired representations. Columns in C are interpreted as representations of nodes when regarded as contexts. 2.2 Problem Definition Directly factorizing matrices of large-scale networks can be unacceptably time-costly. Therefore, we propose SMF, a new optimization problem, as a trade-off between speed and accuracy. To implement separability, SMF divides the problem with a partition over the node set, and correspondingly partitions the proximity matrix. Below is a formal definition of separated matrix factorization (SMF) in NE scenarios. SMF takes a network G = (V, E), |V | = n, its proximity matrix M, and a partition setup f : V 7 V, V = {V1, , Vs} as inputs. The task is to derive representations (W1, , Ws) and (C1, , Cs) for the partitioned sets that optimally reconstruct M. The loss of the reconstruction is defined the same as Problem (1). 2 Without loss of generosity, we permute and partition M according to V as M11 M1s ... ... ... Ms1 Mss where Mij indicates the proximities between Vi and Vj. To achieve independence between sections, SMF restricts that the embedding section of every set is conducted (i) with only the proximities related to itself (the section of embedding Vi can leverage only Mij and Mji, j = 1, , s), and (ii) without any outcomes of other sections. 2.3 Method Partitioning the nodes is the first step to separability. However, representations in the partitioned sets can be incomparable due to the limitation over the access to proximities. In another word, representations of different sets do not have a unified constraint that bounds them in the same linear space. To implement comparability, we establish landmarks with highly interactive nodes in the network, which serve as invariant references for different subsets. For the factorization process, preserving only local information is a simple way to reconstructs micro-structures of networks at a loss of global references. Combining landmark solves the comparability problem, while it still ignores the interactions between different subsets. Therefore, we elaborate a novel objective function for SMF that preserves local, global and landmark information, which achieves state-of-the-art performances. Local information. The proximities in the partitioned matrix are naturally divided into two types, namely local information and global information. Local information refers to the proximities within every set, or sub-matrices on the diagonal; global information refers to the proximities between all pairs of different sets, or the off-diagonal sub-matrices. 2Frobenius forms are usually adopted in SMF, since the Frobenius form of a matrix is additive of all its entries, and therefore can be decomposed into sums of all the Frobenius forms of its partitioned sub-matrices. We start modeling SMF with a na ıve simplification that preserves only local information by factorizing s matrices on the diagonal: min Wi,Ci Mii W T i Ci , i = 1, , s. This primitive approach discards all interactions between different sets, leading to incomparable representations. Landmark information. To implement comparability, we resort to a third type of information, landmark information. Landmark information indicates the proximities between subsets and manually established landmarks, a special set of nodes (denoted as V0) that are chosen as references for different subsets. The improved approach sets a unified constraint over landmark information in different sets and solves the problem in two stages, formulated as min Φ,Ψ M00 ΦT Ψ , (2) M00 M0i Mi0 Mii ΦT Ψ ΦT Ci W T i Ψ W T i Ci i = 1, , s. (3) The first stage embeds the landmarks (W0 = Φ, C0 = Ψ) in Problem (2), and the second stage derives representations for rest sets by solving Problem (3) with calculated Φ and Ψ. If Frobenius forms are used, the loss in Problem (3) can be explicitly decomposed into local and landmark loss as Llc i (W, C) = 1 2 Mii W T C 2 F , (4) Llm i (W, C) = 1 2 M0i ΦT C 2 F 2 Mi0 W T Ψ 2 F . (5) Global information. To further combine global information into the objective function, we elaborate a global loss by first transforming Problem (3) into an equivalent form. We denote k := |V0| and assume calculated Φ, Ψ R(d k) are of rank d. 3 Wi, Ci R(d |Vi|) can then be represented as linear combinations of columns in Φ, Ψ, formulated in matrix denotation as Wi = ΦAi Ci = ΨBi, i = 1, 2, , s, where Ai, Bi R(k |Vi|) are the coefficient matrices. Consider a simple case where s = 2, V0 is the set of landmarks and V1 is the target subset to embed. After the transformation, global information is preserved through min A1,B2 M12 AT 1 ΦT ΨB2 (6) min A2,B1 M21 AT 2 ΦT ΨB1 . (7) Problem (6)(7) are not separable, for the results of embedding V2 (A2,B2) exist in V1 problem. However, a surprising 3This can be guaranteed with SVD decomposition if k d. property emerges after the transformation, that ΦT ΨB2 = W T 0 C2 = M02 can be well-approximated with M02 if representations of V2 are required to preserve landmark information, AT 2 ΦT Ψ similarly. Therefore, Problem (6)(7) can be substituted as min A1 M12 AT 1 M02 (8) min B1 M21 M20B1 , (9) and separability is achieved. The idea can be generalized to any given s and Vi by simply substituting all V2-related variables to V i-related ones, where V i = S j / {0,i} Vj. The approximation still holds if landmark information is preserved in all sets. For any set Vi, the global loss function is defined as Lgb i (A, B) = 1 2( Mi i AT M0 i 2 F + M ii M i0B 2 F ). 2.4 Final Optimization Problem Combined with λ-scaled global loss and regularization over A, B, the final loss function of SMF becomes Li(A, B) = Llc i (A, B) + Llm i (A, B) (10) +λLgb i (A, B) + η 2( A 2 F + B 2 F ), where Llc i and Llm i are redefined in A, B-denotation, namely Llc i (A, B) = 1 2 Mii AT ΦT ΨB 2 F , (11) Llm i (A, B) = 1 2 M0i ΦT ΨB 2 F 2 Mi0 AT ΦT Ψ 2 F . (12) Accordingly, the final optimization problem of SMF is formulate as W0 = Φ, C0 = Ψ. Wi = ΦAi, Ci = ΨBi, Ai, Bi = arg min A,B Li(A, B), i = 1, 2, , s. 3 Sep NE: Separated Network Embedding In this section, we propose Sep NE, a simple and separable NE approach based on SMF. A general framework of Sep NE is presented in Algorithm 1. We then illustrate the details of Sep NE, including the partition setups, landmark-selecting approaches and optimization method. Sep NE takes a given network as input and outputs node representations. In the preparation stage, landmarks are selected and embedded, and rest nodes are partitioned under a certain setup. In the second stage, partitioned sets are independently embedded by optimizing the SMF problem. The second stage is designed in a separable manner, so that if a small proportion of nodes are requested, Loop 4 in Algorithm 1 can be conducted only on the sets containing these nodes. Besides, separability allows cycles in the loop to be run distributedly. Algorithm 1 General framework of Sep NE. Input: G = (V, E), |V | = n. Output: Node embeddings for partitioned sets of nodes. 1: Partition rest nodes in set V into s subsets as V; 2: Sample k landmarks as set V0; 3: Conduct SVD on calculated proximity matrix M00 and calculate Φ, Ψ: M00 = UdΣd V T d , Φ = Ud Σd, Ψ = Vd Σd; 4: W0 = Φ; 5: for i = 1, 2, , s do 6: Calculate relevant proximity matrices M0i, Mi0, Mii, Mi i, M ii; 7: Optimize the loss functions: Ai, Bi = arg min A,B Li(A, B); 8: Calculate embeddings for set Vi: Wi = ΦAi, Ci = ΨBi 9: end for 10: return (W0, W1, , Ws) For the proximity matrix to be factorized, two metrics are adopted in Sep NE. The first metric defines M = A + A2, where A is the transition matrix of Page Rank (1999) 4. The second metric simplifies the first one with M = I + A. Our metrics are similar to TADW (2015), which proved an equivalency between factorizing A and A + A2 and Deep Walk (2014) with very short walks. A more instinctive understanding of the metrics can be derived from the perspective of proximity orders. A can be interpreted as a measurement of first-order proximity, and A + A2 a combination of first-order and second-order proximity. These concepts were proposed and further discussed in (Tang et al. 2015). 3.1 Partition Setups We propose three different partition setups for Sep NE in this paper. Sep NE-LP (Louvain Partition) partitions a network according to its communities using Louvain (2008). Leveraging community structures conforms matrix local information to network local information 5, which serves as an empirical approach to improve performance. However, as Sep NE leverages all the information of the proximity matrix, community-based partitions are not necessary. We further propose Sep NE-RP (Random Patition) which randomly assigns nodes to sets, and Sep NE-IO (Interested Only) which simply puts the requested nodes into one or more sets and ignores all unrequested ones. 3.2 Landmark Selection Landmark-selecting approaches influence not only representations of the landmarks, but also the loss of all sets in the entire SMF problem. As the key intention of setting landmarks is to establish references for different sets, landmarks 4Aij = 1/di if (i, j) E and 0 otherwise, where di is the degree of node i. 5Which refers to connections within real-world communities. are expected (i) to have as much connection with rest nodes as possible; (ii) to have the connection cover as many sets as possible. Approaches that select nodes with high degrees generally work well if k is loosely controlled. However, on realworld networks, nodes with the highest degrees tend to distribute in a few giant and highly connected communities. When k is strictly confined, choosing these nodes actually limits the number of sets these landmarks adjoin. To relieve this problem, we propose GDS (Greedy Dominating Set), an approach that greedily maximizes the number of nodes the landmarks adjoins. GDS first forms a maximum heap using degrees of nodes and initialize the landmark set as empty. After initialization, GDS iteratively examines the top of the heap. The top is simply removed if dominated by the current landmark set, otherwise added into the set and then removed. The process continues until the heap is empty or the size reaches k. Experiments show that GDS well captures the informative structure of a network when k is strictly confined. While serving as good references, landmarks selected with GDS are completely one-hop isolated. As a consequence, if only one-hop proximity is leveraged, M00 of GDS is supposed to be a diagonal matrix. Furthermore, if k > d, SVD will generate null representations for some landmarks. Therefore, we only use GDS when higher order proximity is adopted or k = d. Otherwise, we implement degree-based approaches. 3.3 Optimization The optimization problem in Sep NE is solved similarly to (Yu et al. 2014), where A, B are iteratively optimized as A(t+1) = arg min A Li(A, B(t)), B(t+1) = arg min B Li(A(t+1), B). As explicit calculation of the loss function value involves large matrix multiplications, A, B are calculated by solving the gradient-minimization problem L B = 0 in each iteration. Cholesky decomposition is adopted as the matrix in the gradient problem is always positive-definite. 3.4 Complexity Analysis The complexity of the preparation stage is O(nlogn + k3), while empirically the time expenses are low especially when random partitions are implemented. If M = I + A, the average complexity of each section is O(k (deg + iter k) ni), including both the time in calculating proximity matrices and in optimization. With separability implemented, Sep NE is available to only embed a proportion of nodes. Besides, when small proportions of nodes are requested, the complexity of Sep NE in the second stage is irrelevant to the entire scale of the network. This property yields strong scalability of Sep NE to super-scale networks. 4 Experiments We evaluated Sep NE on several publicly available realworld networks with different sizes and topics, including Dataset Directed Nodes Links Wiki directed 2,405 17,981 Cora directed 2,708 5,429 Citeseer directed 3,312 4,732 Flickr undirected 1,715,255 22,613,981 Youtube undirected 1,157,827 4,945,382 Wiki-Gen directed 2i 7.5 2i Table 1: Statistics of datasets used in this paper. three document networks and two social networks. Performances over three benchmark tasks were evaluated: (i) matrix reconstruction on document networks, (ii) multi-class classification tasks on document networks and (iii) multilabel classification tasks on social networks. 4.1 Experiment Setups Datasets. Five real-world networks were used in this paper. Wiki, Cora and Citeseer are thousand-level document networks. 6 Wiki contains Wikipedia pages of 19 classes; Cora contains machine learning papers from 7 classes and Citeseer contains publications from 6 classes. Links between documents are pointers or citations. Flickr and Youtube are million-level social networks. 7 Users and their relationships are represented as nodes and links on the networks, and real-world communities are available. Wiki-Gens are a series of networks generated by implementing Kronfit (2010) on Wiki. Table 1 shows the statistics of all datasets. Comparison Algorithms and Parameters. Algorithms and their parameters are briefly introduced below. We did not compare Sep NE with algorithms that are not scalable to large networks. Except otherwise noted, the representation dimension for all algorithms was d = 128. SVD was conducted over the full proximity matrices. As SVD theoretically generates the optimal rank-d approximations in F-norm, it was proposed as the strongest possible baseline for matrix reconstruction. Nystr om Method (Drineas and Mahoney 2005) is a fast monte-carlo method of approximating matrix multiplications. It was taken as a representative of probabilistic MF algorithms. The number of landmarks in Nystr om method is set the same as Sep NE for fair comparison. LINE (2015) embeds a network by optimizing an objective function of edge reconstruction. The parameters were set the same as the original paper, namely ρ0 = 0.025, negative sampling K = 5 and sample size T = 1010. Deep Walk (2014) embeds nodes by regarding them in random walk sequences as words in sentences. The parameters were set as window size win = 10, walk length t = 40 and walks per node γ = 40. Sep NE was evaluated under all three partition setups (Sep NE-LP, Sep NE-RP and Sep NE-IO). On document networks, parameters were set as iter = 100, λ = 0.4 6Available at https://linqs.soe.ucsc.edu/data 7Available at http://socialnetworks.mpi-sws.org/datasets.html Table 2: Running time comparison over flickr. Sep NE-IO 6.2mins Sep NE-RP 43.8mins Sep NE-LP 68.8mins LINE(1st) 138.1mins Deep Walk >24hrs 0 2 4 6 8 Number of edges ( 107) 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 Sep NE-IO Sep NE-RP Sep NE-LP 0 600 1200 1800 Time (secs) Preparation Time 0 2 4 6 8 Number of edges ( 107) 0 5 10 15 20 Time (secs) Optimization Time Figure 2: Scalability of Sep NE, demonstrated on Wiki-Gens. and η = 0.1, M = A + A2 and k = 200; on social networks, parameters were iter = 5, λ = 50, η = 1, M = I + A and k = 1000. 4.2 Running Time To demonstrate the speed advantage of Sep NE, we compared the running time of Sep NE, LINE and Deep Walk on Flickr network. 8 The nodes in the five biggest communities of Flickr (75,338 nodes, 4.39%) were regarded as interested for Sep NE-IO. Results are presented in Table 2. With better performance (introduced below), our method was significantly faster than LINE (for 50.2%) and Deep Walk even in embedding the entire network; when requested to embed only the nodes of interest, Sep NE completed the task in a very short time. 9 We also evaluated the trend of running time with network scales increasing on Wiki-Gens and the number of requested nodes for Sep NE-IO fixed as 10,000. Figure 2 (a) shows a linear trend in all three setups. Figure 2 (b)(c) show the trends of preparation and optimization time in Sep NE-IO. Preparation time increased linearly mainly due to the time used in reading data, while optimization time remained invariant. The results all corroborate that Sep NE is scalable to super-large networks. 4.3 Matrix Reconstruction The performance of reconstructing proximity matrix is a direct metric of representation quality. We evaluated matrix reconstruction performances of different algorithms on all document datasets. As the results were similar, we took Wiki as a representative. Two metrics were used, including the R2 8All efficiency experiments were conducted on a single machine with 128GB memory, 32 cores 2.13GHz CPU with 16 workers. 9Data are saved as edge lists in the experiments for fair comparison. If adjacent lists are available, the time of Sep NE-IO can still be reduced significantly. 200 300 400 500 600 landmark size k Nystrom SVD Sep NE 200 300 400 500 600 landmark size k Nystrom SVD Sep NE 200 240 280 320 360 400 landmark size k GDS DD DP UF 200 240 280 320 360 400 landmark size k GDS DD DP UF Figure 3: Performances of matrix preserving on Wiki network: (a)(b). rall and rnz of different algorithms versus k; (c)(d). rall of different landmark-selecting approaches versus k (d = 128 in (c) and 200 in (d)). score over all entries (rall) and non-zero entries (rnz): rall = 1 M M 2 F M 2 F , rnz = 1 ( M M) B 2 F M 2 F , where indicates element-wise multiplication and Bij = 1 if Mij = 0 otherwise 0. We evaluated Nystr om method, SVD and Sep NE with different k over both metrics. We then compared different landmark-selecting approaches with different k and d, including four: DD (Degree Deterministic) picking nodes with the k highest degrees; DP (Degree Probabilistic) sampling landmarks using degrees as weights; UF (Uniform) uniformly selecting landmarks; and GDS. According to Figure 3 (a)(b), Sep NE significantly outperforms Nystr om method for up to 38.3% and shows competent performance compared with SVD. SVD shows its advantage on rall, while preserving non-zero entries can be more important than zero entries on real-world networks due to the existence of unobserved links. When k is large enough, Sep NE outperforms SVD on rnz. This is because non-zero entries are more densely distributed inside communities and better reconstructed by Sep NE. In Figure 3 (c)(d), when k is slightly larger than d, or to say that k is strictly confined, GDS shows its significant advantages. However, the rall of GDS decreases when k > 2d due to the null representations generated in SVD. At the same time, degree-based approaches gradually get rid of their biases as k increases, and therefore show continuous improvements of performance. 4.4 Classification We implemented two types of classification tasks on document and social networks. A simple logistic regression was 0 2 4 6 8 10 iterations Performance (micro-f1) d=64 d=128 d=256 500 1000 1500 landmark size k d=64 d=128 d=256 0 50.0 100.0 150.0 200.0 Performance (micro-f1) d=64 d=128 d=256 2.0 4.0 6.0 8.0 10.0 d=64 d=128 d=256 Figure 4: Performances of multi-label classification under different parameter settings on Flickr network. used as the classifier for both tasks. The representations were all normalized before used as features. All results were averaged over 10 runs. Multi-class classification. We implemented multi-class classification on three document networks which predicts the subject category a given document is in. Table 3 reports the performances. Macro F1 results are not shown due to the similarity. Despite the minimum information Sep NE leverages, it outperforms Deep Walk in the majority (4 out of 6) of cases. This is because Sep NE incorporates a more robust and elegant way to leverage proximities between nodes. LINE is struggling to capture information on smaller networks, while Sep NE is as well competent. Multi-label classification. The multi-label classification task on social networks was defined as predicting whether a given node is in each community. The five largest in Flickr and communities with more that 1,000 members in Youtube were extracted as labels. As labels were sparse, we conducted training and predicting processes over the nodes that have at least one label. The training percentage was varied from 1% to 90%. Table 4 and Table 5 show the results. Sep NE shows significant advantages on Flickr. Using 10% training data, Sep NE-LP outperforms LINE and Deep Walk using 90%. Representations from Sep NE are more predictive than Deep Walk even if only one-hop proximity is leveraged. The reason may be that as Flickr has relatively high average degree, the larger window size of Deep Walk actually encumbers it in determining the importance of information. All three setups of Sep NE show good performances, while Sep NE-LP shows its advantage over the other two setups. This shows the effectiveness of the empirical method of partitioning networks according to communities, while the time cost of Sep NE-LP is significantly higher than the other two simplified setups. The task on Youtube is more challenging as both the network and labels are much sparser. Deep Walk outperforms both LINE and Sep NE due to its ability in leveraging remote proximities with its larger window size, which successfully relieves the problem of spar- Table 3: Multi-class prediction results over document networks (micro-averaged F1 scores). Best performances are bolded. Wiki Cora Citeseer %train 10% 90% 10% 90% 10% 90% LINE(1st) 0.4488 0.5937 0.4657 0.6009 0.3206 0.4259 LINE(2nd) 0.3298 0.4787 0.2637 0.3297 0.2221 0.2561 Deep Walk 0.5737 0.6893 0.7509 0.8187 0.5086 0.5813 Sep NE 0.5764 0.6867 0.7365 0.8220 0.5157 0.6072 Table 4: Multi-label prediction results over Flickr network (micro-averaged F1 scores). Best performances are bolded. %train 1% 3% 5% 10% 20% 30% 50% 90% LINE(1st) 0.3683 0.4118 0.4165 0.4219 0.4270 0.4273 0.4296 0.4274 LINE(2nd) 0.3450 0.3824 0.3955 0.3973 0.4032 0.4056 0.4069 0.4068 Deep Walk 0.4072 0.4353 0.4433 0.4481 0.4518 0.4564 0.4585 0.4592 Sep NE-IO 0.4065 0.4341 0.4477 0.4562 0.4582 0.4607 0.4630 0.4622 Sep NE-RP 0.4061 0.4388 0.4502 0.4601 0.4628 0.4634 0.4636 0.4658 Sep NE-LP 0.4269 0.4468 0.4562 0.4623 0.4645 0.4656 0.4674 0.4677 Table 5: Multi-label prediction results over Youtube network (micro-averaged F1 scores). Best two performances are bolded. %train 1% 3% 5% 10% 20% 30% 50% 90% LINE(1st) 0.1031 0.2322 0.2745 0.3141 0.3410 0.3520 0.3594 0.3673 LINE(2nd) 0.0782 0.1839 0.2158 0.2643 0.2987 0.3159 0.3280 0.3350 Deep Walk 0.2037 0.3397 0.3739 0.4105 0.4355 0.4438 0.4501 0.4556 Sep NE-IO 0.2035 0.3325 0.3574 0.3885 0.4041 0.4129 0.4170 0.4216 Sep NE-RP 0.2256 0.3355 0.3633 0.3920 0.4115 0.4157 0.4214 0.4273 Sep NE-LP 0.2253 0.3361 0.3620 0.3882 0.4118 0.4170 0.4218 0.4277 sity at the cost of much higher time expenses. Sep NE outperforms both LINE(1st) and LINE(2nd), which again corroborates its stronger ability to leverage near proximities. 4.5 Parameter Sensitivity Figure 4 shows the effect of iter, k, λ, η and d. Iter, k and η do not show significant influences. The performance of Sep NE is good with even one iteration, probably indicating that local information on Flickr is less important. Figure 4 (c) shows that larger λ generally leads to better performance, converging with λ 100. The higher performances of larger λs, particularly compared with λ = 0, show the effectiveness of the elaborated global loss. 5 Related Work There are massive literature proposed over NE problems. Traditional dimension reduction approaches (Roweis and Saul 2000; Tenenbaum, de Silva, and Langford 2000; Belkin and Niyogi 2002) are applicable on network data through Graph Laplacian Eigenmaps or proximity MF. Recently, various skip-gram-based NE models and applications were proposed (Perozzi, Al-Rfou, and Skiena 2014; Grover and Leskovec 2016; Du et al. 2018). Besides, the pioneering work of (Levy and Goldberg 2014) proved an equivalency between skip-gram models and matrix factorization, which further leads to new proximity metrics under the proximity MF framework (Yang et al. 2015; Cao, Lu, and Xu 2015; Qiu et al. 2018). Edge reconstruction algorithms (Tang et al. 2015) were proposed to gain scalability on large networks. Neural networks, including autoencoders (Wang, Cui, and Zhu 2016; Tu et al. 2018) and CNNs (Kipf and Welling 2016; Hamilton, Ying, and Leskovec 2017) were also leveraged in NE problems. There is also a new trend (Ribeiro, Saverese, and Figueiredo 2017) that leverages structural information instead of proximity in NE. The most similar work to ours is (Ahmed et al. 2013), in which a similar partition was adopted to achieve separability, while other parts of the work had major differences with ours. Besides, it focused mainly on technical issues in distributed learning and preserved only link information, while Sep NE is more generalized idea with a more elaborated optimization goal. 6 Conclusion In this paper, we formalized the problem of separated matrix factorization, based on which we proposed Sep NE, an separable network embedding method which outperforms strong baselines in both efficiency and performance. The key contribution of Sep NE is providing a novel perspective of evaluating network embedding methods: separability. A separable method is stronger than a distributable one, as partly conducting a separable task provides meaningful outputs. This property provides an option of embedding only a proportion of nodes and yields strong significance in distributed learning, super-large network embedding and dynamic network embedding. Furthermore, SMF reduces the complexity of MF from cubic to linear with the generalizability over all MF-based algorithms. Sep NE is still a simple framework. For future work, one intriguing direction is to incorporate more complex information into the framework without loss of efficiency. Also, a theoretical proof of a lower bound over the loss in matrix reconstruction, and a more theoretical explanation of SMF can be extremely informative. Should there be such work, it will be theoretically-founded to apply SMF on all matrixfactorization-based algorithms. Ahmed, A.; Shervashidze, N.; Narayanamurthy, S.; Josifovski, V.; and Smola, A. J. 2013. Distributed large-scale natural graph factorization. In Proceedings of the 22th International Conference on World Wide Web (WWW 13), 37 48. Belkin, M., and Niyogi, P. 2002. Laplacian eigenmaps and spectral techniques for embedding and clustering. In Advances in Neural Information Processing Systems 14. 585 591. Blondel, V. D.; Guillaume, J.-L.; Lambiotte, R.; and Lefebvre, E. 2008. Fast unfolding of communities in large networks. Journal of Statistical Mechanics: Theory and Experiment 2008(10):P10008. Cao, S.; Lu, W.; and Xu, Q. 2015. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management (CIKM 15), 891 900. Cui, P.; Wang, X.; Pei, J.; and Zhu, W. 2018. A survey on network embedding. IEEE Transactions on Knowledge and Data Engineering. Drineas, P., and Mahoney, M. 2005. On the nystr om method for approximating a gram matrix for improved kernel-based learning. Journal of Machine Learning Research 6:2153 2175. Du, L.; Wang, Y.; Song, G.; Lu, Z.; and Wang, J. 2018. Dynamic network embedding : An extended approach for skipgram based network embedding. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI 18), 2086 2092. Grover, A., and Leskovec, J. 2016. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 16), 855 864. Hamilton, W.; Ying, Z.; and Leskovec, J. 2017. Inductive representation learning on large graphs. In Advances in Neural Information Processing Systems 30. 1024 1034. Kipf, T. N., and Welling, M. 2016. Semi-supervised classification with graph convolutional networks. Co RR abs/1609.02907. Leskovec, J.; Chakrabarti, D.; Kleinberg, J.; Faloutsos, C.; and Ghahramani, Z. 2010. Kronecker graphs: An approach to modeling networks. Journal of Machine Learning Research 11:985 1042. Levy, O., and Goldberg, Y. 2014. Neural word embedding as implicit matrix factorization. In Advances in Neural Information Processing Systems 27. 2177 2185. Mislove, A.; Marcon, M.; Gummadi, K. P.; Druschel, P.; and Bhattacharjee, B. 2007. Measurement and analysis of on- line social networks. In Proceedings of the 5th ACM/Usenix Internet Measurement Conference (IMC 07). Page, L.; Brin, S.; Motwani, R.; and Winograd, T. 1999. The pagerank citation ranking: Bringing order to the web. Technical report, Stanford University. Perozzi, B.; Al-Rfou, R.; and Skiena, S. 2014. Deepwalk: online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 14), 701 710. Qiu, J.; Dong, Y.; Ma, H.; Li, J.; Wang, K.; and Tang, J. 2018. Network embedding as matrix factorization: Unifying deepwalk, line, pte, and node2vec. In Proceedings of the 11th ACM International Conference on Web Search and Data Mining (WSDM 18), 459 467. Ribeiro, L. F.; Saverese, P. H.; and Figueiredo, D. R. 2017. Struc2vec: Learning node representations from structural identity. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 17), 385 394. Roweis, S., and Saul, L. 2000. Nonlinear dimensionality reduction by locally linear embedding. Science 2323 2326. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; and Mei, Q. 2015. Line:large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web (WWW 15), 1067 1077. Tenenbaum, J. B.; de Silva, V.; and Langford, J. C. 2000. A global geometric framework for nonlinear dimensionality reduction. Science 2319 2323. Tu, K.; Cui, P.; Wang, X.; Yu, P. S.; and Zhu, W. Deep recursive network embedding with regular equivalence. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 18). Tu, K.; Cui, P.; Wang, X.; Wang, F.; and Zhu, W. 2018. Structural deep embedding for hyper-networks. In Proceedings of the 32nd AAAI s Conference on Artificial Intelligence (AAAI 18). Wang, D.; Cui, P.; and Zhu, W. 2016. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 16), 1225 1234. Yang, C.; Liu, Z.; Zhao, D.; Sun, M.; and Chang, E. 2015. Network representation learning with rich text information. In Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI 15), 2111 2117. Yu, H.-F.; Jain, P.; Kar, P.; and Dhillon, I. 2014. Large-scale multi-label learning with missing labels. In Proceedings of the 31st International Conference on Machine Learning (ICML 14), 593 601. Zhang, Z.; Cui, P.; Wang, X.; Pei, J.; Yao, X.; and Zhu, W. 2018. Arbitrary-order proximity preserved network embedding. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 18).