# multiview_graph_clustering_via_nodeguided_contrastive_encoding__39f812b8.pdf Multi-View Graph Clustering via Node-Guided Contrastive Encoding Yazhou Ren 1 2 Jun Long Ke 1 Zichen Wen 1 Tianyi Wu 1 Yang Yang 1 Xiaorong Pu 1 2 Lifang He 3 Multi-view clustering has gained significant attention for integrating multi-view information in multimedia applications. With the growing complexity of graph data, multi-view graph clustering (MVGC) has become increasingly important. Existing methods primarily use Graph Neural Networks (GNNs) to encode structural and feature information, but applying GNNs within contrastive learning poses specific challenges, such as integrating graph data with node features and handling both homophilic and heterophilic graphs. To address these challenges, this paper introduces Node-Guided Contrastive Encoding (NGCE), a novel MVGC approach that leverages node features to guide embedding generation. NGCE enhances compatibility with GNN filtering, effectively integrates homophilic and heterophilic information, and strengthens contrastive learning across views. Extensive experiments demonstrate its robust performance on six homophilic and heterophilic multi-view benchmark datasets. 1. Introduction Multi-view clustering (MVC) is an important unsupervised learning task in the machine learning community (Li et al., 2014; Xu et al., 2022; Yin et al., 2020; Tan et al., 2023; Zhang et al., 2023; Liu et al., 2020; Chao et al., 2024; Guo & Ye, 2019; Zhang et al., 2024; Liu et al., 2019; Wu et al., 2020). It is designed to integrate diverse and complementary information across multiple views for clustering(Ke et al., 2024; Huang et al., 2021; Ren et al., 2024a; Wu et al., 2024; Ren et al., 2024b). Recent advances in unsupervised learning for graph data have increasingly emphasized the 1School of Computer Science and Engineering, University of Electronic Science and Technology of China 2Shenzhen Institute for Advanced Study, University of Electronic Science and Technology of China 3Department of Computer Science and Engineering, Lehigh University. Correspondence to: Junlong Ke . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). distinction between the data associated with the ego node and that of its edges, which delineate relationships with neighboring nodes (Tang et al., 2022; Xiao et al., 2022; Chen et al., 2022). However, some methods overly emphasize edge and neighborhood data, often neglecting the intrinsic information of the ego node. This imbalance has led to suboptimal performance on certain datasets, sometimes even underperforming compared to Multi-Layer Perceptrons (MLPs) (Tang et al., 2022; Xiao et al., 2022). On the other hand, some approaches rely heavily on the ego node s attributes to maintain model performance after multiple layers of processing in Graph Neural Networks (GNNs). While this preserves some robustness, it limits the ability to leverage contextual information from neighboring nodes, leading to subpar results on specific datasets (Chen et al., 2022). In the realm of graph contrastive learning for multi-view graph clustering (MVGC), capturing node differences and diversity within graphs plays a crucial role in designing effective contrastive learning methodologies (Hassani & Khasahmadi, 2020). Traditional approaches have largely relied on the homophily assumption (Mc Pherson et al., 2001), but more recent studies have begun to explore the effectiveness of these methods in the context of heterophilic graphs. For example, COMPLETER (Contrastive Prediction for Incomplete Multi-view Clustering) (Lin et al., 2021) has been introduced for heterophilic graphs. It integrates reconstruction, cross-view contrastive learning, and cross-view dual prediction to simultaneously achieve data recovery and consistency learning for incomplete multi-view data. (Hassani & Khasahmadi, 2020) proposes contrastive multi-view representation learning on graphs (MVGRL) by performing representation learning by contrasting two diffusion matrices transformed from the adjacency matrix. HGRL (Chen et al., 2022) leverages the preservation of original node features and the capturing of non-local neighbors to enhance node representations on heterophilic graphs. GREET (Liu et al., 2023) distinguishes between homophilic and heterophilic edges, employing low-pass and high-pass filters to seize the corresponding information. Building upon generation methods, DSSL (Xiao et al., 2022) decouples diverse patterns in local neighborhood distribution to capture both homophilic and heterophilic information. NWR-GAE (Tang et al., 2022) underscores the significance of topological structure Multi-View Graph Clustering via Node-Guided Contrastive Encoding within graphs, reconstructing neighborhoods based on local structure and features. MUSE (Yuan et al., 2023) innovates by creating two subviews on the original view, applying an information fusion controller for view utilization. However, these methodologies either lack efficacy in heterophilic graphs or underperform in homophilic scenarios. We attribute this limitation to an explicit differentiation between homophilic and heterophilic information through decoupling or view reconstruction, or an overemphasis on one aspect over the other. This leads to an inadequate interaction between homophilic and heterophilic information during contrastive analysis. A more natural approach would be to integrate both aspects within a unified representation framework, raising the key question: How can we encode both types of information without compromising either? Our approach involves guiding view feature encoding with node features, followed by adaptive weighting that accounts for graph homophily and joint information aggregation using a similarity matrix derived from refined node embeddings. Recognizing the strengths of traditional GNNs in graph structure mining, as well as the challenges posed by GNN transformations, we propose encoding the original graph structure with GNNs while integrating node features to adapt representations for traditional GNNs. This approach eliminates the need to explicitly distinguish between homophilic and heterophilic graph components. We begin by constructing a node correlation matrix and employing pseudo-labeling to quantify the homophily of the original adjacency matrix, thereby allocating weights to selectively preserve original structural information. Additionally, we introduce a random mask to the original node feature information, generating an additively noisy feature matrix. Subsequently, we leverage GNNs message passing and neighborhood aggregation mechanisms to recover the noisy features, with a contrastive loss applied between the recovered and original feature representations. Finally, we utilize inter-view contrastive learning to foster complementary learning across views. The main contributions of this paper can be summarized as follows. We propose NGCE, a novel contrastive learning model for homophilic and heterophilic graphs. To the best of our knowledge, NGCE is the first graph contrastive learning method guided by the principle that homophilic and heterophilic information in graph data should not be isolated, but rather processed within a unified node-guided framework to preserve its interactive essence in MVGC. We develop an effective joint process for MVGC that integrates several key components: an edge and node embedding similarity matrix sensitive to graph homophily, a contrastive learning-guided graph encoding mechanism driven by the recovery of noise-enhanced node features, and a contrastive fusion mechanism across views. Together, these elements form a comprehensive process for node-guided multi-view co-coding. We conduct extensive experiments on six homophilic and heterophilic benchmark datasets to evaluate the performance of NGCE. Our results show that the proposed NGCE adeptly accommodates both homophilic and heterophilic datasets within the multi-view graph clustering domain, achieving state-of-the-art performance metrics. 2. Related Work 2.1. Multi-View Graph Clustering With the rapid progress of GNNs, there has been a growing interest among researchers in leveraging graph structural information for multi-view clustering (MVGC). By combining with graph convolutional encoders, MVGC aims to learn the graph s underlying semantic information and divide the nodes into different clusters. In recent years, a plethora of methods addressing MVGC have surfaced. O2MAC (Fan et al., 2020) can be credited as a pioneer in applying GNNs to MVGC. Their approach involves encoding multi-view graphs into a lower-dimensional space using a single-view graph convolutional encoder and a multi-view graph structure decoder. Building on this foundation, (Cheng et al., 2020) introduces a novel design of two-pathway graph encoders, facilitating the mapping of graph embedding features and the acquisition of view-consistent information. Meanwhile, (Hassani & Khasahmadi, 2020) proposes an innovative GNN-based solution specifically tailored for multiview self-supervised learning, aiming to acquire both node and graph-level representations. In a similar vein, (Pan & Kang, 2021) applies contrastive learning techniques to unearth shared geometric and semantic patterns, enhancing the learning of a consensus graph. (Xia et al., 2022b) takes a systematic approach, exploring cluster structures through a graph convolutional encoder trained to learn the self-expression coefficient matrix. However, a common challenge across these methods lies in their high sensitivity to the homophily of graphs. In other words, when these models are applied to graphs exhibiting strong heterophily, their effectiveness may fall short of expectations. 2.2. Contrastive Learning Contrastive learning has garnered considerable attention in the realm of unsupervised learning, showcasing remarkable performance across a multitude of tasks, particularly on images (Hjelm et al., 2018; Chen et al., 2020b; Grill et al., 2020; Zbontar et al., 2021; Zhong et al., 2021; Yang et al., 2024) and graphs (Velickovic et al., 2019; You et al., 2020; Zhu et al., 2020; Xia et al., 2022a; Bielak et al., 2022; Hu et al., 2021). The underlying principle involves maximizing Multi-View Graph Clustering via Node-Guided Contrastive Encoding the similarity among positive pairs while minimizing the distance between negative pairs, as outlined by (Hadsell et al., 2006). In general, positive pairs consist of augmented versions of the same instance, while pairs involving different instances are considered negatives. Various loss functions have been introduced for this purpose, including the triplet loss (Chopra et al., 2005), noise contrastive estimation (NCE) loss (Gutmann & Hyv arinen, 2010), and normalized temperature-scaled cross-entropy loss (NT-Xent) (Chen et al., 2020a). Presently, numerous research endeavors leverage contrastive learning methods in the realm of clustering tasks, significantly contributing to the progress of clustering and unsupervised representation learning. (Zhong et al., 2020) suggests converting the maximization of mutual information into minimizing contrast loss, resulting in notable enhancements when applying contrast learning to diminish within-class variability. However, it fails to accommodate the prevalent graph data. (Hassani & Khasahmadi, 2020) proposes contrastive multi-view representation learning on graphs (MVGRL) by performing representation learning by contrasting two diffusion matrices transformed from the adjacency matrix. However, it relies heavily on data augmentation techniques and primarily utilizes graph structure information, specifically the adjacency matrix. We leverage node feature information to guide graph-contrastive encoding, thereby obtaining enhanced graph representations, harmonizing homophilic and heterophilic information in MVGC, and preserving their interactive essence. As shown in Fig. 1, the proposed approach consists of two components: i) Node-Guided Contrastive Encoding Module: This module introduces an innovative encoding strategy that achieves graph joint encoding across different views, guided by the node s original features. This approach comprises two key modules: an adaptive weighted sum based on similarity graphs and contrastive learning, implemented after reconstructing node features through GNN under the influence of additive noise. The former is tailored to integrate node features into the joint encoding output adaptively, whereas the latter aims to guide the encoding process by enabling the joint encoding to exhibit restorative properties that are based on the noise-augmented original node features. This process is designed to ensure that the integrity of homophilic and heterophilic information exchange within the graph data is preserved, thereby yielding a universally applicable output encoding; ii) Cross-View Contrast Module: To enhance the learning of view consistency and complementarity, we adapt the centroids of clusters and the k-nearest neighbor samples as additional positive and negative samples, respectively, significantly improving the efficacy of view integration. 3.1. Node-Guided Contrastive Encoding Module 3.1.1. GRAPH JOINT PROCESS To enhance the proposed framework s capability for effective neighborhood integration, it is critical to ensure that a substantial portion of nodes adjacent to the encoded embeddings share the same class as the node in focus since the original node features will be unavailable for later processing. Considering that nodes within the same category have comparable feature vectors due to identical labels, our methodology prioritizes exploiting this feature similarity for joint graph encoding that combines edge and node relationships. Initially, we employ autoencoders to distill and refine the feature vectors of nodes within the graph: Zn = f n(σ(X; θn)), (1) Xn pred = gn(σ(Zn; φn)), (2) where, for the n-th view, Zn refers to the node embeddings, θn and φn refers to the encoder and decoder s learnable parameters that are unshared, respectively, X refers to the original node features, while σ( ) refers to the activation function. Subsequently, to uncover node correlations, we evaluate the cosine similarity between node features, leading to the formulation of the correlation matrix Sn for the n-th view: Sn = Sim(Zn, Zn T), (3) where the function Sim( , ) refers to cosine similarity in vector space. Using Sn as the basis for encoding the graph does improve its homophily to some extent. However, focusing solely on node features while ignoring the inherent structural aspects of the graph may not optimally serve the encoding goals. Therefore, we argue for a strategic inclusion of the structural nuances of the original graph in our encoding efforts to ensure an increased level of homophily. Considering that the higher homophily degree indicates that the similarity information on the graph is dominant, and vice versa for the dissimilarity information, this involves the development of an adaptive graph encoding strategy where we aim to evaluate and weigh the degree of homophily of the original graph. Naturally, we utilize the homophily degree to assign weights to b An and Sn. However, the homophily degree of a given graph is difficult to measure due to the lack of true labels in the unsupervised scenario. Thus, we propose to learn the pseudo-label information (Arazo et al., 2020) using the consensus embedding H of Eq. (12) of the last iteration, and use the pseudo-label information and adjacency relationship information to approximate the homophily weight ωn for Multi-View Graph Clustering via Node-Guided Contrastive Encoding Joint Process 𝐀𝑛= 𝝎𝑛𝐀𝑛+ 𝐒𝑛 Contrastive (Add Noise By Mask ) Node-Guided Contrastive Encoding Module Node-Guided Contrastive Encoding Module Node-Guided Contrastive Encoding Module Contrastive Cross-View Contrast Module 𝐀 Node-Guided Graph Embedding ഥ𝐗 Recovered Node 𝐗 Original Node Features 𝐀 Original Adjacency Matrix 𝐙 Refined Node Features ഥ𝐘 Pseudo-Label 𝐇 k-means ഥ𝒀 𝝎𝑛= 𝐓 𝐈𝑖,𝑗) σ𝑖,𝑗( 𝐀𝑖,𝑗 Cross-view Fusion 𝝎 Homophily Wight 𝐡 Node-Guided Joint Embedding 𝐇 Consensus Embedding Figure 1. The illustration of NGCE graph clustering framework. The inputs to each view are the node feature matrix X and the adjacency matrix A of views. The output of the framework is the consensus embedding H, after which H is used as input for K-means clustering. Dashed arrows represent the backward propagation. the n-th view: i,j(An i,j Yi YT j Ii,j) P i,j(An i,j Ii,j) , (4) where, ωn refers to the weight reflective of the homophily degree from the preceding iteration, with indicating the Hadamard product. Y {0, 1}N c representing the onehot encoded pseudo labels from the clustering of the consensus embedding H, where N refers to the number of nodes and c refers to the number of clusters. I refers to the identity matrix. Ultimately, this leads to graph joint encoded embeddings that integrate both the homophily assessment and the original structural framework: b An = Sn + ωn An, (5) In order to guide the autoencoder f n( ), we apply reconstruction loss on Xn pred and X as follows: Ln Rec = l(gn(Xn pred; X) = l(gn(σ(f n(σ(X; θn)); φn)); X), (6) where l( ; ) denotes loss function. 3.1.2. NODE-GUIDED CONTRASTIVE ENCODING Furthermore, we design the node-guided optimization process, aiming to harness the true features of nodes to steer the encoding phase within the joint process. Inspired by the Denoising Autoencoder (DAE) (Vincent et al., 2008) and Graph MAE (Hou et al., 2022), we replaced the encoder and decoder in DAE with parameter-free GNNs. In this adaptation, node features are treated as denoising coding objects, which facilitates the training of the node-guided graph embedding b A. In this process, b A works analogous to the parameters of the encoder and decoder. More precisely, we actively encourage the post-encoding aggregation procedure, guided by the encoded graph, to exhibit restorative capabilities towards the original features after noise is masked. This strategy, as substantiated by subsequent experimental validations, proves to be compatible with both homophilic and heterophilic graph data. Through such a design, we not only maintain the interaction between homophilic and heterophilic information within the graph data but also guide an effective excavation of this information. This methodology underscores the importance of preserving the intrinsic data structure while facilitating a nuanced representation of the graph data, accommodating both similarity and diversity within the graph structure. Multi-View Graph Clustering via Node-Guided Contrastive Encoding We first add a random mask as noise to the original matrix X of node features. In practice, we randomly drop some node features in the graph with a probability of pc by adding a random mask as noise to the original matrix X. Specifically, we sample a binary masking matrix E {0, 1}N d from the Bernoulli distribution with a probability of (1 pc), i.e., eij Bernoulli (1 pc), i, j {1, , N}, and perform element-wise multiplication with the node features matrix: e X = X E, (7) where denotes the Hadamard product, X represents the original node features and e X refers to the node features with random mask (noisy features). Then, we utilize GCN s neighborhood aggregation mechanism to recover the feature information following the addition of noise, which can be represented from X e X X n, where X n = GCN( b An, e X) refers to the node features recovered by the aggregation mechanism of parameter-free GCN: X n = GCN( b An, e X), which can be formulated as: H(0) = e X, H(l+1) = σ D 1 2 H(l) + e X, (8) where H(l) refers to the output of l-th layer while D refers to the degree matrix of b A. Residual connections have been integrated to avoid potential gradient problems. The depth of our GCN is defined in terms of order, which is set as a fixed hyperparameter in our framework. We force the n-th view s recovered node features matrix X n to be equal to the original node features X as formulated: Ln C1 = 1 N 2 X (X n X)2 = 1 N 2 X j (X n ij Xij)2 j ((X n ij e Xij) (Xij e Xij))2 = 1 N 2 ( X j I1 ij(X n ij e Xij)2 j I0 ij(X n ij Xij)2), (9) where, the I0 ij and I1 ij functions work as indicators, with the former being 1 when Eij = 1 and 0 otherwise, and the latter vice versa. Specifically, in Eq. (9), the first term minimizes the agreement between the reconstructed node features and the features of the masked nodes in the encoder graph, since there is no valid information correlation in these nodes; the second term drives the GCN with graph joint encoding embeddings to retain critical unmasked node information. In other words, given the variation component as X n ij e Xij, we can reformulate the loss function such that: Ln C1 = 1 N 2 ( X j I1 ij(X n ij e Xij)2 j I0 ij((X n ij e Xij) Xij)2), (10) where the first term minimizes the change in features when node features are preserved, while the second term forces nodes to conform to their original features when node features are not preserved, essentially establishing an adaptive mechanism for selecting positive and negative samples. This involves selecting unpreserved features as positive samples and the variation component of preserved features themselves as negative samples. For the n-th view, considering the dimensionality of nodeguided graph embedding b An is excessively large for crossview fusion processes, we employ parameter-free GCNs once as a dimensionality reduction mechanism. This process results in compressed representations hn that are dimensionally aligned with node features Zn: hn = GCN( b An, Zn), (11) we obtain node-guided joint embeddings hn for the n-th view. 3.2. Cross-View Contrast Module In the context of multi-view tasks, it is acknowledged that distinct views offer varying levels of information, characterized by both consistency and complementarity (Jia et al., 2020; Wu et al., 2019). Initially, we obtain node-guided embeddings hn for n-th view. To optimally leverage this diverse yet complementary information across views, our approach aims to synthesize a consensus embedding rich in information by amalgamating the node-guided joint embeddings hn from each individual view (Jia et al., 2022). Acknowledging the differential informational value across views necessitates the implementation of a weighting mechanism, predicated on an assessment of each view s informational quality. This approach is designed to ensure that contributions from various views to the overarching consensus embedding are proportionate to their evaluated significance. In essence, a view s embedding that exhibits a high degree of similarity to the consensus embedding H is deemed to possess significant information, thereby warranting a higher weight allocation, and the inverse applies for lesser similarity. The consensus embedding H is derived as follows: n=1 ωn hhn, (12) where ωn h represents the weight allocated to the node-guided Multi-View Graph Clustering via Node-Guided Contrastive Encoding joint embeddings of the n-th view, determined through: ωn h = ( evan max (eva1, eva2, , eva V ))ρ. (13) This weight, ωn h, is deduced from an evaluation function designed to measure the congruence between the consensus embedding H and each view s embedding hn, i.e., evan = evaluation(hn, H). The hyperparameter ρ facilitates the modulation of view weights, either enhancing or diminishing their impact. To culminate, the consensus embedding H undergoes k-means clustering to ascertain the clustering outcomes. Inspired by the methodologies outlined in (Yuan et al., 2023; Hassani & Khasahmadi, 2020), we refine the joint node representation H by integrating a contrastive loss component. This component aims to further enhance the learning of view complementarity reflected in H. We employ a variant of NT-Xent loss (Chen et al., 2020a) as our contrastive learning loss function. Crucially, for an embedding hi of node i within the consensus embedding H, we select the embeddings of its Nknn nearest neighbor nodes within H as positive samples. For the selection of negative samples, we adopt a method inspired by that described in (Chao et al., 2024). Specifically, after a predetermined number of training epochs, once the cluster centers are considered sufficiently stable to represent high-confidence and disparate class prototypes, we opt for representations from cluster centers of different classes (Ck where k = ci), which are identified from H in the previous iteration and represent high-confidence, disparate class prototypes. This process can be formulated as follows: LC2 = A1[t T1] m=1 e S(hi,pi,m). m=1 e S(hi,pi,m) + k=1 1[k =ci]e S(hi,Ck) ! (14) where hi refers to the feature of node i within the consensus embedding H. pi,m refers to the feature representation of the m-th nearest neighbor of node i, drawn from P(hi), which denotes the set of Nknn nearest neighbors of hi in H. Nknn indicates the number of nearest neighbors selected, which is fixed at 20 in our implementation. S( , ) denotes the cosine similarity function. N refers to the total number of nodes, while c represents the total number of clusters. A denotes the loss coefficient, and we have empirically set its value to 1 N(c 1) based on experimental results. The indicator function 1[ ] yields a value of 1 if the condition specified in [ ] is met, and 0 otherwise. Herein, t denotes the current training epoch and T1 is a predefined hyperparameter; this loss component is thereby activated only when t T1. Ck represents the cluster center for the k-th class, and ci Datasets Clusters Nodes Features Graphs (hr) ACM 3 3025 1870 G1(0.82), G2(0.64) DBLP 4 4057 334 G1(0.80), G2(0.67), G3(0.32) IMDB 3 4780 1232 G1(0.48), G2(0.62), G3(0.40) Texas 5 183 1703 G1(0.09), G2(0.09) Chameleon 5 2277 2325 G1(0.23), G2(0.23) Wisconsin 5 251 1703 G1(0.19), G2(0.19) Table 1. The statistics information of the six graph datasets. (hr) refers to the homophily rate calculated by true labels. indicates the class assigned to node i. Both Ck and ci are derived from the k-means clustering result applied to H from a previous iteration. 3.3. Model Optimization Learning from previous deep clustering studies (Ren et al., 2025), we design the overall optimization objective for NGCE, which can be formulated as: Ltotal = LC + LRec n=1 Ln C1 + Ln C2 + Ln Rec. (15) Ltotal can be divided into two main sections: unsupervised node embedding learning and contrastive learning. The unsupervised node embedding learning component LRec drives the optimization of refined node features Z. The contrastive learning component LC includes the loss functions LC1 and LC2. In the former, we apply noisy masks as perturbations to node features to drive the optimization of node-guided graph embedding b A, while in the latter, we utilize contrastive learning based on cluster centroids to optimize cross-view consensus embedding H. For the sake of cross-dataset robustness and simplicity in hyperparameter tuning, we omitted the weight of the loss term. 4. Experiments 4.1. Experimental Setup 4.1.1. DATASETS. We used six MVG datasets. The homophilous graph datasets include: ACM (Fan et al., 2020), DBLP (Fan et al., 2020) and IMDB (Fan et al., 2020). The heterophilous graph datasets include: Texas, Chameleon (Rozemberczki et al., 2021) and Wisconsin (Pei et al., 2020). In the appendix, we provide details and sources of the datasets. Table 1 summarizes the statistics of these six datasets. The top half of the table shows the statistics information for the homophilous graph datasets and the bottom half shows the Multi-View Graph Clustering via Node-Guided Contrastive Encoding Methods ACM (hr 0.82 & 0.64) DBLP (hr 0.80 & 0.67 & 0.32) IMDB (hr 0.48 & 0.62 & 0.40) NMI% ARI% ACC% F1% NMI% ARI% ACC% F1% NMI% ARI% ACC% F1% VGAE (2016) 49.1 54.4 82.2 82.3 69.3 74.1 88.6 87.4 0.4 0.9 44.2 35.7 DAEGC (2019) 63.8 70.1 89.0 88.9 30.8 33.4 66.5 65.6 0.6 1.0 37.9 35.3 AGE (2020) 73.5 78.9 92.4 92.4 45.0 47.6 75.3 74.6 4.4 4.6 43.2 42.2 O2MAC (2020) 69.2 73.9 90.4 90.5 72.9 77.8 90.7 90.1 0.3 0.2 40.2 35.4 Mv AGC (2020) 67.4 72.1 89.8 89.9 77.2 82.8 92.8 92.3 1.3 -1.8 48.5 28.2 AGCN (2021) 68.4 74.2 90.6 90.6 39.7 42.5 73.3 72.8 0.3 1.4 54.5 31.1 MCGC (2021) 71.3 76.3 91.5 91.6 83.0 77.5 93.0 92.5 5.2 10.3 58.3 38.8 DCRN (2022) 71.6 77.6 91.9 91.9 49.0 53.6 79.7 79.3 0.2 0.1 53.4 25.5 Dua LGR (2023) 73.2 79.4 92.7 92.7 75.5 81.7 92.4 91.8 6.2 12.5 52.0 44.7 CMGEC (2023) 69.1 72.3 90.9 90.7 72.4 78.6 91.0 90.4 5.1 4.7 48.4 51.0 BMGC (2024) 78.4 83.3 94.1 94.2 80.1 85.4 94.0 93.6 5.5 4.9 44.1 40.7 SMVC (2024) 72.4 78.0 92.3 92.0 76.1 81.6 92.4 92.0 8.0 7.2 41.3 37.2 VGMGC (2025) 76.3 81.9 93.6 93.6 78.3 83.7 93.2 92.7 0.8 3.2 52.6 32.8 NGCE (ours) 80.5 85.0 94.7 94.8 79.1 84.0 93.3 92.8 5.6 12.7 54.6 43.4 Table 2. The results of clustering on homophilous graph datasets. We express all evaluative metrics as percentages. The best and runner-up results are highlighted with bold and underline, respectively. Methods Texas (hr 0.09 & 0.09) Chameleon (hr 0.23 & 0.23) Wisconsin (hr 0.19 & 0.19) NMI% ARI% ACC% F1% NMI% ARI% ACC% F1% NMI% ARI% ACC% F1% VGAE (2016) 12.7 21.7 55.3 29.5 15.1 12.4 35.4 29.6 10.5 13.7 49.3 34.1 DAEGC (2019) 6.4 2.6 31.7 25.0 9.1 5.6 32.2 31.2 10.6 3.4 32.7 28.3 AGE (2020) 7.5 7.3 36.6 36.6 8.6 7.6 32.4 32.4 9.3 1.3 31.1 31.1 O2MAC (2020) 8.7 14.6 46.7 29.1 12.3 8.9 33.5 28.6 11.0 8.9 40.0 27.9 Mv AGC (2020) 5.4 1.1 54.3 19.8 10.8 3.3 29.2 24.3 8.1 4.8 47.7 20.6 AGCN (2021) 15.4 18.1 61.8 43.0 6.7 6.1 32.5 20.4 6.4 6.8 49.8 24.9 MCGC (2021) 12.7 12.9 51.9 32.5 9.5 5.9 30.0 19.1 12.9 5.9 51.8 30.7 DCRN (2022) 10.7 15.1 55.2 27.6 8.7 5.7 30.9 21.9 10.8 16.0 50.2 34.1 Dua LGR (2023) 32.6 26.0 54.3 46.4 19.5 16.0 41.1 37.7 34.1 28.8 56.4 47.1 BMGC (2024) 29.1 15.8 42.5 38.3 9.4 5.9 30.8 30.7 34.0 24.3 51.5 40.8 VGMGC (2025) 35.4 26.0 55.2 46.9 22.4 13.4 40.1 39.5 41.6 34.8 56.6 49.6 NGCE (ours) 47.8 54.9 77.6 46.2 22.6 19.3 42.2 38.4 46.8 46.4 73.7 47.2 Table 3. The results of clustering on heterophilous graph datasets. We express all evaluative metrics as percentages. The best and runner-up results are highlighted in bold and underlined, respectively. statistics information for the heterophilous graph datasets. 4.1.2. BASELINES. Several baselines are replicated for comparison with our model. VGAE (Kipf & Welling, 2016) is a classical singleview clustering method. O2MAC (Fan et al., 2020) is the method that learns from both node features and graphs. Mv AGC (Lin & Kang, 2021) and MCGC (Pan & Kang, 2021) are two methods based on graph filters to learn a consensus graph for clustering. Dual GR (Ling et al., 2023) leverages soft-label and pseudo-label to guide the graph refinement and fusion process for clustering. CMGEC (Wang et al., 2023) utilizes a multi-graph attention fusion encoder with multi-view mutual information maximization mod- ule, aiming to encode more complementary information from multiple views and depict data more comprehensively. BMGC (Shen et al., 2024) addresses view imbalance in multi-relational graphs by proposing unsupervised dominant view mining and dual signals guided representation learning. SMVC (Chen et al., 2024) is a structural deep multi-view clustering method that integrates top-level abstraction with underlying details to jointly optimize cluster assignments and feature embeddings. VGMGC (Chen et al., 2025) introduces a variational graph generator to infer a reliable consensus graph from multiple graphs, aiming to better utilize both view-specific and view-common information for multiview graph clustering. For all baselines, we use results reported in the original papers or conduct experiments with their default parameters setting. Multi-View Graph Clustering via Node-Guided Contrastive Encoding Methods ACM (hr 0.82 & 0.64) Chameleon (hr 0.23 & 0.23) NMI% ARI% ACC% F1% NMI% ARI% ACC% F1% NGCE (w/o LRec) 62.4 65.9 86.8 86.9 15.5 11.6 33.0 27.7 NGCE (w/o LC) 75.6 81.3 92.1 92.2 20.4 17.5 39.9 35.9 NGCE (w/o LC1) 79.4 83.9 93.7 93.7 20.7 17.5 40.2 36.4 NGCE (w/o LC2) 77.6 83.1 93.4 93.5 21.4 18.7 41.3 37.2 NGCE (w/o b A) 61.6 67.6 80.5 81.3 10.1 6.6 27.1 20.3 NGCE (w/o Z) 68.3 71.4 89.8 90.9 17.1 15.6 35.1 31.3 NGCE 80.5 85.0 94.7 94.8 22.6 19.3 42.2 38.4 Table 4. The ablation study results of NGCE on ACM and Chameleon. The original results are shown in bold. w/o b A denotes the replacement of joint process with adjacency matrix A, w/o Z denotes the replacement of refined node features Z with an affinity matrix built directly form original node features. 1 3 5 7 9 11 13 15 order ACM Chameleon 1 3 5 7 9 11 13 15 order ACM Chameleon Figure 2. Parameter sensitivity analysis for order in homophilic graph dataset ACM and heterophilic graph dataset Chameleon. 4.2. Overall Results Tables 2 and 3 present the results of all compared methods on homophilous and heterophilous graph datasets, from which we have the following observations. The experimental results on the homophilous graph dataset ACM demonstrate that NGCE is highly competitive with the state-of-theart models, with significant improvements in all evaluative metrics. Moreover, NGCE outperforms other baselines on heterophilous graph datasets, i.e., Chameleon and Wisconsin, which are typically challenging for most models. Our model performs better than the baselines in most of the evaluation metrics. 4.3. Ablation Studies and Analysis 4.3.1. EFFECT OF EACH LOSS. To understand the importance of the unsupervised node embeddings learning and contrastive learning, we removed reconstruction loss LRec of the autoencoder and the contrastive learning loss LC for the proposed NGCE. Furthermore, for the contrastive learning loss LC, ablation study were also separately carried out on its components LC1 and LC2. We removed each loss in homophilic graph dataset ACM and heterophilic graph dataset Chameleon, respectively, to observe the changes in performance. The experimental results are shown in Table 4. As can be seen, LRec and LC dominate the total losses in terms of their impact on model performance. To be more precise, the effects of LC1 and LC2 on the performance metrics of the model are remarkably similar. Implementing these components resulted in an increase in ACC of 1.0% and 1.3% on the ACM dataset and 2.0% and 1.0% on the Chameleon dataset. Furthermore, in comparison, LRec has a more significant impact on the performance metrics than LC, with the application of these losses leading to an increase in ACC of 7.9% and 2.6% on the ACM dataset, and 9.1% and 2.2% on the Chameleon dataset, respectively. Through ablation studies, we have validated the effectiveness of LC1 and LC2 within the contrastive learning loss LC. This confirms the efficacy of our dual contrastive learning designs: contrasting the noised node features processed by GCN with their original node features, and cross-view node-guided joint embeddings comparisons. These results demonstrate the effectiveness of including specific contrasting elements to improve the model s ability to generate strong and meaningful embeddings, utilizing both intra-view effectiveness and inter-view complementarity. 4.3.2. EFFECT OF EACH COMPONENT. Since our model is based on the notion of node-guided graph encoding and consensus embedding encoding, the fusion of graph in joint process and node representation, along with the refinement of node features, is a critical component of NGCE. We conducted ablation studies to evaluate their impact on the performance of NGCE, where we replace graph joint encoded embeddings b A with adjacency matrix A, refined node features Z with affinity matrix based on original features X, the results are detailed in Table 4. The results highlight that omitting either joint processing or node feature refinement significantly degrades model performance. Specifically, compared to the baseline model, the ablation experiments on the ACM dataset resulted in a decrease in ACC of 14.2% and 4.9%; on the Chameleon dataset, there was a decrease in ACC of 15.1% and 7.1%. Results high- Multi-View Graph Clustering via Node-Guided Contrastive Encoding light the importance of refining node features, which leads to a meaningful representation in the node-guided process and validates the effectiveness of the strategy involving the weighted sum of the refined representational node similarity matrix with the original adjacency matrix. 4.3.3. PARAMETER SENSITIVITY ANALYSIS The sensitivity analysis for order hyperparameter in homophilic graph dataset ACM and heterophilic graph dataset Chameleon is shown in Fig. 2. From the spatial domain point of view, order controls the aggregation order of the graph filter. The higher order enables nodes to aggregate information from more distant ones, while nodes can only access feature information of closer nodes in lower orders. As can be seen from Fig. 2, ACC on the ACM dataset exhibits fluctuations around 93% when order is less than 11, while on the Chameleon dataset ACC remains relatively stable at around 42% under the same condition. However, for order values greater than 15, there is a sharp deterioration in the performance metrics for both datasets, dropping to around 45.4% for ACM and around 26.1% for Chameleon. 4.3.4. COMPLEXITY ANALYSIS Let M be the representation of the maximum number of neurons embedded within the hidden layers of the autoencoder, Z denotes the maximum dimensionality of the embedding features, T denotes the number of iterations for the outer loop, and K, V , and N stand for the numbers of clusters, views, and examples respectively. The K-means and target distribution computations are performed once per outer loop iteration. Thus, their contribution to the total complexity will be T O(NZK), giving us O(TNZK). The autoencoder adheres to a complexity of O(TNV M 2). The operation of GNN can be seen as multiplying an N N matrix by an N Z matrix, with a time complexity of O(N 2Z) in an iteration. So, the total time complexity of our method is O(TN(ZK + V M 2 + NZ)). The complexity of NGCE can be further reduced by methods such as anchor graphs, but this is beyond the scope of the proposed method. 4.3.5. LIMITATION OF THE PROPOSED METHOD Despite its demonstrated capabilities, NGCE depends on the availability of node features. This characteristic inherently limits its direct applicability to datasets where such features are sparse or absent. However, we posit that this limitation is not insurmountable and presents avenues for future investigation. To extend our approach to such scenarios, one potential strategy involves leveraging random walks to generate node embeddings. This technique, frequently employed in established graph embedding methods like Deep Walk (Perozzi et al., 2014), can capture structural information from the graph, thereby creating feature representations for nodes. Another promising direction is developing and integrating a dedicated encoder that learns node features directly from the graph s topology. This would allow the model to function effectively even without pre-defined node attributes, significantly broadening its applicability. We plan to explore these extensions in future work. 5. Conclusion In this paper, we introduce a novel graph-based contrastive learning framework for multi-view graph clustering, which effectively handles both homophily and heterophily in graph data. We propose that the homophilic and heterophilic information of graph data should not be separated, but rather processed within a unified framework to preserve its interactive essence. Consequently, we have developed a multi-view graph clustering method called Node-Guided Contrastive Encoding (NGCE). Specifically, NGCE includes a joint process that integrates a graph and node embedding similarity matrix sensitive to the degree of graph homophily, a contrastive learning-guided graph encoding mechanism driven by the degree of recovery of noise-enhanced node features, and a mechanism for contrastive fusion across views. The experimental results show that the proposed NGCE adeptly accommodates both homophilic and heterophilic datasets within the multi-view graph clustering domain, achieving state-of-the-art performance metrics. Acknowledgements This work was supported in part by National Key Research and Development Program of China (Nos. 2024YFC2310800 and 2024YFC2310801), National Natural Science Foundation of China (No. 62476052), Sichuan Science and Technology Program (No. 2024NSFSC1473), and Central Guidance for Local Science and Technology Development Fund Projects (No. 2024ZYD0268). Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. Arazo, E., Ortego, D., Albert, P., O Connor, N. E., and Mc Guinness, K. Pseudo-labeling and confirmation bias in deep semi-supervised learning. In IJCNN, pp. 1 8, 2020. Bielak, P., Kajdanowicz, T., and Chawla, N. V. Graph barlow twins: A self-supervised representation learning Multi-View Graph Clustering via Node-Guided Contrastive Encoding framework for graphs. KBS, 256:109631, 2022. Chao, G., Jiang, Y., and Chu, D. Incomplete contrastive multi-view clustering with high-confidence guiding. In AAAI, pp. 11221 11229, 2024. Chen, B., Xu, S., Xu, H., Bian, X., Guo, N., Xu, X., Hua, X., and Zhou, T. Structural deep multi-view clustering with integrated abstraction and detail. Neural Netw., 175 (C), July 2024. ISSN 0893-6080. Chen, J., Zhu, G., Qi, Y., Yuan, C., and Huang, Y. Towards self-supervised learning on graphs with heterophily. In CIKM, pp. 201 211, 2022. Chen, J., Ling, Y., Xu, J., Ren, Y., Huang, S., Pu, X., Hao, Z., Yu, P. S., and He, L. Variational graph generator for multiview graph clustering. TNNLS, pp. 1 14, 2025. Chen, M., Wei, Z., Huang, Z., Ding, B., and Li, Y. Simple and deep graph convolutional networks. In ICML, pp. 1725 1735. PMLR, 2020a. Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. A simple framework for contrastive learning of visual representations. In ICML, pp. 1597 1607. PMLR, 2020b. Cheng, J., Wang, Q., Tao, Z., Xie, D., and Gao, Q. Multiview attribute graph convolution networks for clustering. In IJCAI, pp. 2973 2979, 2020. Chopra, S., Hadsell, R., and Le Cun, Y. Learning a similarity metric discriminatively, with application to face verification. In CVPR, pp. 539 546, 2005. Cui, G., Zhou, J., Yang, C., and Liu, Z. Adaptive graph encoder for attributed graph embedding. In KDD, pp. 976 985, 2020. Fan, S., Wang, X., Shi, C., Lu, E., Lin, K., and Wang, B. One2multi graph autoencoder for multi-view graph clustering. In WWW, pp. 3070 3076, 2020. Grill, J.-B., Strub, F., Altch e, F., Tallec, C., Richemond, P., Buchatskaya, E., Doersch, C., Avila Pires, B., Guo, Z., Gheshlaghi Azar, M., et al. Bootstrap your own latent-a new approach to self-supervised learning. Neur IPS, 33: 21271 21284, 2020. Guo, J. and Ye, J. Anchors bring ease: An embarrassingly simple approach to partial multi-view clustering. In AAAI, pp. 118 125, 2019. Gutmann, M. and Hyv arinen, A. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In AISTATS, pp. 297 304. JMLR Workshop and Conference Proceedings, 2010. Hadsell, R., Chopra, S., and Le Cun, Y. Dimensionality reduction by learning an invariant mapping. In CVPR, pp. 1735 1742, 2006. Hassani, K. and Khasahmadi, A. H. Contrastive multi-view representation learning on graphs. In ICML, pp. 4116 4126. PMLR, 2020. Hjelm, R. D., Fedorov, A., Lavoie-Marchildon, S., Grewal, K., Bachman, P., Trischler, A., and Bengio, Y. Learning deep representations by mutual information estimation and maximization. In ICLR, 2018. Hou, Z., Liu, X., Cen, Y., Dong, Y., Yang, H., Wang, C., and Tang, J. Graphmae: Self-supervised masked graph autoencoders. In KDD, pp. 594 604, 2022. Hu, Y., You, H., Wang, Z., Wang, Z., Zhou, E., and Gao, Y. Graph-mlp: node classification without message passing in graph. ar Xiv preprint ar Xiv:2106.04051, 2021. Huang, Z., Ren, Y., Pu, X., and He, L. Non-linear fusion for self-paced multi-view clustering. In ACM MM, pp. 3211 3219, 2021. Jia, X., Jing, X.-Y., Zhu, X., Chen, S., Du, B., Cai, Z., He, Z., and Yue, D. Semi-supervised multi-view deep discriminant representation learning. TPAMI, 43(7):2496 2509, 2020. Jia, X., Jing, X.-Y., Zhu, X., Cai, Z., and Hu, C.-H. Coembedding: a semi-supervised multi-view representation learning approach. Neural Computing and Applications, pp. 1 21, 2022. Ke, J., Wen, Z., Yang, Y., Cui, C., Ren, Y., Pu, X., and He, L. Integrating vision-language semantic graphs in multi-view clustering. In IJCAI, pp. 4273 4281, 2024. Kipf, T. N. and Welling, M. Variational graph auto-encoders. ar Xiv preprint ar Xiv:1611.07308, 2016. Li, S.-Y., Jiang, Y., and Zhou, Z.-H. Partial multi-view clustering. In AAAI, 2014. Lin, Y., Gou, Y., Liu, Z., Li, B., Lv, J., and Peng, X. Completer: Incomplete multi-view clustering via contrastive prediction. In CVPR, pp. 11174 11183, 2021. Lin, Z. and Kang, Z. Graph filter-based multi-view attributed graph clustering. In IJCAI, pp. 2723 2729, 2021. Ling, Y., Chen, J., Ren, Y., Pu, X., Xu, J., Zhu, X., and He, L. Dual label-guided graph refinement for multi-view graph clustering. In AAAI, pp. 8791 8798, 2023. Liu, J., Cao, F., Gao, X.-Z., Yu, L., and Liang, J. A clusterweighted kernel k-means method for multi-view clustering. In AAAI, pp. 4860 4867, 2020. Multi-View Graph Clustering via Node-Guided Contrastive Encoding Liu, X., Zhu, X., Li, M., Tang, C., Zhu, E., Yin, J., and Gao, W. Efficient and effective incomplete multi-view clustering. In AAAI, pp. 4392 4399, 2019. Liu, Y., Tu, W., Zhou, S., Liu, X., Song, L., Yang, X., and Zhu, E. Deep graph clustering via dual correlation reduction. In AAAI, pp. 7603 7611, 2022. Liu, Y., Zheng, Y., Zhang, D., Lee, V., and Pan, S. Beyond smoothing: Unsupervised graph representation learning with edge heterophily discriminating. In AAAI, pp. 4516 4524, 2023. Mc Pherson, M., Smith-Lovin, L., and Cook, J. M. Birds of a feather: Homophily in social networks. Annual review of sociology, 27(1):415 444, 2001. Pan, E. and Kang, Z. Multi-view contrastive graph clustering. Neur IPS, 34:2148 2159, 2021. Pei, H., Wei, B., Chang, K. C. C., Lei, Y., and Yang, B. Geom-gcn: Geometric graph convolutional networks. In ICLR, 2020. Peng, Z., Liu, H., Jia, Y., and Hou, J. Attention-driven graph clustering network. In ACM MM, pp. 935 943, 2021. Perozzi, B., Al-Rfou, R., and Skiena, S. Deepwalk: online learning of social representations. In KDD, pp. 701 710, 2014. Ren, Y., Chen, X., Xu, J., Pu, J., Huang, Y., Pu, X., Zhu, C., Zhu, X., Hao, Z., and He, L. A novel federated multiview clustering method for unaligned and incomplete data fusion. Information Fusion, 108:102357, 2024a. Ren, Y., Pu, J., Cui, C., Zheng, Y., Chen, X., Pu, X., and He, L. Dynamic weighted graph fusion for deep multi-view clustering. In IJCAI, pp. 4842 4850, 2024b. Ren, Y., Pu, J., Yang, Z., Xu, J., Li, G., Pu, X., Yu, P. S., and He, L. Deep clustering: A comprehensive survey. TNNLS, 36(4):5858 5878, 2025. Rozemberczki, B., Allen, C., and Sarkar, R. Multi-scale attributed node embedding. J. Complex Netw., 9(2): cnab014, 2021. Shen, Z., He, H., and Kang, Z. Balanced multi-relational graph clustering. In ACM MM, pp. 4120 -4128, 2024. Tan, Y., Liu, Y., Wu, H., Lv, J., and Huang, S. Metric multi-view graph clustering. In AAAI, pp. 9962 9970, 2023. Tang, M., Li, P., and Yang, C. Graph auto-encoder via neighborhood wasserstein reconstruction. In ICLR, 2022. Velickovic, P., Fedus, W., Hamilton, W. L., Li o, P., Bengio, Y., and Hjelm, R. D. Deep graph infomax. ICLR, 2(3):4, 2019. Vincent, P., Larochelle, H., Bengio, Y., and Manzagol, P.-A. Extracting and composing robust features with denoising autoencoders. In ICML, pp. 1096 1103, 2008. Wang, C., Pan, S., Hu, R., Long, G., Jiang, J., and Zhang, C. Attributed graph clustering: A deep attentional embedding approach. In IJCAI, pp. 3670 3676, 2019. Wang, Y., Chang, D., Fu, Z., and Zhao, Y. Consistent multiple graph embedding for multi-view clustering. TMM, 25:1008 1018, 2023. Wu, F., Jing, X.-Y., Zhou, J., Ji, Y., Lan, C., Huang, Q., and Wang, R. Semi-supervised multi-view individual and sharable feature learning for webpage classification. In WWW, pp. 3349 3355, 2019. Wu, J., Xie, X., Nie, L., Lin, Z., and Zha, H. Unified graph and low-rank tensor learning for multi-view clustering. In AAAI, pp. 6388 6395, 2020. Wu, S., Zheng, Y., Ren, Y., He, J., Pu, X., Huang, S., Hao, Z., and He, L. Self-weighted contrastive fusion for deep multi-view clustering. TMM, 2024. Xia, J., Wu, L., Chen, J., Hu, B., and Li, S. Z. Simgrace: A simple framework for graph contrastive learning without data augmentation. In WWW, pp. 1070 1079, 2022a. Xia, W., Wang, S., Yang, M., Gao, Q., Han, J., and Gao, X. Multi-view graph embedding clustering network: Joint self-supervision and block diagonal representation. Neural Networks, 145:1 9, 2022b. Xiao, T., Chen, Z., Guo, Z., Zhuang, Z., and Wang, S. Decoupled self-supervised learning for graphs. Neur IPS, 35:620 634, 2022. Xu, J., Li, C., Ren, Y., Peng, L., Mo, Y., Shi, X., and Zhu, X. Deep incomplete multi-view clustering via mining cluster complementarity. In AAAI, pp. 8761 8769, 2022. Yang, X., Hu, X., Zhou, S., Liu, X., and Zhu, E. Interpolation-based contrastive learning for few-label semi-supervised learning. TNNLS, 35(2):2054 2065, 2024. Yin, M., Huang, W., and Gao, J. Shared generative latent representation learning for multi-view clustering. In AAAI, pp. 6688 6695, 2020. You, Y., Chen, T., Sui, Y., Chen, T., Wang, Z., and Shen, Y. Graph contrastive learning with augmentations. Neur IPS, 33:5812 5823, 2020. Multi-View Graph Clustering via Node-Guided Contrastive Encoding Yuan, M., Chen, M., and Li, X. Muse: Multi-view contrastive learning for heterophilic graphs. In ACM CIKM, pp. 3094 3103, 2023. Zbontar, J., Jing, L., Misra, I., Le Cun, Y., and Deny, S. Barlow twins: Self-supervised learning via redundancy reduction. ICML, pp. 12310 12320, 2021. Zhang, C., Li, H., Lv, W., Huang, Z., Gao, Y., and Chen, C. Enhanced tensor low-rank and sparse representation recovery for incomplete multi-view clustering. In AAAI, pp. 11174 11182, 2023. Zhang, C., Jia, X., Li, Z., Chen, C., and Li, H. Learning cluster-wise anchors for multi-view clustering. In AAAI, pp. 16696 16704, 2024. Zhong, H., Chen, C., Jin, Z., and Hua, X.-S. Deep robust clustering by contrastive learning. ar Xiv preprint ar Xiv:2008.03030, 2020. Zhong, H., Wu, J., Chen, C., Huang, J., Deng, M., Nie, L., Lin, Z., and Hua, X.-S. Graph contrastive clustering. In ICCV, pp. 9224 9233, 2021. Zhu, Y., Xu, Y., Yu, F., Liu, Q., Wu, S., and Wang, L. Deep graph contrastive representation learning. ar Xiv preprint ar Xiv:2006.04131, 2020. Multi-View Graph Clustering via Node-Guided Contrastive Encoding A. Details of Experiments A.1. Metrics Following previous works, four commonly used metrics, i.e., normalized mutual information (NMI), adjusted rand index (ARI), accuracy (ACC), and F1 scores (F1), are adopted to evaluate the clustering performance. A.2. Visualization of learned consensus embedding To unveil the inherent clustering structure, we have also employed t-SNE algorithm for the visualization of all six datasets to depict the distribution of the learned consensus embedding H. As depicted in Fig. 3, the visual results highlight that NGCE exhibits an enhanced clustering structure. A.3. Details of Datasets The experiments were conducted on six MVG datasets: ACM, DBLP, IMDB, Texas, Chameleon, and Wisconsin. Details and sources of the datasets are presented subsequently. Homophilous graph datasets: ACM is a paper network from the ACM database1 and is composed of two graphs: the co-paper network and the co-subject network; DBLP is a network sourced from the DBLP database2, consists of three graphs: co-author, coconference, and co-term; IMDB is a movie network that originates from the IMDB dataset3, including graphs of both co-actors and co-directors. Heterophilous graph datasets: Texas is a webpage graph from Web KB4; Chameleon is a subset of the Wikipedia network5; Wisconsin is webpage graph from Web KB6. Since Texas, Chameleon and Wisconsin are single-view graph data, we copy the graph as the second view. B. Algorithm The proposed NGCE model is summarized in Algorithm 1. Algorithm 1 presents a comprehensive breakdown of the code flow. C. Preliminaries In this section, we introduce a brief background of multiview graph clustering and GCNs. 1https://dl.acm.org/ 2https://dblp.uni-trier.de/ 3https://www.imdb.com/ 4http://www.cs.cmu.edu/afs/cs.cmu.edu/project/theo11/www/wwkb 5https://github.com/benedekrozemberczki/MUSAE/ 6http://www.cs.cmu.edu/afs/cs.cmu.edu/project/theo11/www/wwkb Algorithm 1 Multi-View Graph Clustering via Node Guided Contrastive Encoding (NGCE) Input: The adjacency matrices {An}V n=1 and node feature matrices {Xn}V n=1 Initialization phase Apply the k-means algorithm to original node feature matrices {Xn}V n=1 for the pseudo label matrix Y of the first iteration Train phase while not reaching the maximum iterations or not reaching exit condition do for n in 1, 2, ..., V do Calculate Zn and Xn pred based on Eq. (3) and (4) Calculate Sn based on Eq. (5) If it is the first iteration do Calculate ωn based on Eq. (6) with pseudo label matrix YT from Initialization phase Calculate ωn based on Eq. (6) with pseudo label matrix Y from last iteration Calculate b An and e Xn based on Eq. (7) and (9) Calculate X n by X n = GCN( b A, e X) based on Eq. (10) Calculate hn and ωn h based on Eq. (13) and (16) end for Update H based on Eq. (15) for v in 1, 2, ..., V do Update Y with H base on k-means end for end while Output: Consensus embedding H In the task of multi-view graph clustering, the objective is to group a set of n nodes into k clusters. To achieve this, we utilize G = (V, E) to denote a graph. Here, V represents the nodes set, and the set of all nodes belonging to class i is represented as Vi, with N = |V|, and E V V represents the edge set with selfloops. The feature matrix for the nodes is denoted as X RN d, and the symmetric adjacency matrix of the graph G is represented by A RN N, with elements aij = 1 indicating the presence of an edge between node i and node j, and aij = 0 otherwise. Additionally, we define the degree matrix of A as Dii = P j an ij, enabling the normalization of each view s A to e A = (D) 1A. The normalized graph Laplacian matrix is defined as e L = I e A, with I representing the identity matrix. The Graph Convolutional Networks (GCNs) leverages the principle of spectral graph theory to facilitate learning on graph-structured data. The core of this method is the utilization of the normalized graph Laplacian matrix to capture the graph s structural information. The GCNs operates by propagating node features across the graph s structure, thereby enabling each node to aggregate features from its neighbors, Multi-View Graph Clustering via Node-Guided Contrastive Encoding (a) ACM (b) Chameleon (c) DBLP (d) IMDB (e) Texas (f) Wisconsin Figure 3. Visualization of learned consensus embedding H in all six datasets. (a) epoch=200 (b) epoch=400 (c) epoch=600 (d) epoch=800 (e) epoch=250 (f) epoch=500 (g) epoch=750 (h) epoch=1000 Figure 4. Visualization of learned consensus embedding H in the ACM dataset (Figure 4(a), 4(b), 4(c) and 4(d)) and DBLP dataset (Figure 4(e), 4(f), 4(g) and 4(h)) across epochs. At the start of the training stage, the embedded features are non-separable. As the training progresses, the clustering structures of the embedded features become more apparent while their centroids gradually separate. effectively capturing both local and global graph structure. In this paper, we use GCN without learnable parameters : H(l+1) = σ D 1 2 H(l) , (16) 14 Multi-View Graph Clustering via Node-Guided Contrastive Encoding Notations Meaning G G = (V, E), graph with node set V and edge set E. X X RN d, the feature matrix of the nodes. c The number of clusters. A A RN N, the adjacency matrix with self-loop. D(n) Degree matrix in GCN. I Identity matrix. N N = |V|, the number of nodes. d The dimension of original node features. V The number of views. f n( ), gn( ) The encoder and decoder in the autoencoders. dl Dimension of the distilled node embeddings. dh Hidden layer dimension of the autoencoder. nl The number of hidden layers. Zn Distilled and refined node embeddings from f n( ). Xn pred Output of the decoder gn( ). Sn Node correlation matrix for the n-th view. S( , ), Sim( , ) Cosine similarity. ωn Homophily wight for the n-th view. Y Y {0, 1}N c, one-hot encoded pseudo labels. b An Graph joint encoded embeddings for the n-th view. E E {0, 1}N d, binary random masking matrix. pc Probability of Bernoulli distribution of random mask. e X Node features with random mask (noisy features). X n Recovered node features matrix for the n-th view. H(l) Output of l-th layer GCN. order Depth of GCN in NGCE. hn Node-guided joint embeddings of the n-th view. 1[ ], I0 ij, I1 ij Indicator function. Hadamard product. Ck Cluster center for the k-th class of H. ci The Class assigned to node i in Y. H Consensus embedding. ρ Cross-view weight modulation parameter ωn h Cross-view fusion wight allocated to hn. evan Function to measure congruence between H and hn. Ln Rec Reconstruction loss on Xn pred and X of the n-th view. LC1, LC2 Contrastive learning loss. η Learning rate. λwd Weight decay regularization parameter. Table 5. Overall notations.