# flexible_infinitewidth_graph_convolutional_neural_networks__655680c7.pdf Published in Transactions on Machine Learning Research (05/2025) Flexible Infinite-Width Graph Convolutional Neural Networks Ben Anson ben.anson@bristol.ac.uk School of Mathematics University of Bristol Bristol, United Kingdom Edward Milsom edward.milsom@bristol.ac.uk School of Mathematics University of Bristol Bristol, United Kingdom Laurence Aitchison laurence.aitchison@bristol.ac.uk School of Computer Science University of Bristol Bristol, United Kingdom Reviewed on Open Review: https: // openreview. net/ forum? id= Q2M4yij KSo A common theoretical approach to understanding neural networks is to take an infinitewidth limit, at which point the outputs become Gaussian process (GP) distributed. This is known as a neural network Gaussian process (NNGP). However, the NNGP kernel is fixed and tunable only through a small number of hyperparameters, thus eliminating the possibility of representation learning. This contrasts with finite-width NNs, which are often believed to perform well because they are able to flexibly learn representations for the task at hand. Thus, in simplifying NNs to make them theoretically tractable, NNGPs may eliminate precisely what makes them work well (representation learning). This motivated us to understand whether representation learning is necessary in a range of graph tasks. We develop a precise tool for this task, the graph convolutional deep kernel machine. This is very similar to an NNGP, in that it is an infinite width limit and uses kernels, but comes with a knob to control the amount of flexibility and hence representation learning. We found that representation learning gives noticeable performance improvements for heterophilous node classification tasks, but less so for homophilous node classification tasks. 1 Introduction A fundamental theoretical method for analyzing neural networks involves taking an infinite-width limit of a randomly initialized network. In this setting, the outputs converge to a Gaussian process (GP), and this GP is known as a neural network Gaussian process or NNGP (Neal, 1996; Lee et al., 2018; Matthews et al., 2018). NNGPs have been used to study various different neural network architectures, ranging from fullyconnected (Lee et al., 2018) to convolutional (Novak et al., 2018; Garriga-Alonso et al., 2018) and graph networks (Walker & Glocker, 2019; Niu et al., 2023). However, it is important to understand whether these NNGP results apply in practical neural network settings. One way of assessing the applicability of NNGP results is to look at performance. Specifically, Published in Transactions on Machine Learning Research (05/2025) if infinite-width NNGPs perform comparably with finite-width NNs, we can reasonably claim that infinitewidth NNGPs capture the essence of what makes finite-width NNs work so well. In contrast, if finite-width NNs perform better, that would indicate that NNGPs are missing something. Why might NNGPs underperform relative to NNs? One key property of the infinite-width NNGP limit is that it eliminates representation learning: the kernel of the GP is fixed, and cannot be tuned except through a very small number of hyperparameters. This fixed kernel makes it straightforward to analyze the behaviour of networks in the NNGP limit. However, the elegant fixed kernel comes at a cost. In particular, the top-layer representation/kernel is highly flexible in finite networks, and this flexibility is critical to the excellent performance of deep networks (Bengio et al., 2013; Le Cun et al., 2015). This indicates that in some cases, the infinite-width NNGP limit can throw the baby out with the bathwater : in trying to simplify the system to enable theoretical analysis, the NNGP limit can eliminate some of the most important properties of deep networks that lead to them performing well. In the setting of convolutional networks for CIFAR-10, this is precisely what was found: namely, that infinitewidth NNGPs underperform finite-width NNs (Adlam et al., 2023; Garriga-Alonso et al., 2018; Shankar et al., 2020), and it has been hypothesised that this difference arises due to the lack of representation learning in NNGP (Aitchison, 2020; Mac Kay et al., 1998). In this paper, we consider the question of NNGP performance in the Graph Convolutional Network (GCN) setting. Surprisingly, prior work has shown that the graph convolutional NNGP performs comparably to the GCN (Niu et al., 2023), perhaps suggesting that representation learning is not important in graph settings. However, our work indicates that this picture is incomplete. In particular, we show that while the graph convolutional NNGP is competitive on some datasets, it is much worse than GCNs on other datasets. This would suggest a hypothesis that representation learning in graph tasks is dataset dependent, with some datasets needing representation learning, and others not. Importantly, testing this hypothesis rigorously is difficult, as there are many differences between the infinitewidth, fixed representation graph convolutional NNGP and the finite-width, flexible representation GCN, not just that the GCN has representation learning. Perhaps the best approach to testing the hypothesis would be to develop a variant of the NNGP with representation learning, and to see how performance changed as we altered the amount of representation learning allowed by this new model. However, this is not possible with the traditional NNGP framework. Instead, we need to turn to recently developed deep kernel machines (DKMs) (Yang et al., 2023; Milsom et al., 2023). DKMs, like NNGPs, are obtained via an infinite-width limit, and in practice work entirely in terms of kernels. The key difference is that DKMs allow for representation learning, while NNGPs do not. Specifically, DKMs have a tunable parameter, ν, which controls the degree of representation learning. For small values of ν, DKM representations are highly flexible and learned from data. In contrast, as ν , flexibility is eliminated, and the DKM becomes equivalent to the NNGP. The graph convolutional DKM therefore forms a precise tool for studying the need for representation learning in graph tasks. Using the graph convolutional DKM, we examined the need for representation learning in a variety of graph tasks. Node classification tasks in particular exist on a spectrum from homophilous, where adjacent nodes in a graph tend to have similar labels, to heterophilous, where adjacent nodes tend to be dissimilar. We found that homophilous tasks tended not to require representation learning, while heterophilous tasks did require representation learning. Concretely, our contributions are as follows: We develop a graph convolutional variant of deep kernel machines (Section 4). We propose two scalable inducing-point approximation schemes for the graph convolutional DKM (Section 4.2, Appendix B). We analyze the graph convolutional DKM in the linear setting, and provide a closed-form solution for the learned representations (Section 5). Published in Transactions on Machine Learning Research (05/2025) By considering the performance of graph convolutional NNGPs relative to GCNs and by tuning ν in the graph convolutional DKM, we find that representation learning improves performance in heterophilous node classification tasks but not homophilous node classification tasks. 2 Related Work Graph convolutional networks are a type of graph neural network (Scarselli et al., 2008; Kipf & Welling, 2017; Bronstein et al., 2017; Velicković et al., 2017) originally introduced by Kipf & Welling (2017), motivated by a localised first-order approximation of spectral graph convolutions. They use the adjacency matrix to aggregate information from a node s neighbours at each layer of the neural network. When first published, graph convolutional networks outperformed existing approaches for semi-supervised node classification by a significant margin. Homophily, a measure of similarity between connected nodes in a graph, has been studied extensively in the context of GNNs (Luan et al., 2024; Zhu et al., 2020; 2021; Maurya et al., 2021; Platonov et al., 2023), mostly to point out that lack of homophily (meaning that connected nodes tend to be dissimilar) is detrimental to performance in GNNs. Here, we examine an adjacent but different matter: the relationship between homophily and representation learning. Infinite-width neural networks were first considered in the 1990s (Neal, 1996; Williams, 1996) where they only considered shallow networks at initialisation. Cho & Saul (2009) derived an equivalence between deep Re LU networks and arccosine kernels. This was later generalised to arbitrary activation functions by Lee et al. (2018), who used a Bayesian framework to show that any infinitely-wide deep neural network is equivalent to a Gaussian process, which they dubbed the neural network Gaussian process (NNGP) (Lee et al., 2020; 2018; Matthews et al., 2018). The NNGP has since been extended, for example to accommodate convolutions (Garriga-Alonso et al., 2018; Novak et al., 2018), graph structure (Niu et al., 2023; Hu et al., 2020a) and more (Yang, 2019). Using a similar infinite-width limit, but on SGD-trained networks rather than Bayesian networks, Jacot et al. (2018) studied the dynamics of neural network training under gradient descent, deriving the neural tangent kernel (NTK), which describes the evolution of the infinite-width network over time. The NTK suffers from an analogous problem to the NNGP s lack of representation learning, since the NTK limit implies that the intermediate features do not evolve over time (Yang & Hu, 2021; Bordelon & Pehlevan, 2023; Vyas et al., 2023). Alternative limits, such as the recent µ-P parameterisation (Yang & Hu, 2021) fix this problem by altering the scaling of parameters. However, the µ-P line of work only tells us that feature/representation learning will happen, not what the learned features/representations will be. In contrast the DKM framework gives us the exact learned representations. Applying fixed kernel functions to graphs is a well explored endeavour (Shervashidze & Borgwardt, 2009; Kashima et al., 2003; Shervashidze et al., 2009). Kernels can be defined directly for graphs and applied in a shallow fashion, shallow kernel can be stacked (Achten et al., 2023), or kernels can be used as a component of a larger deep learning algorithm (Cosmo et al., 2021; Yanardag & Vishwanathan, 2015). Our work differs fundamentally in that kernels themselves are learned, rather than learning features and then applying a fixed kernel. 3 Background We give a background on deep Gaussian processes (DGPs), and how under an infinite-width limit they become neural network Gaussian processes (NNGPs). We review how modifying this infinite-width limit gives flexible DKMs, in contrast to NNGPs which have fixed kernels. We also give an overview of the graph convolutional network (GCN) and the graph convolutional NNGP. These ingredients allow us to define a DKM in the graph domain, a so-called graph convolutional DKM , in Section 4. Published in Transactions on Machine Learning Research (05/2025) 3.1 Neural Networks as Deep Gaussian Processes Consider a fully-connected NN with inputs X RP ν0, outputs/labels Y RP νL+1, and intermediate-layer features Fℓ RP Nℓ. By P we refer to the number of datapoints, ν0 and νL+1 the number of input and output features respectively. The width of each layer is denoted Nℓ= Nνℓ, where νℓis fixed, and N allows us to scale the width of all layers simultaneously. We compute features at each layer using the previous layer via Fℓ= ϕ(Fℓ 1)Wℓ, where Wℓ RNℓ 1 Nℓare the weights and ϕ is the pointwise nonlinearity (e.g. Re LU). With a Gaussian prior on the weights, W ℓ ij N(0, 1 Nℓ 1 ), the conditional distribution of features is given by, P(Fℓ| Fℓ 1) = λ=1 N(f ℓ λ; 0, Kfeatures(Fℓ 1)), (1a) P(Y | FL) = λ=1 N(yλ; 0, Kfeatures(FL) + σ2I), (1b) where Kfeatures(X) = 1 Nℓϕ(X)ϕ(X)T , and the subscript λ refers to the λth feature/column, so that Fℓ= (f ℓ 1 f ℓ 2 f ℓ Nℓ). Eq. (1) is equivalent to a Bayesian NN, but in the above form it is usually called a deep Gaussian process (DGP) (Damianou & Lawrence, 2013; Salimbeni & Deisenroth, 2017), albeit with an unusual kernel. We can generalize Eq. (1) to allow alternative kernels Kfeatures( ) such as the squared exponential kernel. 3.2 Deep Gaussian Processes in Terms of Gram Matrices For certain kernel functions (e.g. arccosine kernels (Cho & Saul, 2009) and RBF), features are unnecessary, and inter-layer dependencies in a DGP can be summarised entirely by Gram matrices Gℓ= 1 NℓFℓ(Fℓ)T RP P . For these kernels, we are able to compute the kernel from the Gram matrix alone, and thus there is a function over Gram matrices K( ) : RP P RP P such that Kfeatures(Fℓ) = K(Gℓ) 1. We can therefore rewrite the DGP in Eq. (1) as, P(Fℓ| Gℓ 1) = λ=1 N(f ℓ λ; 0, K(Gℓ 1)), (2a) P(Y | GL) = λ=1 N(yλ; 0, K(GL) + σ2I), (2b) where all inter-layer dependencies in Eq. (2) are expressed using only Gram matrices. We can remove features altogether, and reframe the model purely in terms of Gram matrices (Aitchison et al., 2021), P(Gℓ| Gℓ 1) = W(Gℓ; Gℓ 1/Nℓ, Nℓ), (3a) P(Y | GL) = λ=1 N(yλ; 0, K(GL) + σ2I), (3b) where W(V, n) is the Wishart distribution with scale matrix V and degrees of freedom n (Gupta & Nagar, 2018). 3.3 Wide Deep Gaussian Processes Consider a deep Re LU network. Bayesian inference over weights/features is not tractable, but remarkably, there is a tractable closed form solution if we take an infinite width limit. To see this, first observe that we 1For avoidance of doubt, Kfeatures( ) : RP Nℓ RP P computes the kernel from the features, while K( ) : RP P RP P equivalently computes the same kernel from the corresponding Gram matrix. Published in Transactions on Machine Learning Research (05/2025) can calculate the kernel non-linearity without features; Cho & Saul (2009) showed that in the limit Nℓ , Re LU kernels are equivalent to arccosine kernels: ϕ = Re LU = lim Nℓ 1 Nℓϕ(Fℓ)ϕT (Fℓ) = Karccos(Gℓ). (4) Further, the arccosine kernel is easily computable, Karccos(G)ij = π (sin θij + (π θij) cos θij) , where cos θij = Gij p Gii Gjj . (5) Secondly, Gram matrices become deterministic in wide networks, since they can be understood as an average of IID features, Gℓ= 1 Nℓ PNℓ λ=1f ℓ λ(f ℓ λ)T . By Eq. (1) or Eq. (2), the expectation of Gℓconditioned on the previous layer is, E[Gℓ|Gℓ 1] = K(Gℓ 1), while the variance scales with 1/Nℓ. Thus, as we take an infinitewidth limit, Nℓ , the variance goes to zero, and the prior over the Gram matrix at layer ℓ, Gℓ, becomes deterministic and concentrated at its expectation, K(Gℓ 1). Thus, the kernel at every layer is a fixed deterministic function of the inputs that can be computed recursively, Knetwork(Fℓ) = (K K) | {z } ℓtimes ν0 XXT ), (6) where 1 ν0 XXT is the kernel for the inputs, and F0 = X. Setting ℓ= L, we see that Eq. (6) holds true at the output layer, implying that the outputs of a NN or DGP are GP distributed with a fixed kernel in the infinite-width limit. This is known as the neural network Gaussian process (NNGP) (Lee et al., 2018). 3.4 Deep Kernel Machines The clean analytic form of the NNGP kernel (Eq. 6) unfortunately has a big weakness: it underperforms finite-width counterparts empirically in many tasks (Aitchison, 2020; Pleiss & Cunningham, 2021). The NNGP kernel is a fixed function of inputs, meaning that no representation learning can occur, and the kernel cannot be shaped by the output labels. This is a big problem, since representation learning is widely understood to be central to the success of modern deep learning systems. DKMs (Yang et al., 2023; Milsom et al., 2023) are a solution to the lack of representation learning in infinitewidth NNGPs. DKMs can still be understood as an infinite-width limit of a DGP/NN, but the limit is altered to retain flexibility and allow learning of representations. In a DKM, Gram matrices are treated as learned parameters, rather than random variables as in a DGP or NNGP. Yang et al. (2023) derive the following DKM objective for fully-connected networks that can be optimized with respect to Gram matrices Gℓ, L(G1, . . . , GL) = log P(Y | GL) ℓ=1 νℓDKL N(0, Gℓ) || N(0, K(Gℓ 1)) . (7) Here, the likelihood term P(Y | GL) measures how well the final layer representation GL performs at the task. Under the derivation of Yang et al. (2023), P(Y | GL) is a likelihood for a Gaussian process distribution over the outputs with kernel K(GL). The KL divergence terms are a measure of how much the Gram matrices G1, . . . , GL deviate from the NNGP. This can be seen due to the fact that a negative KL divergence is maximized exactly when the two distributions it is measuring are equal, i.e. Gℓ= K(Gℓ 1). The DKM objective trades off the marginal likelihood against the KL divergence terms, with the KL terms acting as a regularizer towards the NNGP. The amount of regularization depends on the νℓcoefficients, and as we send νℓ , the DKM becomes equivalent to the standard infinite-width NNGP. As νℓbecomes smaller, we allow more flexibility in the Gram representations. Thus, through νℓ, the DKM gives us a knob to tune the amount of representation learning allowed in our model. 3.5 Graph Convolutional Networks and their Wide counterparts Kipf & Welling (2017) derived Graph Convolutional Networks (GCNs) by applying approximations to spectral methods on graphs. In a GCN layer, an adjacency matrix A P P is used to convolve (or mixup ) Published in Transactions on Machine Learning Research (05/2025) intermediate layer nodes. Mathematically, GCN post-activations are given by Hℓ= ϕ(ˆAHℓ 1Wℓ), where ˆA is a normalized version of A. Similarly to Section 3.2, it is possible to construct an NNGP for the graph domain. The pre-activations Fℓ of a graph convolution layer are Fℓ= ˆAHℓ 1Wℓ. Again, placing an IID Gaussian prior over our weights W ℓ µλ N(0, 1 Nℓ 1 ), the NNGP construction tells us that as we make the layers wide (Nℓ ), the preactivations become GP distributed with kernel matrix K = E[f ℓ λ(f ℓ λ)T ]. We show in Appendix C that this expectation has closed-form, K = ˆAΦℓ 1 ˆAT , (8) where Φℓ 1 = E[hℓ 1 λ (hℓ 1 λ )T ] is the NNGP kernel of a fully connected network (e.g. the arccosine kernel in the case of a Re LU network). Using this kernel in the NNGP defined in Eq. 6 gives the graph convolutional NNGP recursion, Gℓ= ˆAK(Gℓ 1) ˆAT = ˆAK( ˆAK( ˆAK( 1 ν0 XXT )ˆAT )ˆAT ) ˆAT . (9) GCNs and graph convolutional NNGPs can be applied easily to node classification problems, since features or nodes are modelled directly. They can also be used for graph classification by applying mean pooling at the output layer. In Sections 5 and 6 we consider homophily in graphs and its effect on graph convolutional NNGP performance. By homophily, we refer to edge homophily h [0, 1], which is calculated as the proportion of edges (j, k) E that are between nodes with the same class. In Section 3 we have seen how graph convolutional NNGP kernels can be calculated. We have also seen how DKMs add flexibility to fully-connected NNGPs. We now combine these two ideas to obtain a flexible infinite-width graph network the graph convolutional DKM and develop an inducing point scheme to allow training on large datasets. 4.1 Graph Convolutional Deep Kernel Machines The graph convolutional NNGP suffers from the fixed representation problem that plagues all NNGP models its kernel is a fixed transformation of the inputs, regardless of the target labels. This means that there is little flexibility (apart from kernel hyperparameters) to learn suitable features for the task at hand. To solve this problem we develop the graph convolutional DKM, which has learnable kernel representations at each layer. We arrive at a graph convolutional DKM by considering the Gram matrices in a DGP with graph mixups, P(Fℓ| Fℓ 1) = λ=1 N(f ℓ λ; 0, ˆAK(Gℓ 1)ˆAT ), (10) where Gℓ 1 = 1 Nℓ 1 Fℓ 1(Fℓ 1)T . By taking the representation learning limit of the graph convolutional DGP, we show in Appendix A that the posterior over Gram matrices is point distributed; in other words, the Gram matrices are deterministic. Moreover, the Gram matrices maximize the graph DKM objective, L(G1, . . . , GL) = log P(Y | GL) ℓ=1 νℓDKL N(0, Gℓ) || N(0, ˆAK(Gℓ 1)ˆAT ) . (11) The objective in Eq. (11) can be interpreted as a fully-connected DKM, but with a modified kernel. Again, just as in Section 3.4, we trade-off the likelihood with the KL divergence terms. That is, we trade-off Published in Transactions on Machine Learning Research (05/2025) the suitability of the top-layer representation for explaining the labels Y versus divergence from the graph convolutional NNGP. When νℓ in the DKM objective (Eq. 7) we recover the NNGP, and similarly for the graph DKM objective taking νℓ gives us the graph convolutional NNGP. Decreasing the coefficients νℓdecreases the importance of the KL divergence terms, giving Gram matrices flexibility to deviate from the graph convolutional NNGP. In our experiments, we treat the νℓas knobs for controlling the amount of representation learning. By measuring performance for different νℓ, we are able to cleanly demonstrate the importance of learned representations in various tasks. 4.2 Inducing-point schemes In practice it is computationally infeasible to optimize the Gram matrices G1, . . . , GL in the graph DKM objective for anything but modestly sized datasets, due to quadratic memory costs of storing the Gram matrices Gℓ, and the cubic time costs. The KL divergence terms, for example, are expensive because of matrix inversions and log determinants: DKL N(0,Gℓ) || N(0, ˆAK(Gℓ 1)ˆAT ) = 1 2 Tr((ˆAK(Gℓ 1)ˆAT ) 1Gℓ) log det((ˆAK(Gℓ 1)ˆAT ) 1Gℓ) P . (12) We resolve computational issues by developing an inducing point scheme and enabling linear scaling with dataset size. Specifically, we suppose that the graph nodes X RP ν0 can be summarized sufficiently well by a smaller inducing set Xi RPi ν0, where the size of the inducing set Pi is fixed and chosen to be much smaller than the full set of nodes (so that Pi P). These inducing inputs form an initial Gram matrix that we call G0 ii = 1 ν0 Xi XT i RPi Pi. We also have inducing Gram representations at each layer, G1 ii, . . . , GL ii, with all of these inducing representations being learned. The inducing representations are propagated alongside training/test Gram representations G1 tt, . . . , GL tt, so that, Gℓ= Gℓ ii Gℓ it Gℓ ti Gℓ tt The inducing and training/test points interact via the similarity matrix Gℓ ti = (Gℓ it)T RPt Pi, and Gℓ ti and Gℓ tt are predicted according to our inducing point scheme, in a similar way to how training/test features are predicted in an inducing point GP. We refer readers to Algorithm 1 for computational details. The form of the inducing Gram matrices depends on the choice of inducing point scheme. The modeller must make a choice as to how the inducing graph nodes interact with the training/test nodes. Broadly, we can either consider intra-domain inducing points, which belong to the same domain as the data, or inter-domain inducing points which do not. We consider one scheme of each type: an intra-domain scheme that assumes adjacency information between inducing points/nodes and training/test nodes, and an inducing scheme that treats the inducing points as unconnected nodes in a graph. We connect the inducing nodes to each other with an inducing adjacency matrix Aii RPi Pi, and connect the inducing nodes to the training/test nodes with the adjacency Ati RPt Pi. We assume that the adjacency between train/test points is provided in the dataset, which we denote Att for consistency. The two schemes are characterized by, (1) (intra-domain) sampling a random subset S = {s1, . . . , s Pi} of nodes from the dataset and treating these as inducing points. The inducing adjacencies become (Aii)jk = Asj,sk, (Ati)jk = Aj,sk; (2) (inter-domain) treating inducing nodes as independent of all other nodes, such that Aii = I, Ati = 0. We give forward propagation rules for each inducing-point scheme, and an ELBO objective (which is used for optimizing the inducing points and replaces the full-rank objective function from Eq. 11) in Appendix B.2. The computational cost of the intra-domain scheme is slightly higher, due to having to store Aii and Ati and perform matrix multiplications with them. However this cost is negligible compared to storing and computing with the Gram matrices themselves, and further mitigated by the fact that the adjacencies are sparse. Published in Transactions on Machine Learning Research (05/2025) initialization Figure 1: Normalized Gram matrices for a 2-layer linear graph DKM, providing verification of Eq. (14). We use input data, labels, and adjacency from a randomly selected 100 datapoint subset of Cora, and set the adjacency to be ˆA0.5. The leftmost and rightmost kernels are the input and label kernels respectively. The middle two columns are the hidden Gram representations. The top row shows Gram matrices after a Wishart initialization, the middle row shows the analytic solution, and the bottom row shows the results after training from this initialization by gradient descent with Adam. We empirically compare both schemes in Appendix G across several node classification dataset, finding that neither is definitively the best. However, for some datasets (e.g. Citeseer) the inter-domain scheme is markedly better. Given that intra-domain inducing points are not applicable in a multi-graph setting, and that the inter-domain scheme has a simpler implementation, the inter-domain is preferable overall. 5 Analysis of Representation Learning in Linear Graph Convolutional Deep Kernel Machines Remarkably, it is possible to study graph convolutional DKMs analytically in the linear kernel case, which gives insight into representation learning in graphs. Specifically, if we use a linear kernel (K(G) = G), set the regularization to 1 (νℓ= 1 for all ℓ), we show in Appendix D that the Gram matrix representations optimizing the graph convolutional DKM objective at each layer ℓare given by, Gℓ= ˆAℓ 1((ˆA LGL+1 ˆA L)(ˆAG0 ˆA) 1)ℓ/(L+1)(ˆAG0 ˆA)ˆAℓ 1. (14) Here G0 and GL+1 are the input and output kernels respectively (and are assumed to be known). We provide numerical verification of Eq. (14) by comparing it to a solution obtained via gradient descent in Figure 1. To ensure that the normalized adjacency is invertible, we compute Eq. (14) using ˆAλ in place of ˆA, which interpolates with the identity matrix via a parameter λ [0, 1], ˆAλ = λI + (1 λ)ˆA. (15) Published in Transactions on Machine Learning Research (05/2025) Figure 2: Final layer kernels for 2-layer linear graph convolutional DKMs (top row) and graph convolutional NNGPs (bottom row). The models are fit on a toy dataset with Gaussian random inputs and an Erdős Rényi adjacency (50 nodes, and edge probability 0.1). The label kernel is shown in the right column. The remaining columns show the effect of varying λ. = 1, CKA=0.53 = 10, CKA=0.51 = 1000, CKA=0.30 = , CKA=0.27 Roman Empire = 1, CKA=0.25 = 10, CKA=0.24 = 1000, CKA=0.23 = , CKA=0.20 Figure 3: Final layer kernels Cora and Roman Empire kernels for different regularizations ν. The kernels are formed from 400 nodes sampled at random, and the true label kernel is shown on the rightmost column. It is very natural to introduce this modified adjacency matrix in the graph setting, which incorporated by other prior work (Luan et al., 2020). In particular, with λ = 0, the graph mixup/convolution performs an equally weighted average over itself and all adjacent nodes. This can be problematic. For instance, consider a network consisting of two connected nodes. After a single graph mixup, the two nodes will have the same features, which is unlikely to be desirable. To fix this issue, we can use λ > 0, which implies that the graph mixup is now a weighted average, with more weight on self-features. This fixes the issue with our simple example network with two nodes: the resulting features are different even after a graph mixup. We sought to compare the linear graph convolutional DKM kernel to the linear graph convolutional NNGP kernel, Gℓ GCNNGP = ˆAℓG0 ˆAℓ. (16) While it is not practical to calculate Eq. (14) for large datasets (due to the inversion of a large adjacency matrix), a comparison is still possible with small datasets. We compare the two kernels on a randomly generated an Erdős-Rényi graph (Erdos & Rényi, 1963) dataset of 50 nodes (with Gaussian features), and insert edges between nodes with probability 0.1. Half of the nodes are given a positive label and the Published in Transactions on Machine Learning Research (05/2025) 10 1 101 103 Cora (h=0.81) 10 1 101 103 Pubmed (h=0.8) 10 1 101 103 98 Reddit (h=0.76) 10 1 101 103 Citeseer (h=0.74) 10 1 101 103 Minesweeper (h=0.68) 10 1 101 103 Arxiv (h=0.65) 10 1 101 103 Tolokers (h=0.6) 10 1 101 103 Penn94 (h=0.47) 10 1 101 103 Blog Catalog (h=0.4) 10 1 101 103 Amazon Ratings (h=0.38) 10 1 101 103 35 Chameleon (h=0.24) 10 1 101 103 Flickr (h=0.24) 10 1 101 103 Squirrel (h=0.22) 10 1 101 103 Roman Empire (h=0.05) Figure 4: Validation accuracy at different regularization strengths (ν) for each node classification dataset. The datasets are arranged by homophily ratio (denoted by h). The error bands denote 1 standard deviation in the validation accuracy over several random seeds. For each subplot, we ensure the y-axis range is at least 8, so that the effect of ν is comparable between dataset. remainder a negative label. We calculated the output layer kernels for λ {0.1, 0.3, 0.5, 1}, and depict results in Figure 2. Note that the Erdős-Rényi adjacency is not homophilous, because an edge is equally likely between both nodes with the same class and with a different class. We see in Figure 2 (bottom row) that none of the graph convolutional NNGP kernels are aligned to the output kernel, YYT . This is because the graph convolutional NNGP kernel is a deterministic function of the inputs and adjacency, and is not learned based on the labels, Y. In contrast, the DKM is well-aligned with the output (Figure 2, top row). Note we find that increasing λ improves alignment of the graph convolutional Published in Transactions on Machine Learning Research (05/2025) DKM kernels to the output kernel. As λ 0, the adjacency matrix becomes singular, and it seems that this also involves a loss of representation learning, as the graph structure dominates the representation. 6 Experiments In this section we investigate the effect of representation learning in graph networks by training graph convolutional DKMs on several datasets. We train on node classification datasets with varying homophily, as well as graph classification datasets. Our experiments show that representation learning helps to shape output kernels, and that datasets with lower homophily benefit more from representation learning than datasets with high homophily. We also compare the performance of graph convolutional DKMs to GCNs and graph convolutional NNGPs, and find that the extra flexibility afforded by the DKM framework closes the performance gap between graph convolutional NNGPs and GCNs. In all experiments we use the Re LU/arccosine kernel non-linearity, and set the regularization νℓ= ν such that it constant for all layers. 6.1 Deep Kernel Machines Allow Shaping of Kernel Representations Similarly to the fully-connected DKM (Yang et al. (2023), Section 4.6), we demonstrate in Figure 3 that the graph convolutional DKM shapes final layer kernel representations towards the label kernel, YYT . We show alignment both qualitatively via the top-layer kernels, and quantitatively with the Centered Kernel Alignment (CKA) (Cortes et al., 2012) metric. CKA is measured by, CKA(K1, K2) = Tr(K 1K 2) p Tr(K 1K 1) Tr(K 2K 2) , (17) where K 1/2 are centered version of the kernels K1/2 to be compared. Higher CKA values indicate greater similarity between representations, with CKA(K1, K2) [0, 1]. For this experiment, we trained graph convolutional DKMs for different values of ν {100, 101, 103, } on Cora and Roman Empire with the same random seed for 300 epochs. The models were trained using a simple 2-layer architecture with no residual connections or normalization/centering layers, and we used normalized adjacency ˆAλ=0.3 for the graph convolutions/mixups. The most suitable inducing point scheme was determined by a hyperparameter sweep over schemes for each dataset (see Section 6.2 and Appendix E for details). The kernels shown were obtained by taking the kernel representations of a randomly sampled 400 node subset of the last hidden layer, and the CKA similarity measure was calculated on the full final hidden layer kernel. When ν is 1 or 10, we clearly see block-diagonal structure mirroring YYT , indicating kernel alignment . As ν increases to 1000 or , flexibility decreases and the kernel alignment appears to become weaker, and is confirmed by the CKA statistics. Interestingly, this breakdown seems to involve all outputs becoming correlated, which is indicative of the rank-collapse-like phenomenon described by Oono & Suzuki (2019). 6.2 Investigating the Relationship Between Homophily and Representation Learning The graph convolutional NNGP kernel has a natural inductive bias for homophilous datasets; the graph convolutional NNGP pools representations from adjacent nodes, therefore in the extreme case that the connected components of the graph correspond to exactly the class labels, we expect its kernel to align with the labels. We expect the graph convolutional NNGP to be less effective for heterophilous datasets, because the graph structure is not informative for the labels. However, representation learning should help in the heterophilous case because we showed in Section 6.1 that increasing flexibility helps shape the kernel representations. We test the effect of representation learning empirically, training graph convolutional DKMs on several node classification benchmark datasets that exhibit varied levels of homophily. The dataset statistics can be found in Table 3. We trained with different regularization strengths, ν {0, 10 2, 10 1, 100, 101, 102, 103} Published in Transactions on Machine Learning Research (05/2025) Table 1: Mean test accuracies (%) one standard deviation on node classification datasets. Datasets are sorted by homophily, with most homophilous datasets at the top. We provide results for graph convolutional DKMs (GCDKM), graph convolutional NNGPs (GCNNGP) and GCNs. The best accuracy for each dataset is bolded. For datasets with multiple splits, the standard deviation is calculated using the means over splits. In some instances, the best graph convolutional DKM had ν = , which is denoted by ( ) (hence the sparse GCNNGP results are identical), and ( ) denotes graph convolutional NNGP results from Niu et al. (2023). GCDKM sparse GCNNGP GCNNGP GCN (no dropout) GCN Cora 81.1 0.2 80.8 0.1 82.8 80.6 0.2 81.1 0.3 Pubmed 79.8 0.2 79.8 0.2 79.6 79.4 0.2 79.5 0.1 Reddit 96.2 0.0 93.5 0.0 94.7 0.0 95.3 0.1 95.8 0.1 Citeseer 71.7 0.4 71.7 0.4 69.5 71.7 0.2 71.7 0.2 Minesweeper 85.9 0.4 84.3 0.3 85.6 0.4 85.6 0.4 Arxiv 70.8 0.2 68.9 0.2 70.1 0.1 70.0 0.1 70.6 0.2 Tolokers 81.4 0.4 79.0 0.1 81.6 0.7 82.2 0.7 Penn94 83.0 0.2 75.2 0.3 81.2 0.5 81.5 1.8 Blog Catalog 92.9 0.5 92.8 0.5 93.6 0.6 94.1 0.6 Amazon Ratings 49.2 0.5 46.4 0.3 48.5 0.5 49.8 0.6 Flickr 85.7 0.6 81.5 2.2 81.3 2.0 82.6 0.7 Chameleon 68.9 1.4 65.1 1.1 65.3 2.6 65.3 2.6 Squirrel 56.8 1.4 38.6 1.2 57.3 1.4 57.7 1.4 Roman Empire 79.2 0.4 73.6 0.3 79.9 0.5 82.8 0.5 Table 2: Average test accuracies (%) one standard deviation on graph classification datasets. Error bars are not statistically significant. GCN accuracies are sourced from Zhang et al. (2019)( ) and Yang et al. (2020) ( ). GCDKM sparse GCNNGP GCN Mutag 86.6 3.8 80.8 5.5 85.6 5.8 NCI1 75.3 1.0 66.5 1.0 76.3 1.8 NCI109 75.2 1.0 65.9 1.3 75.9 1.8 Proteins 72.2 1.2 67.6 1.4 75.2 3.6 Mutagenicity 79.1 0.7 72.7 1.1 79.8 1.6 to control the amount of representation learning. We also trained with two inducing point schemes (detailed in Section 4.2), and for each dataset we selected the optimal scheme by calculating mean validation accuracy for each ν and scheme. We used a 2-layer architecture, with the adjacency renormalization described by Kipf & Welling (2017), and no residual connections or normalization layers. Figure 4 illustrates the empirical differences between NNGPs and DKMs, which we find to be highly datasetdependent. For homophilous datasets (see top row of Figure 4) like Cora, performance tends to remain relatively stable or even improve with larger ν, suggesting that the standard architectural inductive biases are well-suited to these tasks. In contrast, heterophilous datasets such as Roman Empire benefit greatly from smaller ν (i.e. increased flexibility). This aligns with our theoretical understanding, as DKMs converge to NNGPs in the limit ν . The lack of representation learning in NNGPs becomes particularly problematic for small h (bottom row), leading to dramatic declines in accuracy of at least 10% when comparing ν= 103 to ν= 10 2. This suggests that for heterophilous tasks, where standard architectural inductive biases are less appropriate, the graph convolutional DKM s ability to learn representations allows it to compensate, while the graph convolution NNGP s inability to do so results in poor performance. Published in Transactions on Machine Learning Research (05/2025) 6.3 Final Results To assess overall performance of the graph convolutional DKM, we optimized hyperparameters for each dataset with a series of grid searches on a range of node and graph classification datasets. We kept the number of layers fixed at 2, and searched over inducing scheme, ν, architecture, kernel centering, and finally the number of inducing points. The final test accuracies are shown in Table 1 and Table 2. We provide details of the datasets and the training procedure in the Appendix E.2. In Table 1, we see that the graph convolutional DKM performs similarly to the graph convolutional NNGP for homophilous node classification datasets, but less so for heterophilous datasets. We also see that the graph convolutional DKM performs similarly to the GCN for most datasets, closing the performance gap between infinite-width and finite-width networks. We attribute the instances where the GCN outperforms the graph convolutional DKM to dropout, as test accuracy of the graph convolutional DKM is similar to the GCN without dropout. We also provide full-rank graph convolutional NNGP accuracies (where applicable) from Niu et al. (2023). Their graph convolutional NNGP results are slightly different ours, most likely because ours is a sparse model with learned inducing points, whereas theirs is treated like a conventional GP. In Table 2, we find that graph convolutional DKM performance is competitive with the GCN on graph classifications benchmarks, and uniformly better than the NNGP. The superior performance versus the NNGP is perhaps expected; the graph classification datasets are all molecule datasets, which tend to exhibit a mixture of homophily and heterophily (Ye et al., 2022), and thus the flexibility of the DKM ought to be very helpful. Overall, we find that the graph convolutional DKM closes the gap between the NNGP and the GCN, supporting the hypothesis that representation learning is the key element lacking from the NNGP. This appears to be true regardless of task (e.g., node classification or graph classification). Further, our results demonstrate that the NNGP is well-suited to homophilous tasks, thus explaining why it is possible to obtain strong NNGP performance on some benchmark graph tasks (Niu et al., 2023), despite NNGPs generally being considered inferior GCNs. 7 Limitations The key limitations of graph convolutional DKMs, like kernel methods in general, is scaling to large dataset sizes due to the cubic cost of naive methods. While we implemented an inducing point scheme to circumvent this limitation, it may be possible to improve on these methods. Our current analysis considers standard graph convolutional architectures. While it is theoretically possible to develop DKM extensions by kernelizing alternative graph architectures (Bo et al., 2021; Chien et al., 2020; Velicković et al., 2017), these extensions are non-trivial and beyond the scope of this work. We speculate that such extensions would improve performance across all methods (graph convolutional DKM/NNGP, and GCN) by incorporating additional inductive biases. The impact might be most pronounced for NNGPs, which lack representation learning capabilities and thus rely more heavily on built-in inductive biases. In contrast, graph convolutional DKMs and GCNs can compensate for imperfect inductive biases through representation learning. 8 Conclusion and Discussion By leveraging the DKM framework, we have developed a flexible variant of an infinite-width graph convolutional network. The graph convolutional DKM provides a hyperparameter that allows us to interpolate between a flexible kernel machine and the graph convolutional NNGP with fixed kernel. This feature of the DKM enabled us to examine the importance of representation learning on graph node classification tasks, and show competitive performance on graph classification tasks. Remarkably, we found that some tasks benefit more from representation learning than others. This finding helps explain the fact that NNGPs have previously been shown to perform well on graph tasks, but not in other domains. Moreover, the fact that our graph DKM model performs similarly to a GCN without dropout suggests that the kind of flexibility Published in Transactions on Machine Learning Research (05/2025) introduced by the DKM is the sensible kind, and may be used to advance kernel methods in the era of deep learning. Despite using an inducing point scheme, computational cost of the graph convolutional DKMs is still relatively high. We expect that improving computational efficiency can help improve performance, as it would allow for easier architecture exploration, and training of bigger models. We also expect that more careful optimization could boost performance even further since efficiency gains could enable larger models and longer training. However, we leave these issues for future work. Published in Transactions on Machine Learning Research (05/2025) Sonny Achten, Francesco Tonin, Panagiotis Patrinos, and Johan AK Suykens. Semi-supervised classification with graph convolutional kernel machines. ar Xiv preprint ar Xiv:2301.13764, 2023. Ben Adlam, Jaehoon Lee, Shreyas Padhy, Zachary Nado, and Jasper Snoek. Kernel regression with infinitewidth neural networks on millions of examples. ar Xiv preprint ar Xiv:2303.05420, 2023. Laurence Aitchison. Why bigger is not always better: on finite and infinite neural networks. In ICML, 2020. Laurence Aitchison, Adam Yang, and Sebastian W Ober. Deep kernel processes. In International Conference on Machine Learning. PMLR, 2021. Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8):1798 1828, 2013. Deyu Bo, Xiao Wang, Chuan Shi, and Huawei Shen. Beyond low-frequency information in graph convolutional networks. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pp. 3950 3957, 2021. Blake Bordelon and Cengiz Pehlevan. Self-consistent dynamical field theory of kernel evolution in wide neural networks. Journal of Statistical Mechanics: Theory and Experiment, 2023(11):114009, 2023. Michael M Bronstein, Joan Bruna, Yann Le Cun, Arthur Szlam, and Pierre Vandergheynst. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4):18 42, 2017. Eli Chien, Jianhao Peng, Pan Li, and Olgica Milenkovic. Adaptive universal generalized pagerank graph neural network. ar Xiv preprint ar Xiv:2006.07988, 2020. Youngmin Cho and Lawrence Saul. Kernel methods for deep learning. Advances in neural information processing systems, 2009. Corinna Cortes, Mehryar Mohri, and Afshin Rostamizadeh. Algorithms for learning kernels based on centered alignment. The Journal of Machine Learning Research, 13:795 828, 2012. Luca Cosmo, Giorgia Minello, Michael Bronstein, Emanuele Rodolà, Luca Rossi, and Andrea Torsello. Graph kernel neural networks. ar Xiv preprint ar Xiv:2112.07436, 2021. Andreas Damianou and Neil Lawrence. Deep gaussian processes. In Artificial Intelligence and Statistics, 2013. Paul Erdos and Alfréd Rényi. Asymmetric graphs. Acta Math. Acad. Sci. Hungar, 14(295-315):3, 1963. Matthias Fey and Jan E. Lenssen. Fast graph representation learning with Py Torch Geometric. In ICLR Workshop on Representation Learning on Graphs and Manifolds, 2019. Adrià Garriga-Alonso, Carl Edward Rasmussen, and Laurence Aitchison. Deep convolutional networks as shallow gaussian processes. ar Xiv preprint ar Xiv:1808.05587, 2018. Arjun K Gupta and Daya K Nagar. Matrix variate distributions. Chapman and Hall/CRC, 2018. Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. Advances in neural information processing systems, 2017. Jilin Hu, Jianbing Shen, Bin Yang, and Ling Shao. Infinitely wide graph convolutional networks: semisupervised learning via gaussian processes. ar Xiv preprint ar Xiv:2002.12168, 2020a. Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118 22133, 2020b. Published in Transactions on Machine Learning Research (05/2025) Arthur Jacot, Clément Hongler, and Franck Gabriel. Neural tangent kernel: Convergence and generalization in neural networks. In Neur IPS, pp. 8580 8589, 2018. Hisashi Kashima, Koji Tsuda, and Akihiro Inokuchi. Marginalized kernels between labeled graphs. In Proceedings of the 20th international conference on machine learning (ICML-03), pp. 321 328, 2003. Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. International Conference on Learning Representations, 2017. Yann Le Cun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. nature, 521(7553):436 444, 2015. Jaehoon Lee, Yasaman Bahri, Roman Novak, Samuel S Schoenholz, Jeffrey Pennington, and Jascha Sohl Dickstein. Deep neural networks as gaussian processes. ar Xiv preprint ar Xiv:1711.00165, 2017. Jaehoon Lee, Yasaman Bahri, Roman Novak, Samuel S Schoenholz, Jeffrey Pennington, and Jascha Sohl Dickstein. Deep neural networks as gaussian processes. International Conference on Learning Representations, 2018. Jaehoon Lee, Samuel Schoenholz, Jeffrey Pennington, Ben Adlam, Lechao Xiao, Roman Novak, and Jascha Sohl-Dickstein. Finite versus infinite neural networks: an empirical study. Advances in Neural Information Processing Systems, 33:15156 15172, 2020. Derek Lim, Felix Hohne, Xiuyu Li, Sijia Linda Huang, Vaishnavi Gupta, Omkar Bhalerao, and Ser Nam Lim. Large scale learning on non-homophilous graphs: New benchmarks and strong simple methods. Advances in neural information processing systems, 34:20887 20902, 2021. Sitao Luan, Mingde Zhao, Chenqing Hua, Xiao-Wen Chang, and Doina Precup. Complete the missing half: Augmenting aggregation filtering with diversification for graph convolutional networks. ar Xiv preprint ar Xiv:2008.08844, 2020. Sitao Luan, Chenqing Hua, Minkai Xu, Qincheng Lu, Jiaqi Zhu, Xiao-Wen Chang, Jie Fu, Jure Leskovec, and Doina Precup. When do graph neural networks help with node classification? investigating the homophily principle on node distinguishability. Advances in Neural Information Processing Systems, 36, 2024. David JC Mac Kay et al. Introduction to gaussian processes. NATO ASI series F computer and systems sciences, 168:133 166, 1998. Alexander G de G Matthews, Mark Rowland, Jiri Hron, Richard E Turner, and Zoubin Ghahramani. Gaussian process behaviour in wide deep neural networks. ar Xiv preprint ar Xiv:1804.11271, 2018. Sunil Kumar Maurya, Xin Liu, and Tsuyoshi Murata. Improving graph neural networks with simple architecture design. ar Xiv preprint ar Xiv:2105.07634, 2021. Edward Milsom, Ben Anson, and Laurence Aitchison. Convolutional deep kernel machines, 2023. R. M. Neal. Bayesian Learning for Neural Networks, Vol. 118 of Lecture Notes in Statistics. Springer-Verlag, 1996. Zehao Niu, Mihai Anitescu, and Jie Chen. Graph neural network-inspired kernels for gaussian processes in semi-supervised learning. ar Xiv preprint ar Xiv:2302.05828, 2023. In press. Roman Novak, Lechao Xiao, Jaehoon Lee, Yasaman Bahri, Greg Yang, Jiri Hron, Daniel A Abolafia, Jeffrey Pennington, and Jascha Sohl-Dickstein. Bayesian deep convolutional networks with many channels are gaussian processes. ar Xiv preprint ar Xiv:1810.05148, 2018. Kenta Oono and Taiji Suzuki. Graph neural networks exponentially lose expressive power for node classification. ar Xiv preprint ar Xiv:1905.10947, 2019. Oleg Platonov, Denis Kuznedelev, Michael Diskin, Artem Babenko, and Liudmila Prokhorenkova. A critical look at the evaluation of gnns under heterophily: are we really making progress? ar Xiv preprint ar Xiv:2302.11640, 2023. Published in Transactions on Machine Learning Research (05/2025) Geoff Pleiss and John P Cunningham. The limitations of large width in neural networks: A deep gaussian process perspective. Advances in Neural Information Processing Systems, 34:3349 3363, 2021. Hugh Salimbeni and Marc Deisenroth. Doubly stochastic variational inference for deep gaussian processes. Advances in neural information processing systems, 2017. Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. The graph neural network model. IEEE transactions on neural networks, 20(1):61 80, 2008. Vaishaal Shankar, Alex Fang, Wenshuo Guo, Sara Fridovich-Keil, Jonathan Ragan-Kelley, Ludwig Schmidt, and Benjamin Recht. Neural kernels without tangents. In International conference on machine learning, pp. 8614 8623. PMLR, 2020. Nino Shervashidze and Karsten Borgwardt. Fast subtree kernels on graphs. Advances in neural information processing systems, 22, 2009. Nino Shervashidze, SVN Vishwanathan, Tobias Petri, Kurt Mehlhorn, and Karsten Borgwardt. Efficient graphlet kernels for large graph comparison. In Artificial intelligence and statistics, pp. 488 495. PMLR, 2009. Petar Velicković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graph attention networks. ar Xiv preprint ar Xiv:1710.10903, 2017. Nikhil Vyas, Alexander Atanasov, Blake Bordelon, Depen Morwani, Sabarish Sainathan, and Cengiz Pehlevan. Feature-learning networks are consistent across widths at realistic scales. ar Xiv preprint ar Xiv:2305.18411, 2023. Ian Walker and Ben Glocker. Graph convolutional gaussian processes. In International Conference on Machine Learning, pp. 6495 6504. PMLR, 2019. Christopher Williams. Computing with infinite networks. In M.C. Mozer, M. Jordan, and T. Petsche (eds.), Advances in Neural Information Processing Systems, volume 9. MIT Press, 1996. URL https://proceedings.neurips.cc/paper_files/paper/1996/file/ ae5e3ce40e0404a45ecacaaf05e5f735-Paper.pdf. Pinar Yanardag and SVN Vishwanathan. Deep graph kernels. In Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, 2015. Adam X. Yang, Maxime Robeyns, Edward Milsom, Nandi Schoots, Ben Anson, and Laurence Aitchison. A theory of representation learning in deep neural networks gives a deep generalisation of kernel methods. International Conference on Machine Learning, 2023. In press. Greg Yang. Scaling limits of wide neural networks with weight sharing: Gaussian process behavior, gradient independence, and neural tangent kernel derivation. ar Xiv preprint ar Xiv:1902.04760, 2019. Greg Yang and Edward J. Hu. Tensor programs IV: feature learning in infinite-width neural networks. In ICML, volume 139 of Proceedings of Machine Learning Research, pp. 11727 11737. PMLR, 2021. Yiding Yang, Zunlei Feng, Mingli Song, and Xinchao Wang. Factorizable graph convolutional networks. Advances in Neural Information Processing Systems, 33:20286 20296, 2020. Wei Ye, Jiayi Yang, Sourav Medya, and Ambuj Singh. Incorporating heterophily into graph neural networks for graph classification. ar Xiv preprint ar Xiv:2203.07678, 2022. Zhen Zhang, Jiajun Bu, Martin Ester, Jianfeng Zhang, Chengwei Yao, Zhi Yu, and Can Wang. Hierarchical graph pooling with structure learning. ar Xiv preprint ar Xiv:1911.05954, 2019. Jiong Zhu, Yujun Yan, Lingxiao Zhao, Mark Heimann, Leman Akoglu, and Danai Koutra. Beyond homophily in graph neural networks: Current limitations and effective designs. Advances in neural information processing systems, 33:7793 7804, 2020. Published in Transactions on Machine Learning Research (05/2025) Jiong Zhu, Ryan A Rossi, Anup Rao, Tung Mai, Nedim Lipka, Nesreen K Ahmed, and Danai Koutra. Graph neural networks with heterophily. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pp. 11168 11176, 2021. Published in Transactions on Machine Learning Research (05/2025) A Full Rank Graph Convolutional Deep Kernel Machines Here, we outline a derivation for the graph convolutional DKM objective. We start with a DGP with graph mixups, P(Fℓ| Fℓ 1) = λ=1 N(f ℓ λ; 0, ˆAKfeatures(Fℓ 1) ˆA T ), (18) where f ℓ λ are columns of features, Fℓ= (f ℓ 1 f ℓ Nℓ). (19) We take P to be the number of datapoints, and Nℓthe number of features at each layer ℓ, so that Fℓ RP Nℓ. The log-posterior over Gram matrices Gℓ= 1 NℓFℓ(Fℓ)T is equivalent to, log P(G1, . . . , GL | X, Y) = log P(Y | GL) + ℓ=1 log P(Gℓ| Gℓ 1), (20) with G0 the being a kernel formed from the inputs X. Following Section 3.2, we assume that Kfeatures( ) is a kernel function that can be calculated in terms of Gram matrices, i.e., K such that K(Gℓ) = Kfeatures(Fℓ). Examples of suitable kernels are the arccosine kernel (Cho & Saul, 2009) or the squared exponential kernel. Since Fℓis multivariate Gaussian, Gℓis Wishart distributed by its definition, and we can write down its log density, log P(Gℓ|Gℓ 1) = Nℓ P 1 2 log|ˆAK(Gℓ 1)ˆAT | Nℓ 2 Tr{(ˆAK(Gℓ 1)ˆAT ) 1Gℓ} + αℓ. (21) The constant αℓis the normalizing constant, and we assume that ˆA is full rank. Yang et al. (2023) showed that αℓsatisfies the following scaling law as we increase width: lim Nℓ αℓ/Nℓ= const. Therefore if Nℓ= Nνℓ, it follows that, lim N 1 N log P(Gℓ|Gℓ 1) = const + νℓDKL N(0, Gℓ) || N(0, ˆAK(Gℓ 1)ˆAT ) . (22) The Bayesian representation learning limit (Yang et al., 2023) dictates that we scale the final layer alongside the hidden layers (this is unlike the usual NNGP limit (Lee et al., 2017) which only scales the intermediate layers). To do so, we duplicate the outputs so that there are N copies of the labels, i.e. Y = (Y Y). Since channels are assumed to be i.i.d., we have, P( Y | GL) = P(Y | GL)N. (23) We now consider the posterior of this network with wide hidden layers, wide outputs, and labels Y. Combining Eqs. (20), (22), and (23), and letting Nℓ= Nνℓ, we have that the scaled log-posterior over Gram matrices satisfies, lim N 1 N log P(G1, . . . , GL | X, Y) = log P(Y | GL) + ℓ=1 νℓDKL N(0, Gℓ) || N(0, ˆAK(Gℓ 1)ˆAT ) + const := L(G1, . . . , GL). (24) Finally, assuming that the features are sufficiently well-behaved, we note that the Gram matrices are deterministic in the limit N , since they are a sum of i.i.d. terms, λ=1 f ℓ(f ℓ)T = lim N Gℓ= const. (25) Published in Transactions on Machine Learning Research (05/2025) The consequence of Eq. (25) is that the limiting posterior of the Gram matrices is a point distribution. But we know that the Gram matrices on which this point distribution is centered on should maximize the posterior, and hence the scaled log-posterior. Therefore, we seek to maximize Eq. (24), the scaled log-posterior. This is exactly the graph convolutional DKM objective we declared in Section 4.1, up to an additive constant. B Inducing Point Approximations The graph convolutional DKM objective is difficult to optimize directly due to quadratic memory costs and cubic time complexity. To avoid these large costs, we use an inducing point approximation similar to the one described in (Yang et al., 2023). We review in this Appendix how to construct an inducing point objective in a fully-connect setting, and then extend to the graph setting. B.1 An Inducing Point Objective Sparse DKMs as derived by Yang et al. (2023) are inspired by the sparse DGP literature (Damianou & Lawrence, 2013; Salimbeni & Deisenroth, 2017). In essence, we place an approximate posterior on a set of sparse inducing-points/features and derive an objective to optimize these inducing points by finding a lower bound on the evidence with variational inference (VI). In this subsection we describe the procedure. We introducing some inducing-points/features Fℓ i RPi Nℓin addition to the real test/training features Fℓ t RPt Nℓat each layer in the following prior model, Fℓ: = Fℓ i Fℓ t P(Fℓ| Fℓ 1) = λ=1 N(f ℓ λ; 0, Kfeatures(Fℓ 1)), (26b) λ=1 N(wλ; 0, I), (26c) P(Y | KL, W) = i=1 Categorical(yi; softmax(KL tichol(KL ii) 1W)i), (26d) where KL is the kernel applied to the final hidden layer features FL, with KL ii being the inducing/inducing sub-block and KL ti being the training/test-inducing sub-block. Eq. (26) is similar to the DGP in Eq. (1), but it has a different top-layer. The motivation for using the top-layer in Eq. (26) is efficiency; it is more efficient than a categorical GP under the parameterization described in Appendix I. To perform variational inference, we place a Gaussian approximate posterior on the inducing hidden-layer features, λ=1 N((f ℓ i )λ; 0, Gii), (27) and the following approximate posterior on the top-layer, λ=1 N(wλ; µλ, Σ). (28) Thus, G1 ii, . . . , GL ii, µ1, . . . , µνL+1, and Σ are the variational parameters. We can write down the distribution of the test/training features conditional on the inducing features using standard multivariate Gaussian rules, P(Fℓ t | Fℓ i , Fℓ 1) = N Fℓ t; Kti K 1 ii Fℓ i , Ktt Kti K 1 ii Kit , (29) Published in Transactions on Machine Learning Research (05/2025) where by Kii, Kti, and Ktt we refer to inducing-inducing, test/train-inducing, and test/train-test/train blocks respectively of K(Fℓ 1). We use Eq. (29) to construct an approximate posterior over the test/training features. The prior at each layer is equivalent to, P(Fℓ| Fℓ 1) = P(Fℓ t | Fℓ i , Fℓ 1) P(Fℓ i | Fℓ 1 i ), (30) and we select the full approximate posterior at each hidden layer to be Q(Fℓ| Fℓ 1) = P(Fℓ t | Fℓ i , Fℓ 1) Q(Fℓ i ). (31) This choice of approximate posterior is a classic one; the common term, P(Fℓ t | Fℓ i , Fℓ 1), is going to lead to cancellation in the ELBO. At the top layer, we use an i.i.d. Gaussian approximate posterior for the weights W, λ=1 N(wλ; µλ, Σ). (32) Similarly to Eq. (23) in the full-rank derivation, we duplicate the entries in the top-layer N times and make an i.i.d. assumption, such that Y = (Y Y) RPt NνL+1 and W = (W W) RPi NνL+1. Note that both the prior and the approximate posterior factorize across layers, which allows us to simplify the ELBO, log P( Y, X, F1, . . . , FL, W) Q(F1, . . . , FL, W) log P( Y, W | FL) Q( W) + ℓ=1 log P(Fℓ| Fℓ 1) log P( Y, W | KL) Q( W) log P(Fℓ i | Fℓ 1 i ) Q(Fℓ i ) In the limit N , where Nℓ= Nνℓ, the covariances at each layer, N 1 ℓ Fℓ i (Fℓ i )T , collapse to become equal to Gℓ ii. It can be shown that the hidden-layer terms in the ELBO simplify, log P(Fℓ i | Fℓ 1 i ) Q(Fℓ i ) νℓDKL N(0, Gℓ ii) || N(0, K(Gℓ 1 ii )) . (34) Additionally, N the term involving Y satisfies, log P( Y, W | KL) Q( W) log P(Y | KL, W) + log P(W) = EQ[log P(Y | KL, W)] λ=1 DKL (N(µλ, Σ) || N(0, I)) , (35b) which can easily be estimated by sampling W, since in the limit the kernel KL = K(GL) is constant. We define the sparse DKM objective to be the limiting ELBO scaled by N 1, lim N 1 N ELBO := Lsparse = EQ[log P(Y | K(GL), W)] λ=1 DKL (N(µλ, Σ) || N(0, I)) ℓ=1 νℓDKL N(0, Gℓ ii) || N(0, K(Gℓ 1 ii )) . Published in Transactions on Machine Learning Research (05/2025) Notice that the likelihood term in Eq. (36) requires the full kernel. We calculate this recursively. Eq. (29) tells us that, Fℓ t = Kti K 1 ii Fℓ i + (Ktt Kti K 1 ii Kit)1/2Ξ, (37) where Ξ is a matrix of i.i.d. standard Gaussian noise, and K = K(Gℓ 1) = [Kii Kit; Kti Ktt]. Therefore, Gℓ ti = lim N 1 ℓ Fℓ t(Fℓ i )T = Kti K 1 ii Gℓ ii, (38) Gℓ tt = lim N 1 ℓ Fℓ t(Fℓ t)T = Ktt Kti K 1 ii Kit + Kti K 1 ii Gℓ ii K 1 ii Kit. (39) B.2 Sparse Graph Convolutional Deep Kernel Machines To use a graph convolutional DKM, we need to be able to compute the graph kernel G 7 AK(G)AT with our inducing points. We already have adjacency information regarding the training/test nodes, but not the inducing nodes. Therefore we assume we have an adjacency between inducing points, Aii, as well as adjacency information between inducing points and training/test points, Ait = AT ti. For notational consistency, we denote the adjacency of the data as Att. The graph convolutional kernel over Gram matrices when using an inducing scheme becomes KGC(G) = Aii Ait Ati Att K(G) Aii Ait Ati Att where K(G) is a base kernel (e.g. the arccosine kernel) applied to G. We derive this kernel for an graph convolutional NNGP concretely in Appendix C. Substituting this graph kernel into the intermediate layers of the sparse DKM in Eq. (36) gives us a sparse graph convolutional DKM objective, LGC-sparse = EW Q[log P(Y | K(GL), W)] λ=1 DKL (N(µλ, Σ) || N(0, I)) ℓ=1 νℓDKL N(0, Gℓ ii) || N(0, Aii K(Gℓ 1 ii )AT ii ) , Again, the Gram matrices are calculated recursively in the forward pass with, Gℓ ti = Kti K 1 ii Gℓ ii, (42) Gℓ tt = Ktt Kti K 1 ii Kit + Kti K 1 ii Gℓ ii K 1 ii Kit, (43) and we then apply the graph convolutional kernel KGC( ) to Gℓ= [Gℓ ii Gℓ it; Gℓ ti Gℓ tt]. We considered two choices for Aii and Ati/Ait, namely (1) (intra-domain) using adjacency information from inducing points sampled from the training set, (2) (inter-domain) treating inducing points as independent of other inducing points and the training/test points, i.e. Aii = I, Ati = 0. In the case where the dataset is a single graph, we use a minibatch size of 1, and the minibatch becomes all nodes in the graph. In the case where the dataset contains several graphs, the graphs can be batched together. Naively, this would involve creating a single large adjacency matrix that describes adjacency information for graphs in the minibatch, but in practice this can be efficiently represented with a sparse tensor. The inter-domain scheme has the advantage that it can be used for datasets with a single graph or multiple graphs, whereas the intra-domain scheme is not expected to work when we change the graph batch (in the multi-graph case). Published in Transactions on Machine Learning Research (05/2025) C Graph Convolutional NNGP Kernel Derivations We show how to derive the kernel/covariance expressions given in Eq. (40) (for the sparse inducing point case). Eq. (8) (for the full-rank case) is the same, but with inducing points dropped. We take the graph nodes to have adjacency information as described in Section 4.2 and Appendix B. In a graph convolutional NNGP, pre-activation features Fℓ i , Fℓ t for the inducing and test/train blocks can be written in terms of the post-activations Hℓ 1 i , Hℓ 1 t , Fℓ i Fℓ t = Aii Ait Ati Att Hℓ 1 i Hℓ 1 t where W RNℓ 1 Nℓhas i.i.d. Gaussian elements with mean zero and variance 1/Nℓ 1. Remember that each column λ of Fℓis i.i.d., so the covariance of Fℓis, Kii Kit Kti Ktt = E f i;ℓ λ f t;ℓ λ f i;ℓ λ f t;ℓ λ = Aii Ait Ati Att Hℓ 1 i Hℓ 1 t E[wλw T λ ] Hℓ 1 i Hℓ 1 t T Aii Ait Ati Att = Aii Ait Ati Att Hℓ 1 i Hℓ 1 t Hℓ 1 i Hℓ 1 t T # Aii Ait Ati Att This is the general form of the graph convolutional kernel. However, if the activation function is Re LU, i.e. Hℓ= Re LU(Fℓ), then in the limit Nℓ 1 , N 1 ℓ 1Hℓ 1(Hℓ 1)T Φ(Kℓ 1), where Φ( ) is arccos kernel (Cho & Saul, 2009) and Kℓ 1 is the previous layer s kernel. This gives the recursive relation Kℓ ii Kℓ it Kℓ ti Kℓ tt = Aii Ait Ati Att Φ Kℓ 1 ii Kℓ 1 it Kℓ 1 ti Kℓ 1 tt Aii Ait Ati Att We call this kernel the graph convolutional kernel. C.1 Sparse Graph Convolutional Deep Kernel Machine Algorithm For completeness, we provide a Algorithm 1 to demonstrate node prediction with the sparse graph convolutional DKM. We discuss how to parameterize the learned inducing points for training in Appendix I. The algorithm can be modified for graph classification by adding a mean-pool layer immediately before sampling. For the sample step at the top-layer, we have two main options. The first, most conventional method, is to sample with a sparse GP (similar to Milsom et al. (2023)). The second, which is the approach taken in this paper, is to sample weights W from the approximate posterior (Eq. 28), which in turn allows sampling logits by calculating Ktichol(Kii) 1W RPt νL+1. Published in Transactions on Machine Learning Research (05/2025) Algorithm 1 Graph convolutional DKM node classification Parameters: {νℓ}L ℓ=1 Train/test inputs, train labels: Xt, Yt Inducing and train/test adjacencies: Aii, Ati, Att Inducing inputs and inducing Gram matrices: Xi, {Gℓ ii}L+1 ℓ=1 Variational parameters for the output layer: µ1, . . . , µνL+1, Σ Initialize full Gram matrix G0 ii G0 it G0 ti G0 tt Xi XT i Xi XT t Xt XT i Xt XT t for ℓin (1, . . . , L) do Apply kernel non-linearity Φ and perform graph convolution Kii Kit Kti Ktt Aii Ait Ati Att Φ Gℓ 1 ii Gℓ 1 it Gℓ 1 ti Gℓ 1 tt Aii Ait Ati Att Propagate the test-test and test-inducing blocks Gℓ ti Kti K 1 ii Gℓ ii Gℓ tt Ktt Kti K 1 ii Kit + Kti K 1 ii Gℓ ii K 1 ii Kit end for Calculate output kernel Kii Kit Kti Ktt Φ GL ii GL it GL ti GL tt Perform prediction using a sparse GP (or alternative model) with output kernel ˆY sample P(Y | Kii, Kti, Ktt) D Closed-form Solution for a Linear Graph Convolutional Deep Kernel Machine For some node regression problems with a linear kernel, K(G) = G, we are able to find an optimum for the DKM analytically. Assuming a constant regularizer coefficient for each layer, νℓ= 1, and no observation noise, the objective L for a linear graph convolutional DKM is, L(G1, . . . , GL) = ℓ=1 DKL(N(0, Gℓ) || N(0, ˆAK(Gℓ 1)ˆAT )) (47a) = const + ν h log|( ˆAGℓ 1 ˆA T ) 1Gℓ| Tr(( ˆAGℓ 1 ˆA T ) 1Gℓ) i . (47b) A slight abuse of notation follows, in that superscripts ℓon Gℓand Zℓrefer to values at the ℓ th layer, but elsewhere, superscripts for example on ˆAℓrefer to powers of ˆA, so that ˆAℓ= ˆA ˆA. For objective Eq. (47) to make sense, we must assume that the normalized adjacency matrix, ˆA, is invertible. In practice, ˆA may not be invertible, but we can resolve this by interpolating it with the identity, ˆAλ = λI + (1 λ)ˆA, λ (0, 1). (48) With this adjustment to the adjacency matrix, we can simplify the objective: L(G1, . . . , GL) = const ν ℓ=1 Tr(( ˆAGℓ 1 ˆA T ) 1Gℓ). (49) From here, we drop the transposes because ˆA and Gℓare symmetric. We take gradients: L(G1, . . . , GL) 2 Gℓ h Tr(( ˆAGℓ 1 ˆA) 1Gℓ+ Tr(( ˆAGℓˆA) 1Gℓ+1) i (50a) h (ˆAGℓ 1 ˆA) T + ((Gℓ) 1 ˆA 1Gℓ+1 ˆA 1(Gℓ) 1)T i . (50b) Published in Transactions on Machine Learning Research (05/2025) Then set those gradients to zero: = GℓˆA 1(Gℓ 1) 1 ˆA 1 = ˆA 1Gℓ+1 ˆA 1(Gℓ) 1. (51) For convenience, define, Zℓ= GℓˆA 1(Gℓ 1) 1, then, Eq. 51 = Z2 = ˆAZ1 ˆA 1 (52a) Z3 = ˆAZ2 ˆA 1 = ˆA2Z1 ˆA 2 (52b) ... (52c) Zℓ= ˆAℓ 1Z1 ˆA ℓ+1. (52d) By the definition of Zℓ, we have that, ℓ =1 Zℓ = (GℓˆA 1(Gℓ 1) 1)(Gℓ 1 ˆA 1(Gℓ 2) 1) ((G2 ˆA 1(G1) 1)(G1 ˆA 1(G0) 1) (53a) = GℓˆA ℓ(G0) 1. (53b) Also, by Eq. (52d), we have ℓ =1 Zℓ = (ˆAℓ 1Z1 ˆA ℓ+1)(ˆAℓ 2Z1 ˆA ℓ+2) (ˆA1Z1 ˆA 1)(ˆA0Z1 ˆA0) (54a) = ˆAℓ 1(Z1 ˆA 1)ℓˆA. (54b) By setting ℓ= L + 1, and combining equations 53b and 54b, we can solve for Z1, GL+1 ˆA (L+1)(G0) 1 = ˆAL(Z1 ˆA 1)L+1 ˆA (55a) = Z1 = (ˆA LGL+1 ˆA (L+1)(G0) 1 ˆA 1)1/(L+1) ˆA. (55b) Matrix roots are not unique, but we calculate roots by performing an eigendecomposition. Finally we use this expression for Z1 to obtain a relationship between the kernel representation at layer ℓ, Gℓ, and the input and label kernels, G0 and GL+1, Gℓ= ˆAℓ 1((ˆA LGL+1 ˆA L)(ˆAG0 ˆA) 1)ℓ/(L+1)(ˆAG0 ˆA)ˆAℓ 1. (56) D.1 Validity of the Solution Note that the solution given by Eq. (56) is only valid if Gℓis symmetric and positive definite. The latter property, positive definiteness, can be shown by the fact that Gℓis a product of positive definite matrices. We show symmetricity with the following argument. Let C = (ˆA LGL+1 ˆA L)(ˆAG0 ˆA) 1. Since C is the product of two symmetric PSD matrices, it is PSD (though not necessarily symmetric). Therefore we write C using its eigendecomposition, C = VDV 1, where D is the diagonal matrix of the eigenvalues of C. In particular, D 0. We write GL+1 in terms of V and D, and use the fact that GL+1 is a kernel matrix, GL+1 = ˆALVDV 1(ˆAG0 ˆA)ˆAL (57a) = ˆAL(ˆAG0 ˆA)V T DVT ˆAL = (GL+1)T (57b) = VDV 1 = (ˆAG0 ˆA)V T DVT (ˆAG0 ˆA) 1. (57c) This allows us to show that the powers of C have the following property, = Cn = (VDV 1)n = VDn V 1 = (ˆAG0 ˆA)V T Dn VT (ˆAG0 ˆA) 1. (58) Published in Transactions on Machine Learning Research (05/2025) 2000 3000 4000 5000 6000 7000 8000 9000 10000 train iter Figure 5: Losses during SGD training of the linear DKM. The loss of the analytic solution is shown by the dashed red line. In particular, we have Cℓ/(L+1) = (ˆAG0 ˆA)V T Dℓ/(L+1)VT (ˆAG0 ˆA) 1. (59) Gℓ= ˆAℓ 1Cℓ/(L+1) ˆAG0 ˆAℓ (60) = ˆAℓ 1VDℓ/(L+1)V 1(ˆAG0 ˆA)ˆAℓ 1 (61) = ˆAℓ 1(ˆAG0 ˆA)V T Dℓ/(L+1)VT ˆAℓ 1 (62) = (Gℓ)T (63) as desired. D.2 Experimental Confirmation of Linear Analytic Results Figures 1 and 5 demonstrate an exact match between our closed-form solution (Eq. 56) and when optimizing with gradient descent. We selected a random 200 node subset of Cora and (a) calculated the Gram matrices in closed-form, and (b) optimized a linear 2-layer graph convolutional DKM with Adam for 10000 epochs. We used a polynomial learning rate schedule, initialized at 0.1 and decaying with power 0.7. Figure 1 visualizes the Gram matrix representations themselves, and we can see an exact match between the closed-form and gradient descent solutions. Figure 5 shows that the training loss of the gradient descent solution approaches the loss of the analytic solution (that is, the value of the linear graph convolutional DKM objective when plugging in the analytic solution). E Experimental Details For final performance metrics, we benchmarked the graph convolutional DKM on 14 node classification datasets and 5 graph classification datasets. We also benchmarked a graph convolutional NNGP and GCN on the same datasets to obtain a suitable baselines. We describe details of the datasets and the experimental method below. E.1 Dataset Descriptions Here, we describe the datasets used in experiments qualitatively. The majority of the datasets we used for benchmarking were node classification datasets, where the task is to predict the labels of all nodes in a single Published in Transactions on Machine Learning Research (05/2025) graph, given labels for only a limited number of nodes. Cora, Citeseer, and Pubmed (Planetoid datasets) are all citation networks, where nodes represent scientific publications, and edges are citations between those publications; one-hot keyword encodings are provided as features, and the goal is to place documents into predefined categories. Chameleon and Squirrel are both Wikipedia networks on their respective topics, so that nodes correspond to articles, and edges correspond to links between articles; features are again onehot as in the citation networks, but represent the presence of relevant nouns, and the task is to categories pages based on their average daily traffic. Penn94 (Lim et al., 2021) is an online social network where nodes represent users, and edges represent social connections. The task is to predict the reported gender of the users. Flickr is a social network where nodes represent users, and edges correspond to friendships among users. The labels represent the interest groups of the users. Blog Catalog is a network dataset with social relationships of bloggers from the Blog Catalog website, where nodes attributes are constructed by the keywords of user profiles. The labels represent the topic categories provided by the authors. Several of the remaining datasets were introduced by Platonov et al. (2023): Roman Empire is derived entirely from the Roman Empire article on Wikipedia. Nodes are words in the text, and the words are connected if they are either adjacent in a sentence or if they are dependent on each other syntactically. Features are word embeddings, and the task is to label each word s syntactic role. Tolokers is a dataset of workers from the Toloka platform. The nodes are workers, and nodes are connected if workers have worked on the same task. Features are statistics about each worker, and the task is to predict if a worker has been banned. Minesweeper is a synthetic dataset. The graph has 10,000 nodes corresponding to the cells in the 100 by 100 grid in the Minesweeper game. Nodes are connected if the associated cells are adjacent. Features are one-hot encoded numbers of neighbouring mines, plus a extra feature for when the number of neighbouring mines is unknown. The task is to predict which nodes are mines. Amazon Ratings is a graph where nodes are Amazon products, and nodes are connected if they are commonly purchased together. Features are text embeddings of the product descriptions. The task is to predict the average user rating. The two remaining node classification datasets, Reddit and Arxiv, are the largest (roughly one order of magnitude larger than the next biggest). In the Reddit dataset (Hamilton et al., 2017), the nodes are posts, the features are Glo Ve embeddings of the posts, the labels represent subreddits, and posts are connected if a user comments on both posts. The Arxiv dataset (Hu et al., 2020b) is another citation network; nodes are computer science Arxiv papers, and papers are connected if one cites the other. Node features are Word2Vec embeddings of a paper s title and abstract, and the labels are subject areas. We also perform experiments with graph classification datasets. The graph classification datasets all contain molecules. NCI1 and NCI109 both contain graphs of chemical compounds (nodes representing atoms, with nodes connected if there are chemical bonds between them), where the goal is to predict the effect of the compounds versus lung and ovarian cancer. Proteins is a dataset of proteins molecules, where the task is to classify as enzymes or non-enzymes. Mutag is a dataset of compounds, and the task is to predict mutagenicity on Salmonella typhimurium. Finally, Mutagenicity is a dataset of drug compounds, and the task is to categorize into mutagen or non-mutagen. E.2 Details of Training, Hyperparameter Selection and Final Benchmarking E.2.1 Node classification datasets For node classification datasets, we used train/validation/test splits from the torch_geometric library (Fey & Lenssen, 2019) for most datasets; the exceptions were Arxiv and Reddit for which we adapted code from Niu et al. (2023). For data preprocessing, we (a) normalized all adjacency matrices as in Kipf & Published in Transactions on Machine Learning Research (05/2025) Welling (2017), and (b) we scaled input features such that the associated linear kernel had entries in [ 1, 1], i.e. X iλ = Xiλ P µ XiµXiµ . (64) We found that this transformation helped avoid numerical errors when training the models, as it ensures G0 = X (X )T has elements all on roughly the same scale. To select graph DKM hyperparameters, we started with a base model analogous to the GCN described by Kipf & Welling (2017), but with an extra Gram layer before the first kernel non-linearity. From this base model, we performed 4 sweeps to select (1) ν and inducing point scheme, (2) architecture, (3) centering parameters, and finally (4) the number of inducing points. The sweeps were, Sweep (1): ν {0, 0.01, 0.1, 1, 10, 100, 1000, }, scheme {inter, intra}, Sweep (2): arch {kipf, kipf + ˆAλ, kipf + ˆAλ + resid, platinov}. Here, kipf refers to the default architecture, +ˆAλ means that ˆAλ was used instead of the usual renormalized adjacency matrix, +resid refers to the addition of residual blocks, and platinov refers to an architecture similar to the one used in Platonov et al. (2023). The platinov architecture is similar to a residual 2-layer GCN, but with two linear layers in each residual block. Sweep (3) searched over centering or no centering, and whether to include a learned scale and bias parameter (described in Appendix F). Sweep (4): Pi {50, 100, 200, 300, 400}. We trained for 300/200 epochs in sweep (1), and in the remaining sweeps we trained for 200/150 for the smaller/bigger datasets (where Arxiv and Reddit and the bigger datasets). For the smaller datasets one epoch means a single full-batch gradient descent step; for the bigger datasets it means two mini-batched gradient descent steps. We had a numerical stability issue with Squirrel when selecting hyperparameters. To resolve this, we set arch=kipf, rather than using arch=platinov, which was the best according to sweep (2). We repeated the above sweeps, but with DKM regularization fixed to ν = (equivalent to removing Gram layers) to obtain final performance metrics for a sparse graph convolutional NNGP. To find suitable hyperparameters for the GCN, we used a single grid search starting with the base GCN model of Kipf & Welling (2017). We searched over width {100, 200}, dropout {0, 0.5}, batchnorm {yes, no}, and row_normalization {yes, no}, arch {kipf, platinov}. We used the Adam optimizer with a two-stage learning rate schedule for all training runs. We increase the learning rate linearly from 10 3 to 10 2 for the first quarter of the epochs, and after that use a cosine schedule with a minimum learning rate of 10 5. The models were written in Pytorch, and we trained on a cluster containing RTX 2080 s, RTX 3090 s and A100s. E.2.2 Graph classification datasets For graph classification datasets, we constructed our own cross-validation splits (10 splits), since no default split was provided by torch_geometric. We adopted a very similar approach for hyperparameter selection as in Section E.2.1. The main differences being that we set scheme=inter (since an intra-domain inducing-point scheme is not applicable in the multi-graph setting), and we trained for 300 batches on all datasets. To perform graph classification with node-level features/kernels of multiple graphs, we (a) stack graphs together in batches of 1024, and (b) perform a mean aggregation over nodes in each graph at the top layer of the network, thus obtaining logits for each graph. E.3 Dataset Statistics We include dataset statistics for node classification datasets in Table 3 and Table 4 respectively. Published in Transactions on Machine Learning Research (05/2025) Table 3: Node classification dataset statistics. Dataset # nodes # edges Homophily ratio # features # classes Roman Empire 22,662 32,927 0.05 300 18 Squirrel 5,201 198,353 0.22 2,089 5 Flickr 7,575 239,738 0.24 12,047 9 Chameleon 2,227 31,371 0.23 2,325 5 Amazon Ratings 24,492 93,050 0.38 300 5 Blog Catalog 5,196 171,743 0.40 8,189 6 Penn94 41,554 1,362,229 0.47 4,814 2 Tolokers 11,758 519,000 0.59 10 2 Arxiv 169,343 1,157,799 0.65 128 40 Minesweeper 10,000 39,402 0.68 7 2 Citeseer 3,327 4,552 0.74 3,703 6 Reddit 232,965 57,307,946 0.76 602 41 Pubmed 19,717 44,324 0.80 500 3 Cora 2,708 5,278 0.81 1,433 7 Table 4: Graph classification dataset statistics. Dataset # graphs avg. # nodes avg. # edges # features # classes Proteins 1,113 39.1 72.8 5 2 NCI1 4,110 29.9 32.3 38 2 NCI109 4,127 29.7 32.1 39 2 Mutag 188 17.9 19.8 8 2 Mutagenicity 4,337 30.3 30.8 15 2 E.4 Details of the Other Experiments For Figure 1 and Figure 5, we trained a linear graph convolutional DKM for 10,000 epochs on a randomly selected 200 node subset of Cora. For Figure 2, we calculate the linear graph convolutional DKM kernels on randomly generated Erdős Rényi graphs, with 50 nodes and edge probability 0.1. When calculating the kernels, we use ˆAλ=0.5 for the graph mixup. The input kernels are standard Wishart samples with 50 degrees of freedom. The nodes are randomly assigned one of two classes, with an even split. For Figure 3 we trained graph convolutional DKMs with the arc-cosine kernel non-linearity for 300 epochs, with regularization ν {0, 1, 10, 1000}, and ˆAλ=0.3 for graph mixup. The kernel alignment (CKA) figures were computed using the full kernels, without any normalization. For all of the Figures described above in this subsection, we used a simple 2-layer architecture, similar to Kipf & Welling (2017). All kernel plots are normalized to ensure entries are between [ 1, 1]. Finally, to produce Figure 4 we used data from the first hyperparameter sweep. Published in Transactions on Machine Learning Research (05/2025) F Kernel Centering Layer As part of our grid search, we considered adding a kernel centering layer . When enabled, the centering layer performs a centering similar to batchnorm, (Ft) iλ = (Ft)iλ 1 j (Ft)jλ, (65) where K = Ft FT t is the kernel to be centered. Additionally, we include the option of learning a scale parameter γ and bias parameter β, (Ft) iλ = γ(Ft) iλ + β. (66) G Inducing-point ablations Table 5 shows validation accuracies for different inducing point schemes. We obtained these accuracies during Sweep (1), so that ν has been tuned for each inducing-point scheme (but not other hyperparameters). We find that inter-domain is a better scheme overall, but intra-domain can be beneficial for a few datasets (most notably Penn94). Table 5: Validation accuracies for the two inducing point scheme on a range of node classification datasets. Inter-domain Intra-domain Cora 78.9 0.2 58.2 0.5 Pubmed 79.4 0.4 74.4 1.6 Reddit 94.2 0.0 94.5 0.0 Citeseer 70.8 0.2 23.4 0.2 Minesweeper 80.2 0.2 80.1 0.3 Arxiv 69.6 0.1 70.2 0.1 Tolokers 79.9 0.4 80.5 0.5 Penn94 69.5 0.3 81.3 0.4 Blog Catalog 76.0 0.8 72.8 1.3 Amazon Ratings 46.3 0.3 47.8 0.5 Flickr 59.2 1.6 55.8 1.6 Chameleon 66.5 2.0 66.0 2.0 Squirrel 49.9 0.7 54.4 0.9 Roman Empire 51.2 0.7 51.7 0.8 H Deep Graph Convolutional DKMs We investigated how the graph convolutional DKM performs as we increase the number of layers, with results shown in Figure 6. Similarly to the inducing point ablations (Section G), we show validation accuracies using settings found in Sweep (1), so that ν and inducing-point scheme have been tuned (but not other hyperparameters). We find that increasing depth tends not to help performance (though it can occassionally help in some cases, e.g. Citeseer). This finding is expected, since GCN also loses performance as depth is scaled (Oono & Suzuki, 2019). Note that we encountered a numerical instability when running the Reddit depth experiment, so the relevant data points have been omitted (depth=6 and 8). Published in Transactions on Machine Learning Research (05/2025) 2 4 6 8 10 70 2 4 6 8 10 86 2 4 6 8 10 68 2 4 6 8 10 72 88 Minesweeper 2 4 6 8 10 62 2 4 6 8 10 72 Blog Catalog Amazon Ratings 2 4 6 8 10 Depth 2 4 6 8 10 Depth 2 4 6 8 10 Depth 2 4 6 8 10 Depth Roman Empire Figure 6: Validation accuracy at different depths for different node classification datasets. Error bands are 1 standard deviation. I Parameterization of the Deep Kernel Machine Even though the inducing point scheme significantly reduces the computational complexity of training a DKM, computing the objective and forward equations is still expensive and has the potential to be unstable. In particular, the KL regularization terms DKL N(0,Gℓ) || N(0, K(Gℓ 1)) = 1 2 P log det(K(Gℓ 1) 1Gℓ) + Tr(K(Gℓ 1) 1Gℓ) (67) Published in Transactions on Machine Learning Research (05/2025) involve an unpleasant inverse and log-determinant. When implementing convolutional DKMs, Milsom et al. (2023) performed these calculations naively, which necessitates the use of double precision to avoid stability issues. The use of double precision is inefficient. We propose the following parameterization to avoid the log det, while still computing the regularization term exactly. At each layer, we learn the parameter Lℓ= cholesky(K(Gℓ 1 ii ) 1Gℓ ii). Due to the fact that Lℓis lower triangular, the log determinant term can easily be computed as, log det(K(Gℓ 1 ii ) 1Gℓ ii) = log det(LLT ) = 2 j=1 log Lℓ jj. (68) The forward equations given by Yang et al. (2023) are Gℓ ti = Kti K 1 ii Gℓ ii, (69a) Gℓ tt = Ktt Kti K 1 ii Kit + Kti K 1 ii Gℓ ii K 1 ii Kit. (69b) We suggest using the Nyström approximation Ktt Kti K 1 ii Kit to avoid computing Kti K 1 ii Kit. Additionally, we know that there exist Fi and Ft such that Gℓ ii Gℓ it Gℓ ti Gℓ tt = Fi FT i Fi FT t Ft FT i Ft FT t Therefore the Gram forward equations (Eq. 69) with a Nystrom approximation for Ktt are equivalent to Hi = cholesky(Kii), (71a) Fi = Hi Lℓ, (71b) Ft = Kti H 1 i Lℓ. (71c) Notably, the Kti H 1 i computation can be achieved using a triangular solve, since Hi is lower triangular. The trade-off of propagating with Eq. (71) rather than the original scheme (Eq. 69) is that in exchange for using the Nyström approximation we get (a) efficient objective computation avoiding log det, (b) a triangular solve instead of a cholesky solve, and (c) fewer matrix multiplies.