# generalised_implicit_neural_representations__ac860b1e.pdf Generalised Implicit Neural Representations Daniele Grattarola EPFL Lausanne, Switzerland daniele.grattarola@epfl.ch Pierre Vandergheynst EPFL Lausanne, Switzerland pierre.vandergheynst@epfl.ch We consider the problem of learning implicit neural representations (INRs) for signals on non-Euclidean domains. In the Euclidean case, INRs are trained on a discrete sampling of a signal over a regular lattice. Here, we assume that the continuous signal exists on some unknown topological space from which we sample a discrete graph. In the absence of a coordinate system to identify the sampled nodes, we propose approximating their location with a spectral embedding of the graph. This allows us to train INRs without knowing the underlying continuous domain, which is the case for most graph signals in nature, while also making the INRs independent of any choice of coordinate system. We show experiments with our method on various real-world signals on non-Euclidean domains. Figure 1: Given a continuous signal on a non-Euclidean domain, we observe a discrete graph realisation of it. A generalised implicit neural representation is a neural network fθ trained to map a spectral embedding of each node to the corresponding signal value. 1 Introduction Implicit neural representations (INRs) are a class of techniques to parametrise signals using neural networks [48, 41, 45, 34]. INRs are trained to map each point in a given domain to the corresponding value of a signal at that point. For example, INRs for images learn to map the 2D coordinates of pixels to their corresponding RGB values. INRs can also be conditioned on a latent vector, typically learned end-to-end as part of the model, that allows them to represent different signals on the same domain. INRs have been successfully applied to model complex signals like images [48, 45], signed distance functions [41], and radiance fields [40]. By expressing a signal as an INR, we obtain a continuous approximation of the signal on the whole domain. This allows us to compute a higher-resolution approximation of the signal by sampling more points on the domain (e.g., a finer grid of pixels). Additionally, since INRs are fully differentiable, 36th Conference on Neural Information Processing Systems (Neur IPS 2022). the latent vector of a conditional INR can be optimised with backpropagation and gradient descent to obtain a signal with some desired characteristics. For example, a conditional INR for signed distance functions can be used to design surfaces with a target aerodynamic profile [42]. So far, the literature on INRs has only focused on signals on Euclidean domains. However, signals defined on non-Euclidean domains are ubiquitous in nature and are especially relevant in artificial intelligence, as demonstrated by the recent rise of geometric deep learning [12]. In this paper, we propose an extension of the INR setting to signals on arbitrary non-Euclidean domains. Contributions We formulate the generalised INR problem as the task of learning an implicit representation for a signal on an arbitrary topological space T where, instead of observing the signal sampled on a regular lattice, we observe a graph (i.e., a discretisation of T ) and the corresponding graph signal. In most practical cases, T is unknown and we cannot represent the sampled vertices in a coordinate system. Even when T is known (as in the Euclidean case), training an INR on a fixed coordinate system means that the model will depend on this choice. We solve both these issues by identifying the sampled nodes with an intrinsic spectral embedding obtained from the eigenvectors of the graph Laplacian. We then train a neural network to map the spectral embeddings to the corresponding signal values. Figure 1 shows a schematic view of the method. Since the eigenvectors of the graph Laplacian are a discrete approximation of the continuous eigenfunctions of the Laplace-Beltrami operator on T (when appropriately rescaled) [5, 10, 7], at inference time we can map arbitrary points on T to the corresponding approximation of the signal, as long as the discrete graph signal is sampled consistently. This allows us to compute higher-resolution signals or to estimate the signal on different graph realisations of the same phenomenon (e.g., different social networks with a similar structure). Results In the experiments section, we show concrete examples of learning INRs for signals on graphs and manifolds, using real-world data from biology and social networks. We also show that INRs trained on one instance of a graph can be transferred to different graph realisations of the underlying domain. Then, we show applications of conditional generalised INRs to model different signals on the same irregular domain, and also different signals on different domains. Finally, we conclude the paper with an experiment that consolidates all our results, modelling real-world meteorological signals on the spherical surface of the Earth. 2 Related works INRs have recently attracted the attention of the machine learning community for their ability to represent complex signals, especially in computer vision, 3D rendering, and image synthesis applications [41, 38, 2, 22, 40, 46, 32]. Recent work has highlighted that INRs benefit from computing sinusoidal transformations of the coordinates, typically called Fourier features or positional encodings (PEs) [52, 50, 40, 53, 8]. Since the PEs typically used in INRs are also the eigenfunctions of the Laplace operator in Euclidean domains, our method is a generalisation of this approach. However, these transformations are not applicable to non-Euclidean domains and still depend on a choice of coordinate system. A proposal for making INRs equivariant under the SO(3) group has been presented by Deng et al. [15], although this technique does not apply in general to non-Euclidean domains. The field of geometric deep learning has also explored the use of graph PEs similar to the ones we use in this paper, showing their usefulness as node features for graph neural networks [47, 36, 17, 16]. Recent research has also investigated the idea of learnable graph PEs [18, 13]. Related to our work, Belkin and Niyogi [5], Boguna et al. [11] and Levie et al. [30] have studied the geometrical link between graphs and topological spaces. Keriven et al. [25, 26] investigated the convergence, in the limit, of graph neural networks trained on large graphs. An alternative view of graphs as discrete samples of continuous spaces is the concept of graphon, which can be seen both as a generative model and as the limit of a sequence of graphs [37]. We also mention the work of Koestler et al. [28], who propose a similar idea to ours for learning fields on meshes. In this work, the authors train a neural network to map the eigenfunctions of the Laplacian to an RGB texture. By contrast, we focus on a wider range of tasks and experiments, including transferring to completely different graphs, conditioning the INRs, and modelling dynamical systems. Our scheme to compute generalised embeddings is also different from Koestler et al. s, not relying on knowing node coordinates. Standard setting In the standard INR setting, we consider a signal f : X Y with X Rd and Y Rp. We observe a discrete realisation of the signal f(xi) for i = 1, . . . , n, where points xi are sampled on a regular lattice on X. Then, we train a neural network fθ : X Y, with parameters θ, on input-output pairs (xi, f(xi)). Since we know that the signal domain is a subset of Rd, at inference time we can sample points anywhere on X to compute the approximated value of the signal at those points. For example, an INR for images maps equispaced points in the unit square (pixel coordinates) to points in the unit cube (RGB values normalised between 0 and 1). The specific image on which we train the INR is a realisation of one such signal at a given resolution, and at inference time we can sample a finer lattice to super-resolve the image [48, 45]. Generalised setting In the generalised setting, we consider a continuous signal f : T Y, with T an arbitrary topological space. We observe a discrete graph signal f(vi) on an undirected graph G = (V, E), with node set V = {vi} for i = 1, . . . , n and edge set E V V . Note that the graph can be weighted if a metric is available on T . The meaning of sampling G from T is intuitive in the case of geometric meshes or other physical structures like proteins (in which case we also know the coordinates of vi), but the same reasoning also applies to more complex domains with an abstract meaning (e.g., the space of all possible papers and their citations). We generally assume that the graph describes some measure of closeness between uniformly sampled points on T , be it a function of the coordinates (e.g., a kernel) or some logical or functional relation that is given as part of the data. In general, we don t assume to know the true T from which G is sampled or the coordinates of vi. We only observe G and the associated graph signal. For an in-depth discussion on the relation between graphs and topological spaces, see references [5], [11], and [30]. Proposed method We approach the generalised INR problem by mapping vi to f(vi) through a spectral embedding obtained from the graph Laplacian. Let A Rn n be the weighted adjacency matrix of graph G, D the diagonal degree matrix, and L = D A the combinatorial Laplacian. The eigendecomposition of the Laplacian yields an orthonormal basis of eigenvectors {uk Rn} for k = 1, . . . , n, with a canonical ordering given by their associated eigenvalues λ1 λ2 λn. We define a generalised spectral embedding of size k for node vi as ei = n [u1,i, . . . , uk,i] Rk. (1) 0.0 0.5 1.0 i/n n = 100 n = 1000 0.0 0.5 1.0 i/n Figure 2: Laplacian eigenvectors (left) vs. rescaled spectral embeddings (right) of a path graph, for k = 3 and n = 100, 1000, 2000. The curves on the right are shifted to improve visualisation. The generalised spectral embeddings are a discrete approximation of the eigenfunctions of the continuous Laplace-Beltrami operator on T and converge to them for n . See the discussion in references [5], [10], [9], and [7] for a formal analysis. In practice, rescaling the eigenvectors by n ensures that the embeddings for graphs of different sizes will have the same range component-wise and that similar nodes will have similar embeddings regardless of the graph size. As an example, we show a comparison between the eigenvectors and the generalised spectral embeddings of a path graph in Figure 2. We train a neural network fθ : ei 7 f(vi) on the observed graph signal. Since the generalised spectral embeddings are an intrinsic property of the graph, the INR will not depend on any choice of coordinate system but only on the topology of the underlying continuous domain. At inference time, we can compute the approximated value of the signal at an arbitrary location on T by computing the associated spectral embedding. If we know T , or if we estimate it, we can sample new vertices directly from the domain and apply the same procedure used to construct the training graph. If T is unknown (e.g., in the case of natural graphs that describe some abstract concept like citations, social interactions, or biological relations), then we just assume to observe a similar graph sampled from T . An important difference between generalised INRs and Euclidean INRs is that generally we must observe the full graph in order to compute the inputs. This is necessary because, while in the Euclidean case we completely know the domain of the signal a priori, in the generalised case we need to estimate the topology of the domain by sampling. Figure 3: Ground truth signals vs. signals predicted by the INR. Best viewed in colour. Complexity Although computing the full eigendecomposition of the Laplacian has a complexity of O(n3), here we are only interested in the first k eigenvectors. Also, the Laplacian of most graphs is very sparse, with nodes having an average degree d n. We can therefore use the implicitly restarted Lanczos method for eigendecomposition implemented by the ARPACK software, which has a complexity of O( dn2) and can be easily parallelised [29]. In practice, all computations for this paper were easily managed on a commercial laptop with 10 CPU cores, scaling up to graphs in the order of 105 nodes. We also note that at inference time, depending on how the edges are constructed, it could be possible to add new nodes and edges to the training graph instead of sampling an entirely new graph from T . In this case, the spectral embeddings for the new nodes can be estimated using the Nyström method without needing to compute the full eigendecomposition [3, 9]. Alternative approaches The generalised Laplacian embeddings used in this paper are only one of many possibilities to represent nodes sampled from T . To name a few, locally linear embeddings [43], Isomap [51], Laplacian eigenmaps [6] and diffusion maps [14] are all based on the idea of embedding data using the first few principal eigenvectors of a similarity matrix. Here we focus on the Laplacian since it is well known, sparse, and easy to compute, and its eigendecomposition is stable to graph perturbations. One disadvantage of Laplacian eigenvectors is that they are only unique up to sign, leading to 2k possible eigenbases for a given T and the consequent ambiguity when transferring an INR to a different graph realisation. Additionally, eigenvalues of multiplicity greater than 1 also introduce ambiguity in the eigenvectors, since all rotations and reflections of the associated eigenspace are valid choices. However, simple heuristics can be used to eliminate sign ambiguity, and we did not encounter any other practical issue in this regard. Alternative ways to resolve the ambiguity is to use PEs based on random walks [18, 39, 31], heat kernels [49, 19], or the more recent sign-and-basis invariant neural networks [33]. We leave the exploration of these alternatives to future work, since they do not significantly impact our main contributions. 4 Experiments Setting We implement the generalised INR as a SIREN multi-layer perceptron [45]. The model has 6 layers with 512 hidden neurons and a skip connection from the input to the middle layer. We use the same hyperparameters and initialisation scheme suggested by Sitzmann et al. [45]. We train the model using Adam [27] with a learning rate of 10 4 and an annealing schedule that halves the learning rate if the loss does not improve for 1000 steps. At each step, we sample 5000 nodes randomly from the graph as a mini-batch. We use spectral embeddings of size k = 100, a value that we found empirically as a good trade-off between complexity and performance. We report an analysis on the effect of k in Section 4.1. We ran all experiments on an Nvidia Tesla V100 GPU. The code to reproduce our results and the high-resolution version of all figures are available at https://github.com/danielegrattarola/GINR. All deviations from the default setting are documented in the supplementary material, where we also report an additional experiment on solving differential equations with generalised INRs. 4.1 Learning generalised INRs We begin by verifying that a typical INR architecture can successfully learn arbitrary graph signals in the generalised setting. We consider three datasets as test cases. Bunny We generate a texture on the Stanford bunny mesh1 using the Gray-Scott reaction-diffusion model given by: a = Da La a b b + F(1n a); b = Db Lb + a b b (F + K)b (2) for state vectors a, b Rn and parameters Da, Db, F, K R, where indicates elementwise product [21]. The mesh has 34834 nodes and 104288 edges. We configure the system to yield coral patterns2 and evolve the simulation for 104 steps starting from random positive values for a, b. At the end of the evolution, we use vector a as target graph signal. Table 1: R2 and mean squared error for the bunny, protein, and US election signals. Bunny Protein US Election R2 1.000 1.000 0.999 MSE 9.14 10 8 1.17 10 10 1.45 10 3 Protein As a real-world domain, we consider the solvent excluded surface of a protein structure.3 The continuous signal is the value of the electrostatic field generated by the amino acid residues at the surface. Protein function can be largely understood by studying their surfaces and the corresponding chemical features, like the electrostatic charge, so this represents a potentially interesting application of INRs to biology. We use a combination of the MSMS [44] and APBS [24] software packages to estimate the surface, sample the nodes, and get the corresponding signal, following the same workflow of Gainza et al. [20]. The final graph has 11966 nodes and 35892 edges. Table 2: Performance comparison (R2) of typical INRs and generalised INRs (GINRs), using Re LU (r) or SIREN (s) activation. INR (r) INR (s) GINR (r) GINR (s) Bunny 0.000 0.919 0.932 0.885 Protein 0.799 0.275 0.921 0.916 US election Finally, we consider a social network dataset introduced by Jia and Benson [23], in which nodes represent United States counties and edges are estimated using the Facebook Social Connectedness Index. The target signal represents the county-wise outcome of the 2012 US presidential election, as values in the range [ 1, 1] (indicating the proportion of votes in favour of one candidate vs. the other). The graph has 3106 nodes and 22574 edges. Results In all cases, the generalised INR achieves an R2 close to 1 indicating that the model can indeed learn non-trivial signals on non-Euclidean domains. We report the R2 and mean squared error for all datasets in Tab. 1. We compare the true graph signals and those learned by the INR, for the bunny and protein, in Fig. 3. As a second test, we compare typical INRs trained on node coordinates with generalised INRs trained on the spectral embeddings. We consider the bunny and protein since they have input coordinates for the typical INRs. We also test different activation functions Re LU or SIREN. We report in Tab. 2 the R2 for a held-out set of nodes (to evaluate whether the INRs are overfitting instead of learning a meaningful representation).4 All results are averaged over 5 runs and have negligible standard deviations. We see that the generalised INRs perform almost always better than the typical ones. Note that these experiments only verify that the model can indeed learn the target signals. We will test the transferability of the model in-depth in Sec. 4.2. In the following sections, we mostly focus on known spaces T or synthetic graphs so that we can better answer other research questions by having full control of the data. Size of the spectral embeddings Typical INRs take as input the d-dimensional coordinates of points on the sampled lattice. In our general setting, the dimension k of the spectral embeddings is a hyperparameter of the method. To evaluate the impact of k on the performance of the INR, we train the same model for different values of k on the Stanford bunny. We report the results in Figure 4, showing that the model fails to learn a meaningful representation for k 3. This is reasonable because at least 3 non-trivial eigenvectors are needed to correctly distinguish the main structural features of the bunny (ears, tail, etc.), as shown in Figure 5. We also see that, while the INR has no issues in learning the signal for k = 100, for lower values of k the model struggles to represent the signal on the ears. 1Available at https://graphics.stanford.edu/data/3Dscanrep/ 2Da = 0.64, Db = 0.32, F = 0.06, K = 0.062 3Protein Data Bank identifier: 1AA7 4We report training details in the supplementary material. Figure 4: Left: R2 vs. k; Right: signals learned by the INR for k = 2, 3, 4. Figure 5: Spectral drawing (blue) of the bunny using the first three non-trivial eigenvectors. In grey, 2d projections of the 3d drawing. Figure 5 again provides a possible explanation for this result, since the first few eigenvectors tend to collapse on the ears and higher frequency eigenvectors are needed to distinguish points towards the narrow tips. We will show another consequence of this in Section 4.2. We also mention that the drop in performance for lower values of k on the other two datasets is less pronounced than on the bunny, likely because of their lack of any high-frequency features. 4.2 Transferability of generalised INRs Because the generalised spectral embeddings are a discrete approximation of the continuous eigenfunctions of the Laplace Beltrami operator on T , we can apply a trained INR to a different graph realisation of T as long as its spectral structure is consistent with the training data. In practical terms, we need to know 1) how to sample new nodes and 2) how the nodes are connected. If we know T , then the problem simply boils down to choosing a sampling strategy for the points and a suitable measure of closeness (e.g. a kernel or edge-generating function like the k-nearest neighbours algorithm). If T is unknown, then we assume to observe a new graph through the same process that generated the training data (e.g., we observe new social interactions among the same or a similar set of people). Note that, when sampling a new graph from T , the spectral embeddings will be a slightly different (possibly better) approximation of the continuous eigenfunctions. As this effect is more evident on high-frequency eigenvectors, we empirically observed that training the INR with a smaller k and Re LU activations improved its transferability (cf. Table 2). See the supplementary material for more details. We investigate the transferability of generalised INRs in two settings. Transferring to similar graphs As a first example, we study the case in which T is unknown and we only observe the discrete graphs. We consider a stochastic block model (SBM) with two communities of equal size (in total, n = 1000), parametrised by an inter-connection probability p and an intra-connection probability r. We train an INR on a graph with a strong community structure (p = 0.1, r = 0.5) and a graph signal indicating the community of each node (effectively a node classification task). Then, we test the INR on graphs sampled in the ranges p [0.1, 1] and r [0.1, 1] to test the robustness of the model to changes in the community structure. We report our results in Figure 6. First, we note that perfectly recovering the correct clusters is only theoretically guaranteed in certain regions of the parameter space [1], delimited with an orange line in the figure.5 We see that the trained INR transfers perfectly to all graphs sampled in this region and that its performance sharply drops only for p > r, where graphs have no community structure (see Figure 6 a, b). Additionally, we see that the INR maintains a good performance also in the region between the theoretical boundary and the line p = r, i.e., a region in which graphs still exhibit some community structure although perfect recovery is not guaranteed. Overall, this indicates that training the INR on an instance of a graph allows us to transfer it to graphs with a similar structure. 5The region is above curve r = ln(n)/n np/ ln(n) 2 [1]. (a) p = 0.1, r = 0.5 (b) p = 0.5, r = 0.1 Figure 6: Left: normalised mutual information at test time, evaluating on graphs with different interand intra-connection probabilities. We trained the INR on p = 0.1 and r = 0.5, marked with on the plot. The orange line indicates the theoretical boundary above which it is possible to perfectly recover the true signal. Best viewed in colour. Right: the first two non-trivial eigenvectors of two SBMs, respectively exhibiting strong and no community structure. Figure 7: Zoomed-in comparison between the nodes of the training graph (left) and its super-resolved version (right), plotted on a black surface to aid visualisation. An interactive version in vector graphics of this figure is available in the supplementary material. 0.0 0.2 0.4 Squared error Figure 8: Distribution of the error in the super-resolution experiment. Super-resolution As a second test case, we consider the super-resolution of a signal. We consider again the Stanford bunny mesh with the same texture graph signal as Section 4.1 and we apply Loop s method for mesh subdivision to obtain a higher-resolution mesh with 139122 nodes and 416929 edges [35]. Then, we compute the spectral embeddings for the new graph and predict the signal with the INR trained on the low-resolution graph. Because the unit-norm eigenvectors are only unique up to sign, the INR will transfer to a different graph only if its eigenvectors are aligned to the training ones (i.e., similar nodes must have similar spectral embeddings in the previous experiment, we did not have to deal with this problem due to symmetry in the SBM). However, in many cases, a simple comparison of the eigenvectors histograms provides a useful heuristic for automatically aligning them. For simplicity, here we also verify the alignment of the two eigenbases manually. We show in Figure 7 a qualitative comparison between the training and super-resolved signals, from which we see that the INR can predict the graph signal correctly. Quantitatively, since mesh subdivision does not remove the original nodes from the mesh, we can compute the R2 between the training signal and part of the super-resolved signal. The model achieves an R2 of 0.39 compared to a training R2 of 0.99. However, this apparent drop in performance is mostly due to the high-frequency ears of the bunny, for which the model evidently overfits and cannot tolerate even small changes in the eigenvectors. Figure 8 shows the distribution of the squared error between the ground truth and predicted signal, from which we see that the error is close to zero for most of the nodes (note that the y-axis is logarithmic). By ignoring the nodes above the 90th percentile of the squared error distribution, the R2 score is 0.94. Increasing k resulted in a better prediction of the ears region (cf. Figure 4) at the cost of worse overall transferability. (a) t = 625 (b) t = 875 (c) t = 1125 (d) t = 1375 (e) t = 1625 Figure 9: Signals predicted by the conditional INR fθ(ei, t) at equispaced time steps t [625, 1625]. All time steps shown were not part of the training set. Also, each test time step is as far as possible from its two closest training samples (i.e., training samples are at t 0 (mod 10) while test samples are at t 5 (mod 10)). An animated version of this figure with 600 predicted samples is available in the supplementary material. 4.3 Conditional generalised INRs We also consider the setting in which a generalised INR is conditioned on a latent vector z Rq to allow the representation of different signals by the same neural network fθ(ei, z). Unlike the Euclidean setting, here we can investigate two cases: 1) learning a conditional INR for different signals on the same domain (the typical setting), and 2) learning a conditional INR for different signals on different domains. Reaction-diffusion process For the first case, we train a generalised INR to approximate the Gray-Scott reaction-diffusion process that we used to generate the graph signal in Section 4.1, given by Equation (2). We consider a signal f(vi, t), where t N is the number of updates since the start of the simulation. We evolve the system for 3000 steps starting from the same random initialisation used in Section 4.1 and we train the INR fθ(ei, t) using 300 time steps sampled at t 0 (mod 10). The graph is fixed throughout the evolution, so the INR must learn to exploit the conditioning input t to correctly predict the signal at different times. Figure 10: R2 vs. size of the dataset |D|. At test time, we predict the signal for all the remaining time steps and compare the prediction with the ground truth signals. The INR achieves a test R2 of 0.999 0.006 indicating that it successfully learned to approximate the true dynamics. We report in Figure 9 a few samples predicted by the INR at test time. By comparison, a simple baseline using linear interpolation between training samples achieves a similar R2, while requiring to keep all 300 training steps in memory. The INR, on the other hand, represents the signal and its evolution with a single set of weights and a scalar control parameter. Multi-protein INR For the second case, we consider a dataset D = {(Tm, fm)} of protein surfaces Tm with associated electrostatic signals fm : Tm R for m = 1, . . . , |D|, obtained following the same protocol as Section 4.1. We configure the neural network fθ(ei, zm) as an autodecoder [41] where zm Rz is a vector of free parameters learned end-to-end alongside the weights of the neural network. We assign a separate vector zm to each training sample, allowing the model to find the best latent representation for the dataset. Since the continuous surface and the associated signals are unique to each protein in the dataset, the latent vector zm must capture both aspects. We set the latent vector size z = 8, which we empirically found to give consistently good performance across datasets. Larger values of z did not significantly improve the performance. We test the ability of the model to represent datasets of different sizes, by sampling random proteins from the dataset used in reference [20]. Figure 10 shows the R2 of the model for different values of |D|. Understandably, the performance decreases as the number of samples grows, although the model achieves R2 0.9 for datasets of up to 500 different proteins. Overall, these results indicate that the model can represent many domains and signals with a single set of weights. (a) Temperature Figure 11: Top: prediction of the INR for t = 4.5 (randomly chosen, not in the training data) at the original resolution. Bottom: The super-resolved versions of the top row. Note that the apparently lighter colours are simply artifacts of the plotting software. Best viewed in colour and high resolution. An animated and uncompressed version of this figure is available in the supplementary material. 4.4 Weather modelling As a final experiment, we consolidate all results from previous sections to showcase an application of generalised INRs in a real-world setting. We consider signals f(vi, t) representing hourly meteorological measurements on the spherical domain of the Earth s surface. Our goal is to train a conditional INR at a given spatial and temporal resolution and then use it to super-resolve the signal over space and time. Data We collected data from the National Oceanic and Atmospheric Administration (NOAA) Operational Model Archive and Distribution System, specifically from the Global Forecast System (GFS). The data consists of three different signals representing: the dew point temperature 2m above-ground measured in K (dpt2m), the atmosphere s total cloud cover percentage (tcdcclm), and the surface wind speed in m/s (gustsfc). We consider a 24-hour period sampled at 1-hour increments over an equiangular spherical grid sampled at 1 increments.6 We generate a spherical mesh from the 65160 locations of the grid by computing their convex hull with the Qhull software [4], for a total of 193320 edges. Results We train a conditional INR fθ(ei, t) at the given spatial and temporal resolution, for t = 0, . . . , 23. Then, we evaluate its predictions on a high-resolution spherical mesh with 258480 nodes and 773280 edges (obtained with the same procedure of Section 4.2), and at time increments of 30 minutes. We report in Figure 11 the prediction of the INR for a random test time, on both the original mesh and the high-resolution one. We refer the reader to the animated version of the figure available in the supplementary material for a better appreciation of the results. Overall, our results confirm that generalised INRs can model complex real-world signals on non-Euclidean domains, learning a realistic continuous approximation of the signals from low-resolution samples. 5 Conclusion We have presented the problem of learning generalised implicit neural representations for signals on non-Euclidean domains. Our method learns to map a spectral embedding of the domain to the value of the signal, without relying on a choice of coordinate system. We have shown applications of our method on biological, social, and meteorological data, highlighting the potential usefulness of such INRs in a wide range of scientific fields. We hope that future work will explore more real-world applications of our method. 6Specifically the period from 00:00 to 23:59 of May 9, 2022. All times t given in the text are in hours relative to the start of the period. [1] Emmanuel Abbe. Community detection and stochastic block models: recent developments. The Journal of Machine Learning Research, 18(1):6446 6531, 2017. 6 [2] Matan Atzmon and Yaron Lipman. Sal: Sign agnostic learning of shapes from raw data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2565 2574, 2020. 2 [3] Christopher TH Baker. The numerical treatment of integral equations. Oxford University Press, 1977. 4 [4] C Bradford Barber, David P Dobkin, and Hannu Huhdanpaa. The quickhull algorithm for convex hulls. ACM Transactions on Mathematical Software (TOMS), 22(4):469 483, 1996. 9 [5] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps and spectral techniques for embedding and clustering. Advances in Neural Information Processing Systems, 14, 2001. 2, 3 [6] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation, 15(6):1373 1396, 2003. 4 [7] Mikhail Belkin and Partha Niyogi. Convergence of laplacian eigenmaps. Advances in Neural Information Processing Systems, 19, 2006. 2, 3 [8] Nuri Benbarka, Timon Höfer, Andreas Zell, et al. Seeing implicit neural representations as fourier series. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 2041 2050, 2022. 2 [9] Yoshua Bengio, Jean-françcois Paiement, Pascal Vincent, Olivier Delalleau, Nicolas Roux, and Marie Ouimet. Out-of-sample extensions for lle, isomap, mds, eigenmaps, and spectral clustering. Advances in Neural Information Processing Systems, 16, 2003. 3, 4 [10] Yoshua Bengio, Pascal Vincent, Jean-François Paiement, Olivier Delalleau, Marie Ouimet, and Nicolas Le Roux. Spectral clustering and kernel PCA are learning eigenfunctions, volume 1239. Citeseer, 2003. 2, 3 [11] Marian Boguna, Ivan Bonamassa, Manlio De Domenico, Shlomo Havlin, Dmitri Krioukov, and M Serrano. Network geometry. Nature Reviews Physics, 3(2):114 135, 2021. 2, 3 [12] Michael M Bronstein, Joan Bruna, Yann Le Cun, Arthur Szlam, and Pierre Vandergheynst. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34 (4):18 42, 2017. 2 [13] Benjamin Chamberlain, James Rowbottom, Davide Eynard, Francesco Di Giovanni, Xiaowen Dong, and Michael Bronstein. Beltrami flow and neural diffusion on graphs. Advances in Neural Information Processing Systems, 34, 2021. 2 [14] Ronald R Coifman, Stephane Lafon, Ann B Lee, Mauro Maggioni, Boaz Nadler, Frederick Warner, and Steven W Zucker. Geometric diffusions as a tool for harmonic analysis and structure definition of data: Diffusion maps. Proceedings of the National Academy of Sciences, 102(21): 7426 7431, 2005. 4 [15] Congyue Deng, Or Litany, Yueqi Duan, Adrien Poulenard, Andrea Tagliasacchi, and Leonidas J Guibas. Vector neurons: A general framework for so (3)-equivariant networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 12200 12209, 2021. 2 [16] Vijay Prakash Dwivedi and Xavier Bresson. A generalization of transformer networks to graphs. ar Xiv preprint ar Xiv:2012.09699, 2020. 2 [17] Vijay Prakash Dwivedi, Chaitanya K Joshi, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. Benchmarking graph neural networks. ar Xiv preprint ar Xiv:2003.00982, 2020. 2 [18] Vijay Prakash Dwivedi, Anh Tuan Luu, Thomas Laurent, Yoshua Bengio, and Xavier Bresson. Graph neural networks with learnable structural and positional representations. In International Conference on Learning Representations, 2022. 2, 4 [19] Or Feldman, Amit Boyarski, Shai Feldman, Dani Kogan, Avi Mendelson, and Chaim Baskin. Weisfeiler and leman go infinite: Spectral and combinatorial pre-colorings. ar Xiv preprint ar Xiv:2201.13410, 2022. 4 [20] Pablo Gainza, Freyr Sverrisson, Frederico Monti, Emanuele Rodola, D Boscaini, MM Bronstein, and BE Correia. Deciphering interaction fingerprints from protein molecular surfaces using geometric deep learning. Nature Methods, 17(2):184 192, 2020. 5, 8 [21] Peter Gray and Stephen K Scott. Autocatalytic reactions in the isothermal, continuous stirred tank reactor: Oscillations and instabilities in the system a+ 2b 3b; b c. Chemical Engineering Science, 39(6):1087 1097, 1984. 5 [22] Amos Gropp, Lior Yariv, Niv Haim, Matan Atzmon, and Yaron Lipman. Implicit geometric regularization for learning shapes. In International Conference on Machine Learning, pages 3789 3799. PMLR, 2020. 2 [23] Junteng Jia and Austin R Benson. A unifying generative model for graph learning algorithms: Label propagation, graph convolutions, and combinations. SIAM Journal on Mathematics of Data Science, 4(1):100 125, 2022. 5 [24] Elizabeth Jurrus, Dave Engel, Keith Star, Kyle Monson, Juan Brandi, Lisa E Felberg, David H Brookes, Leighton Wilson, Jiahui Chen, Karina Liles, et al. Improvements to the apbs biomolecular solvation software suite. Protein Science, 27(1):112 128, 2018. 5 [25] Nicolas Keriven, Alberto Bietti, and Samuel Vaiter. Convergence and stability of graph convolutional networks on large random graphs. Advances in Neural Information Processing Systems, 33:21512 21523, 2020. 2 [26] Nicolas Keriven, Alberto Bietti, and Samuel Vaiter. On the universality of graph neural networks on large random graphs. Advances in Neural Information Processing Systems, 34, 2021. 2 [27] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. International Conference on Learning Representations, 2015. 4 [28] Lukas Koestler, Daniel Grittner, Michael Moeller, Daniel Cremers, and Zorah Lähner. Intrinsic neural fields: Learning functions on manifolds. ar Xiv preprint ar Xiv:2203.07967, 2022. 2 [29] Richard B Lehoucq, Danny C Sorensen, and Chao Yang. ARPACK users guide: solution of large-scale eigenvalue problems with implicitly restarted Arnoldi methods. SIAM, 1998. 4 [30] Ron Levie, Wei Huang, Lorenzo Bucci, Michael Bronstein, and Gitta Kutyniok. Transferability of spectral graph convolutional neural networks. Journal of Machine Learning Research, 22 (272):1 59, 2021. 2, 3 [31] Pan Li, Yanbang Wang, Hongwei Wang, and Jure Leskovec. Distance encoding: Design provably more powerful neural networks for graph representation learning. Advances in Neural Information Processing Systems, 33:4465 4478, 2020. 4 [32] Yunzhu Li, Shuang Li, Vincent Sitzmann, Pulkit Agrawal, and Antonio Torralba. 3d neural scene representations for visuomotor control. In Conference on Robot Learning, pages 112 123. PMLR, 2022. 2 [33] Derek Lim, Joshua Robinson, Lingxiao Zhao, Tess Smidt, Suvrit Sra, Haggai Maron, and Stefanie Jegelka. Sign and basis invariant networks for spectral graph representation learning. ar Xiv preprint ar Xiv:2202.13013, 2022. 4 [34] Yaron Lipman. Phase transitions, distance functions, and implicit neural representations. In International Conference on Machine Learning, pages 6702 6712. PMLR, 2021. 1 [35] Charles Loop. Smooth subdivision surfaces based on triangles. 1987. 7 [36] Andreas Loukas. What graph neural networks cannot learn: depth vs width. In International Conference on Learning Representations, 2019. 2 [37] László Lovász. Large networks and graph limits, volume 60. American Mathematical Soc., 2012. 2 [38] Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3d reconstruction in function space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4460 4470, 2019. 2 [39] Grégoire Mialon, Dexiong Chen, Margot Selosse, and Julien Mairal. Graphit: Encoding graph structure in transformers. ar Xiv preprint ar Xiv:2106.05667, 2021. 4 [40] Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In European Conference on Computer Vision, pages 405 421. Springer, 2020. 1, 2 [41] Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 165 174, 2019. 1, 2, 8 [42] Edoardo Remelli, Artem Lukoianov, Stephan Richter, Benoît Guillard, Timur Bagautdinov, Pierre Baque, and Pascal Fua. Meshsdf: Differentiable iso-surface extraction. Advances in Neural Information Processing Systems, 33:22468 22478, 2020. 2 [43] Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323 2326, 2000. 4 [44] Michel F Sanner, Arthur J Olson, and Jean-Claude Spehner. Reduced surface: an efficient way to compute molecular surfaces. Biopolymers, 38(3):305 320, 1996. 5 [45] Vincent Sitzmann, Julien Martel, Alexander Bergman, David Lindell, and Gordon Wetzstein. Implicit neural representations with periodic activation functions. Advances in Neural Information Processing Systems, 33:7462 7473, 2020. 1, 3, 4 [46] Ivan Skorokhodov, Savva Ignatyev, and Mohamed Elhoseiny. Adversarial generation of continuous images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10753 10764, 2021. 2 [47] Balasubramaniam Srinivasan and Bruno Ribeiro. On the equivalence between positional node embeddings and structural graph representations. In International Conference on Learning Representations, 2020. 2 [48] Kenneth O Stanley. Compositional pattern producing networks: A novel abstraction of development. Genetic Programming and Evolvable Machines, 8(2):131 162, 2007. 1, 3 [49] Jian Sun, Maks Ovsjanikov, and Leonidas Guibas. A concise and provably informative multiscale signature based on heat diffusion. In Computer graphics forum, volume 28, pages 1383 1392. Wiley Online Library, 2009. 4 [50] Matthew Tancik, Pratul Srinivasan, Ben Mildenhall, Sara Fridovich-Keil, Nithin Raghavan, Utkarsh Singhal, Ravi Ramamoorthi, Jonathan Barron, and Ren Ng. Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, 33:7537 7547, 2020. 2 [51] Joshua B Tenenbaum, Vin de Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290(5500):2319 2323, 2000. 4 [52] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in Neural Information Processing Systems, 30, 2017. 2 [53] Lior Yariv, Yoni Kasten, Dror Moran, Meirav Galun, Matan Atzmon, Basri Ronen, and Yaron Lipman. Multiview neural surface reconstruction by disentangling geometry and appearance. Advances in Neural Information Processing Systems, 33:2492 2502, 2020. 2 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Sections 3 and 4 (c) Did you discuss any potential negative societal impacts of your work? [N/A] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] At the beginning of Section 4 and in Section 4.1 (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]