# calibrated_physicsinformed_uncertainty_quantification__2b1c7888.pdf Calibrated Physics-Informed Uncertainty Quantification Vignesh Gopakumar 1 2 Ander Gray 3 Lorenzo Zanisi 2 Timothy Nunn 2 Daniel Giles 1 Matt J. Kusner 4 5 Stanislas Pamela 2 Marc Peter Deisenroth 1 Simulating complex physical systems is crucial for understanding and predicting phenomena across diverse fields, such as fluid dynamics and heat transfer, as well as plasma physics and structural mechanics. Traditional approaches rely on solving partial differential equations (PDEs) using numerical methods, which are computationally expensive and often prohibitively slow for real-time applications or large-scale simulations. Neural PDEs have emerged as efficient alternatives to these costly numerical solvers, offering significant computational speed-ups. However, their lack of robust uncertainty quantification (UQ) limits deployment in critical applications. We introduce a model-agnostic, physics-informed conformal prediction (CP) framework that provides guaranteed uncertainty estimates without requiring labelled data. By utilising a physics-based approach, we can quantify and calibrate the model s inconsistencies with the physics rather than the uncertainty arising from the data. Our approach utilises convolutional layers as finite-difference stencils and leverages physics residual errors as nonconformity scores, enabling data-free UQ with marginal and joint coverage guarantees across prediction domains for a range of complex PDEs. We further validate the efficacy of our method on neural PDE models for plasma modelling and shot design in fusion reactors. 1. Introduction Numerical PDE solvers are essential tools in scientific and engineering simulations (Danabasoglu et al., 2020; Giudi- 1Centre for Artificial Intelligence, University College London 2Computing Division, UK Atomic Energy Authority 3Heudiasyc Laboratory 4Polytechnique Montr eal 5Mila - Quebec AI Institute. Correspondence to: Vignesh Gopakumar . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). Figure 1. Neural PDE framework: Neural PDE solvers use data from traditional numerical solvers to quickly approximate PDEs across various conditions (shown by black arrows). To ensure reliability, these models incorporate uncertainty quantification (UQ) methods. If the predicted error exceeds a coverage threshold ϵ, the numerical solver is utilised, further adding to the training data; otherwise, predictions are used as output (shown by red arrows). This paper focuses on developing a new UQ method for assessing confidence in neural PDE models, emphasising the grey-shaded region of the framework. celli et al., 2024), but their computational demands and environmental impact pose significant challenges (Horwitz, 2024). Machine learning approaches have emerged as efficient alternatives for modelling/emulating PDEs (Bertone et al., 2019; Karniadakis et al., 2021), successfully deployed across weather forecasting (Kochkov et al., 2024; Meyer et al., 2022; Giles et al., 2024), fluid dynamics (Jiang et al., 2020; Pfaff et al., 2021), and nuclear fusion applications (Poels et al., 2023; Carey et al., 2024; Gopakumar & Samaddar, 2020). Neural PDE solvers provide rapid approximations, but present a critical cost-accuracy trade-off. While generating outputs consistently, their solutions may violate physical constraints or produce misleading results with high confidence (Gopakumar et al., 2023a). A typical neural PDE framework (see Figure 1) trains surrogate models on numerical simulation data to predict the evolution of PDE under various conditions, with uncertainty quantification (UQ) methods reverting to numerical solvers when predictions fail coverage thresholds. However, current UQ methods lack statistical guarantees (Zou et al., 2024), require extensive simulation data (Gopakumar et al., 2024a), or require architectural modifications (Abdar et al., 2021). To address these limitations, we propose a framework combining PDE residuals over neural PDEs with Conformal Prediction (CP) to provide uncertainty estimates that guarantee coverage. Our approach evaluates Physics Residual Errors Calibrated Physics-Informed UQ (PRE) from neural PDE solver predictions and performs calibration using marginal and joint CP formulations. Our method provides statistically valid coverage within the residual space (Vovk et al., 2005), offering error bounds against the violation of physical conservation laws. The framework is model-agnostic (applicable to all types of surrogate models) and does not require additional data generated from expensive numerical simulations. It yields interpretable uncertainty bounds indicating the model s physics-guided inconsistencies, addressing the (over)confidence issue of neural PDE solvers (Zou et al., 2024). Our contributions: Calibrated physics-informed UQ: A novel physicsinformed nonconformity metric using PDE residuals. This quantifies uncertainty through physical inconsistency rather than training data variation, providing input-independent prediction sets while relaxing exchangeability restrictions. Marginal and Joint CP: Our approach guarantees coverage bounds both marginally (univariate per dimension) and jointly (multivariate across the entire prediction domain), enabling the identification of volatile predictions and creating a rejection-acceptance criteria. 2. Related Work Recently, CP, as a method of performing UQ, has been gaining popularity for usage with spatio-temporal data (Sun, 2022). Several works have explored the inductive CP framework for spatial and sequential data (Stankeviciute et al., 2021; Xu & Xie, 2021; Xu et al., 2023), including in the operator space (Ma et al., 2024). Gopakumar et al. (2024a) extends the marginal-CP framework to pre-trained as well as to fine-tuned surrogate models for physical system modelling across an infinite-dimensional setting. Alternatively, error bounds for PDE surrogates have been devised by Gray et al. (2025) using set propagation to project the singular value decomposition of the prediction error onto the prediction space. Figure 2. Schematic of physics-informed uncertainty quantification workflow. Initial conditions generate neural PDE predictions autoregressively, over which physics residual errors are estimated. Calibration via marginal and joint conformal prediction yields error bars - pointwise for marginal-CP and domain-wide for joint-CP. The usage of PDE residuals under the guise of Physics Informed Machine Learning (PIML) (Karniadakis et al., 2021) was made popular as an optimisation strategy for Physics-Informed Neural Networks (PINNs) (Raissi et al., 2019) and has found application in optimising neural operators (Li et al., 2024) and soft/hard enforcement of the physical constraints to deep learning models (Du et al., 2024; Chalapathi et al., 2024). However, they have rarely been used as a tool for providing UQ to surrogate models, and where they have found application, UQ remained uncalibrated (Zhu et al., 2019). The majority of the literature in UQ for neural PDE solvers has been looking at Bayesian methods, such as dropout, Bayesian neural networks, and Monte Carlo methods (Geneva & Zabaras, 2020; Zou et al., 2024; Psaros et al., 2023), which lack guarantees or are computationally expensive. 3. Background 3.1. Neural PDE Solvers Consider the generic formulation of a PDE modelling the spatio-temporal evolution of n field variables u Rn across a range of initial conditions: D = Dt(u) + DX(u) = 0, X Ω, t [0, T], (1) u(X, t) = g, X Ω, (2) u(X, 0) = a(λ, X). (3) Here, X defines the spatial domain bounded by Ω, [0, T] the temporal domain, and DX and Dt the composite operators of the associated spatial and temporal derivatives. The PDE is further defined by the boundary condition g and initial condition a, which can be parameterised by λ. The set of solutions of field variables is expressed as u U. Neural PDE solvers as surrogate models learn the behaviour governed by Equation (1) using a parameterised neural network NN θ. Starting from the initial conditions, the network is trained to solve the spatio-temporal evolution of the fields given by Ω [0, T]. Neural operators NOθ are a special class of neural networks that learn the operator mapping from the function space of the PDE initial conditions a A to the function space of solutions u U. A neural operator for solving an initial-value problem can be expressed as U = NOθ(A), u(X, t) = NOθ u(X, 0), t , X Ω, t [0, T]. (4) A Fourier Neural Operator (FNO) is an autoregressive neural operator that learns the spatio-temporal evolution of PDE solutions by leveraging the Fourier transform as the kernel integrator (Li et al., 2021). The field evolution is learned using tunable weight matrices of the network, parameterised directly in the Fourier space of the PDE solutions. Since CP and our extension of it provide a post-hoc measure of quantifying the uncertainty of a neural PDE, it remains Calibrated Physics-Informed UQ agnostic to model choice and training conditions. Considering the model independence of our approach, we restrict our experiments to modelling PDEs with an FNO. The FNO is chosen due to its cost-accuracy trade-off and efficiency as demonstrated by de Hoop et al. (2022) and Gopakumar et al. (2023b). CP over a range of neural-PDE solvers has been applied by Gopakumar et al. (2024a), who also demonstrate that the coverage guarantees are upheld irrespective of the model choice, further necessitating us to not experiment with various model architectures. 3.2. Conformal Prediction Conformal prediction (CP) (Vovk et al., 2005; Shafer & Vovk, 2008) is a statistical framework that addresses the accuracy of a predictive model. Consider a machine learning model ˆf : X Y trained on a dataset (Xi, Yi)N i=1, that can be used to predict the next true label Yn+1 at query point Xn+1. CP extends the point prediction P : Yn+1 to a prediction set Cα, ensuring that P(Yn+1 Cα) 1 α. (5) This coverage guarantee, a function of the user-defined confidence level α, holds irrespective of the chosen model and training dataset. The only condition is that the calibration samples and the prediction samples are exchangeable. Traditional inductive CP partitions the labelled data into training and calibration sets (Papadopoulos, 2008). The performance of the model on the latter, measured using a nonconformity score, is used to calibrate the model and obtain prediction sets. Conventionally, nonconformity scores act on the model predictions and a labelled dataset (Kato et al., 2023). For deterministic models, they are often formulated as the Absolute Error Residual (AER) of the model predictions ˆf(X) and targets Y . For probabilistic models, the score function (STD) is the absolute value of the z-score (with the prediction mean, standard deviation and target given as ˆfµ(X), ˆfσ(X), Y respectively. Having obtained a distribution of nonconformity scores ˆs of the calibration dataset (Xi, Yi)n i=1, a quantile ˆq corresponding to the desired coverage 1 α is estimated from its cumulative distribution function Fˆs (Papadopoulos, 2008): ˆqα = F 1 ˆs (n + 1)(1 α) The quantile estimates the error bar associated with the desired coverage and is combined with the new prediction to obtain the prediction sets. The nonconformity score functions and their prediction sets for AER and STD are given in Table 1. Both AER and STD are data-intensive CP methods, requiring calibration data that converges to a beta distribution (Angelopoulos & Bates, 2023). This data dependence Table 1. Overview of nonconformity metrics AER, STD, and PRE and their corresponding score functions and prediction sets. Score Function (ˆs) Prediction Sets (Cα) | ˆf(Xi) Yi| n i=1 ˆf(Xn+1) ˆqα | ˆ fµ(Xi) Yi| i=1 ˆfµ(Xn+1) ˆqα ˆfσ(Xn+1) |D( ˆf(Xi))| n restricts their application to domains where sufficient data exists a priori or can be easily obtained 4. Physics Residual Error (PRE) We introduce a novel data-free nonconformity score based directly on the PDE for surrogate models. The Physics Residual Error (PRE) is defined as the PDE residual (Saad & Schultz, 1986) estimated over the discretised PDE solution obtained from the surrogate model. For an abstract PDE as in Equation (1), the PDE residual is the evaluation of the composite differential operator D over a field(s) u under the influence of an external force b D(u) b = 0, (7) The PDE residual is treated as a score function by taking its L1 norm as indicated in Table 1. While well-defined PDEs have solutions obeying Equations (1) to (3), numerical solutions often fail to converge to the true solution (Pinder, 2018). Neural PDEs, trained on approximate numerical data, are further prone to non-convergent predictions. In numerical analysis, the norm of the PDE residual is often used as a criterion for stability, convergence, and accuracy (Iserles, 2009). The PRE typically represents the violation of conservation laws associated with the physical system. Using the residual error as a nonconformity score quantifies the neural PDE solver s non-convergence to the physical ground truth of the PDE. By further using conformal prediction, we obtain coverage bounds over this conservative, residual space with guaranteed coverage. The norm |D(NOθ(u))| of the residual operator itself provides a measure of UQ for the neural PDE. However, it is limited by the accuracy of the gradient estimation method and can become computationally expensive when exploring a vast solution space (Tolsma & Barton, 1998). By using the residual norm as a nonconformity score, we further calibrate the approximate physics residual error that an inexpensive and coarse differential operator obtains. CP using PRE provides statistically valid and guaranteed error Calibrated Physics-Informed UQ bars across the PDE s residual space, incorporating physical information into the calibration procedure and providing a calibrated measure of the physical misalignment of the surrogate model. PRE as a nonconformity score enables data-free conformal prediction. The estimated scores rely only on the neural PDE predictions and not on the target as in AER and STD. The only criterion is that the calibration and prediction domains arise from exchangeable PDE conditions. As shown in Table 1, PRE gives prediction sets independent of the prediction inputs (see Appendix B for formalism). Traditional CP methods rely on calibration using labelled data to construct prediction intervals that contain the true values with a specified confidence level α. These methods guarantee that the true solution will lie within the estimated error bounds based on this empirical calibration. Contrastingly, our CP-PRE formulation takes a fundamentally different approach. Instead of requiring target data for calibration, we leverage the unique property of PDEs where the true solution in the residual space exists at zero. This eliminates the need for empirical calibration data. Our method focuses on ensuring that predictions themselves fall within coverage bounds Cα, rather than guaranteeing that the true solution lies within these bounds. This allows us to validate prediction sets without access to ground truth data a significant advantage over traditional CP approaches. We formalise this novel property in our theoretical framework presented in Appendix A. 4.1. Marginal-CP The CP formulation was initially conceptualised for calibrating univariate functions with single-point outputs (Vovk et al., 2005). It has recently been extended to spatiotemporal data, with multi-dimensional outputs with an immutable tensor structure (Gopakumar et al., 2024a). Within such spatio-temporal settings, CP has been implemented to provide marginal coverage, i.e. the calibration procedure provides independent error bars for each cell within the spatio-temporal domain. For an output tensor Y RNx Ny Nt, where Nx, Ny, Nt represent the spatiotemporal discretisation of the domain, marginal-CP uses the non-conformity scores outlined in Table 1 across each cell of Y to obtain error bars, which will be compliant with Equation (5) for each cell. Marginal-CP using PRE helps indicate spatio-temporal regions within individual predictions that lie outside the calibrated bounds of physics violation and require specific attention, treating those prediction regions with caution. 4.2. Joint-CP The joint-CP formulation constructs a calibration procedure that provides coverage bands for multivariate functions. These coverage bands expand across the entire simulation domain Ω [0, T] (discretised as RNx Ny Nt) rather than an individual cell within it. For a coverage band Cα, the joint-CP formulation ensures that 1 α predictions/solutions lie within the bounds. For performing joint-CP, the non-conformity scores are modified to reflect the supremum of the score functions ˆs in Table 1. They are modulated by the standard deviation σ of the calibration scores (Diquigiovanni et al., 2021) to obtain prediction bands with varying widths based on local behaviour (Diquigiovanni et al., 2022). The modifications of the score functions and prediction sets to perform CP are given by Sjoint = sup X Ω, t [0,T ] Cα joint = p(Xn+1) ˆqα σ(ˆs), (9) where ˆs and p(Xn+1) are the formulations of the nonconformity scores and the prediction at Xn+1 used for marginal CP as shown in Table 1. Joint-CP becomes particularly useful in identifying predictions that fail to fall within coverage, allowing us to accept or reject a prediction based on a predetermined probability. Similar to that demonstrated by Casella et al. (2004), our framework can perform acceptancerejection using a CP-based criterion. The acceptance probability is based on confidence level α. If a prediction is rejected, the PDE parameters that led to those predictions could be provided to the expensive physics-based numerical PDE solver for further evaluation as indicated in Figure 1. 4.3. Differential Operator: Finite-Difference Stencils as Convolutional Kernels Calibrating neural PDEs using PRE nonconformity scores requires frequent evaluations of the composite differential operator D in Equation (1). For PDEs, this involves estimating spatio-temporal gradients across the discretised domain, ranging from millions in simple cases to billions of gradient operations for complex physics. To address this computational challenge, we developed a scalable gradient estimation method for evaluating PRE. We use convolution operations with Finite Difference (FD) stencils as convolutional kernels for gradient estimation (Actor et al., 2020; Chen et al., 2024a;b). For instance, the 2D Laplacian operator 2, using a central difference scheme with discretisation h, can be approximated by 0 1 0 1 4 1 0 1 0 and used as a kernel. This approach is justified by the mathematical equivalence of FD approximations and discrete convolutions. Both represent matrix-vector multiplications of a block Toeplitz matrix with a field vector (Strang, 1986; Calibrated Physics-Informed UQ Fiorentino & Serra, 1991). The efficiency of this method stems from the optimised implementation of convolution operations in machine learning libraries like Py Torch (Paszke et al., 2019) and Tensor Flow (Abadi et al., 2015). 1 The FD approximation offers several advantages over Automatic Differentiation (AD) for our application. It is compatible with CP as a post-hoc measure, requires no architectural modifications, and is model-agnostic. Furthermore, FD implemented via convolutions is more memory-efficient than AD, which requires storing the entire computational graph. As we focus on the (mis)alignment of neural PDEs with Equation (1), we disregard boundary conditions, but demonstrate that they can be accounted for with convolutional padding in Appendix E (Alguacil et al., 2021). Utilising finite difference schemes, the residuals are estimated up to the truncation errors of the Taylor approximation (Mac Kinnon & Johnson, 1991). Although discretisation plays a role in the width of the error bars, CP-PRE still guarantees coverage, and this is further explored in Appendix D.1. 5. Experiments Figure 3. Validation plots demonstrating coverage guarantee detailed in Equation (5) obtained by performing CP using PRE across experiments. The average empirical coverage obtained experimentally is given on the y-axis (ranging from 0 to 1, with 1 representing 100% coverage), while the theoretical coverage is represented on the x-axis. We obtain guaranteed coverage while using marginal-CP formulation and near-to-ideal coverage for the joint-CP formulation. CP-PRE experiments comprise two campaigns. First, we benchmark CP-PRE within standard neural PDEs (Section 5.1 to 5.3). The calibration process (Figure 2) involves: (a) sampling inputs to the model to generate predictions, (b) calculating PRE(s) scores, and (c) calibrating physical error using marginal and joint-CP formulations. Validation uses the same PDE condition bounds as calibration. Within this campaign, we compare our method (CP-PRE) with other methods of providing uncertainty estimation for neural- 1The Basic Linear Algebra Subroutines (BLAS) in these libraries leverage vectorisation and efficient memory access for significant performance gains. Our experiments demonstrate a 1000x speedup using torch.nn.functional.conv3d versus an equivalent numpy finite difference implementation on standard CPU. PDEs. We compare various Bayesian methods along with multivariate inductive conformal prediction to our method as shown in Table 2. We demonstrate that our method is capable of providing valid coverage guarantees for both in and out-of-distribution testing without a significant increase in evaluation times (including time taken for sampling in Bayesian methods, data generation in data-driven CP and subsequent calibration). Methods that attain the required coverage are emboldened. Table 2 provides a qualitative comparison of our method against the other benchmarks and highlights that our method is data-free, requires no modification or sampling and provides guaranteed coverage in a physics-informed manner. We confine our comparison studies in Table 3, 4, and 5 to the Wave, Navier-Stokes and Magnetohydrodynamic equations. The second campaign (Section 5.4, 5.5) applies CP-PRE to fusion applications. We enhance tokamak plasma behaviour surrogate models to identify erroneous dispersion regions (Section 5.4) and integrate CP-PRE with tokamak design surrogates to identify viable designs and areas needing additional simulations (Section 5.5). This campaign demonstrates the utility of CP-PRE in complex, practical applications. One-dimensional PDE experiments demonstrating CP-PRE are demonstrated in Appendix F and can be reproduced using code in the supplementary material. 5.1. Wave Equation The two-dimensional wave equation is given by t2 = c2 2u x2 + 2u We solve the field u given in Equation (11) within the domain x, y [ 1, 1], t [0, 1.0] with wave velocity c = 1.0 using a spectral solver with periodic boundary conditions (Canuto et al., 2007). The initial conditions are parameterised by the amplitude and position of a Gaussian field. A 2D FNO is trained on this data to predict 20-time steps autoregressively from a given initial state. Figure 4 compares the model predictions against ground truth, showing the PRE and confidence bounds from marginal and joint-CP at 90% coverage. The PRE reveals noise artefacts in regions where the field should be zero, highlighting physical inconsistencies despite apparently accurate predictions. Joint-CP bounds are necessarily larger than marginal-CP bounds as they span across the spatio-temporal domain as opposed to being cell-wise. As demonstrated in Figure 3, both CP approaches achieve the expected coverage guarantees. Table 3 highlights the superior performance of CP-PRE over other methods of UQ, where it guarantees coverage when evaluated for in and out-of-distribution without requiring any additional data, drastically reducing the evaluation time. Marginal-CP shows linear coverage due to cell-wise averag- Calibrated Physics-Informed UQ Method Data-Free Modification-Free Sampling-Free Guaranteed Coverage Physics-Informed MC Dropout (Gal & Ghahramani, 2016) ! % % % % Deep Ensemble (Lakshminarayanan et al., 2017) ! % % % % BNN (Mac Kay, 1992) ! % % % % SWA-G (Maddox et al., 2019) ! % % % % CP-AER (Gopakumar et al., 2024a) % ! ! ! % CP-PRE (Ours) ! ! ! ! ! Table 2. Comparing features across various UQ measures. Our method is data-free, does not require any modifications or sampling, and helps obtain guaranteed coverage bounds in a physics-informed manner. Figure 4. Wave: (From left to right) neural PDE (FNO) prediction at the last time instance, physics residual error of the prediction, Upper error bars obtained by performing marginal-CP and joint-CP respectively (90% coverage). For brevity, we have only shown the upper error bars of the symmetric prediction sets. mod represents the modulation function in Table 1. The physical inconsistencies within the residual space of the prediction are calibrated and bounded using CP-PRE. ing. At the same time, joint-CP exhibits coverage variations that depend on the calibration dataset (see Table 3 for stability analysis across multiple calibration sets). Additional experimental details are provided in Appendix J. 5.2. Navier-Stokes Equation Consider the two-dimensional Navier-Stokes equations v = 0, (12) (Continuity equation) t + ( v ) v = ν 2 v P, (13) (Momentum equation) where we are interested in modelling the evolution of the velocity vector ( v = [u, v]) and pressure (P) field of an incompressible fluid with kinematic viscosity (ν). For data generation, Equations (12) and (13) are solved on a domain x [0, 1], y [0, 1], t [0, 0.5] using a spectral-based solver (Canuto et al., 2007). A 2D multi-variable FNO (Gopakumar et al., 2024b) is trained to model the evolution of velocity and pressure autoregressively up until the 20th time instance. Unlike the previous example, the Navier-Stokes case has two equations with corresponding PRE estimates: continuity (Equation (12)) and momentum (Equation (13)) for mass and momentum conservation. Our CP-PRE method calibrates model deviation from physical ground truth for each equation. Table 4 shows that Bayesian methods fail to provide coverage for FNO modelling, while multivariate CP (CP-AER) provides coverage but requires substantially higher evaluation time arising from the generation of simulation data for calibration. Figures 20 and 21 show PRE bounds from both equations. Having two PDE residuals enables easier rejection of predictions violating both bounds. Physics details, parameterisation, and training are in Appendix K. Within the scope of this paper, we limit ourselves to measuring the deviation of the model with the PDE residual. The CP-PRE formulation can be extended to obtain bounds for both the initial and boundary conditions, further explored within Appendix E. 5.3. Magnetohydrodynamics Consider the magnetohydrodynamic (MHD) equations t + (ρ v) = 0, (Continuity equation) (14) t + v v = 1 µ0 B ( B) P, (15) (Momentum equation) = 0, (Energy equation) (16) t = ( v B), (Induction equation) (17) B = 0, (Gauß law for magnetism) (18) Calibrated Physics-Informed UQ in-distribution out-of-distribution Time UQ MSE Coverage (95%) MSE Coverage (95%) Train (hr) Eval (s) Deterministic 1.77e-05 3.69e-07 - 2.46e-03 2.00e-05 - 0:38 22 MC Dropout 1.44e-04 3.26e-06 97.31 0.03 2.12e-03 2.60e-05 89.83 0.07 0:52 120 Deep Ensemble 8.76e-06 2.43e-07 98.02 0.04 2.42e-03 1.58e-05 83.44 0.12 3:10 112 BNN 1.92e-04 1.92e-06 97.10 0.09 2.67e-03 1.26e-05 91.76 0.10 0:53 118 SWA-G 1.41e-05 1.74e-06 94.55 3.25 2.55e-03 2.82e-05 81.90 3.31 0:47 113 CP-AER 1.76e-05 4.40e-07 95.70 0.21 2.46e-03 1.41e-05 95.59 0.14 0:38 2022 CP-PRE (Ours) 1.78e-05 4.61e-07 95.52 0.21 2.46e-03 1.25e-05 95.39 0.12 0:38 32 Table 3. Wave Equation CP-PRE guarantees coverage across distributions while providing the quickest evaluation time. in-distribution out-of-distribution Time UQ MSE Coverage (95%) MSE Coverage (95%) Train (hr) Eval (s) Deterministic 1.05e-04 6.91e-06 - 3.67e-03 5.30e-05 - 3:22 25 MC Dropout 5.96e-04 2.30e-05 82.21 0.22 4.30e-03 8.05e-05 44.05 0.26 3:34 153 Deep Ensemble 1.22e-04 3.95e-06 91.31 0.08 3.67e-03 3.52e-05 30.74 0.19 16:22 147 BNN 6.90e-03 1.31e-04 89.91 0.20 6.95e-03 1.31e-04 85.19 0.23 3:39 152 SWA-G 1.96e-04 1.15e-05 84.22 2.37 3.63e-03 1.37e-04 31.00 2.85 3:28 146 CP-AER 1.05e-04 6.58e-06 95.56 0.40 3.66e-03 2.81e-05 95.54 0.15 3:22 20026 CP-PRE (Ours) 1.07e-04 5.18e-06 95.44 0.22 3.70e-03 4.23e-05 95.57 0.14 3:22 134 Table 4. Navier-Stokes Equations CP-PRE guarantees coverage across distributions while providing the quickest evaluation time. where the density (ρ), velocity vector ( v = [u, v]) and the pressure of plasma is modelled under a magnetic field ( B = [Bx, By]) across a spatio-temporal domain x, y [0, 1]2, t [0, 5]. µ0 is the magnetic permeability of free space. Equations (14) to (18) represent the ideal MHD equations as a combination of the Navier-Stokes equations for fluid flow with Maxwell s equations of electromagnetism (Alfv en, 1942; Gruber & Rappaz, 1985; Mocz et al., 2014). The equations assume perfect conductivity (no magnetic diffusivity) and no viscosity. We focus our experiment on the modelling of the Orszag-Tang vortex of a turbulent plasma (Orszag & Tang, 1979) with the data being generated using a finite volume method (Eymard et al., 2000). A 2D FNO is trained to model the evolution of all 6 variables over a dataset generated by parameterised initial conditions. Equations (14) to (18) provide us with five measures of estimating the PRE of the MHD surrogate model. Each PRE estimate depends on a different set of variables associated with the system, allowing us to infer errors contributed to each variable accordingly. Table 5 demonstrates that the Bayesian methods fail miserably in providing valid error bars over the MHD case. The CP-AER provides guaranteed coverage but is heavily dependent on simulation data, adding to the computational expense. CP-PRE using induction (Equation (17)) and energy (Equation (16)) are shown for 90% coverage (α = 0.1), sliced at y = 0.5m. The plots show PRE along the x-axis with marginal and joint bounds at a specific simulation time. Marginal bounds are significantly tighter, while joint bounds are wider but useful for identifying predictions that violate conservation equations. Additional CP plots using other residuals (Figures 27 and 28) and physics/surrogate model details are in Appendix L. 5.4. Plasma Modelling within a Tokamak In (Gopakumar et al., 2024b), the authors model the evolution of plasma blobs within a fusion reactor (known as a tokamak) using an FNO. They explore the case of electrostatic modelling of reduced magnetohydrodynamics with data obtained from the JOREK code (Hoelzl et al., 2021). In the absence of magnetic pressure to confine it, the plasma, driven by kinetic pressure, moves radially outward and collides with the wall of the reactor. The plasma is characterised by density ρ, electric potential ϕ and Temperature T, and the FNO models their spatio-temporal evolution autoregressively. Borrowing that pre-trained model and utilising the reduced-MHD equations within the toroidal domain, we demonstrate obtaining calibrated error bars using CP-PRE at scale. The FNO demonstrated in (Gopakumar et al., 2024b) can model the plasma six orders of magnitude faster than traditional numerical solvers, and by providing calibrated error bars over the predictions, a wider range of plasma configurations can be validated. We focus on the temperature equation within reduced-MHD (equation 3 within (Gopakumar et al., 2024b)) as it comprises all the variables associated with the plasma. As shown Calibrated Physics-Informed UQ in-distribution out-of-distribution Time UQ MSE Coverage (95%) MSE Coverage (95%) Train (hr) Eval (s) Deterministic 2.20e-03 5.20e-03 - 4.71e-02 1.06e-03 - 5:00 40 MC Dropout 3.29e-02 5.86e-04 41.13 0.19 2.09e-01 1.38e-03 16.91 0.06 5:30 240 Deep Ensemble 3.59e-03 3.51e-04 78.15 0.16 3.41e-01 3.15e-02 39.63 0.31 26:25 235 BNN 4.20e-03 4.08e-05 90.24 0.10 4.63e-02 8.98e-04 62.37 0.46 5:40 240 SWA-G 2.61e-03 9.68e-05 48.50 3.81 4.53e-02 6.64e-04 14.22 1.35 5:22 236 CP-AER 2.20e-03 4.38e-05 95.61 0.26 4.69e-02 8.18e-04 95.60 0.27 5:00 40042 CP-PRE (Ours) 2.20e-03 4.96e-03 95.54 0.18 4.71e-02 1.06e-03 95.67 0.22 5:00 482 Table 5. Magnetohydrodynamic Equations CP-PRE guarantees coverage across distributions with only a marginal increase in evaluation time, arising from residual evaluation over a larger family of PDEs (see Equations (14) to (18)). Absolute Error: (T) (a) Abs. Error PRE: D( , , T) Figure 5. Reduced MHD: CP-PRE using the Temperature equation (Eqn. 3 in (Gopakumar et al., 2024b)) of reduced-MHD to bound the plasma surrogate models. The PRE captures the model error relatively well, allowing us to provide lower and upper error bars corresponding to our required coverage. in Figure 5, our method can capture the model error across a range of predictions and devise error bars that provide guaranteed coverage without additional data. In figure 5(a), we demonstrate the absolute error in the model prediction of the temperature evolution, correlating that with the PRE over the temperature equations in figure 5(b). By obtaining bounds on the PRE, we can determine the efficacy of the surrogate model in evaluating plasma evolution and identify conditions under which it fails, running the MHD code JOREK under those failed conditions. 5.5. Magnetic Equilibrium in a Tokamak Tokamaks confine plasma within a toroidal vessel using magnetic fields to achieve nuclear fusion. The plasma at high temperatures is contained by magnetic fields that counterbalance its kinetic pressure. This equilibrium state, a function of magnetic coil configurations and plasma parameters, is governed by the Grad-Shafranov (GS) equation (Somov, 2012) z2 = µ0r2 dp where ψ represents the poloidal magnetic flux, p the kinetic pressure, F = r B the toroidal magnetic field, and µ0 the magnetic permeability given in a 2D toroidal coordinate system characterised by r and z. While traditional numerical solvers like EFIT++ and Free GSNKE (Lao et al., 1985; Amorisco et al., 2024) are used for equilibrium reconstruction, their computational cost has motivated neural network alternatives (Joung et al., 2023; Jang et al., 2024). However, these surrogate models lack UQ capabilities, making their deployment within the control room challenging. Figure 6. Grad-Shafranov: The PRE for a specific poloidal field coil configuration is indicated on the left, and the lower and upper bars for 50% are displayed adjacent to it. Aside from guaranteeing coverage, the CP-PRE framework allows us to discard physically inconsistent equilibria predicted by the surrogate model. We implement an auto-encoder that maps poloidal magnetic flux across the poloidal cross-section for given tokamak architectures, conditioned on poloidal field coil locations under constant plasma current. While this accelerates simulation by 4 orders of magnitude, it lacks physical guarantees. By incorporating Equation (19) within the CP-PRE framework, we identify physically stable equilibria and obtain statistically valid error bounds. Figure 6 shows the PRE over a surrogate model prediction with lower and upper error bars for 50% coverage. Further details about the problem setting and the model can be found in Appendix N. Calibrated Physics-Informed UQ 6. Extensions of CP-PRE beyond PDEs Our framework applies to any prediction case where the forward problem can be formulated as a residual with an equality constraint (Appendices A and B). The framework works in any scenario where the forward model can be expressed in the standard canonical form Ax b = 0, (20) where x is the model prediction, A is a differential or algebraic operator governing the dynamics, and b is a nonhomogeneous term such as a function or a constant. This extends our applications beyond PDEs to ODEs and algebraic equations found in control problems (Jiang & Jiang, 2014), chemical reactions (Th oni et al., 2025), biological systems (Wang et al., 2019), and financial scenarios (Liu et al., 2019). We deliberately focus on PDEs as they represent the most comprehensive and challenging case multi-dimensional domains with complex spatio-temporal dependencies and unique computational challenges. Success with PDEs implicitly validates applicability to simpler systems. 7. Discussion If All models are wrong, but some are useful (Box, 1976), through this work, we explore a novel framework for providing a measure of usefulness of neural PDEs. We deploy a principled method of evaluating the accuracy of the solution, i.e. its (calibrated) obedience to the known physics of the system under study. As opposed to other methods of UQ for neural PDEs, our method has the unique advantage of being physics-informed. This allows us to study the physical inconsistencies of the model predictions with coverage guarantees provided by conformal prediction. We conclude with a discussion of the strengths, limitations and potential improvements. Strengths PRE estimates the violation of conservation laws in neural PDE predictions, guaranteed error bounds over the physics deviation. This post-hoc uncertainty quantification is modeland physics-agnostic, scaling linearly with model complexity and quasi-linearly with PDE complexity due to the additive nature of differential operators. Our framework reformulates CP to be data-free, expressing model inaccuracy solely through PRE, not requiring a labelled dataset. This approach reduces calibration costs and loosens exchangeability restrictions as we can modify the calibration and, hence, the prediction domain by simply reformulating the PRE accordingly. The PRE formulation (Section 4, Appendix B) yields input-independent prediction sets, allowing for the identification of weak prediction regions within single simulations (marginal-CP) and across multiple predictions (joint-CP). The latter enables a rejection criterion for a set of predictions, potentially serving as an active-learning pipeline for neural PDE solvers (Musekamp et al., 2025). CP-PRE provides guaranteed coverage irrespective of the model, chosen discretisation, or the PDE of interest; however, the width of the error bar indicates quantitative features in the model quality. A well-trained model will exhibit tighter error bars as opposed to a poorer fit model; see Appendix I. Limitations Our method s coverage bounds exist in the PDE residual space rather than the Euclidean space of physical variables. Transforming to physical space involves challenging set propagation through integral operations, which may require approximations (Teng et al., 2023) or expensive Monte Carlo sampling (Andrieu et al., 2003). The data-free approach lacks a grounding target for calibration, though we argue that a large sample of model outputs provides a statistically significant overview of uncertainty. The sampling cost from the neural-PDE solver for calibration involves intensive gradient evaluations. PRE estimation using finitedifference stencils also introduces the errors associated with Taylor expansion. The current formulation is limited to regular grids with fixed spacing, though extensions to unstructured grids via graph convolutions are possible (Eliasof & Treister, 2020). CP-PRE does not help differentiate between aleatoric and epistemic uncertainty. It aligns with conformal prediction s characterisation of predictive uncertainty. From one perspective, this could be viewed as aleatoric uncertainty since we construct confidence intervals relative to a specific probability distribution (the distribution from which calibration data, i.e. initial conditions/PDE coefficients, are sampled). Alternatively, it could be considered epistemic uncertainty since we model the neural network s error through confidence intervals (typically used for unknown but fixed quantities). While we believe the latter interpretation is more appropriate, we acknowledge that the traditional aleatoric/epistemic dichotomy may not be directly applicable to our framework. This distinction is most valuable when both uncertainty types coexist and require separate treatment (Ferson & Ginzburg, 1996). 8. Conclusion We address the problem of the reliability of neural-PDE solvers by proposing CP-PRE, a novel conformal prediction framework. Our method provides guaranteed and physics-informed uncertainty estimates for each cell within a prediction, identifying erroneous regions while discerning physically inconsistent predictions across the entire spatiotemporal domain. Our work enhances the reliability of neural PDE solvers, potentially broadening their applicability in science and engineering domains where robust uncertainty quantification is crucial. Calibrated Physics-Informed UQ Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none of which we feel must be specifically highlighted here. Calibrated Physics-Informed UQ Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., Corrado, G. S., Davis, A., Dean, J., Devin, M., Ghemawat, S., Goodfellow, I., Harp, A., Irving, G., Isard, M., Jia, Y., Jozefowicz, R., Kaiser, L., Kudlur, M., Levenberg, J., Man e, D., Monga, R., Moore, S., Murray, D., Olah, C., Schuster, M., Shlens, J., Steiner, B., Sutskever, I., Talwar, K., Tucker, P., Vanhoucke, V., Vasudevan, V., Vi egas, F., Vinyals, O., Warden, P., Wattenberg, M., Wicke, M., Yu, Y., and Zheng, X. Tensor Flow: Largescale machine learning on heterogeneous systems, 2015. URL https://www.tensorflow.org/. Software available from tensorflow.org. Abdar, M., Pourpanah, F., Hussain, S., Rezazadegan, D., Liu, L., Ghavamzadeh, M., Fieguth, P., Cao, X., Khosravi, A., Acharya, U. R., Makarenkov, V., and Nahavandi, S. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Information Fusion, 76:243 297, 2021. ISSN 1566-2535. doi: https://doi.org/10.1016/j.inffus.2021.05. 008. URL https://www.sciencedirect.com/ science/article/pii/S1566253521001081. Actor, J., Fuentes, D. T., and Riviere, B. Identification of kernels in a convolutional neural network: connections between the level set equation and deep learning for image segmentation. In Landman, B. A. and Iˇsgum, I. (eds.), Medical Imaging 2020: Image Processing. SPIE, March 2020. Alfv en, H. Existence of electromagnetic-hydrodynamic waves. Nature, 150(3805):405 406, Oct 1942. ISSN 1476-4687. doi: 10.1038/150405d0. URL https:// doi.org/10.1038/150405d0. Alguacil, A., Pinto, W. G., Bauerheim, M., Jacob, M. C., and Moreau, S. Effects of boundary conditions in fully convolutional networks for learning spatio-temporal dynamics. In Dong, Y., Kourtellis, N., Hammer, B., and Lozano, J. A. (eds.), Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track, pp. 102 117, Cham, 2021. Springer International Publishing. ISBN 978-3-030-86517-7. Amorisco, N. C., Agnello, A., Holt, G., Mars, M., Buchanan, J., and Pamela, S. Freegsnke: A python-based dynamic free-boundary toroidal plasma equilibrium solver. Physics of Plasmas, 31(4):042517, Apr 2024. ISSN 1070-664X. doi: 10.1063/5.0188467. URL https: //doi.org/10.1063/5.0188467. Andrieu, C., de Freitas, N., Doucet, A., and Jordan, M. I. An introduction to mcmc for machine learning. Machine Learning, 50(1):5 43, Jan 2003. ISSN 1573- 0565. doi: 10.1023/A:1020281327116. URL https: //doi.org/10.1023/A:1020281327116. Angelopoulos, A. N. and Bates, S. Conformal prediction: A gentle introduction. Found. Trends Mach. Learn., 16 (4):494 591, mar 2023. ISSN 1935-8237. doi: 10.1561/ 2200000101. URL https://doi.org/10.1561/ 2200000101. Bartolucci, F., de Bezenac, E., Raonic, B., Molinaro, R., Mishra, S., and Alaifari, R. Representation equivalent neural operators: a framework for alias-free operator learning. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https: //openreview.net/forum?id=7LSEkv EGCM. Bertone, G., Deisenroth, M. P., Kim, J. S., Liem, S., Ruiz de Austri, R., and Welling, M. Accelerating the bsm interpretation of lhc data with machine learning. Physics of the Dark Universe, 24:100293, 2019. Box, G. E. P. Science and statistics. Journal of the American Statistical Association, 71(356):791 799, 1976. doi: 10. 1080/01621459.1976.10480949. Canuto, C., Hussaini, M., Quarteroni, A., and Zang, T. Spectral Methods: Evolution to Complex Geometries and Applications to Fluid Dynamics. Scientific Computation. Springer Berlin Heidelberg, 2007. ISBN 9783540307280. URL https://books.google. co.uk/books?id=7COg Ew5_EBQC. Carey, N., Zanisi, L., Pamela, S., Gopakumar, V., Omotani, J., Buchanan, J., and Brandstetter, J. Data efficiency and long term prediction capabilities for neural operator surrogate models of core and edge plasma codes, 2024. URL https://arxiv.org/abs/2402.08561. Casella, G., Robert, C. P., and Wells, M. T. Generalized accept-reject sampling schemes. Lecture Notes-Monograph Series, 45:342 347, 2004. ISSN 07492170. URL http://www.jstor.org/ stable/4356322. Chalapathi, N., Du, Y., and Krishnapriyan, A. S. Scaling physics-informed hard constraints with mixtureof-experts. In The Twelfth International Conference on Learning Representations, 2024. URL https:// openreview.net/forum?id=u3d X2CEIZb. Chen, B., Heaney, C. E., Gomes, J. L., Matar, O. K., and Pain, C. C. Solving the discretised multiphase flow equations with interface capturing on structured grids using machine learning libraries. Computer Methods in Applied Mechanics and Engineering, 426:116974, 2024a. ISSN 0045-7825. doi: https://doi.org/10.1016/j.cma.2024.116974. Calibrated Physics-Informed UQ URL https://www.sciencedirect.com/ science/article/pii/S0045782524002305. Chen, B., Heaney, C. E., and Pain, C. C. Using ai libraries for incompressible computational fluid dynamics, 2024b. URL https://arxiv.org/abs/2402.17913. Crank, J. and Nicolson, P. A practical method for numerical evaluation of solutions of partial differential equations of the heat-conduction type. Mathematical Proceedings of the Cambridge Philosophical Society, 43(1):50 67, 1947. doi: 10.1017/S0305004100023197. Danabasoglu, G., Lamarque, J.-F., Bacmeister, J., Bailey, D. A., Du Vivier, A. K., Edwards, J., Emmons, L. K., Fasullo, J., Garcia, R., Gettelman, A., Hannay, C., Holland, M. M., Large, W. G., Lauritzen, P. H., Lawrence, D. M., Lenaerts, J. T. M., Lindsay, K., Lipscomb, W. H., Mills, M. J., Neale, R., Oleson, K. W., Otto-Bliesner, B., Phillips, A. S., Sacks, W., Tilmes, S., van Kampenhout, L., Vertenstein, M., Bertini, A., Dennis, J., Deser, C., Fischer, C., Fox-Kemper, B., Kay, J. E., Kinnison, D., Kushner, P. J., Larson, V. E., Long, M. C., Mickelson, S., Moore, J. K., Nienhouse, E., Polvani, L., Rasch, P. J., and Strand, W. G. The community earth system model version 2 (cesm2). Journal of Advances in Modeling Earth Systems, 12(2):e2019MS001916, 2020. doi: https://doi.org/10.1029/2019MS001916. URL https://agupubs.onlinelibrary.wiley. com/doi/abs/10.1029/2019MS001916. e2019MS001916 2019MS001916. de Hoop, M. V., Huang, D. Z., Qian, E., and Stuart, A. M. The cost-accuracy trade-off in operator learning with neural networks, 2022. URL https://arxiv.org/ abs/2203.13181. Diquigiovanni, J., Fontana, M., and Vantini, S. The importance of being a band: Finite-sample exact distributionfree prediction sets for functional data, 2021. URL https://arxiv.org/abs/2102.06746. Diquigiovanni, J., Fontana, M., and Vantini, S. Conformal prediction bands for multivariate functional data. Journal of Multivariate Analysis, 189:104879, 2022. ISSN 0047259X. doi: https://doi.org/10.1016/j.jmva.2021.104879. URL https://www.sciencedirect.com/ science/article/pii/S0047259X21001573. Du, Y., Chalapathi, N., and Krishnapriyan, A. S. Neural spectral methods: Self-supervised learning in the spectral domain. In The Twelfth International Conference on Learning Representations, 2024. URL https: //openreview.net/forum?id=2Db Veuoa6a. Eliasof, M. and Treister, E. Diffgcn: Graph convolutional networks via differential operators and algebraic multigrid pooling. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 18016 18027. Curran Associates, Inc., 2020. URL https://proceedings.neurips. cc/paper_files/paper/2020/file/ d16a974d4d6d0d71b29bfbfe045f1da7-Paper. pdf. Eymard, R., Gallou et, T., and Herbin, R. Finite volume methods. In Lions, J. L. and Ciarlet, P. (eds.), Solution of Equation in Rn (Part 3), Techniques of Scientific Computing (Part 3), volume 7 of Handbook of Numerical Analysis, pp. 713 1020. Elsevier, 2000. ISBN 9780444503503. doi: 10.1016/S1570-8659(00)070058. Ferson, S. and Ginzburg, L. R. Different methods are needed to propagate ignorance and variability. Reliability Engineering & System Safety, 54(2):133 144, 1996. ISSN 0951-8320. doi: https://doi.org/10.1016/S0951-8320(96)00071-3. URL https://www.sciencedirect.com/ science/article/pii/S0951832096000713. Treatment of Aleatory and Epistemic Uncertainty. Fiorentino, G. and Serra, S. Multigrid methods for toeplitz matrices. CALCOLO, 28(3):283 305, Sep 1991. ISSN 1126-5434. doi: 10.1007/BF02575816. URL https: //doi.org/10.1007/BF02575816. Gal, Y. and Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In International Conference on Machine Learning, 2016. Geneva, N. and Zabaras, N. Modeling the dynamics of pde systems with physics-constrained deep auto-regressive networks. Journal of Computational Physics, 403:109056, 2020. ISSN 00219991. doi: https://doi.org/10.1016/j.jcp.2019.109056. URL https://www.sciencedirect.com/ science/article/pii/S0021999119307612. Giles, D., Briant, J., Morcrette, C. J., and Guillas, S. Embedding machine-learnt sub-grid variability improves climate model precipitation patterns. Communications Earth & Environment, 5(1):712, Nov 2024. ISSN 26624435. doi: 10.1038/s43247-024-01885-8. URL https: //doi.org/10.1038/s43247-024-01885-8. Giudicelli, G., Lindsay, A., Harbour, L., Icenhour, C., Li, M., Hansel, J. E., German, P., Behne, P., Marin, O., Stogner, R. H., Miller, J. M., Schwen, D., Wang, Y., Munday, L., Schunert, S., Spencer, B. W., Yushu, D., Recuero, A., Prince, Z. M., Nezdyur, M., Hu, T., Miao, Y., Jung, Y. S., Matthews, C., Novak, A., Langley, B., Truster, T., Nobre, N., Alger, B., Andrˇs, D., Kong, F., Carlsen, R., Slaughter, A. E., Calibrated Physics-Informed UQ Peterson, J. W., Gaston, D., and Permann, C. 3.0 - MOOSE: Enabling massively parallel multiphysics simulations. Software X, 26:101690, 2024. ISSN 23527110. doi: https://doi.org/10.1016/j.softx.2024.101690. URL https://www.sciencedirect.com/ science/article/pii/S235271102400061X. Gopakumar, V. and Samaddar, D. Image mapping the temporal evolution of edge characteristics in tokamaks using neural networks. Machine Learning: Science and Technology, 1(1):015006, 2020. doi: 10.1088/2632-2153/ ab5639. URL https://dx.doi.org/10.1088/ 2632-2153/ab5639. Gopakumar, V., Pamela, S., and Samaddar, D. Loss landscape engineering via data regulation on PINNs. Machine Learning with Applications, 12:100464, 2023a. ISSN 2666-8270. doi: https://doi.org/10.1016/j.mlwa.2023.100464. URL https://www.sciencedirect.com/ science/article/pii/S2666827023000178. Gopakumar, V., Pamela, S., Zanisi, L., Li, Z., Anandkumar, A., and Team, M. Fourier neural operator for plasma modelling, 2023b. URL https://arxiv.org/abs/ 2302.06542. Gopakumar, V., Gray, A., Oskarsson, J., Zanisi, L., Pamela, S., Giles, D., Kusner, M., and Deisenroth, M. P. Uncertainty quantification of surrogate models using conformal prediction, 2024a. URL https://arxiv.org/ abs/2408.09881. Gopakumar, V., Pamela, S., Zanisi, L., Li, Z., Gray, A., Brennand, D., Bhatia, N., Stathopoulos, G., Kusner, M., Deisenroth, M. P., Anandkumar, A., the JOREK Team, and Team, M. Plasma surrogate modelling using fourier neural operators. Nuclear Fusion, 64(5):056025, 4 2024b. doi: 10.1088/1741-4326/ad313a. URL https://dx. doi.org/10.1088/1741-4326/ad313a. Gray, A., Gopakumar, V., Rousseau, S., and Destercke, S. Guaranteed confidence-band enclosures for pde surrogates, 2025. URL https://arxiv.org/abs/ 2501.18426. Gruber, R. and Rappaz, J. The Ideal MHD Model, pp. 34 41. Springer Berlin Heidelberg, Berlin, Heidelberg, 1985. ISBN 978-3-642-86708-8. doi: 10.1007/ 978-3-642-86708-8 3. URL https://doi.org/10. 1007/978-3-642-86708-8_3. Hoelzl, M., Huijsmans, G. T. A., Pamela, S. J. P., B ecoulet, M., Nardon, E., Artola, F., Nkonga, B., Atanasiu, C. V., Bandaru, V., Bhole, A., Bonfiglio, D., Cathey, A., Czarny, O., Dvornova, A., Feh er, T., Fil, A., Franck, E., Futatani, S., Gruca, M., Guillard, H., Haverkort, J. W., Holod, I., Hu, D., Kim, S. K., Korving, S. Q., Kos, L., Krebs, I., Kripner, L., Latu, G., Liu, F., Merkel, P., Meshcheriakov, D., Mitterauer, V., Mochalskyy, S., Morales, J. A., Nies, R., Nikulsin, N., Orain, F., Pratt, J., Ramasamy, R., Ramet, P., Reux, C., S arkim aki, K., Schwarz, N., Verma, P. S., Smith, S. F., Sommariva, C., Strumberger, E., van Vugt, D. C., Verbeek, M., Westerhof, E., Wieschollek, F., and Zielinski, J. The jorek non-linear extended mhd code and applications to largescale instabilities and their control in magnetically confined fusion plasmas. Nuclear Fusion, 61(6):065001, 2021. doi: 10.1088/1741-4326/abf99f. URL https: //dx.doi.org/10.1088/1741-4326/abf99f. Horwitz, J. A. K. Estimating the carbon footprint of computational fluid dynamics. Physics of Fluids, 36(4):045109, 04 2024. ISSN 1070-6631. doi: 10.1063/5.0199350. URL https://doi.org/10.1063/5.0199350. Iserles, A. A first course in the numerical analysis of differential equations. Cambridge university press, 2009. Jang, B., Kaptanoglu, A. A., Gaur, R., Pan, S., Landreman, M., and Dorland, W. Grad shafranov equilibria via data-free physics informed neural networks. Physics of Plasmas, 31(3):032510, Mar 2024. ISSN 1070-664X. doi: 10.1063/5.0188634. URL https: //doi.org/10.1063/5.0188634. Jiang, C. M., Esmaeilzadeh, S., Azizzadenesheli, K., Kashinath, K., Mustafa, M., Tchelepi, H. A., Marcus, P., Prabhat, and Anandkumar, A. Meshfreeflownet: A physics-constrained deep continuous space-time superresolution framework. ar Xiv:2005.01463, 2020. Jiang, Y. and Jiang, Z. P. Robust adaptive dynamic programming and feedback stabilization of nonlinear systems. IEEE Transactions on Neural Networks and Learning Systems, 25(5):882 893, 2014. doi: 10.1109/TNNLS. 2013.2294968. Jord a, M., Valero-Lara, P., and Pe na, A. J. Performance evaluation of cudnn convolution algorithms on nvidia volta gpus. IEEE Access, 7:70461 70473, 2019. doi: 10.1109/ACCESS.2019.2918851. Joung, S., Ghim, Y.-C., Kim, J., Kwak, S., Kwon, D., Sung, C., Kim, D., Kim, H.-S., Bak, J. G., and Yoon, S. W. Gs-deepnet: mastering tokamak plasma equilibria with deep neural networks and the grad shafranov equation. Scientific Reports, 13(1):15799, Sep 2023. ISSN 20452322. doi: 10.1038/s41598-023-42991-5. URL https: //doi.org/10.1038/s41598-023-42991-5. Karniadakis, G. E., Kevrekidis, I. G., Lu, L., Perdikaris, P., Wang, S., and Yang, L. Physics-informed machine learning. Nature Reviews Physics, 3(6):422 440, 2021. ISSN 2522-5820. doi: 10.1038/ Calibrated Physics-Informed UQ s42254-021-00314-5. URL https://doi.org/10. 1038/s42254-021-00314-5. Kato, Y., Tax, D. M., and Loog, M. A review of nonconformity measures for conformal prediction in regression. In Papadopoulos, H., Nguyen, K. A., Bostr om, H., and Carlsson, L. (eds.), Proceedings of the Twelfth Symposium on Conformal and Probabilistic Prediction with Applications, volume 204 of Proceedings of Machine Learning Research, pp. 369 383. PMLR, 13 15 Sep 2023. URL https://proceedings.mlr.press/ v204/kato23a.html. Kingma, D. P. and Ba, J. Adam: A method for stochastic optimization. In Bengio, Y. and Le Cun, Y. (eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. URL http: //arxiv.org/abs/1412.6980. Kochkov, D., Yuval, J., Langmore, I., Norgaard, P., Smith, J., Mooers, G., Kl ower, M., Lottes, J., Rasp, S., D uben, P., Hatfield, S., Battaglia, P., Sanchez-Gonzalez, A., Willson, M., Brenner, M. P., and Hoyer, S. Neural general circulation models for weather and climate. Nature, 632(8027):1060 1066, Aug 2024. ISSN 14764687. doi: 10.1038/s41586-024-07744-y. URL https: //doi.org/10.1038/s41586-024-07744-y. Lakshminarayanan, B., Pritzel, A., and Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles, 2017. Lao, L., St. John, H., Stambaugh, R., Kellman, A., and Pfeiffer, W. Reconstruction of current profile parameters and plasma shapes in tokamaks. Nuclear Fusion, 25(11):1611, nov 1985. doi: 10.1088/0029-5515/25/ 11/007. URL https://dx.doi.org/10.1088/ 0029-5515/25/11/007. Li, S., Jiang, H., Ren, Z., and Xu, C. Optimal tracking for a divergent-type parabolic pde system in current profile control. Abstract and Applied Analysis, 2014:1 8, 06 2014. doi: 10.1155/2014/940965. Li, Z., Kovachki, N. B., Azizzadenesheli, K., liu, B., Bhattacharya, K., Stuart, A., and Anandkumar, A. Fourier neural operator for parametric partial differential equations. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum? id=c8P9NQVtmn O. Li, Z., Zheng, H., Kovachki, N., Jin, D., Chen, H., Liu, B., Azizzadenesheli, K., and Anandkumar, A. Physicsinformed neural operator for learning partial differential equations. ACM / IMS J. Data Sci., 1(3), may 2024. doi: 10.1145/3648506. URL https://doi.org/10. 1145/3648506. Liu, S., Oosterlee, C. W., and Bohte, S. M. Pricing options and computing implied volatilities using neural networks. Risks, 7(1), 2019. ISSN 2227-9091. doi: 10.3390/risks7010016. URL https://www.mdpi. com/2227-9091/7/1/16. Ma, Z., Azizzadenesheli, K., and Anandkumar, A. Calibrated uncertainty quantification for operator learning via conformal prediction. ar Xiv preprint ar Xiv:2402.01960, 2024. Mac Kay, D. J. C. A practical bayesian framework for backpropagation networks. Neural Computation, 4(3):448 472, 05 1992. ISSN 0899-7667. doi: 10.1162/neco. 1992.4.3.448. URL https://doi.org/10.1162/ neco.1992.4.3.448. Mac Kinnon, R. J. and Johnson, R. W. Differentialequation-based representation of truncation errors for accurate numerical simulation. International Journal for Numerical Methods in Fluids, 13(6):739 757, 1991. doi: https://doi.org/10.1002/fld.1650130606. URL https://onlinelibrary.wiley.com/doi/ abs/10.1002/fld.1650130606. Maddox, W. J., Izmailov, P., Garipov, T., Vetrov, D. P., and Wilson, A. G. A simple baseline for bayesian uncertainty in deep learning. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alch e-Buc, F., Fox, E., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips. cc/paper_files/paper/2019/file/ 118921efba23fc329e6560b27861f0c2-Paper. pdf. Mc Cabe, M., Harrington, P., Subramanian, S., and Brown, J. Towards stability of autoregressive neural operators. Transactions on Machine Learning Research, 2023. ISSN 2835-8856. URL https://openreview. net/forum?id=RFf UUt KYOG. Meyer, D., Hogan, R. J., Dueben, P. D., and Mason, S. L. Machine learning emulation of 3d cloud radiative effects. Journal of Advances in Modeling Earth Systems, 14(3):e2021MS002550, 2022. doi: https://doi.org/10.1029/2021MS002550. URL https://agupubs.onlinelibrary.wiley. com/doi/abs/10.1029/2021MS002550. e2021MS002550 2021MS002550. Meyer, H. and Team, S. Plasma burn mind the gap. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 382 (2280):20230406, 2024. doi: 10.1098/rsta.2023.0406. URL https://royalsocietypublishing. org/doi/abs/10.1098/rsta.2023.0406. Calibrated Physics-Informed UQ Mocz, P., Vogelsberger, M., and Hernquist, L. A constrained transport scheme for MHD on unstructured static and moving meshes. Monthly Notices of the Royal Astronomical Society, 442(1):43 55, 06 2014. ISSN 0035-8711. doi: 10.1093/mnras/stu865. URL https: //doi.org/10.1093/mnras/stu865. Musekamp, D., Kalimuthu, M., Holzm uller, D., Takamoto, M., and Niepert, M. Active learning for neural PDE solvers. In The Thirteenth International Conference on Learning Representations, 2025. URL https:// openreview.net/forum?id=x4Zm Qaum Rg. Orszag, S. A. and Tang, C.-M. Small-scale structure of twodimensional magnetohydrodynamic turbulence. Journal of Fluid Mechanics, 90(1):129 143, 1979. doi: 10.1017/ S002211207900210X. Papadopoulos, H. Inductive conformal prediction: Theory and application to neural networks. In Fritzsche, P. (ed.), Tools in Artificial Intelligence, chapter 18. Intech Open, Rijeka, 2008. doi: 10.5772/6078. Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., K opf, A., Yang, E., De Vito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., and Chintala, S. Pytorch: An imperative style, high-performance deep learning library, 2019. URL https://arxiv.org/abs/1912.01703. Pfaff, T., Fortunato, M., Sanchez-Gonzalez, A., and Battaglia, P. Learning mesh-based simulation with graph networks. In International Conference on Learning Representations, 2021. URL https://openreview. net/forum?id=ro Nq YL0_XP. Pinder, G. F. Numerical methods for solving partial differential equations : a comprehensive introduction for scientists and engineers. John Wiley and Sons, Inc. : Wiley, Hoboken, NJ, 2018. Poels, Y., Derks, G., Westerhof, E., Minartz, K., Wiesen, S., and Menkovski, V. Fast dynamic 1d simulation of divertor plasmas with neural pde surrogates. Nuclear Fusion, 63(12):126012, sep 2023. doi: 10.1088/1741-4326/ acf70d. URL https://dx.doi.org/10.1088/ 1741-4326/acf70d. Psaros, A. F., Meng, X., Zou, Z., Guo, L., and Karniadakis, G. E. Uncertainty quantification in scientific machine learning: Methods, metrics, and comparisons. Journal of Computational Physics, 477:111902, 2023. ISSN 00219991. doi: https://doi.org/10.1016/j.jcp.2022.111902. URL https://www.sciencedirect.com/ science/article/pii/S0021999122009652. Raissi, M., Perdikaris, P., and Karniadakis, G. Physicsinformed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational Physics, 378:686 707, 2019. ISSN 0021-9991. doi: https://doi.org/10.1016/j.jcp.2018.10. 045. URL https://www.sciencedirect.com/ science/article/pii/S0021999118307125. Saad, Y. and Schultz, M. H. Gmres: A generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM Journal on Scientific and Statistical Computing, 7(3):856 869, 1986. doi: 10.1137/0907058. URL https://doi.org/10.1137/0907058. Shafer, G. and Vovk, V. A tutorial on conformal prediction. Journal of Machine Learning Research, 9(3), 2008. Somov, B. V. Plasma Equilibrium in Magnetic Field, pp. 403 427. Springer New York, New York, NY, 2012. ISBN 978-1-4614-4283-7. doi: 10.1007/ 978-1-4614-4283-7 19. URL https://doi.org/ 10.1007/978-1-4614-4283-7_19. Stankeviciute, K., M. Alaa, A., and van der Schaar, M. Conformal time-series forecasting. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 6216 6228. Curran Associates, Inc., 2021. URL https://proceedings.neurips. cc/paper_files/paper/2021/file/ 312f1ba2a72318edaaa995a67835fad5-Paper. pdf. Strang, G. A proposal for toeplitz matrix calculations. Studies in Applied Mathematics, 74(2):171 176, 1986. doi: https://doi.org/10.1002/sapm1986742171. URL https://onlinelibrary.wiley.com/doi/ abs/10.1002/sapm1986742171. Sun, S. Conformal methods for quantifying uncertainty in spatiotemporal data: A survey, 2022. Teng, J., Wen, C., Zhang, D., Bengio, Y., Gao, Y., and Yuan, Y. Predictive inference with feature conformal prediction. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview. net/forum?id=0u Rm1Ym FTu. Th oni, A. C. M., Robinson, W. E., Bachrach, Y., Huck, W. T. S., and Kachman, T. Modelling chemical reaction networks using neural ordinary differential equations, 2025. URL https://arxiv.org/abs/2502.19397. Tolsma, J. E. and Barton, P. I. On computational differentiation. Computers and Chemical Engineering, 22(4):475 490, 1998. ISSN 0098-1354. Calibrated Physics-Informed UQ doi: https://doi.org/10.1016/S0098-1354(97)00264-0. URL https://www.sciencedirect.com/ science/article/pii/S0098135497002640. Vovk, V., Gammerman, A., and Shafer, G. Algorithmic Learning in a Random World. Springer, 2005. Wang, S., Fan, K., Luo, N., et al. Massive computational acceleration by using neural networks to emulate mechanism-based biological models. Nature Communications, 10:4354, 2019. doi: 10.1038/ s41467-019-12342-y. URL https://doi.org/10. 1038/s41467-019-12342-y. Xu, C. and Xie, Y. Conformal prediction interval for dynamic time-series. In Meila, M. and Zhang, T. (eds.), Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pp. 11559 11569. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr.press/ v139/xu21h.html. Xu, C., Xie, Y., Vazquez, D. A. Z., Yao, R., and Qiu, F. Spatio-temporal wildfire prediction using multi-modal data. IEEE Journal on Selected Areas in Information Theory, 4:302 313, 2023. doi: 10.1109/JSAIT.2023. 3276054. Zhu, Y., Zabaras, N., Koutsourelakis, P.-S., and Perdikaris, P. Physics-constrained deep learning for high-dimensional surrogate modeling and uncertainty quantification without labeled data. Journal of Computational Physics, 394:56 81, 2019. ISSN 0021-9991. doi: https://doi.org/10.1016/j.jcp.2019.05. 024. URL https://www.sciencedirect.com/ science/article/pii/S0021999119303559. Zou, Z., Meng, X., Psaros, A. F., and Karniadakis, G. E. Neuraluq: A comprehensive library for uncertainty quantification in neural differential equations and operators. SIAM Review, 66(1):161 190, 2024. doi: 10.1137/ 22M1518189. URL https://doi.org/10.1137/ 22M1518189. Calibrated Physics-Informed UQ A. Theorem: Data-Free CP Preliminaries: Let D : Rm Rm be a physics residual operator mapping a function to its PDE residual value, where: {Xi}n i=1 is the calibration set, ˆf is the model, ˆqα is estimated as the (n + 1)(1 α) /n -quantile of {|D( ˆf(Xi))|}n i=1 Theorem A.1. If the residuals {D( ˆf(Xi))}n+1 i=1 are exchangeable random variables, then for any significance level α (0, 1) and any new input Xn+1 we have the following coverage guarantee: P(|D( ˆf(Xn+1))| Cα) 1 α ; Cα = [ ˆqα, ˆqα] Proof. Let Ri = |D( ˆf(Xi))| for i = 1, . . . , n + 1. We have, by assumption, (R1, . . . , Rn, Rn+1) is an exchangeable sequence. Define the rank π of Rn+1 w.r.t. all other residuals: π(Rn+1) = |{i = 1, . . . , n + 1 : Ri Rn+1}| By exchangeability, the rank π(Rn+1) is uniformly distributed over {1, . . . , n + 1}. Therefore, P(π(Rn+1) (n+1)(1 α) ) = (n + 1)(1 α) By construction of ˆqα we have that, {π(Rn+1) (n + 1)(1 α) } {Rn+1 ˆqα}. Putting this together, P(|D( ˆf(Xn+1))| ˆqα) = P(Rn+1 ˆqα) 1 α, which completes the proof. B. PRE: Score Function and Prediction Sets For a general nonconformity score S, the prediction set for a new input Xn+1 is typically defined as: Cα(Xn+1) = {y : S(Xn+1, y) ˆqα}, where ˆqα is the (1 α)-quantile of the nonconformity scores on the calibration set. For AER and STD, the nonconformity scores depend on both the input X and the output (target) Y : SAER(X, Y ) = | ˆf(X) Y |, (21) SST D(X, Y ) = | ˆfµ(X) Y | ˆfσ(X) . (22) The resulting prediction sets are: Cα AER(Xn+1) = [ ˆf(Xn+1) ˆqα, ˆf(Xn+1) + ˆqα], (23) Cα ST D(Xn+1) = [ ˆfµ(Xn+1) ˆqα ˆfσ(Xn+1), ˆfµ(Xn+1) + ˆqα ˆfσ(Xn+1)]. (24) These prediction sets clearly depend on the input Xn+1. For PRE, the nonconformity score depends only on the model output and not on the target: SP RE( ˆf(X)) = |D( ˆf(X)) 0|, where D is the PDE residual operator. The key difference is that the true output Y for PRE, irrespective of the PDE is always 0 and does not depend on the input X. PRE is a measure of how well the model output satisfies the physics rather than how it fits certain data. Hence, we can formulate a nonconformity score that is data-free and eventually leads to input-independent prediction sets as given below. For PRE, we can reframe the prediction set definition: Cα P RE = { ˆf(X) : |D( ˆf(X))| ˆqα}. This set is not defined in terms of the true Y values but in terms of the allowable model outputs ˆf(X) that satisfy the PDE residual constraint. Thus, the prediction set can be expressed as: Cα P RE = [ ˆqα, ˆqα]. This formulation is independent of the input X, as it only depends on the quantile ˆqα derived from the calibration set as given in Equation (6). To validate predictions using PRE: 1. For a new input Xn+1, compute ˆf(Xn+1). 2. Calculate the residual: r = |D( ˆf(Xn+1))|. Calibrated Physics-Informed UQ 3. Check if r [ ˆqα, ˆqα] for a given α. If the condition in step 3 is satisfied, the error bounds dictated by [ ˆqα, ˆqα] is considered valid according to the CP framework, regardless of the specific input Xn+1. C. Algorithmic Procedure 1. Set up the Neural PDE Solver (a) Define the PDE system of interest with its governing equations in a numerical solver (b) Train a neural network (e.g., Fourier Neural Operator) to approximate solutions to the PDE (c) Ensure the model can make predictions on new initial conditions / PDE coefficients 2. Define the Physics Residual Error (PRE) (a) For a given PDE case with operator D, the physics residual is defined as |D( ˆf(X))|, where ˆf is the neural PDE prediction (b) Create a differential operator using finite difference stencils implemented as convolutional kernels. This operator evaluates how well the predicted solution satisfies the underlying physics 3. Generate Calibration Set (a) Sample a set of initial conditions X1, X2, ..., Xn from the domain of interest (b) Run the neural PDE solver on these initial conditions to get predictions ˆf(X1), ˆf(X2), ..., ˆf(Xn) (c) Compute the PRE nonconformity scores |D( ˆf(X1))|, |D( ˆf(X2))|, ..., |D( ˆf(Xn))| for each prediction 4. Calibration Process (a) For a desired confidence level 1-α (e.g., 90% for α = 0.1), compute the empirical quantile: (b) ˆqα = (n + 1)(1 α) /n quantile of |D( ˆf(X1))|, |D( ˆf(X2))|, ...., |D( ˆf(Xn))| (c) This quantile represents the threshold for the physics residual that will provide the desired coverage 5. Apply to New Predictions (a) For a new initial condition Xn+1, obtain the neural PDE prediction ˆf(Xn+1) (b) Compute the physics residual |D( ˆf(Xn+1))| (c) The prediction set is defined as Cα = [ ˆqα, ˆqα] (d) If |D( ˆf(Xn+1))| falls within Cα, the prediction is considered valid at the 1 α confidence level 6. Interpretation of Results (a) For marginal conformal prediction: Apply steps 1-5 cell-wise across the spatial-temporal domain to get localized error bounds (b) For joint conformal prediction: Calculate the supremum of the normalized residuals across the entire domain to obtain global error bounds (c) The width of the error bounds indicates the model s physical consistency - tighter bounds suggest better alignment with the physics 7. Decision Framework (Optional) (a) Accept predictions where the physics residual falls within the calibrated bounds (b) Reject and flag predictions where the residual exceeds the bounds, potentially routing these cases to traditional numerical solvers (c) This approach guarantees that, with probability at least 1 α, the physics residual of new predictions will fall within the computed bounds, ensuring that the model maintains physical consistency while providing faster solutions than traditional numerical methods. Calibrated Physics-Informed UQ D. Conv Operator: Convolutional Kernels for Gradient Estimation Within the code base for this paper, we release a utility function that constructs convolutional layers for gradient estimation based on your choice of order of differentiation and Taylor approximation. This allows for the PRE score function to be easily expressed in a single line of code 2 This section provides an overview of the code implementation and algorithm for estimating the PRE using Convolution operations. We ll use an arbitrary PDE example with a temporal gradient u t and a Laplacian 2 x2 + 2 y2 to illustrate the process. + βu = 0, (25) where u is the field variable, t is time, x and y are spatial coordinates, and α and β are constants. To estimate the PDE residual given by Equation (25), we need to estimate the associated spatio-temporal gradients. First, we use the Conv Operator class from Utils/Conv Ops 2d.py to set up the convolutional layer with kernels 3 taken from the appropriate finite difference stencils: from Conv Ops_2d import Conv Operator # Define each operator within the PDE D_t = Conv Operator(domain= t , order=1) #time-derivative D_xx_yy = Conv Operator(domain=( x , y ) , order=2) #Laplacian D_identity = Conv Operator() #Identity Operator The Conv Operator class is used to set up a gradient operation. It takes in variable(s) of differentiation and order of differentiation as arguments to design the appropriate forward difference stencil and then sets up a convolutional layer with the stencil as the kernel. Under the hood, the class will take care of devising a 3D convolutional layer, and setup the kernel so that it acts on a spatio-temporal tensor of dimensionality: [BS, Nt, Nx, Ny] which expands to batch size, temporal discretisation and the spatial discretisation in x and y. 2The code and associated utility functions can be found in: https://github.com/gitvicky/CP-PRE 3Convolution functions can be set as cross-correlations as it is default in the Py Torch framework, or they could be set up as convolutions by obtaining the complex conjugate in the frequency space.(Jord a et al., 2019) alpha, beta = 1.0, 0.5 # Example coefficients D = Conv Operator() #Additive Kernels D.kernel = D_t.kernel - alpha * D_xx_yy .kernel - beta * D_identity.kernel The convolutional kernels are additive i.e. in order to estimate the residual in one convolutional operation, they could be added together to form a composite kernel that characterises the entire PDE residual. Once having set up the kernels, PRE estimation is as simple as passing the composite class instance D the predictions from the neural PDE surroga te (ensuring that the output is in the same order as the kernel outlined above). y_pred = model(X) PRE = D(y_pred) Only operating on the outputs, this method of PRE estimation is memory efficient, computationally cheap and with the Conv Operator evaluating the PDE residual can be done in a single line of code. Calibrated Physics-Informed UQ D.1. Impact of Discretisation As demonstrated in (Bartolucci et al., 2023), the discretisation of the inputs and hence model outputs plays an important role in the accuracy of the neural-PDE solvers. Though the neural operators are constructed for discretisationinvariant behaviour due to the band-limited nature of the functions, they often exhibit discretisation-convergent behaviour as opposed to discretisation-invariance. This is of particular importance in the temporal dimensions as these neural-PDE models utilise a discrete, autoregressive based time-stepping, baked into the model within its training regime (Mc Cabe et al., 2023). The lack of control in teh temporal discretisation (dt), leads to higher numerical errors within the the PRE estimates. In fig. 7, we visualise the evaluation of finite difference in 2D+time as a 3D convolution. The finite difference stencil i.e. the convolutional kernel has a unit discretisation of dx, dy and dt associated with the problem and is applied over the signal i.e. the output from the neural-PDE u spanning the domain x, y, t, where x [0, X], y [0, Y ], t [0, T]. Though the discretisation deployed within CP-PRE stems from the neural PDE solver, it consistently delivers guaranteed coverage regardless of resolution. Even with coarser discretisation, CP-PRE remains valuable as it allows for: statistical identification of poorer fit regions (marginal formulation), highlights physically inconsistent predictions (joint formulation) and enables relative assessment of physical inconsistencies across predictions. While residuals may be inflated with coarser discretisation, the corresponding bounds reflect this inflation, preserving the relative information about physical inconsistency across a series of predictions. Figure 7. PRE estimation using the 3D convolutions with finite difference stencils as convolutional kernels being applied over the neural-PDE predictions as the signals. 0.0 0.5 1.0 1.5 2.0 x Coarse (dt=0.01) PRE Lower Bound Upper Bound (a) CP-PRE over coarser discretisation 0.0 0.5 1.0 1.5 2.0 x Fine (dt=0.005) PRE Lower Bound Upper Bound (b) CP-PRE over finer discretisation 0.2 0.4 0.6 0.8 1- Empirical Coverage Ideal Coarse Fine (c) Guaranteed coverage irrespective of discretisation error. Figure 8. CP-PRE provides guaranteed coverage irrespective of the discretisation associated with the model outputs., however, the width of the obtained coverage bounds indicates the discretisation error associated with the gradient estimation. Coverage taken for α = 0.1 90% coverage. Calibrated Physics-Informed UQ E. Initial and Boundary Conditions As mentioned in Section 4.3, the focus of our experiments has been in quantifying the misalignment of the model with the PDE in the domain of the problem. A well-defined PDE is characterised by the PDE on the domain, the initial condition across the domain at t = 0 and the boundary conditions, reflecting the physics at the boundary. Within a neural-PDE setting, the initial condition does not need to be enforced or measured for as the neural-PDE is set up as an initial-value problem, taking in the initial state to autoregressively evolve the later timesteps and hence does not come under the purview of the neural-PDE s outputs. The boundary conditions, whether Dirichlet, Neumann or periodic, follows a residual structure as outlined in Equation (2), allowing us to use it as a PRE-like nonconformity score for performing conformal prediction. In all the problems we have under consideration, the PDEs are modelled under periodic boundary conditions: u X = 0; X Ω (26) By deploying the eqn 26 as the PRE across the boundary, we can obtain error bars over the boundary conditions as well. Within fig. 9, we demonstrate the error bars obtained by using the boundary conditions as the PRE nonconformity scores for the Navier-Stokes equations. Figure 9. Error bars obtained over the boundary conditions over the right wall of domain of the Navier-Stokes Equation using Marginal and Joint CP. The empirical coverage obtained using the boundary condition as the PRE nonconformity score is also given. Calibrated Physics-Informed UQ F. Toy Problems: 1D cases F.1. Advection Equation Consider the one-dimensional advection equation x = 0. (27) The state variable of interest u is bounded within the domain x [0, 2], t [0, 0.5] and moves within the domain at a constant velocity v. Each trajectory is generated by solving Equation (27) using a Crank-Nicolson method (Crank & Nicolson, 1947). Data is sampled using a parameterised initial condition that characterises the amplitude and position of the Gaussian field. Generated data is used to train a 1D FNO that takes in the initial condition and autoregressively with a step size of 1, learns to map the next 10 time frames as outlined in Equation (28). A reproducible script is attached to the supplementary material. un+1 = f(un), (28) f represents the FNO and un and un+1 represents the current (input) and the predicted future state of the system (output). Figure 10. Advection Equation: (Left) Comparing the neural PDE (FNO) performance with that of the physics-based numerical solver at the last time instance. (Middle) Upper and lower bounds for 90% coverage obtained by performing marginal-CP. (Right) Upper and lower bounds for 90% coverage obtained by performing joint-CP. Marginal-CP provides tighter bounds for a prediction as opposed to joint-CP, whereas joint-CP provides a method of employing a relative sense of reliability of a prediction within a domain. Figure 10 demonstrates the guaranteed bounds obtained over the residual space of Equation (27) with both the marginal and joint-CP formulations. Being cell-wise, marginal-CP guarantees coverage for each discretised point within the spatio-temporal domain. This allows for tighter bounds and error quantification of interested subdomains within regions but does not provide any guarantee across the entire prediction domain. Joint-CP acts across the entire domain and provides a guarantee as to whether a prediction (instead of a single cell) will fall within the domain or not. Larger bounds are observed as they extend over the multivariate nature of the output. Though this comes with bigger bounds, it provides us with a mechanism to perform rejection criteria for predictions. Within joint-CP, bounds dictating 1 α coverage suggest that approximately, α 100% predictions from the same domain will fall outside the bounds and can be rejected. Further details about the physics, parameterisation of the initial conditions, model and its training can be found in Appendix G. F.2. Burgers Equation Consider the 1D Burgers Equation The state variable of interest u is bounded within the domain x [0, 2], t [0, 1.25]. The field is prescribed by a kinematic viscosity ν = 0.002. Data is generated by solving Equation (29) using a spectral method (Canuto et al., 2007). Data sampled using a parameterised initial condition is used to train a 1D FNO that takes in the initial distribution of the state and learns to autoregressively predict the PDE evolution for the next 30 time frames. Figure 11. Burgers Equation: (Left) Comparing the neural PDE (FNO) performance with that of the physics-based numerical solver at the last time instance. (Middle) Upper and lower bounds for 90% coverage obtained by performing marginal-CP. (Right) Upper and lower bounds for 90% coverage guaranteed by joint-CP over the residual space. Figure 11 illustrates the guaranteed bounds over the residual space of Equation (29) using marginal and joint-CP formulations for 90% coverage. Marginal-CP provides cell-wise coverage, yielding tighter bounds for specific subdomains. Joint-CP provides bounds 50 times larger than that of the marginal-CP as it covers the entire prediction domain. Despite the large bounds, approximately α 100% predictions fall outside it as given in Figure 3. For details on physics, initial condition parameterisation, model, and training, see Appendix H. Calibrated Physics-Informed UQ G. 1D Advection Equation G.1. Physics Consider the one-dimensional advection equation, parameterised by the initial condition: x, x [0, 2], t [0, 0.5], u(x, t = 0) = Ae(x X)2. (30) Here u defines the density of the fluid, x the spatial coordinate, t the temporal coordinate and v the advection speed. initial condition is parameterised by A and X, representing the amplitude and position of a Gaussian distribution. A no-flux boundary condition bounds the system. The numerical solution for the above equation is built using a finite difference solver with a crank-nicolson method implemented in Python. We construct a dataset by performing a Latin hypercube sampling across parameters A, X. Each parameter is sampled from within the domain given in Table 6 to generate 100 simulation points, each with its own initial condition. Each simulation is run for 50-time iterations with a t = 0.01 across a spatial domain spanning [0,2], uniformly discretised into 200 spatial units in the x-axis. Table 6. Domain range of initial condition parameters for the 1D advection equation. Parameter Domain Type Amplitude (A) [50, 200] Continuous Position (X) [0.5, 1.0] Continuous G.2. Model and Training We use a one-dimensional FNO to model the evolution of the convection-diffusion equation. The FNO learns to perform the mapping from the initial condition to the next time instance, having a step size of 1. The model autoregressively learns the evolution of the field up until the 10th time instance. Each Fourier layer has 8 modes and a width of 16. The FNO architecture can be found in Table 7. Considering the field values governing the evolution of the advection equation are relatively small, we avoid normalisations. The model is trained for up to 100 epochs using the Adam optimiser (Kingma & Ba, 2015) with a step-decaying learning rate. The learning rate is initially set to 0.005 and scheduled to decrease by half after every 100 epochs. The model was trained using an LP-loss (Gopakumar et al., 2024b). Table 7. Architecture of the 1D FNO deployed for modelling 1D Advection Equation Part Layer Output Shape Input - (50, 1, 200, 1) Lifting Linear (50, 1, 200, 16) Fourier 1 Fourier1d/Conv1d/Add/GELU (50, 1, 16, 200) Fourier 2 Fourier1d/Conv1d/Add/GELU (50, 1, 16, 200) Fourier 3 Fourier1d/Conv1d/Add/GELU (50, 1, 16, 200) Fourier 4 Fourier1d/Conv1d/Add/GELU (50, 1, 16, 200) Fourier 5 Fourier1d/Conv1d/Add/GELU (50, 1, 16, 200) Fourier 6 Fourier1d/Conv1d/Add/GELU (50, 1, 16, 200) Projection 1 Linear (50, 1, 200, 256) Projection 2 Linear (50, 1, 200, 1) 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper Marginal CP (50% coverage) Figure 12. Advection Equation: Marginal-CP with α = 0.5 G.3. Calibration and Validation To perform the calibration as outlined in Section 5, model predictions are obtained using initial conditions sampled from the domain given in Table 6. The same bounded domain for the initial condition parameters is used for calibration and validation. 100 initial conditions are sampled and fed to the model to obtain and prediction for both the calibration and the validation. 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper Joint CP (50% coverage) Figure 13. Advection Equation: joint-CP with α = 0.5 Calibrated Physics-Informed UQ H. 1D Burgers Equation H.1. Physics Consider the one-dimensional Burgers equation: u(x, t = 0) = sin(απx) + cos( βπx) + 1 cosh(γπx), where u defines the field variable, ν the kinematic viscosity, x the spatial coordinate, t the temporal coordinates. α, β and γ are variables that parameterise the initial condition of the PDE setup. The system is bounded periodically within the mentioned domain. The solution for the Burgers equation is obtained by deploying a spectral solver (Canuto et al., 2007). The dataset is built by performing a Latin hypercube scan across the defined domain for the parameters α, β, γ, sampled for each simulation. We generate 1000 simulation points, each one with its initial condition and use it for training. The physics of the equation, given by the various coefficients is held constant across the dataset generation throughout as given in Equation (31). Each data point, as in each simulation is generated with a different initial condition as described above. The parameters of the initial conditions are sampled from within the domain as given in Table 8. Each simulation is run for 500-time iterations with a t = 0.0025 across a spatial domain spanning [0, 2], uniformly discretised into 1000 spatial units in the x and y axes. The temporal domain is subsampled to factor in every 10th time instance, while the spatial domain is downsampled to every 5th instance. Table 8. Domain range of initial condition parameters for the 1D Burgers equation. Parameter Domain Type α [ 3, 3] Continuous β [ 3, 3] Continuous γ [ 3, 3] Continuous H.2. Model and Training We train a 1D FNO to map the spatio-temporal evolution of the field variables. We deploy an auto-regressive structure that performs time rollouts allowing us to map the initial distribution recursively up until the 30th time instance with a step size of 1. Each Fourier layer has 8 modes and a width of 32. The FNO architecture can be found in Table 9. We employ a linear range normalisation scheme, placing the field values between -1 and 1. Each model is trained for up to 500 epochs using the Adam optimiser (Kingma & Ba, 2015) with a step-decaying learning rate. The learning rate is initially set to 0.005 and scheduled to decrease by half after every 100 epochs. The model was trained using an LP-loss (Gopakumar et al., 2024b). Table 9. Architecture of the 1D FNO deployed for modelling 1D Burgers equation Part Layer Output Shape Input - (50, 1, 200, 1) Lifting Linear (50, 1, 200, 32) Fourier 1 Fourier2d/Conv1d/Add/GELU (50, 1, 32, 200) Fourier 2 Fourier2d/Conv1d/Add/GELU (50, 1, 32, 200) Fourier 3 Fourier2d/Conv1d/Add/GELU (50, 1, 32, 200) Fourier 4 Fourier2d/Conv1d/Add/GELU (50, 1, 32, 200) Fourier 5 Fourier2d/Conv1d/Add/GELU (50, 1, 32, 200) Fourier 6 Fourier2d/Conv1d/Add/GELU (50, 1, 32, 200) Projection 1 Linear (50, 1, 200, 256) Projection 2 Linear (50, 1, 200 1) H.3. Calibration and Validation To perform the calibration as outlined in Section 5, model predictions are obtained using initial conditions sampled from the domain given in Table 8. The same bounded domain for the initial condition parameters is used for calibration and validation. 1000 initial conditions are sampled and fed to the model to perform the calibration and 100 samples are gathered for performing the validation. Calibrated Physics-Informed UQ 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper Marginal CP (75% coverage) Figure 14. Burgers Equation: Marginal-CP with α = 0.75 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 x Residual Lower Upper Joint CP (75% coverage) Figure 15. Burgers Equation: joint-CP with α = 0.75 Calibrated Physics-Informed UQ I. Utilising CP-PRE as a measure of model quality While evaluating the performance of a neural-PDE, it is important to their fit not just to the data but to the underlying physics. CP-PRE will provide guaranteed coverage irregardless of the quality of the model. It will have considerably wider error bounds when the neural-PDE (whether PINN or a Neural Operator) fails to comply with the physics. However, we believe that this is an advantage of our method. In CP-PRE formulation, the bounds are estimated across the PDE residual, where the ground truth for a well-fit solution should always be near zero. If we get wide error bars further away from the 0 for potentially high coverage estimates, it is a strong indication that statistically the model violates the physics of interest. Consider the example with the Advection equation. We have two models, a well-fit (good model) and a poorly fit one (bad model). As shown in fig. 16, though we obtain guaranteed coverage in the case of both the bad and good models, the width of the error bars indicates the quality of the model. Taken for 90 % coverage, the width of the coverage bounds obtained over the bad model is substantially larger than that obtained by the good model. There still could be a concern as to what width can be considered to be within a good range within the residual space. This could be estimated by running the PRE convolution operator(s) across a single numerical simulation of the interested physics, thereby estimating the impact of the operator in estimating the residual. The PRE over the simulation data will allow us to judge what ranges for the coverage width differentiate between a bad and a good model. 0.0 0.5 1.0 1.5 2.0 x Model Performance Ground Truth Good Model Bad Model (a) Marginal advection performance 0.0 0.5 1.0 1.5 2.0 x PRE-CP over Good Model PRE Lower Bound Upper Bound (b) Good marginal advection 0.0 0.5 1.0 1.5 2.0 x PRE-CP over Bad Model PRE Lower Bound Upper Bound (c) Bad marginal advection 0.2 0.4 0.6 0.8 1- Empirical Coverage Ideal Bad Model Good Model (d) Coverage analysis Figure 16. CP-PRE provides guaranteed coverage irrespective of the model performance, however, the width of the obtained coverage bounds indicates the accuracy of the model in obeying the underlying physics. Coverage taken for α = 0.1 90% coverage. Calibrated Physics-Informed UQ J. 2D Wave Equation J.1. Physics Consider the two-dimensional wave equation: = 0, x, y [ 1, 1], t [0, 1], u(x, y, t = 0) = e A (x X)2+(y Y )2 , (32) u(x, y, t = 0) t = 0, u(x, y, t) = 0, x, y Ω, t [0, 1], where u defines the field variable, c the wave velocity, x and y the spatial coordinates, t the temporal coordinates. A, X and Y are variables that parameterise the initial condition of the PDE setup. There exists an additional constraint to the PDE setup that initialises the velocity of the wave to 0. The system is bounded periodically within the mentioned domain. The solution for the wave equation is obtained by deploying a spectral solver that uses a leapfrog method for time discretisation and a Chebyshev spectral method on tensor product grid for spatial discretisation (Gopakumar et al., 2023a). The dataset is built by performing a Latin hypercube scan across the defined domain for the parameters A, X, Y , which accounts for the amplitude and the location of the 2D Gaussian. We generate 1000 simulation points, each one with its initial condition and use it for training. The physics of the equation, given by the various coefficients, is held constant across the dataset generation throughout, as given in Equation (32). Each data point, as in each simulation is generated with a different initial condition as described above. The parameters of the initial conditions are sampled from within the domain as given in Table 10. Each simulation is run for 150-time iterations with a t = 0.00667 across a spatial domain spanning [ 1, 1]2, uniformly discretised into 64 spatial units in the x and y axes. The temporal domain is subsampled to factor in every 5th time instance only. Table 10. Domain range of initial condition parameters for the 2D wave equation. Parameter Domain Type Amplitude (A) [10, 50] Continuous X Position (X) [0.1, 0.5] Continuous Y Position (X) [0.1, 0.5] Continuous J.2. Model and Training We deploy an auto-regressive FNO that performs time rollouts allowing us to map the initial distribution recursively up until the 20th time instance with a step size of 1. Each Fourier layer has 16 modes and a width of 32. The FNO architecture can be found in Table 10. We employ a linear range normalisation scheme, placing the field values between -1 and 1. Each model is trained for up to 500 epochs using the Adam optimiser (Kingma & Ba, 2015) with a step-decaying learning rate. The learning rate is initially set to 0.005 and scheduled to decrease by half after every 100 epochs. The model was trained using an LP-loss (Gopakumar et al., 2024b). The performance of the trained model can be visualised in Figure 17. Table 11. Architecture of the 2D FNO deployed for modelling the 2D wave equation Part Layer Output Shape Input - (50, 1, 64, 64, 1) Lifting Linear (50, 1, 64, 64, 32) Fourier 1 Fourier2d/Conv2d/Add/GELU (50, 1, 32, 64, 64) Fourier 2 Fourier2d/Conv2d/Add/GELU (50, 1, 32, 64, 64) Fourier 3 Fourier2d/Conv2d/Add/GELU (50, 1, 32, 64, 64) Fourier 4 Fourier2d/Conv2d/Add/GELU (50, 1, 32, 64, 64) Fourier 5 Fourier2d/Conv2d/Add/GELU (50, 1, 32, 64, 64) Fourier 6 Fourier2d/Conv2d/Add/GELU (50, 1, 32, 64, 64) Projection 1 Linear (50, 1, 64, 64, 256) Projection 2 Linear (50, 1, 64, 64 1) Figure 17. Wave Equation: Temporal evolution of field associated with the wave equation modelled using the numerical spectral solver (top of the figure) and that of the FNO (bottom of the figure). The spatial domain is given in Cartesian geometry. J.3. Calibration and Validation To perform the calibration as outlined in Section 5, model predictions are obtained using initial conditions sampled from the domain given in Table 10. The same bounded domain for the initial condition parameters is used for calibration and validation. 1000 initial conditions are sampled and fed to the model to perform the calibration and 100 samples are gathered for performing the validation. Calibrated Physics-Informed UQ (a) Ground Truth PRE: D(uval) (b) PRE over Ground Truth (c) Prediction from neural PDE PRE: D(upred) (d) PRE over Prediction Figure 18. Analysing the PRE over the ground truth and the prediction. Though the neural PDE solver is capable of learning seemingly indistinguishable emulation of the physics while exploring the PRE over each tells a different story. As opposed to the smooth Laplacian of the PRE over the ground truth, PRE over the prediction indicates a noisy solution, potentially arising due to the stochasticity of the optimisation process. Calibrated Physics-Informed UQ K. 2D Navier-Stokes Equations K.1. Physics Consider the two-dimensional Navier-Stokes equations: t + ( v ) v = ν 2 v P, with initial conditions: u(x, y, t = 0) = sin(2παy) y [ 1, 1], (34) v(x, y, t = 0) = sin(4πβx) x [ 1, 1], (35) where u defines the x-component of velocity, v defines the y-component of velocity. The Navier-stokes equations solve the flow of an incompressible fluid with a kinematic viscosity ν. The system is bounded with periodic boundary conditions within the domain. The dataset is built by performing a Latin hypercube scan across the defined domain for the parameters α, β, which parameterises the initial velocity fields for each simulation. We generate 500 simulation points, each one with its initial condition and use it for training. The solver is built using a spectral method outlined in Philip Mocz s code. Each data point, as in each simulation is generated with a different initial condition as described above. The parameters of the initial conditions are sampled from within the domain as given in Table 12. Each simulation is run up until wallclock time reaches 0.5 t = 0.001. The spatial domain is uniformly discretised into 400 spatial units in the x and y axes. The temporal domain is subsampled to factor in every 10th time instance, and the spatial domain is downsampled to factor every 4th time instance leading to a 100 100 grid for the neural PDE. Table 12. Domain range of initial condition parameters for the 2D Navier-Stokes equations Parameter Domain Type Velocity x-axis (u0) [0.5, 1.0] Continuous Velocity y-axis (v0) [0.5, 1.0] Continuous K.2. Model and Training We train a 2D multivariable FNO to map the spatio-temporal evolution of the field variables (Gopakumar et al., 2024b). We deploy an auto-regressive structure that performs time rollouts allowing us to map the initial distribution recursively up until the 20th time instance with a step size of 1. Each Fourier layer has 8 modes and a width of 16. The FNO architecture can be found in Table 13. We employ a linear range normalisation scheme, placing the field values between -1 and 1. Each model is trained for up to 500 epochs using the Adam optimiser (Kingma & Ba, 2015) with a step-decaying learning rate. The learning rate is initially set to 0.005 and scheduled to decrease by half after every 100 epochs. The model was trained using an LP-loss (Gopakumar et al., 2024b). The performance of the trained model can be visualised in Figure 19. Table 13. Architecture of the 2D FNO deployed for modelling 2D Navier-Stokes equations Part Layer Output Shape Input - (50, 1, 100, 100, 1) Lifting Linear (50, 1, 100, 100 16) Fourier 1 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 100, 100) Fourier 2 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 100, 100) Fourier 3 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 100, 100) Fourier 4 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 100, 100) Fourier 5 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 100, 100) Fourier 6 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 100, 100) Projection 1 Linear (50, 1, 100, 100, 256) Projection 2 Linear (50, 1, 100, 100 1) K.3. Calibration and Validation To perform the calibration as outlined in Section 5, model predictions are obtained using initial conditions sampled from the domain given in Table 12. The same bounded domain for the initial condition parameters is used for calibration and validation. 1000 initial conditions are sampled and fed to the model to perform the calibration and 100 samples are gathered for performing the validation. Calibrated Physics-Informed UQ (a) Spatio-temporal evolution of the horizontal component of velocity (u) (b) Sptio-temporal evolution of the vertical component of velocity (v) (c) Spatio-temporal evolution of the pressure field (P) Figure 19. Navier-Stokes Equations: Temporal evolution of velocity and pressure modelled using the numerical spectral solver (top of the figure) and that of the FNO (bottom of the figure) Calibrated Physics-Informed UQ PRE: Dmom(u, v, P) (a) PRE: Momentum Equation Marginal CP (+q) (b) Marginal Bounds Joint CP (+q mod) (c) Joint Bounds Figure 20. Navier-Stokes: CP using the Momentum Equation (13) as the PRE for a neural PDE surrogate model trained to model fluid dynamics. Figure 20(a) depicts the PRE, Figure 20(b) depicts the upper error bar, marginal for each cell, while Figure 20(c) indicates the upper error bar obtained across the entire prediction space. Both are estimated for 90% coverage. PRE: Dcont(u, v) (a) PRE of the Continuity Equation over the FNO prediction Marginal CP (+q) (b) Upper error bar indicating 90% coverage with marginal-CP Joint CP (+q mod) (c) Upper error bar indicating 90% coverage with joint-CP Figure 21. Navier-Stokes: CP using the Continuity Equation (12) as the PRE for a neural PDE surrogate model trained to model fluid dynamics. Figure 20(a) depicts the PRE, Figure 20(b) depicts the upper error bar, marginal for each cell, while Figure 20(c) indicates the upper error bar obtained across the entire prediction space. Both are estimated for 90% coverage. Calibrated Physics-Informed UQ L. 2D Magnetohydrodynamics Consider the Ideal MHD equations in 2D: t + (ρ v) = 0, t + v v = 1 µ0 B ( B) P, t = ( v B), with initial conditions: u = sin(2aπY ), (36) v = sin(2bπX), (37) P = γ 4cπ , (38) where the density (ρ), velocity field ( v = [u, v]) and the pressure of plasma is modelled under a magnetic field ( B = [Bx, By]) across a spatio-temporal domain x, y [0, 1]2, t [0, 5]. µ0 is taken to be the magnetic permeability of free space. The system is bounded with periodic boundary conditions within the domain. The dataset is built by performing a Latin hypercube scan across the defined domain for the parameters a, b, c, which parameterises the initial velocity fields for each simulation. We generate 500 simulation points, each one with its initial condition and use it for training. The solver is built using a finite volume method outlined in Philip Mocz s code. Each data point, as in each simulation is generated with a different initial condition as described above. The parameters of the initial conditions are sampled from within the domain as given in Table 12. Each simulation is run up until wallclock time reaches 0.5 with a varying temporal discretisation. The spatial domain is uniformly discretised into 128 spatial units in the x and y axes. The temporal domain is downsampled to factor in every 25th time instance. Table 14. Domain range of initial condition parameters for the 2D MHD equations Parameter Domain Type Velocity x-axis (a) [0.5, 1.0] Continuous Velocity y-axis (b) [0.5, 1.0] Continuous Pressure (c) [0.5, 1.0] Continuous L.1. Model and Training We train a 2D multi-variable FNO to map the spatiotemporal evolution of the 6 field variables collectively. We deploy an auto-regressive structure that performs time rollouts allowing us to map the initial distribution recursively up until the 20th time instance with a step size of 1. Each Fourier layer has 8 modes and a width of 16. The FNO architecture can be found in Table 15. We employ a linear range normalisation scheme, placing the field values between -1 and 1. Each model is trained for up to 500 epochs using the Adam optimiser (Kingma & Ba, 2015) with a step-decaying learning rate. The learning rate is initially set to 0.005 and scheduled to decrease by half after every 100 epochs. The model was trained using an LP-loss (Gopakumar et al., 2024b). The performance of the trained model Calibrated Physics-Informed UQ can be visualised in Figures 22 and 23. Table 15. Architecture of the 2D FNO deployed for modelling 2D MHD equations Part Layer Output Shape Input - (50, 1, 128, 128, 1) Lifting Linear (50, 1, 128, 128 16) Fourier 1 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 128, 128) Fourier 2 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 128, 128) Fourier 3 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 128, 128) Fourier 4 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 128, 128) Fourier 5 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 128, 128) Fourier 6 Fourier2d/Conv2d/Add/GELU (50, 1, 16, 128, 128) Projection 1 Linear (50, 1, 128, 128, 256) Projection 2 Linear (50, 1, 128, 128 1) L.2. Calibration and Validation To perform the calibration as outlined in Section 5, model predictions are obtained using initial conditions sampled from the domain given in Table 12. The same bounded domain for the initial condition parameters is used for calibration and validation. 100 initial conditions are sampled and fed to the model to perform the calibration and 100 samples are gathered for validation. (a) Spatio-temporal evolution of density (ρ) (b) Spatio-temporal evolution of the horizontal component of velocity (u) (c) Spatio-temporal evolution of the vertical component of velocity (v) Figure 22. MHD Equations: Temporal evolution of velocity and pressure modelled using the numerical solver (top of the figure) and that of the FNO (bottom of the figure). (Continued on next page) Calibrated Physics-Informed UQ (a) Spatio-temporal evolution of the pressure field (P) (b) Spatio-temporal evolution of the horizontal component of the magnetic field (Bx) (c) Spatio-temporal evolution of the vertical component of the magnetic field (By) Figure 23. MHD Equations: Temporal evolution of velocity and pressure modelled using the numerical solver (top of the figure) and that of the FNO (bottom of the figure). (Continued from previous page) Calibrated Physics-Informed UQ Figure 24. MHD: Slice plots along the x-axis (sliced at y = 0.5m) indicating the marginal and joint coverage (90%) obtained over the neural PDE modelling the MHD equations using the induction equation Equation (17) (on the left) and the energy equation Equation (16) (on the right). Marginal coverage, evaluated cell-wise, generates tight bounds to the PRE, whereas joint coverage spanning across the spatio-temporal domain introduces wider bounds. PRE: Dinduction(v, B) (a) PRE of the Induction Equation Equation (17) over the FNO prediction Marginal CP (+q) (b) Upper error bar indicating 90% coverage with marginal-CP Joint CP (+q mod) (c) Upper error bar indicating 90% coverage with joint-CP Figure 25. MHD: CP using the Induction Equation (17) as the PRE for a neural PDE surrogate model solving the Ideal MHD equations. The last time instance of the prediction is shown. PRE: Denergy( , v, P, B) (a) PRE of the Energy Equation Equation (16) over the FNO prediction Marginal CP (+q) (b) Upper error bar indicating 90% coverage with marginal-CP Joint CP (+q mod) (c) Upper error bar indicating 90% coverage with joint-CP Figure 26. MHD: CP using the Energy Equation (16) as the PRE for a neural PDE surrogate model solving the Ideal MHD equations. The last time instance of the prediction is shown. Error bars obtained using joint CP are an order of magnitude higher than that obtained by marginal CP as it is measured across the entire spatio-temporal domain rather than for each cell. Calibrated Physics-Informed UQ PRE: Dcont. ( , v) (a) PRE of the Continuity Equation Equation (14) over the FNO prediction Marginal CP (+q) (b) Upper error bar indicating 90% coverage with marginal-CP Joint CP (+q mod) (c) Upper error bar indicating 90% coverage with joint-CP Figure 27. MHD: CP using the Continuity Equation (14) as the PRE for a neural PDE surrogate model solving the Ideal MHD equations. PRE: DGauss(B) (a) PRE of the Divergence Equation Equation (18) over the FNO prediction Marginal CP (+q) (b) Upper error bar indicating 90% coverage with marginal-CP Joint CP (+q mod) (c) Upper error bar indicating 90% coverage with joint-CP Figure 28. MHD: CP using the Gauss s law for magnetism Equation (18) as the PRE for a neural PDE surrogate model solving the Ideal MHD equations. Calibrated Physics-Informed UQ M. Plasma Modelling within a Tokamak M.1. Physics and Problem Setting We evaluate our uncertainty quantification framework on a simplified Magneto-hydrodynamics (MHD) model in toroidal geometry, leveraging the dataset and neural architecture from (Gopakumar et al., 2023b). The dataset is generated using the JOREK code (Hoelzl et al., 2021) with a physics model similar to the Reduced-MHD model but with electrostatic and isothermal constraints. In this setup, only the density ρ, electric potential Φ, and temperature T fields are evolved. The physical system models the dynamics of a toroidally axisymmetric density blob initialized on top of a low background density. Without plasma current to maintain confinement, the pressure gradient in the momentum equation acts as a buoyancy force, causing radial blob motion. This simplified scenario serves as a proxy for studying the evolution of plasma filaments and Edge-Localized Modes (ELMs) in tokamak devices. M.2. Dataset Generation The dataset consists of 120 simulations (100 training, 20 testing) generated by solving the reduced MHD equations using JOREK with periodic boundary conditions. The initial conditions vary in the blob s position, width, and amplitude. Each simulation is performed within the toroidal domain with 1000 timesteps, downsampled to 100 slices and 200 200 bi-cubic finite-element spatial grid, downsampled to 100 100. M.3. Model Architecture and Training We employ the Fourier Neural Operator (FNO) architecture from (Gopakumar et al., 2023b) with the following specifications: Input: 20 time instances of field values on 100 100 grid Autoregressive prediction up to 70 timesteps Output: Next 5 time instances 4 Fourier layers with width 32 and 16 modes Physics Normalisation followed by standard linear range normalization to [-1,1] The training was conducted on a single A100 GPU with the Adam optimizer having an initial learning rate of 0.001 with halving every 100 epochs for 500 epochs total using a relative LP loss function. This model achieves a normalised MSE of 4e 5 on each field variable while providing 0.2 0.4 0.6 0.8 1-alpha Empirical Coverage Ideal Residual Figure 29. Empirical coverage obtained by performing CP-PRE over JOREK-FNO predictions 6 orders of magnitude faster than the numerical solver. For a complete description of the MHD system and additional experimental details, we refer readers to (Gopakumar et al., 2023b). M.4. Calibration and Validation For this system, we compute physics residual errors (PRE) for the temperature equation (equation 3 from (Gopakumar et al., 2024b)), as it comprises all the variables modelled by the neural PDE. Empirical coverage obtained by performing CP-PRE over the JOREK FNO is shown in Figure 29. Calibrated Physics-Informed UQ N. Magnetic Equilibrium in a Tokamak N.1. Physics and Problem Setting A tokamak uses magnetic fields to confine and shape a plasma for nuclear fusion. While the main toroidal field (TF) running the long way around the torus provides the primary confinement, the poloidal field (PF) coils running in loops around the cross-section are crucial for plasma stability and performance. These coils serve multiple functions: they induce the plasma current that generates an additional poloidal field component, shape the plasma cross-section, and provide vertical stability control. Without these carefully controlled poloidal fields, the plasma would be unstable and quickly lose confinement. The structure of the tokamak with emphasis on the magnetic fields and the coils are given in Figure 31. The shape of the plasma cross-section significantly impacts performance, with modern tokamaks typically using a Dshaped plasma with strong elongation and triangularity. This shape, controlled actively by varying currents in different poloidal field coils, is superior to a simple circular crosssection as it allows for higher plasma current at a given magnetic field strength, improving confinement. The Dshape with triangularity also provides better stability against plasma instabilities, particularly edge localized modes, and enables better access to high-confinement operation. Maintaining this shape requires sophisticated feedback control systems since the plasma shape is naturally unstable and needs continuous adjustment, especially for maintaining the separatrix - the magnetic surface that defines the plasma boundary and creates the X-point for the divertor. A sample magnetic equilibrium across the poloidal cross-section is showcased in Figure 30, with the equilibrium shown in the contour plots and the poloidal field coils indicated with the grey block. The structure of the tokamak and the separatrix are indicated in black and red respectively. Considering the impact of the PF coils on the equilibrium and, hence on the plasma efficiency, its placement within the structure is an open design problem for future tokamaks. Traditionally, the design space is explored by obtaining the equilibrium associated with a certain PF coil configuration, which involves solving the Grad-Shafranov equation using numerical solvers, rendering them computationally expensive. As an alternative, we construct a surrogate model to explore the design space significantly faster. N.2. Surrogate Model A conditional auto-encoder, as given in Figure 32, is trained as a surrogate model that can approximate the magnetic equilibrium for a given configuration of PF coils. The auto-encoder takes in as input the spatial domain of the tokamak as the input and outputs the magnetic equilibrium Figure 30. Sample Equilibrium plot showcasing the magnetic equilibrium as the contour plots observed for a given poloidal field (PF) configuration (PF coil locations are indicated in blue blocks). The poloidal cross-section of the tokamak is shown here, with the structural boundary in black and the closed flux surfaces in pink (Meyer & Team, 2024). Our problem looks at mapping the equilibrium across the spatial domain for a given PF coil location. while being conditioned on the locations of the PF coils in the latent space. The performance of the model can be found in Figure 33. The model was trained on 512 simulations of the Grad-Shafranov simulations obtained using the Free GSNKE code (Amorisco et al., 2024). The coil currents are kept constant as we are looking to identify the ideal coil configuration for a steady-state plasma. N.3. Calibration and Validation CP-PRE is performed by evaluating the Grad-Shafranov equation over the solution space of the surrogate model, with inputs generated by sampling within the bounds of the PF coil locations. By using the GS equation, we can identify the predictions from the surrogate model that generate untenable Calibrated Physics-Informed UQ Figure 31. A simple schematic diagram of a generic tokamak with all of the main magnetic components and fields shown (Li et al., 2014). The poloidal field coil magnets (grey) are that which this work aims to optimise. PF Coil Locations Encoder Decoder Figure 32. Conditional auto-encoder developed as a surrogate model for mapping the poloidal field coil locations to the corresponding magnetic equilibrium under constant coil currents. equilibria. This allows us to explore the design space quickly while adding trustworthiness to your surrogate model. Both marginal and coverage obtained using CP-PRE with GS equations over the surrogate model is indicated in Figure 34. Marginal-CP shows smooth coverage as it represents the coverage averaged across each cell. Ground Truth Figure 33. Comparing the ground truth with the prediction from the surrogate model. 0.2 0.4 0.6 0.8 1- Empirical Coverage Ideal Marginal Joint Figure 34. Marginal and joint empirical coverage obtained by performing CP-PRE over the Grad-Shafranov equation