| Issue |
EPJ Nuclear Sci. Technol.
Volume 11, 2025
Special Issue on ‘Overview of recent advances in HPC simulation methods for nuclear applications’, edited by Andrea Zoia, Elie Saikali, Cheikh Diop and Cyrille de Saint Jean
|
|
|---|---|---|
| Article Number | 61 | |
| Number of page(s) | 24 | |
| DOI | https://doi.org/10.1051/epjn/2025055 | |
| Published online | 01 October 2025 | |
https://doi.org/10.1051/epjn/2025055
Regular Article
Multi-output Gaussian processes for the reconstruction of homogenized cross-sections
1
EDF R&D PERICLES, 7 Boulevard Gaspard Monge, 91120 Palaiseau, France
2
Université Paris-Saclay, CEA, Service d’Études des Réacteurs et de Mathématiques Appliquées, 91190 Gif-sur-Yvette, France
3
EDF DQI, 2 Rue Ampère, 93206 Saint-Denis Cedex, France
* e-mail: olivier.truffinet@edf.fr
Received:
1
June
2025
Received in final form:
6
August
2025
Accepted:
7
August
2025
Published online: 1 October 2025
Deterministic nuclear reactor neutronics codes employing the prevalent two-step scheme often generate a substantial amount of intermediate data at the interface of their two subcodes, which can impede the overall performance of the software. The bulk of this data comprises “few-groups homogenized cross-sections” or HXS, which are stored as tabulated multivariate functions and interpolated inside the core code. A number of mathematical tools have been studied for this interpolation purpose over the years, but few meet all the challenging requirements of neutronics computation chains: extreme accuracy, low memory footprint, fast predictions. We here present a new framework to tackle this task, based on multi-output Gaussian processes (MOGP). These smooth and tunable bayesian regressors are able to model several quantities at once, and to capture the correlations between them – a key asset in the modeling of HXS’s, which we show to be highly similar from one another. Several models of this family are discussed, compared, adapted to the case of very numerous HXS’s, and their possible modeling choices are experimented on. These machine learning models enable us to interpolate HXS’s with improved accuracy compared to the current multilinear standard, using only a fraction of its training data – meaning that the amount of required precomputation is reduced by a factor of several dozens. They also necessitate an even smaller fraction of its storage requirements, preserve its reconstruction speed, and unlock new functionalities such as adaptive sampling and facilitated uncertainty quantification. We demonstrate the efficiency of this approach on a rich test case reproducing the VERA benchmark, proving in particular its scalability to datasets of millions of HXS’s.
© O. Truffinet et al., Published by EDP Sciences, 2025
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
1.1. Problem description
Most deterministic neutronics codes make use of the so-called “two-step scheme”, and thus pass processed nuclear data at the interface of their two subcodes (lattice code and core code). This data consists mainly of few-groups homogenized cross-sections (denoted HXS in this article), which quantify the interaction of neutrons with matter. For a given isotope i and nuclear reaction r, a spatial region V and an energy range [Eg − 1;Eg], the corresponding HXS is defined as:
where σi, r(E, T) is the native, “physical” cross-section as delivered by a neutronics data library, Φ the (scalar) neutron flux, and P a set of operating parameters (temperatures, burnup, etc.) on which the flux Φ computed by the lattice code depends. The flux-weighted averaging aims at preserving reaction rates in the given energy range; this definition makes the HXS’s depend on the operating parameters P. They are therefore represented as multivariate functions, of generally 3 to 7 variables1.
![]() |
Fig. 1. Typical aspect of the 2-group thermal absorption XHS of Xenon 135 (homogenized on a fuel cell). The wireframe mesh corresponds to a realistic multivariate grid for multilinear interpolation. (a) Projection on the Bu (burnup) and Tf (fuel temperature) variables. (b) Projection on the Tm (moderator temperature) and Cb (boron concentration) variables. |
To solve a given reactor state, the core solver requires a set of HXS matching the current operating conditions; but these cross-sections can usually not be computed on-the-fly, lest computation times would skyrocket. In current computation chains, they are therefore tabulated in advance, then interpolated inside the core code. A popular tool for this has been multilinear interpolation [1, 2], which is simply a piecewise interpolation method of degree 1; but despite some significant advantages (in particular simplicity, error control and speed), this model is suboptimal in terms of accuracy, data usage efficiency and storage requirements, which is becoming more and more of a burden as the finesse of simulations increases. There is thus a pressing need to develop more efficient interpolation/regression schemes, by exploring three complementary approaches:
-
Finding better function approximators: HXS’s are very regular quantities, so it is questionable to approximate them with local, memory-based methods such as multilinear interpolation when proficient and easily-tunable smooth estimators are available;
-
Getting rid of full grids: HXS’s exhibit weak interactions between variables, so fitting approximation models to them on full grids is very data-inefficient, and sparse or unstructured supports should be used instead;
-
Leveraging redundancy: HXS’s are strongly correlated between them, as most of their variations originate from the homogenization operation, performed with a flux which is common to all. Exploiting the low effective dimensionality of HXS datasets can thus be a powerful way of avoiding useless computation.
These three lines of work have been investigated in the last decades; note that the first two are often intertwined, as the choice of an approximator can put constraints on the interpolation support.
1.2. Existing methodologies
Better and grid-less approximators. Polynomial models, whether global or local, have been especially studied by practitioners. Works on HXS reconstruction with global polynomials include multivariate polynomial regression in [3] (backed by a cubic spline for the burnup variable), quasi-regression performed as an ANOVA analysis in [4], and hierarchical interpolation on sparse grids in [5]. Local polynomials refer to splines, which have been studied thoroughly in [6]. All these methods yielded satisfactory results, with relative interpolation errors of the order of 0.01%2, and significant data and storage savings – especially with global models for which a few dozens of polynomial terms are usually enough to model a HXS. Yet they all have their limitations:
-
They struggle with the build-up of many isotopic concentrations at the start of the reactor, which causes a sharp variation along the burnup axis in many HXS’s3 – as it can be seen in Figure 1a – and require some specific “trick” to account for it (addition of a cubic spline in burnup in [3], refined meshes at low Bu values in [6], etc.);
-
They sometimes suffer from oscillation problems [6] – the well-known Runge phenomenon for polynomial interpolation [7];
-
They still require large numbers of data samples to be fitted on, from 640 in [3] to dozens of thousands in [4];
-
Some of them maintain an inconvenient grid-like structure: full grids for the spline models in [6], sparse grids in [5], factorized (burnup × other variables) support in [3] (which imposes to store and use depletion cross-sections in addition to the branch ones);
-
Some of them rely heavily on hand-tuned hyperparameters, such as the location of spline knots in [6] or the anisotropy vector in [5], which could lead to problems of generality when applied to different datasets.
The first two items highlight a more general fact: low-order polynomial methods are known to struggle with steep or localized variations, so it is likely that future test-cases will put these methods in difficulty, in particular if more variables are added.
By contrast, kernel methods are in principle able to deal with any function shape, and work with any multidimensional set of points. First results have been obtained with them: in [8], Support Vector Regression was investigated and yielded good performance, with relative errors of the order of 0.05% – although with an unspecified number of training data. In [6], the author experimented on kernel regression and also obtained good results, reaching accuracies one order of magnitude higher than with multilinear interpolation for a same number of datapoints. Yet these two references don’t leverage the full potential of kernel-based approximation: they still use Cartesian grids as an interpolation support, and use rather outdated regressors, devoid of automatic hyperparameter tuning in particular. Moreover, with these models the time spent to make a prediction is linear in the number of training points: this is less favorable than with multilinear interpolation, for which this duration is only proportional to 2d with d the number of variables, and can turn out to be a problem if high interpolation speed is required in applications. Therefore, there is still room for improvement in this line of work.
Lastly, Artificial Neural Networks (ANNs) are a trending tool for many regression problems, and they are beginning to be investigated for the approximation of HXS’s. Still in [6], the author experimented on many different ANN architectures, but didn’t manage to reach the accuracy of multilinear interpolation, let alone this of his previous work on kernel methods. In [9], a three-layered multi-layer perceptron (MLP) – one model outputting the 20 HXS’s contained in the dataset – was able to outperform multilinear interpolation, but at the cost of a 20-fold increase in storage requirements4, and of a rather high prediction duration of ≃10−4 s per HXS and point. In [10], the authors also used a (smaller) 3-layered multi-output MLP, coupled with a step of Principal Component Analysis (PCA) for dimensionality reduction, to simultaneously approximate all HXS’s; they obtained good results, with relative errors of the order of 0.1%. The great originality of their approach is to get rid of the burnup variable by taking isotopic concentrations as inputs of the model, which also enables it to account for history effects. Lastly, the ANN of [11] uses a convolutional structure to model the spatial relationships between the equivalence parameters of all the fuel cells in a given assembly. In all these works, ANNs appear less promising than kernel methods, requiring more data to reach a similar accuracy, and making for less efficient computations; this will be discussed in the concluding section. They are also dependent on a profusion of hyperparameters, in particular the number of layers and the size of each one, which significantly affect the predictive performance and thus require systematical testing of many architectures.
Dimensionality reduction. As stated before, as computational data becomes more numerous in neutronics codes, it is also likely to become more redundant; by resorting to dimensionality reduction techniques (compression, projections…), one could therefore deal with this data more efficiently. Recent studies have begun exploring this avenue. In [12], singular values decomposition (SVD) was used to perform lossy compression on HXS’s from distinct fuel cells; this method authorized 90% of memory savings at the expense of losing only 10−6 of the dataset’s variance, which incurs only a few pcm of error on the keff. The same approach was followed in [13] with Boiling Water Reactor HXS’s, with similar results. In [6] and [9], as previously mentioned, multi-output ANNs were able to jointly approximate all HXS’s from a dataset with relatively low errors; these networks being both shallow (3 layers) and narrow (only ∼10 “neurons” in some layers), they constitute a low-dimensional representation of HXS datasets. Dimensionality reduction has also been used on other neutronic quantities: 618-group HXS-flux products in [14], isotopic concentration data in [15], or even local burnup data in [16]. These works are already showing the potential of dimensionality reduction in neutronic computation chains, although they could still be improved in several ways: better choices of variables to compress or neglect, addition of normalization schemes.
2. Description of the model
2.1. Single-output Gaussian processes
The basic theory of Gaussian processes (GP’s) is now well-known in many applied mathematics communities; a good pedagogical introduction can be found in a reference manual on the topic [17]. As stated in definition 2.1 of this book, “A Gaussian process is a collection of random variables, any finite number of which have a jointGaussian distribution”. Performing Gaussian process regression of some data starts by modeling these data points as observed (probabilistic) samples from an underlying Gaussian process5. Using the joint distribution of these observed values and unobserved ones, probabilities can be estimated for the quantity to be interpolated by conditioning this distribution on known (observed) values.
To fix notations, let (xi,yi)1 ≤ i ≤ n be a set of n input-output datapoints, where the xi’s are d-dimensional and yi’s are scalars. The output values are supposed to be noisy observations of the interpolated quantity f(x): y = f(x)+ϵ(x), where ϵ is an i.i.d. (Independent and Identically Distributed) Gaussian noise of magnitude σ2, that is, 𝔼[ϵ(x)] = 0 ∀x and Cov[ϵ(x),ϵ(x′)] = σ2δx, x′ (δ is the Kronecker symbol). We assume that f follows aGaussian process of mean function m(x) and covariance function k(x, x′), where m(⋅) can be any real function and k(⋅, ⋅) is a kernel, i.e a symmetric positive definite function. As for any Gaussian variable, f is entirely specified by its mean and covariance. If x* is an unobserved input point, X = (xi)1 ≤ i ≤ n denotes the training data inputs and y = (yi)1 ≤ i ≤ n their matching outputs, following the Gaussian process assumption, one can write the joint distribution of the unobserved and observed values:
In order to simplify notations, we denote K = k(X, X), kx*X = k(X, x*) and kx*x* = k(x*, x*). One can deduce from Equation (2) the conditional distribution of f(x*) given the observations (Eq. (3)); the mean of this Gaussian distribution constitutes the regression estimator per se, whereas its variance serves as a measure of uncertainty.
One key to good regression accuracy is obviously to use a covariance function (and to a lesser extent a mean function) able to reproduce the observed data well. The strength of the GP methodology lies in the possibility to chose them as members of parametric families of functions – squared exponential kernel for the covariance, linear model for the mean… – and tune their parameters by maximum likelihood estimation, that is, by maximizing the marginal log-likelihood (MLL) of the data given the current model parameters. This MLL has the following expression in the case of a Gaussian noise ϵ and a zero mean function m(⋅):
where Θ denotes model parameters: the noise σ2 and all parameters θ of the covariance function. This optimization procedure can be performed efficiently by modern automatic differentiation tools and is usually well-behaved (monotonic decrease of the loss function).
2.2. Multi-output Gaussian processes
We now want to perform multitask regression, that is, to model simultaneously several correlated quantities – in our case several HXS’s. Gaussian processes can be adapted readily to this setup, as long as a correlation model between outputs (regression tasks) is specified [18]. A sensible framework for representing cross-task correlations is to model these outputs as linear combinations of some unobserved and independent GPs:
with p the number of outputs to model, mj(x) the mean function of the j-th output, ui’s zero-mean and mutually independent Gaussian processes of respective kernels ki, H ∈ ℝp × q the mixing matrix which mixes these processes into observed outputs, and the
task-specific and independent white noise terms6. The parameters of the model are: all coefficients of H, the noises σj’s, all parameters of the kernels ki’s and mean functions mj’s. They can still be tuned by MLL optimization. The underlying Gaussian processes ui are called latent processes, which means that they are a low-dimensional and unobserved embedding of the observations.
The resulting model is dubbed linear model of coregionalization (LMC) and is again described in [18]; unfortunately, the runtime of computing the multi-task versions of Equations (3) and (4) for this LMC is cubic in the number of points n and number of tasksp. Several approximations or simplifications have therefore been developed in recent years, the most prominent being the intrinsic coregionalization model (ICM, [19]) and the variational LMC7 [20]. Another simplification, the Projected LMC, has been introduced by the authors in a yet-unpublished article ([21]). Lastly, a simplified and training-less approach, the Lazy-LMC, is described in the next section. The description of these models, even with little detail, is cumbersome and outside the scope of this article; we only give an overview of their main features and shortcomings in Table 1, as well as references sufficient to grasp their definition and implementation.
Features of the multi-output Gaussian process models studied in this work.
2.3. A convenient, training-less multitask model: the Lazy-LMC
One of the MOGP models used in this work is an ad hoc construction absent from the literature, so we describe it here. The purpose of this model is to be training-less, that is, to have no free parameter to tune: this way, the interpolation of HXS’s can be done directly from the data, without having to resort to an optimization phase and to the associated numerical tools – auto-differentiation library, etc.
Looking at Equation (5) (and the following explanatory paragraph), we can list a number of parameters, which value we have to set without resorting to optimization methods. This job is detailed in Table 2. Several clarifications should be added to this list:
-
A zero mean function is retained for each regression task (HXS to model): mj(⋅)=0 ∀j. This is a very common assumption in Gaussian process modeling, as the kernel-based component of the model is usually sufficient by itself.
-
Setting H to the principal components of the data matrix Y is not an arbitrary choice: there is in fact a strong connection between the LMC model and the Probabilistic Principal Component Analysis8 (see for instance [30]), so we can hope for this method to yield a good estimate of the optimal mixing matrix.
-
It would be delicate to set the parameters of a given kernel to an adequate value without tuning them to the data. Instead, a safer choice is to choose a parameter-free kernel right from the start. To this end, we retained the cubic spline kernel of [31], defined in Equation (6), which in 1D emulates a cubic smoothing spline:
One key feature of this kernel is that it is non-stationary (see Sect. 2.4): this guarantees its flexibility, making it able to match the local roughness of the data despite the absence of parameter tuning.
Description of the Lazy-LMC model. For each parameter featured in the definition of the LMC (Eq. (5)), a procedure is given to set its value to an adequate one. See Section 2.3 for more information.
In the end, the philosophy behind this proposed model – dubbed Lazy-LMC from now on – is straightforward. To make predictions with it, one simply:
-
Computes the truncated Singular Values Decomposition (SVD, equivalent to PCA) of the data, for q principal components which will correspond to q latent processes;
-
Interpolates these principal components with independent Gaussian processes implementing a parameter-free kernel, as in Equation (6);
-
Lifts these interpolations to the original space by applying the inverse PCA transform (multiplication by the matrix of principal directions, i.e in this case the mixing matrix H).
Note however that this procedure is not rigorous, and doesn’t guarantee good statistical properties9. Its predictions themselves should be accurate, but its measures of uncertainty such as the predictive variance are irremediably flawed.
2.4. Application to the neutronics problem
Let’s now specify how this framework can be adapted to the interpolation of HXS’s. A LMC model is here defined at the level of each fuel assembly. Each row of the training inputs X ∈ ℝn × d is a d-dimensional vector describing the state of the considered assembly: in our test case, it is for instance (Bu, Tf, Tm, Cb), with Bu the burnup, Tf and Tm the fuel and moderator temperatures and Cb the boron concentration. Training outputs Y ∈ ℝn × p are stacked vectors of p HXS’s: each row Yi is a complete set of them (for all spatial regions, isotopes, reactions, energy groups…) evaluated at parameter point Xi. Then predictions y* at an unknown point x* (the interpolation task) are obtained by computing p(f(x*)|Y) as in Equation (3): the mean of this Gaussian distribution is the interpolated values of the HXS’s, while its variance represents prediction uncertainty. We further specify below three important choices made in our approach.
Grouping of the outputs. As stated above, we gather all HXS’s from an assembly into the same LMC model: this assumes that they can accurately be decomposed as a linear combination of shared “prototypes” (latent processes). This assumption is not hazardous but rather motivated by prior work on the low effective dimensionality of HXS datasets [32]. In fact, even defining one model by assembly is superfluous, since the cited article shows that HXS’s are as much correlated across different assemblies as within a given one; we only proceed as such for application-related convenience, and because there would be little computational gain in building an all-assembly model. The reasoning behind this last statement – and more generally, the question of the optimal size of a model for our application – is discussed in Appendix B.
Note that the presented methodology doesn’t depart from the standard interpolation pipeline of neutronics codes: a MOGP model can only be used on the very assembly on which it has been trained (same geometry and materials). More elaborate models could be devised in order to generalize to unseen geometries or material composition, for instance by taking fuel enrichment as a parameter, or by modeling the spatial structure of fuel cells. We leave such possibilities for future work.
Choice of the interpolation points. Unlike traditional methods such as multilinear or spline interpolation, Gaussian process regression doesn’t require a specific structure (multi-dimensional grids) for the interpolation support: any set of multi-dimensional points can serve as training data. This is a strong argument in their favor as it authorizes more flexibility and above all more sparsity, full grids being notoriously data-inefficient for modeling functions with weak interactions between variables10 – which is the majority of real-life quantities, HXS’s included. However, we must now select a relevant set of interpolation points, with little guidance as to its structure. A principled choice for this can be low-discrepancy sequences or space-filling designs, quasi-random sampling methods which fill the domain of interest quickly and evenly. These tools have been thoroughly studied for tasks closely related to interpolation (uncertainty quantification, quadrature…); a popular sequence, Sobol’ points, has shown itself to be very effective for these tasks in a large variety of setups [33]. We therefore retained this option in the present work. The good properties of Sobol’ points are only guaranteed if the sample size is a power of two, so we stuck to such data sizes in our tests. Here, we only present results for 256 points: we found indeed model accuracy to strongly improve up to this number, then nearly stall11. Finally, let us mention that Gaussian process methodology authorizes principled adaptive sampling – incremental and intelligent construction of the training set – which we will discuss in the conclusion.
Construction of the covariance function. In Expression (5) of the LMC model, each latent process ui follows a GP with kernel ki; we have not provided so far a recipe for these covariance functions. This encompasses at least two questions: the choice of a mathematical function, and this of a dependence structure between variables. For the first one, the most common choice is this of a stationary kernel: in this case ki is a positive function ψ : ℝ ↦ ℝ+ applied to a custom norm between multi-dimensional points,
, with λj a fitted lengthscale. The choice of a dependence structure between variables refers to kernel composition: a sum or product of kernels being a valid kernel, one can build partial kernels featuring only some of the variables – for instance:
– and then combine them appropriately. Note that it is customary to scale the base kernel function k by a learned positive parameter: k′=α k; this has no effect on the prediction themselves, but directly calibrates the predictive variance, which is linear with respect to α.
In our experiments, we retained for testing four stationary kernels and five groupings of variables, all described in Table 3. The rationale behind our specific choices of groupings was to test several hypotheses, namely, whether it was beneficial to:
-
Add an extra kernel containing the burnup variable, so that it can resolve short-scale variations at reactor startup;
-
Add another variable (Tf or Tm), which we know to be significantly coupled with the burnup (that is, HXS’s are sensitive to the interaction of Tm and Bu), to this extra kernel;
-
Have at least one kernel featuring all four variables, or if on the contrary the “Fuel + moderator” decomposition suggested in [34] was sufficient.
Description of all modeling choices tested with a grid search. See Section 2.4 for explanations of the notation ψ(r) for stationary kernels, the lengthscale parameters λj, and the notations of variable decomposition. The piecewise-polynomial kernel of order 2, taken from [35], is defined as such :
, where
and ( ⋅ )+ = max(⋅, 0) denotes the positive part.
Choice of the mean function. We must also specify a choice of mean function for the regression tasks – the mj’s of Equation (5). This choice usually receives much less attention in the GP literature; in fact, the mean function is generally chosen to be zero, to the point that Equation (3) is in many places written without it. This is a sensible choice, because GPs are so flexible that their predictions are little impacted by their mean function, as long as the training data is abundant enough. When it is not set to zero, this function is often modeled as a linear combination of basis functions (linear terms, low-degree polynomials…), as it is done in classical regression models. In this case, it is written as m(x)=wTϕ(x), with w the weights of the linear regression and ϕ(x) a vector of basis functions evaluated at x. This regression-like approach can be treated in a bayesian manner along with the GP component: this introduces several complications, discussed in chapter 2.7 of [17]. Otherwise, these weights can be treated as deterministic parameters and tuned by MLL optimization, which is the approach retained here. Note that among our four candidate models (these of Tab. 1), the PLMC is not yet compatible with non-zero mean functions, and the Lazy-LMC will never be as its purpose is to avoid any form of optimization12 – see Section 2.3. For the other models, four types of mean functions were tested, which are again given in Table 3.
2.5. Effect of the modeling choices
In order to simplify the presentation of the experimental results, we evacuate as of now the question of the impact of the previously-described modeling choices, by showing that all “reasonable” options yield similar results, and subsequently deciding on a common set of them for all our MOGPs. All options depicted in Table 3 were gathered in a grid search13 – a systematical test of all the combinations. They were implemented for single-output GPs, and for our three trainable MOGPs (ICM, VLMC and PLMC – see Sect. 2.2); the performances of the resulting models were measured on the test set described in Section 3.1. The results of this large experiment are tedious to represent, so we will summarize them shortly:
-
“Advanced” modeling choices yield small benefits, in particular for MOGPs. Choosing a complex mean function or a nonbasic kernel decomposition can only improve accuracy by about 10 − 15%14, and can often be useless (for instance with the PLMC) or even detrimental (with the VLMC). This was not unexpected for MOGPs, as the fact that outputs are linear mixtures of latent GPs can make complex model features pointless: for instance, adding variable-specific kernels with shorter lengthscales is less relevant if each output already combines q GPs with various lengthscales. Moreover, their optimization landscape is intricate, meaning that adding complexity to the model may not guarantee convergence to the global optimum.
-
Modeling choices have different effects from one model to another, and from the SOGP case. For instance, the choice of kernel function has a noticeable impact on the ICM, but not on the other models; complex mean functions hinder the VLMC, but not the other ones. The interactions between these choices also differ for each model.
-
If a specific grouping of variables is used, one sub-kernel should contain all of them. The “Fuel+Mod” decomposition of Table 3, only option not to meet this criterion, yielded catastrophic results for all models.
-
All standard, long-ranged kernels yield similar results, and this is true for all models; by this we refer to the squared exponential, Matérn-5/2 and rational quadratic kernels of Table 3. Here again, the choice of kernel can only impact predictive accuracy by 10 − 20% in relative terms, and their ranking varies from one model to another. The main criterion for choice should therefore be computational efficiency; in particular, polynomial truncations of the exponential-based kernels tried here could be put at test.
-
By contrast, the Piecewise polynomial kernel yielded significantly worse results in all settings, hindering accuracy by more than 30% in relative terms. The aim of testing it was to investigate the use of compactly supported kernels for greater computational efficiency: it can be seen indeed from its expression in Table 3 that it is zero for r ≥ 1, meaning that datapoints far away from the predicted one don’t contribute to the prediction. This experiment thus advises against local interpolation, as long-range terms apparently contribute to prediction accuracy; other findings of ours corroborated this idea15.
All these observations lead to the same conclusion: simple modeling choices should be favored when using MOGPs for HXS approximation. More complex options bring little improvement at best and can behave in unsettling ways – which differ from one model to another, one implementation to another, and perhaps one dataset to another. In the remainder of this work, we therefore stick to the following simple choices for all models16: a zero mean function for all tasks; squared exponential kernels for all latent processes; and no specific variable grouping (each kernel features the four considered variables).
3. Experimental results
3.1. Test-case description
In this talk, we test our models on a full set of microscopic HXS’s, sufficient to model a standard PWR. The test-case reproduces the notorious VERA benchmark [36]; more specifically, it comprises the twelve assembly types which describe the full core of problem no5. HXS’s matching this benchmark were computed with the neutronics code APOLLO3® [37], a deterministic transport neutronic code with multi-channels 1D thermal-hydraulic feedback developed at CEA. They include 40 isotopes, corresponding to actinides, disintegration chains of xenon and samarium, intercycle elements and gadolinium isotopes; only reactions of fission, absorption and scattering were considered. Two discretization schemes were adopted, leading to two increasingly challenging datasets:
-
A standard-discretized dataset, with 2 energy groups, spatial homogenization at the level of the assembly and 1 anisotropy level for scattering cross-sections: this makes for 287 HXS per assembly (3444 in total), and a training dataset weighing about 3 Mb.
-
A finely-discretized dataset, with 20 energy groups, spatial homogenization at the level of the fuel cell and 4 anisotropy levels for scattering cross-sections: this makes for about 360 000 non-zero HXS per assembly (4.3 millions in total), and a training dataset weighing about 8 Gb.
As explained in paragraph 2.4, the retained set of points used for training our MOGPs is a 256-point Sobol’ design in 4 dimensions, which value ranges are specified in Table 4; all variables are defined at the level of the assembly. Burnup values start at 75 MWd/t to match a standard industry practice17. Sobol’ points having Tm > Tf were also discarded18 for being non-physical, so that the final training set contains only 201 input points. The last column of the table describes the grid size used for the multilinear interpolation baseline, to which we compare our method; this grid therefore contains 29 × 7 × 7 × 3 = 4263 training points. We emphasize that this grid is only used by the multilinear model, not by the MOGPs. Finally, a test set was generated to compare methods on: it contains 393 points sampled from a Latin Sampling Hypercube design filling the whole input space (as defined by the value ranges of Tab. 4). More exactly, 512 points were sampled and these with Tm > Tf were discarded as above. We illustrate the three sets of points used in our experiments in Figure 2.
![]() |
Fig. 2. Schematized view of the three sets of input points used in the experiments, projected in the plane (Tf, Cb). Color fading represents the euclidian distance to the projection plane. |
Specification of the input domain used in experiments.
3.2. Experimental specifications
Data postprocessing. Several processing operations can impact the training of the model and/or facilitate the interpretation of error metrics. First, it was noted in [38] that working with
rather than Bu improved interpolation results, as variations are much steeper at low Bu values: we tested and corroborated this claim. Then all input variables were adjusted to the range [ − 1; 1] by an affine transform19, in order to ease kernel computations and align their characteristic lengths. HXS’s were centered and normalized by their maximal absolute value (
, with σi an original HXS), so that each training output yi has zero mean and values in [ − 1; 1]. Centering is crucial as our model has zero mean by construction (we set m(⋅)=0), and normalization is too as HXS amplitudes spread over 10 orders of magnitude: a LMC model fitted over un-normalized sections would only focus on these of the largest magnitude. However, this implies storing the original mean and norm of each HXS to renormalize their interpolated values.
Training protocol. Our models contain several parameters: all kernel parameters, the mixing matrix H, task wise noises σi, parameters of the variational approximation for the VLMC… These must be optimized by maximum likelihood estimation: we therefore have to specify a training protocol. Many variants were tried, and optimization proved to be well-behaved in all “reasonable” setups; the only concerns were convergence speed and training budget (allowing enough iterations for the model to converge to the global minimum). We retained the following scheme, well-suited to all of our experiments. The popular Adam optimizer [39] was used for gradient-based optimization; gradients were computed by automatic differentiation with the python torch library [40]. Although Adam automatically varies the learning rate of each parameter, we found additional learning rate scheduling to speed up training: an effective policy was to decrease this rate linearly from ∼10−2 to ∼10−3 in a few hundred iterations, then keep it constant. We also implemented a simple stopping criterion, ending training when differences between successive values of the loss (MLL, see Eq. (4)) remained below a relative threshold of δℒ = 5 ⋅ 10−5 for a given number of iterations Nfluct = 50. Our implementation of GP models is based on the gpytorch library [41], and available at the address https://github.com/QWERTY6191/projected-lmc. All experiments were coded in python and performed in double precision, except for this run on GPU with the large dataset, for which single precision was used. In the course of our experiments, we tried switching from double to simple precision in various contexts; we found no effect on the quality of results, but occasional impacts on the dynamics of model training.
Error metrics. Model accuracy is assessed through the use of an external test set: we make predictions with the model on the testing dataset described in Section 3.1 (which was not used for training), and compute error metrics on these predictions by comparing them with the true HXS values. Our retained metrics are the R2 coefficient or coefficient of determination, which measures the share of variance in the output that is predictable from the inputs (
, with yij the true value of the normalized HXS j at input location i,
its predicted value, and
the mean of the true values of HXS j), the Root Mean Square Error (
), and the maximal absolute error (
). The R2 coefficients should be as close as possible to 1; the RMSE and Errmax should be comprised in [0; 1], as we are working with unit-sized normalized outputs. Notice that errors computed in this way are expected to be much larger than these reported in the literature: that’s because we work with absolute errors measured on centered and normalized outputs, as opposed to relative errors on raw outputs in the previously-cited sources [3–5, 42]. We proceed as such to obtain results consistent with these of the machine learning community: it can indeed be noted that HXS often vary very little around their average value (usually around a few percents and sometimes much lower), yielding relative interpolation errors of 10−4 − 10−5 which are meaningless from a mathematical point of view. In order to produce error metrics closer to the final observables, we also compute errors at the macro-sections level – that is, we multiply the absolute error on un-normalized microscopic HXS’s by the true concentration of the corresponding isotope at the considered point, to obtain an error in cm−1: Err
.
We also assess the quality of variance estimation with two metrics: the Predictive Variance Adequacy,
, where vij is the estimated variance at point j and task i, which compares predicted variance with actual errors and should be as close as possible to 0; and α-CI, the proportion of test points falling in the 95% confidence interval, which should therefore be as close as possible to 95%. Finally, we display a storage rate Rstor, ratio between the number of floating-point numbers stored by the model and this of the multilinear model (which in this case are only the tabulated HXS values).
Baselines. In order to assess the performances of our GP models for HXS approximation, we wish to compare them with time-tested methods. At the same time, we want to keep easily-implementable methods which require little practitioner knowledge. We therefore selected three reference methods, either directly taken from the HXS literature or inspired by it:
-
Multilinear interpolation (abbreviated as MLI in the following): this is the industry standard (already described in Sect. 1) which we aim to replace, and therefore an obvious baseline to compare to. Of course, this method requires the data to be located on Cartesian grids, contrary to GP models: consequently, the comparison is performed between very different-sized interpolation supports. It can indeed be seen that the grid described in Table 4 features 4263 points, against only 201 for our training Sobol’ set.
-
Bayesian Polynomial Regression (abbreviated as BPR): this is a polynomial regression method with a bayesian treatment of the weights, described in [43]. This model, although not tested in the HXS literature, is similar to existing works with polynomials, in particular [4]; it serves as an intermediate-complexity baseline, more advanced and flexible than MLI (in particular, it is not grid-bound, so it can be compared with MOGPs on the same training data) but less than GPs. Moreover, this bayesian method comes with a variance estimator, which serves as a comparison for this of the GP models. Implementation is done with the BayesianRidge class of the sklearn library.
-
Single-output GPs (abbreviated as SOGP): here, we simply use a collection of SOGPs, one for each modeled HXS. These models are trained as in the above paragraph 3.2. Regarding modeling choices, the same grid search as in Section 2.5 was undertaken on these GPs; the retained option was the combination which yielded the best RMSE on average over all HXS’s. This baseline serves as a best-estimate model: we wish to assert whether using low-dimensional, multi-output models such as MOGPs incurs a loss of accuracy compared to a collection of GPs, each fine-tuned to a specific HXS.
One advantage of these baselines is that they have few hyperparameters and thus require little hand-tuning; MLI, for starters, has none. BPR has four hyperparameters, which relate to priors on the weights of the regression: the original article describing the method [43] suggests setting them to 10−6 to make the prior non-informative, which we did. We also had to provide a maximal degree for the polynomial features: we selected a value Dmax = 7, as the accuracy of the model didn’t improve beyond this value. Lastly, a maximal number of iterations and a tolerance value must be provided for the fitting of the model; these options were found to be non-impactful as long as they were selected large enough and small enough respectively.
3.3. Effect of the latent dimension
Before comparing our MOGPs to the baselines, we must make an important choice: this of a latent dimension, i.e of a number of latent processes (see Eq. (5)). Optimizing it is key for performance, as we will see in Section 4: both the computational cost of predictions and the storage requirements of the model are linear in this parameter q. On the other hand, its value must be large enough for the low-dimensional space to accurately represent the data.
We therefore wish to study how this parameter affects the accuracy of predictions. To this end, for each even value of q ranging from 4 to 40, we train and fit an instance of each MOGP with q latent processes on one assembly of the standard-discretized dataset, make predictions on the external test-set, and compute the RMSE between these predictions and the true HXS values. The results are shown in Figure 3a. They are very clear: for all models, the RMSE decreases sharply until q = 12, then stalls. This of the PLMC even increases beyond this point, showing signs of instability. The perfect agreement of the 5 models between q = 6 and q = 18 is remarkable, and very reassuring, as it tends to indicate that they are all fitted to their best20. In the same vein, it can be observed that the RMSE values displayed on the graph are excellent: the data being normalized to unit size, a RMSE of 0.002 is a synonym of relative errors around 0.2% on average.
![]() |
Fig. 3. Variation with the number of latent processes q of the prediction error (RMSE measured on one assembly of the test dataset) for the studied MOGP models. (a) Standard-discretized dataset. (b) Finely-discretized dataset. |
The stalling of the RMSE above some value of q is surprising: our models seem to be unable to exploit the additional information brought by more latent processes. One possible explanation is that the optimization procedure is not guaranteed to find the global optimum of the MLL for large values of q: an intricate optimization landscape could prevent training from resolving the fine-grain details of low-amplitude latent processes. Another one is that the information gain brought by additional latent processes may be of the same magnitude as the interpolation error made on them: in this case, it could act as “noise” at prediction time and yield no accuracy improvement. In order to test this last hypothesis, we can simply perform truncated SVDs of the considered training data matrix Y with increasing values of the truncation rank r, and plot the resulting compression losses against r. We define this compression loss as 1 − EVR(r), with EVR(r) the cumulative explained variance ratio of the SVD – the sum of the r computed singular values divided by ||Y||F, such that EVR(p)=1. The results are shown in Figure 4. They are in line with our hypothesis: for r = 12, only 10−4 of the data variability is lost to SVD truncation, which is smaller than the accuracy of our SOGP interpolants measured in Table 5. It is therefore plausible that adding latent processes beyond this value only brings “noise” to the model, whether it is at training time or at prediction time. It is important to notice that curves Figures 4a and 4b are almost identical: the finely-discretized dataset contains scarcely more information than the standard-discretized one, as in both cases this information is almost entirely contained in the first 30 principal components.
![]() |
Fig. 4. Compression loss 1 − EVR(r) of the training data (normalized HXS’s of one assembly) in function of the truncation rank r of the SVD, where EVR(r) is its cumulative explained variance ratio. The erratic behavior of the left part of subfigure b/ is due to the compression loss reaching single-float precision (while figure a/ was computed in double-float precision). (a) Standard-discretized dataset. (b) Finely-discretized dataset. |
Performance metrics on all assemblies of the standard-discretized dataset, for the studied MOGP models (all with q = 16 latent processes) and three baselines. The best and worst value for each metric are highlighted in green and red respectively. Definitions of the metrics are given in Section 3.2; they are all averaged across all HXS’s. The ErrΣ metrics were not computed for the polynomial method, which was tested at a time when concentration data was unavailable. The training time Ttrain is defined at the level of one assembly. n is the number of datapoints used for training.
We can also inquire on the optimal value of q for the finely-discretized dataset, containing p ≃ 3.5 ⋅ 105 HXS’s per assembly. We thus perform the same experiment as in Figure 3a on this dataset; results are shown in Figure 3b. Notice that the ICM is absent from it: in available implementations of this model, computational complexities contain a term O(p3), making it intractable for large numbers of HXS’s 21. The main observation is that we still observe a plateau in the progress of the RMSE with q, starting at about the same value. With the PLMC, the same signs of instability as for the smaller dataset (Fig. 3a) are visible. Another interesting observation is that this time, the training-less model Lazy-LMC performs significantly worse than its competitors: its best RMSE is twice larger than the best value for the VLMC.
Returning to our initial question of the optimal choice of q balancing accuracy with computational performance, one important conclusion stands out: even the largest HXS datasets are of extremely small effective linear dimension, which makes it possible to use low-dimensional approximations on them with as few as a dozen of latent components, at the cost of an accuracy loss which is indistinguishable from interpolation error. We therefore retain q = 16 in the next experiments.
3.4. Comparison with the baselines
We start by testing our methodology on the standard-discretized dataset described in Section 3.1, performing a simple comparison between our regressors: we train a multi-output model (or a collection of models for the baselines) for each of the 12 fuel assemblies, make predictions with them on the external test set, compute error metrics and aggregate them. Results are displayed in Table 5; the training time for each assembly-wise LMC model was about 2 minutes on a standard CPU machine. Note that variational GPs, of which our VLMC model is an example, make use of inducing points, pseudo-observations meant to “replace” the more numerous original ones: we retained a number n′=⌊n/1.5⌋ of them, with locations identical for all latent processes and learned by the model22. This choice was once again made for optimal accuracy: putting n′=n fixed inducing points at the locations of the observed data would theoretically be more accurate, but proved less stable in practice. A few main observations can be made from Table 5:
-
All four MOGP models have very similar performances, as already noticed in Figure 3a.
-
Their accuracy is overall excellent: all R2 coefficients are very close to 1 in particular. All error metrics are very close to these of the collection of SOGPs, even though major dimensionality reduction was achieved, and simpler modeling choices retained for MOGPs (squared-exponential kernel, no complex mean or variable decomposition). MOGPs also best the BPR baseline in terms of RMSE and maximal error.
-
GP models reach better accuracy than the multilinear baseline with 20 times less interpolation points 23.
-
Training a MOGP is much faster than training a set of SOGPs – about 500 s in the first case versus 16 600 s in the latter. These SOGPs were however trained sequentially rather than in parallel: if we assume that 40 such models can be trained simultaneously on a single machine with the same training time per model as in the sequential setting24, the training duration of the SOGPs would be reduced to 415 s. Although it is not displayed in Table 5, making predictions with MOGPs is also much faster than making them with a set of SOGPs, as it will be explained in Section 4.1.
-
MOGPs offer major storage savings: a value ℛ = 167 means that the model occupies 167 times less memory than the lookup table required for multilinear interpolation. This is one order of magnitude better than the gains achieved with SOGPs or BPR; this performance is made possible by dimensionality reduction.
-
The performance of Lazy-LMC is surprising: its accuracy rivalizes with or even surpasses this of other MOGPs, while it is the only one of them which doesn’t require training – hence its tiny Ttrain metric, which only corresponds to fitting duration in this case. This makes it a very attractive applicative choice if simplicity is desired.
-
Lastly, the predictive variance of the MOGPs appears to be very ill-calibrated: a PVA of −5.87, for instance, means that predicted uncertainties are on average
times larger than actual errors. In these conditions, all test points fall inside the 95% confidence interval, hence the α-CI values of 1.00 for VLMC, PLMC and Lazy-LMC. For ICM, the situation is more intricate: this model has at the same time a very negative PVA, indicating a significant overall overestimation of uncertainties, and an α-CI well below 0.95, showing that uncertainty is on the contrary underestimated for a significant fraction of all points. We explore the behavior of these variance estimates in the next subsection, to hint whether there is still hope for UQ of neutronics data with MOGPs.
In short, apart from variance estimation, MOGPs are up to their promise: they deliver very accurate predictions – better than multilinear interpolation and polynomial regression, almost as good as the best-performing SOGPs – with a very low training cost, and allow very fast predictions and major storage savings. For the moment, all four of our MOGP models are still in the running, as they all have very similar predictive performances; the absence of training of the Lazy-LMC, however, is a significant advantage, which authorizes in particular extremely simple implementations (no auto differentiation library is needed).
Results for the finely-discretized dataset are very similar to these presented in Table 5, with some subtleties to be discussed; we therefore defer it to Appendix A. We only emphasize that three of our MOGP models (VLMC, PLMC and Lazy-LMC) could be successfully trained and/or fitted on ≃3.5 ⋅ 105 HXS’s, yielding predictive RMSEs of the order of 10−3; for the PLMC, training was completed in only 2 minutes on a GPU. This proves that this methodology is fully scalable to even the largest current HXS datasets.
3.5. Improving the predictive variance
In the previous section, we found MOGPs to yield ill-calibrated variance estimates. This deserves investigation: although uncertainty quantification is not a primary goal of this work, it is nevertheless a growing subject in neutronics, in which GPs could open new avenues. After exploring several leads, we found three main factors impacting the predictive variance, which can become very well-calibrated if the right choices are made.
Fixing conditioning issues. It is a well-known problem in GP literature that kernel matrices tend to be poorly conditioned [44, 45]. This has a natural explanation: for two nearby training points xi and xj, the corresponding rows k(xi, X) and k(xj, X) in the kernel matrix K are very similar if the kernel function is smooth, meaning that K is close to singular. The noise parameter σ2 (see Eq. (3)) acts as a regularization term in the inversion of K, as it adds a positive diagonal term to it; noise being closely connected to variance estimation, one can suspect that wrong predictive variances are linked to conditioning problems. We found indeed the condition numbers of kernel matrices to be very large in our case – up to 1010.
One way in which conditioning can impede the predictive variance is through an eventual noise lower bound. Indeed, it is sensible for a GP implementation to have a noise thresholding mechanism, in order to keep all computations well-conditioned during model training; this is the case in the gpytorch library which we use. In our case, the modeled data is so smooth that we observe fitted values of σ to match this lower bound σB in all situations. We can thus suspect that these fitted noises are constrained to artificially high values, and that better variance estimates could be obtained by decreasing the lower bound and letting noise terms reach their “true” ranges. Experiments confirm this intuition: when we decrease σB, we see all computation times to increase sharply due to worse conditioning, but the predictive variance to become much better-calibrated. There is therefore a trade-off to make between variance estimation and speed. One could also try to improve the conditioning of kernel matrices by other means, in order to be able to lower noise values without sacrificing computation times – for instance by pruning the training data as suggested in Section 4.3, or by resorting to the preconditioning techniques and sparse approximations of kernel matrices presented in [45]. One such improvement is available on-the-shelf for the VLMC: it suffices to decrease the number of inducing points which the variational approximation relies on.
Moving to standard normalization. As stated in Section 3.2, we processed each HXS by normalizing it to its maximal value. The reason for this choice was to work wit unit-sized quantities, so that absolute errors could be easily interpreted; this way we could avoid the computation of relative errors, where problems of divisions by zero occurred. However, for any vector f there stands max(|f|) ≥ std(f), so that HXS’s normalized with this scheme can have any variance between 0 and 1; it happens that for some of them var(f)≪1. We can make the hypothesis that gathering tasks with disparate variances inside a single LMC model is troublesome for its variance estimation; we therefore tried to normalize the data by its standard deviation instead, and to assess the predictive variance of the models in this setting. Once again, a drastic improvement of the predictive variance is observed: we thus conclude that the training data for each task of the MOGP should have variance ≃1 for the predictive variance of the model to function well. It is interesting to note in Table 6 that this change of normalization also improves on the model’s R2; this suggests that the outputs of the smallest variance are correctly addressed in this setting, whereas they were neglected by the model in the previous one25.
Summary of the process of fixing predictive variance for VLMC and PLMC. Each row denotes a modification performed on them and the corresponding results. These modifications are incremental: each one is kept in the following rows.
Making noise-appropriate modeling choices. One last possible cause of poor variance estimation is straightforward: any GP model makes assumptions about the noise in the data, and these assumptions must be appropriate for uncertainty estimation to work well. A discussion of these modeling choices is way beyond the scope of this article, at it delves deep into the mathematics of the studied models. Let us just mention the options which yielded the best results, and refer to the literature for an explanation of them:
-
For the VLMC, the important factors are the chosen variational distribution and loss function – which in this model is not the standard MLL (Eq. (4)), but a lower bound of it. The variational distribution refers to the probability density which is used to parametrize the posterior distribution of the model; we found the most effective choice to be the standard one of using a fully-parametrized multivariate Gaussian, instead of simplified options such as the Dirac distribution (“delta variational distribution”) which we used in previous experiments. Regarding the loss function, several options are available in the literature, the most popular today probably being the ELBO of [26]; but some have been designed which specifically aim at improving the predictive variance, for instance the so-called “Predictive Log-likelihood” [28]. We found indeed that training the model with this PLL rather than the standard ELBO improved its variance estimation, while preserving its accuracy.
-
The PLMC, in its turn, is based upon a restriction of the general noise model of the LMC [21]. This restriction can be made more or less drastic, the harshest choices yielding the best computational performance for the model; yet one can assume that the less-restricted noise models will give the most faithful variance estimation. This is indeed what we observed: the least-simplified variants of the PLMC should therefore be used if the quality of variance estimation is favored over speed.
The above experiments are summarized in Table 6, in a cumulative fashion: each change made to the model is kept in the following rows. It can be seen that the added effects of the suggested changes turn the calibration of predictive variance from catastrophic to very decent: the VLMC exhibits a PVA of 0.484, meaning that the actual errors are on average only
larger than the predicted uncertainties; likewise for the PLMC,
. The α-CI values for these models mean that respectively 89.3% and 99.5% of prediction errors fall within the 95% predicted confidence interval, close to the desired value (95%). Further work on conditioning and modeling choices could still improve on these results.
Case of the ICM and Lazy-LMC. We weren’t able to achieve such good results with the ICM. The change in normalization yielded somewhat acceptable variance estimation, with PVA = −0.70 and α-CI = 0.80. Notice however that these two values point towards opposed conclusions: PVA < 0 suggests overall overestimation of errors, while α-CI < 0.95 indicates the opposite. This means that errors are underestimated for many HXS’s, and very overestimated for others: there is not only a problem of calibration, but also of estimation quality. It is unclear how to improve on this situation; there are hints that this problem may stem from linear algebra issues, as the conditioning problem remains vivid, and as we will see in Section 4.3 that another kind of computations with this implementation of the ICM exhibits inconsistencies. Efficient computations with this model indeed rely on subtle Kronecker-product-based operations and eigenvalue computations, as described in [25], making it sensitive to implement details.
Lastly, no modification had a positive impact on the predictive variance of the Lazy-LMC, which PVA always remained below the catastrophic value of −11. Several reasons can explain this underperformance: the training-less SOGP modeling each latent process already has very poor variance estimates, probably because it doesn’t feature a variance scaling parameter (see Sect. 2.4), and perhaps also because of its peculiar spline kernel, which is not even derivative-continuous. The covariance structure of the latent processes of the Lazy-LMC is also inconsistent, as explained in Section 2.3. There is therefore little hope that this model can ever yield decent uncertainty estimates.
4. Practical considerations
4.1. Computational performance of interpolation
We didn’t include measurements of prediction durations in our experimental results of Section 3.4. This is because such values would be heavily implementation-dependent; the methods tested here being only prototypes written in Python with little consideration of performance, they would be little representative of industry-grade implementations. However, we already mentioned interpolation speed to be critical in neutronics applications; assessing this performance in some manner is thus mandatory.
We therefore propose here a theoretical analysis of the computational cost of MOGP prediction. We focus on the computation of the predictive mean (the interpolated values), assuming that prediction speed is much less critical for variance estimation and UQ-related tasks. In the following paragraphs, we compare how predictions are made with SOGPs, MOGPs and multilinear interpolation, in order to compute their costs and identify their respective sources of over- or under-performance.
Cost of predictions with multilinear interpolation. We start with predictions of the MLI model, as the reasoning behind this analysis will help to grasp the issues posed by the other models. A prediction with MLI at the input point x* is represented by the following operation:
where the
’s are the values of the HXS at the 2d grid points which surround the test location x*. The coefficients αi(x*) – which expressions are given in [46] for d = 226 – are geometrical in nature: they only depend on the coordinates of x*, not on values of y. They are therefore identical for all HXS’s. Consequently, as long as the number of modeled HXS’s is large enough, that is, p ≫ 2d, the cost of computing the αi’s is negligible compared to this of performing the p linear combinations of Equation (7). The total cost of making predictions (defined as the number of required floating-point operations) is then:
Cost of predictions with SOGPs. Gaussian processes being linear estimators – predictions are linear combinations of the known values yi – the cost of making predictions with them follows the same logic as for MLI. A prediction with a zero-mean SOGP which can indeed be written as:
where we recall that xi is the i-th training point, K = k(X, X), and [ ⋅ ]i denotes the i-th component of a vector. The first line of Equation (9) aims at highlighting the similarity with the multilinear case: in this case too, the prediction is a linear combination of the observations yi, which coefficients
are geometrical functions27 of the test location x*. There are however two important differences: this linear combination now has size n instead of 2d, and the coefficients
are different for each HXS, as a distinct kernel has been trained for each.
The second line of Equation (9) highlights a different aspect of this computation: the product K−1y is independent on x*, so it can be precomputed in an offline phase to make a cacheC. The cost of making predictions for an HXS is then the cost of computing the n kernel values k(x*, xi), plus this is the linear combination of size n. If we note ck the cost of a single kernel evaluation, the total cost of making predictions for p HXS’s is then:
This compares unfavorably to MLI, for which 𝒞MLI ≃ 2 × 2dp. For one thing, (ck + 2) is expected to be much greater than 2, as the cost of a kernel evaluation is generally much higher than this of a simple multiplication28, at least by one order of magnitude. Moreover, n is expected to be significantly larger than 2d – all training points are involved in the computation, instead of just the nearest ones: in our case for instance, n = 201 while 2d = 16. The lesson here is clear: even when precomputing all quantities that can be, predictions with SOGPs involve two costly operations: the computation of a kernel value at each of the n training points, and the linear combination of these n values with precomputed coefficients. This applies to any long-ranged, kernel-based regressor; such methods will always have prediction costs at least two orders of magnitude larger than MLI. Fortunately, MOGPs can bypass this bottleneck, as they don’t perform kernel operations for each HXS, but only on a much smaller collection of latent processes.
Cost of predictions with multi-output GPs. Predictions with the VLMC, PLMC and Lazy-LMC29 are performed in two steps: first, predictions are made for the independent latent processes of the MOGP, as in Equation (9); then these latent values are mixed by a multiplication with the mixing matrix H (see Eq. (5)).
where we recall from Section 2.2 that the uj’s are unobserved latent processes of respective kernel kj. Here, Uj denotes a low-dimensional summary of the training data “feeding” the j-th latent process; the expression of this summary differs from one of our models to another, so we refer to the fifth chapter of [24] for a complete exposition. The important point is that it can be precomputed in an offline phase, meaning that the term Cij in Equation (12) is a precomputed cache.
The cost of predictions with such models can be read from Equation (12). The first line is the computation of q SOGP predictions, which represents n(ck + 2)q operations according to the previous paragraph, while the second line is simply a matrix-vector product of size pq × q; the total cost is then:
The key point here is that the costly GP interpolation now only applies to a small number q ≪ p of latent processes. Let us consider the case of large HXS datasets, that is, p ≫ n(ck + 2); in this case 𝒞LMC ≃ 2pq. Remind that 𝒞MLI ≃ 2d + 1p, so that comparing the interpolation speed of MLI and LMC in this setting amounts to compare 2d and the number of latent processes q. We saw in Section 3.3 that q ≃ 10 is optimal even for our finely-discretized dataset; one can expect similar values for other test-cases30. Therefore, for large HXS datasets, the cost of making predictions with the LMC is identical to or even better than this of MLI. This performance is made possible by dimensionality reduction: in this regime, the main cost is not the interpolation task, which is confined to the low-dimensional space, but the lifting of the low-dimensional interpolated values to the original space.
However, this many-HXS regime may not apply to all situations encountered in industry calculations: for instance, in our representative standard-discretized dataset with p = 287 HXS per assembly, assuming n ≃ 200 and ck ≃ 30 (piecewise-polynomial or spline kernel), there stands n(ck + 2)≃6000 ≫ 287. Prediction costs are then one order of magnitude higher than these of MLI. There are however two ways to alleviate this burden:
-
Reducing the number of training data n: the training support can be pruned by downsampling (see Sect. 4.3) or built incrementally from scratch by active learning (Sect. 5). Preliminary experiments in [24] show that values as low as n = 100 can be reached on our test-case without loss of accuracy. Besides, for the VLMC, it is in fact not the training points, but fewer virtual inducing points which are used at prediction time; here again, numbers m < 50 of inducing points were tested in [24] with preserved accuracy.
-
Mutualizing kernel computations: if the same kernel is used across all latent processes, the number of kernel operations goes from n(ck + 2)q to n(ck + 2). This is the case if the spline kernel of Equation (6) is used for instance, as it has no parameter which can differ from one latent process to another. In such a case, the computational cost falls back to the order of this of MLI.
Also mind that this analysis only addresses arithmetic operations, not memory transfers. The operations described here being elementary (small matrix-vector products) and the volume of involved data being large, HXS interpolation is in fact likely to be cache-bound or memory-bound in many settings and implementations. In this case MOGPs would have the edge over MLI, as they work with much smaller data, as we will see it in the next section.
4.2. Model storage and interpretability
One of the objectives of the present work is to reduce the size of HXS data libraries, which can get very large today for best-estimate simulations (up to hundreds of gigabytes). MOGPs achieve this goal in two manners: by reducing the number of training samples to store for interpolation, and by leveraging dimensionality reduction. It can be seen in Tables 5 and A.1 that the volume of HXS libraries can be divided by a factor of more than 200 this way. The only quantities to store, which completely define the model, are: the mixing matrix H ∈ ℝp × q, the matrix U ∈ ℝq × n collecting the quantities Uj introduced in the previous section (low-dimensional summaries the data “seen” by the latent processes), the matrix of training data inputs31X ∈ ℝd × n, and the few parameters of the model (task-wise noises and kernel parameters for the q latent processes). Neglecting the latter32, the number of floating numbers to store is thus pq + qn + nd, to be compared to the np values contained in the matrix of training observations Y.
It is interesting to note that for the PLMC and Lazy-LMC, the form in which the model is stored matches a very common matrix compression framework. For these models, U has a simple expression given in [21]: U = TY, with T some generalized pseudoinverse of the mixing matrix H (i.e, TH = Iq). This means that HT is a projection matrix ((HT)2 = HT), and therefore it stands:
In other words, multiplying together the two stored matrices defining the model –H and U– yields a rank-q projection of the original data Y. Such a factorization is a common manner of compressing a matrix of low effective rank; for instance, SVD-based compression is a staple in many fields. Incidentally, recall from Section 2.3 that the definition of the Lazy-LMC involves Singular Values Decomposition: if UΣVT is the ranq-q truncated SVD of Y, H = UΣ and U = VT. The property Y ≃ HU is beneficial in terms of model interpretation: one can dispense with storing the full (potentially large) data matrix Y, but still recover an optimal rank-q projection of it by a simple multiplication of the stored H and U.
Also note that this format makes for efficient computations. U ∈ ℝq × n being very small, it can be stored in low-level caches during predictions; moreover, X, H and U can each be stored contiguously and are fully read during any interpolation work. By contrast, with multilinear interpolation, the 2d grid points nearest to the prediction location must be searched for each single HXS inside the large tensor of its tabulated values, in a discontiguous manner.
4.3. Closed-form expressions for leave-one-out metrics
Leave-one-out cross-validation (often abbreviated LOO-CV) is a widespread tool in machine learning, which makes it possible to assess the performances of a model without resorting to an external test set. Its principle is straightforward: for each point of the n-sized training dataset, the model is trained and/or fitted on the n − 1 other points, tested on the point left aside, and error metrics are computed on this prediction; then these metrics are aggregated over all n runs. We can give at least three interesting applications of LOO-CV in neutronics applications:
-
Model assessment: any model used for interpolating HXS’s in the core code should have been assessed after its fitting and/or training, in order to guarantee the quality of its predictions. In industry applications, it would be desirable that this validation doesn’t rely on an external test set: otherwise, the computational effort dedicated to generate this set would be similar to this of generating the training data. LOO-CV makes this possible, by estimating the model’s accuracy directly on the training outputs. In the same manner, several candidate models (having various modeling options, hyperparameters, etc.) can be compared this way without any external test set.
-
Outlier detection: HXS data can be home to outliers, non-physical values which may greatly disturb machine learning models incorporating it. There is no unequivocal manner to categorize a datapoint as an outlier, as the variability of the data is rarely known in advance. LOO-CV offers a practical way to define outliers in an application-focused manner: if a few training points yield LOO errors above a predefined threshold, while all the other ones have much smaller errors, they are likely to be outliers and to harm the model if incorporated to it.
-
Training data pruning (downsampling): as explained in Section 4.1, for moderate-size HXS datasets, the number of training datapoints are a key for computational performance. Moreover, redundant training points can hinder the conditioning of kernel computations, for reasons exposed in Section 3.5. Therefore, it is often interesting to eliminate little-informative points from the training set. LOO-CV offers a convenient way to do so: the points with the smallest LOO errors are little useful to the model, as they can be very well predicted from the other ones. This way, one can incrementally downsample the training set, deleting at each iteration the best-approximated point. Preliminary experiments in this direction exhibited promising results: the size of the initial dataset could be divided by 2 with no significant loss of accuracy (see chapter 6.3 of [24]).
Naturally, the LOO-CV methodology can be applied to any machine learning model. But there is a specificity when using it with exact GPs: leave-one-out errors can be computed with closed-form formulas, without the need of fitting the model at each run. Such expressions can be found for instance in chapter 5.4.2 of [17] for SOGPs. We restate them here: if yi is the i-th real data value,
its leave-one-out estimate – that is, the value obtained when fitting the model to X ∖ xi and y ∖ yi and making a prediction on xi – and
the predictive variance of this estimate, then
and
are given by :
where [u]i denotes the i-th component of the vector u, and [M]ii the (i, i)-th coefficient of the matrix M. Naturally, these expressions don’t account for the (slight) modification of the optimal model parameters resulting from the removal of one data point from the training set.
Similar expressions are given in the PhD thesis from which this article is derived ([24]) for the PLMC and Lazy-LMC (which is essentially a subcase of the former). The latent processes of these models being conditionally independent, the LOO formulas of Equation (16) for SOGPs apply to them. It is then straightforward to recover the corresponding LOO estimates at the level of predicted tasks:
LOO-CV expressions are also given for the ICM in the chapter 5.4.2 of [24], but are omitted here because they would require to introduce the formalism of this model33. For the VLMC, no such expressions can be devised: indeed, the principle of this model is to replace some data-relative quantity of the GP posterior by a free-form distribution optimized in a black-box manner. The link between the training data and predictions, which underlies the above LOO expressions, is thus lost in the process; all information about the training data is stored implicitly in the model’s parameters, as in a neural network.
We wished to assess whether these LOO estimates are good proxies for errors measured on an adequate external test set; this turned out to be the case for the PLMC and Lazy-LMC, but not for the ICM – see Appendix C for a quantitative exposition of this statement. In the latter case, the problem seems to arise from numerical issues, in a context of very ill-conditioned kernel matrices (see Sect. 3.5): the variance terms featured in the denominator of LOO error expressions (see Eqs. (16) and (18)) can be very small, and with the ICM they are strongly affected by conditioning issues, for being obtained by eigenvalue computations. A better implementation of the ICM should thus be able to lift this issue quantifies how well the error metrics obtained with closed-form expressions match these computed on an external test set for our MOGP models.
5. Conclusions
In this article, we introduced the use of multi-output Gaussian processes (MOGP) to interpolate homogenized cross-sections (HXS) inside nuclear reactor neutronics codes. We compared several models of this family, discussed the main modeling choices which they offer and the best option is to retain for approximating HXS’s with them. It was shown on a complete and realistic test-case that the desirable properties of these tools – tunable, long-ranged and smooth predictor, sparse interpolation support, dimensionality reduction leveraging the redundancy of the data – make it possible to improve on interpolation accuracy compared with the multilinear standard, with only a fraction of its training data (20 times fewer tabulated points required) and of its storage requirement (HXS library sizes reduced by two orders of magnitude). We also proved these models to be scalable to datasets of millions of HXS, with training durations of the order of the minute, and interpolation speeds shown analytically to match this of the multilinear method. Some of their convenient features – efficient and interpretable storage format, closed-form expressions of leave-one-out errors authorizing easy model assessment and training dataset pruning, etc. – were presented; their respective additional functionalities and limitations were compared in Table 1.
This collection of arguments seems to make MOGPs viable candidates for seamless integration into current neutronics codes, alleviating the burden associated with large HXS libraries in today’s best-estimate calculations. The next step for assessing this methodology will be to implement it in an actual core code, and validate on a variety of benchmarks that it doesn’t hinder the accuracy of simulations. The models should also be tested on more diverse tests-cases: other reactor technologies, more HXS variables (moderator density, xenon concentration, etc.), accidental conditions… Incidentally, one issue not addressed in this work is this of discrete variables, such as the type of control rod inserted in the assembly: while kernels able to handle such inputs exist [47], modeling HXS data with them has not been tested yet34.
Apart from validation, the main developments to bring to this line of work are uncertainty quantification and active learning. Uncertainty quantification is a growing subject in neutronics, and Gaussian processes would be valuable additions to UQ pipelines: they make it possible to propagate uncertainties trivially35, non-linearly and explicitly, contrary to other machine learning models like neural networks for which uncertainty propagation is an ongoing field of research36 [48]. We showed in the present article that although sensitive to several factors, good-quality variance estimates could be obtained with MOGPs on HXS data; much more thorough investigation of their statistical properties should now be undertaken. Active learning, for its part, refers to the data-informed construction of an optimal training dataset; such methodologies could further reduce the computational burden of generating HXS libraries, and make it possible to parametrize HXS’s with many more variables for instance. First results were obtained in this direction in the PhD thesis associated with this work [24], but much more ambitious developments could be envisioned thanks to the large amount of literature dedicated to active learning with Gaussian processes [49].
Glossary
ANN: Artificial Neural Network
BPR: Bayesian Polynomial Regression
HXS: Homogenized Cross-Section
ICM: Intrinsic Coregionalization Model
LMC: Linear Model of Coregionalization
LOO-CV: Leave-One-Out Cross-Validation
MLI: Multi-Linear Interpolation
MOGP: Multi-Output Gaussian Process
PCA: Principal Components Analysis
PLMC: Projected Linear Model of Coregionalization
PVA: Predictive Variance Adequacy
R2: coefficient of determination
SOGP: Single-Output Gaussian Process
SVD: Singular Values Decomposition
UQ: Uncertainty Quantification
VLMC: Variational Linear Model of Coregionalization
Math Symbols
n: number of training datapoints
p: number of regression tasks (modeled HXS’s)
q: number or latent processes (see Sect. 2.2)
d: dimension of the inputs, i.e number of variables along which HXS’s are parametrized
H: mixing matrix of the LMC model (see Sect. 2.2), of size p × q
U: low-dimensional summary of the data seen by the latent processes (see paragraph 4.1), of size q × n
X: matrix containing all training data inputs, of size d × n
Y: matrix containing all training data outputs, of size p × n
y: size-n vector containing all training data outputs in the case of a single-output model
σ2: noise term of a GP, added to the diagonal of the kernel matrix
:
lower bound enforced by σ2 in some implementations
x*: input point at which a prediction is made
kX*: vector k(X, x*) (=[k(xi, x*)]1 ≤ i ≤ n)
Acknowledgments
APOLLO3® is registered trademarks of CEA.
Funding
O. Truffinet, K. Ammar and N. Gérard-Castaing thank EDF and Framatome for partial financial support.
Conflicts of interest
The authors have nothing to disclose.
Data availability statement
Data associated with this article cannot be disclosed for reasons of confidentiality.
Author contribution statement
Conceptualization and methodology, O. Truffinet, K. Ammar, B. Bouriquet and J.-P. Argaud; Software, O. Truffinet and N. Gérard Castaing.; Resources, O. Truffinet and K. Ammar; Investigation, Data Curation, Validation, Visualization and Writing – Original Draft Preparation, O. Truffinet; Supervision and Writing – Review & Editing, K. Ammar, B. Bouriquet and J.-P. Argaud.
References
- A. Weiser, E. Zarantonello, A note on piecewise linear and multilinear table interpolation in many dimensions, Math. Comput. 50, 189 (1988) [Google Scholar]
- S. Sánchez-Cervera et al., Optimization of multidimensional cross-section tables for few-group core calculations, Ann. Nucl. Energy 69, 226 (2014) [Google Scholar]
- V. Zimin, A. Semenov, Building neutron cross-section dependencies for few-group reactor calculations using stepwise regression, Ann. Nucl. Energy 32, 119 (2005) [Google Scholar]
- P. Bokov, Automated few-group cross-section parameterization based on quasi-regression, Ann. Nucl. Energy (2009) [Google Scholar]
- D. Botes, P. Bokov, Polynomial interpolation of few-group neutron cross sections on sparse grids, Ann. Nucl. Energy 64, 156 (2014) [Google Scholar]
- E.A. Szames, Amélioration du modèle reconstruction des sections efficaces dans un code de calcul de neutronique à l’échelle cœur, Ph.D. thesis, Université Paris-Saclay 2020 [Google Scholar]
- J.P. Berrut, L.N. Trefethen, Barycentric lagrange interpolation, SIAM Rev. 46, 501 (2004) [Google Scholar]
- K. Tan et al., The assembly homogeneous fewgroup constants generation method for PWR based on machine learning, Ann. Nucl. Energy 165, 108772 (2022) [Google Scholar]
- V.L. Nicolas Martin Zachary Prince, M. Tano-Retamales, Deep learning for multigroup cross-section representation in two-step core calculations, Nucl. Sci. Eng. 197, 1406 (2023) [Google Scholar]
- Y.M. Chan, J. Dufek, A deep-learning representation of multi-group cross sections in lattice calculations, Ann. Nucl. Energy 195, 110123 (2024) [Google Scholar]
- S. Dzianisau, D. Lee, Application of artificial neural network for assembly homogenized equivalence parameter generation, Prog. Nucl. Energy 173, 105285 (2024) [Google Scholar]
- D. Tomatis, A multivariate representation of compressed pin-by-pin cross sections, EPJ Nucl. Sci. Technol. 7, 8 (2021) [Google Scholar]
- M. Yamamoto, T. Endo, A. Yamamoto, Compression of cross-section data size for highresolution core analysis using dimensionality reduction technique, Nucl. Sci. Eng. 195, 33 (2021) [Google Scholar]
- B. Whewell, R.G. McClarren, Data reduction in deterministic neutron transport calculations using machine learning, Ann. Nucl. Energy 176, 109276 (2022) [Google Scholar]
- G. Hua, Y. Li, S. Wang, PWR pinhomogenized cross-sections analysis using big-data technology, Prog. Nucl. Energy 121, 103228 (2020) [Google Scholar]
- D. Tomatis, A. Dall’Osso, Compression of 3D pin-by-pin burnup data, Ann. Nucl. Energy 136, 107058 (2020) [Google Scholar]
- C.E. Rasmussen, C.K.I. Williams, Gaussian Processes for Machine Learning. Adaptive Computation and Machine Learning (MIT Press, Cambridge, Mass., 2006), p. 248 [Google Scholar]
- M.A. Alvarez, L. Rosasco, N.D. Lawrence, Kernels for vector-valued functions: A review, Found. Trends Mach. Learn. 4, 195 (2012) [Google Scholar]
- E.V. Bonilla, K. Chai, C. Williams, Multitask Gaussian process prediction, in Advances in Neural Information Processing Systems, edited by J. Platt et al. (Curran Associates, Inc., 2007), Vol. 20 [Google Scholar]
- M. van der Wilk et al., A framework for interdomain and multioutput Gaussian processes (2020). http://arxiv.org/abs/2003.01115 [Google Scholar]
- O. Truffinet et al., Exact and general decoupled solutions of the LMC Multitask Gaussian Process model (2024). http://arxiv.org/abs/2310.12032 [Google Scholar]
- T.W. Evans, P.B. Nair, Exploiting structure for fast kernel learning, in Proceedings of the 2018 SIAM International Conference on Data Mining (SDM) (2018), pp. 414–422 [Google Scholar]
- S. Van Vaerenbergh et al., Fixed-budget kernel recursive least-squares, in 2010 IEEE International Conference on Acoustics, Speech and Signal Processing (2010), pp. 1882–1885 [Google Scholar]
- O. Truffinet, Machine learning methods for cross-section reconstruction in full-core deterministic neutronics codes, Ph.D. thesis, Université Paris-Saclay, 2024 [Google Scholar]
- B. Rakitsch et al., It is all in the noise: Efficient multi-task Gaussian process inference with structured residuals, in Advances in Neural Information Processing Systems, edited by C. Burges et al. (Curran Associates, Inc., 2013), Vol. 26 [Google Scholar]
- J. Hensman, A. Matthews, Z. Ghahramani, Scalable variational Gaussian process classification, in Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, edited by G. Lebanon, S.V.N. Vishwanathan (PMLR, San Diego, California, USA, 2015), Vol. 38, pp. 351–360. [Google Scholar]
- H. Liu et al., Learning multitask Gaussian process over heterogeneous input domains, IEEE Trans. Syst. Man Cybern. Syst. 53, 6232 (2023) [Google Scholar]
- M. Jankowiak, G. Pleiss, J. Gardner, Parametric Gaussian process regressors, in International Conference on Machine Learning (PMLR, 2020), pp. 4702–4712 [Google Scholar]
- W. Bruinsma et al., Scalable exact inference in multi-output Gaussian processes, in ICML’20: Proceedings of the 37th International Conference on Machine Learning, edited by H. Daume, A. Singh (2020) [Google Scholar]
- M. Gu, W. Shen, Generalized probabilistic principal component analysis of correlated data, J. Mach. Learn. Res. 21, 428 (2020) [Google Scholar]
- G. Wahba, Spline models for observational data (1990). https://doi.org/10.1137/1.9781611970128 [Google Scholar]
- O. Truffinet et al., An EIM-based compressionextrapolation tool for efficient treatment of homogenized cross-section data, Ann. Nucl. Energy 185, 109705 (2023) [Google Scholar]
- N. Dige, U. Diwekar, Efficient sampling algorithm for large-scale optimization under uncertainty problems, Comput. Chem. Eng. 115, 431 (2018) [Google Scholar]
- Y. Li et al., PWR few-group constants parameterization analysis, Prog. Nucl. Energy 88, 104 (2016) [Google Scholar]
- H. Wendland, Piecewise polynomial, positive definite and compactly supported radial functions of minimal degree, Adv. Comput. Math. 4, 389 (1995) [CrossRef] [MathSciNet] [Google Scholar]
- A. Godfrey, Vera Core Physics – Benchmark Progression – Problem Specifications, Tech. Rep., Consortium for Advanced Simulation of LWRs (Oak Ridge National Laboratory, 2014) [Google Scholar]
- P. Mosca et al., APOLLO3®: Overview of the new code capabilities for reactor physics analysis, in M&C 2023 (2023) [Google Scholar]
- E. Szames et al., Few-group cross sections modeling by artificial neural networks, EPJ Web Conf. 247, 06029 (2021) [CrossRef] [EDP Sciences] [Google Scholar]
- D.P. Kingma, J. Ba, Adam: A method for stochastic optimization (2017). http://arxiv.org/abs/1412.6980 [Google Scholar]
- A. Paszke et al., PyTorch: An imperative style, high-performance deep learning library, in Advances in Neural Information Processing Systems (Curran Associates, Inc., 2019), Vol. 32 [Google Scholar]
- J.R. Gardner et al., GPyTorch: Blackbox matrix-matrix Gaussian process inference with GPU acceleration, in Advances in Neural Information Processing Systems (2018) [Google Scholar]
- N. Martin, M. Riedmann, J. Bigot, Latest developments in the ARTEMIS TM core simulator for BWR steady-state and transient methodologies, in MC2017 (2017) [Google Scholar]
- M.E. Tipping, Sparse bayesian learning and the relevance vector machine, J. Mach. Learn. Res. 1, 211 (2001) [Google Scholar]
- J.P. Boyd, The uselessness of the Fast Gauss Transform for summing Gaussian radial basis function series, J. Comput. Phys. 229, 1311 (2010) [Google Scholar]
- K. Cutajar et al., Preconditioning kernel matrices, in Proceedings of the 33rd International Conference on Machine Learning, edited by M.F. Balcan, K.Q. Weinberger (PMLR, New York, New York, USA, 2016), Vol. 48, pp. 2529–2538 [Google Scholar]
- P. Monasse, Extraction of the level lines of a bilinear image, Image Process. Line 9, 205 (2019) [Google Scholar]
- P. Saves et al., A mixed-categorical correlation kernel for Gaussian process, Neurocomputing 550, 126472 (2023) [Google Scholar]
- J. Gawlikowski et al., A survey of uncertainty in deep neural networks, Artif. Intell. Rev. 56, 1513 (2023) [Google Scholar]
- H. Liu, Y.-S. Ong, J. Cai, A survey of adaptive sampling for global metamodeling in support of simulation-based complex engineering design, Struct. Multidiscip. Optim. 57, 393 (2018) [Google Scholar]
Because of definition (1), these variables can be macroscopic and global (boron concentration in the moderator, assembly-level burnup…) even if the notion of cross-section is intuitively attached to an isolated nucleus.
In fact this term doesn’t describe a specific model, but rather the application of inducing points and variational approximations to the LMC framework. To our best knowledge no existing paper is dedicated to describing this integration, but clear descriptions can be found in the provided reference.
Indeed, it implicitly assumes that the latent processes are independent conditionally on the observations, which it not the case here. The Lazy-LMC can be framed as a PLMC model ([21]) which main assumption has been overruled rather than enforced.
Let us explain why on a simplified case, with two non-interacting variables: suppose we approximate a function f(x1, x2)=f1(x1)+f2(x2) on a cartesian grid, with n points on each axis. Then all points from a given x1-aligned row only provide information about one value of f2, as x2 is constant on this row. One could have taken instead the diagonal points of the grid as a training set, retrieving the exact same amount of information with n points instead of n2.
In the course of our experiments, we also trained models on other datasets that these used in this article, like Sobol’ sequences with different direction numbers, or Latin Hypercube Samples. The results were very consistent with these presented. In fact, chapter 6 of the Ph.D. thesis from which this article is derived ([24]) shows that the obtained accuracy is remarkably insensitive to the layout of the training points.
Except for the Lazy-LMC of Section 2.3, for which we explained the necessity of using the spline kernel of Equation (6).
Except for the spline kernel of Section 2.3, which requires its inputs to be in [0; 1].
An alternative implementation lifting this shortcoming is however suggested – but not tested – in the PhD thesis from which this article is derived ([24]).
This is also true for Section 3.3.
Contrary to all of our other metrics, the R2 coefficient relates the error measured on each regression task to the variance of this task, so it is the only one to detect if the model is “lazy” on low-variance ones (it makes low prediction errors, but which are in fact significant compared to this task’s variability).
For higher dimensions, they are rather obtained in a recursive manner, as in [1].
One way around this issue would be to mutualize kernel operations between all SOGPs of the collection, either by enforcing them to share the same kernel parameters, or by resorting to a parameter-less kernel like this of Equation (6). The prediction cost would then go down to 2np + nck.
A complete counting of stored quantities is given in the appendices of [24].
Note that all discussed LOO-CV formulas were verified against LOO errors computed “by hand” – by explicit formation of the n data folds, and fitting (but not re-training) of the model to these folds – on several datasets, both real and synthetic. We found a perfect match between the values obtained by closed-form expressions and the empirical ones.
In particular, our LOO-CV expressions for this model, given in Section 5.5 of [24], were validated by comparison with empirical LOO values on several datasets; perfect matches were observed in each case.
Appendix A
Experimental results on the finely-discretized dataset
Here we present our experimental results on our large dataset (12 assemblies containing 3.5 ⋅ 105 HXS’s each, see Sect. 3.1); we left them aside in Section 3.4, as they are very similar to these of Table 5 but require a few additional explanations. These results are summarized in Table A.1. Here are some points to specify:
-
The ICM is absent from this table, because as mentioned in Section 3.3, current implementations use a full-rank correlation matrix between all p tasks [25], which incurs computational costs in O(p3), intractable in this case. This could be remedied in the future by using a rank-q correlation matrix instead, Kf = HHT which would yield the same correlation structure as with our other LMC models.
-
The MLI baseline is also absent, as we wanted to avoid the generation, storage and manipulation of such a dataset on full Cartesian Grids: it would have weighted about 200 gigabytes.
-
The “best SOGP” baseline is replaced by “Training-less SOGPs”: this is because training millions of SOGP models on a standard machine, at the rythm of a few seconds per model, would be infeasible. These training-less SOGPs are the same model that wraps the latent processes of our Lazy-LMC: their kernel is the already-introduced, parameter-less spline kernel, and their few additional parameters have been neutralized (see Tab. 2).
-
The metrics assessing the calibration of predictive variance are missing for the PLMC. This is because predictive variance could not be computed in this setting with our current implementation of this model: it requires the formation of a full p × p noise matrix, which causes a memory error for large p. This is only a matter of implementation: only the diagonal of this matrix is required for variance computation, so these coefficients could very well be computed alone without overloading memory.
-
It may be surprising that the R2 coefficients of the PLMC and VLMC are respectively subpar and catastrophic, while their other error metrics are excellent. This relates to an observation already made in Section 3.5: if the training data of each regression task is not normalized to have variance ≃1, our trainable MOGP models tend to neglect the low-variance tasks. In other terms, they are “lazy” with these tasks: they achieve small errors on them, which is why all other metrics than the R2 remain flawless, but these errors are in fact significant compared to the small variability of the modeled function. Once again, this advocates for the use of standard deviation-based normalization of each HXS before model training. More recent experiments confirmed that this choice indeed restores all R2 coefficients to excellent values.
Performance metrics on all assemblies of the finely-discretized dataset, for three of the studied MOGP models (all with q = 16 latent processes) and two baselines. The best and worst value for each metric are highlighted in green and red respectively. Definitions of the metrics are given in Section 3.2. The training time Ttrain is defined at the level of one assembly. PLMC and VLMC were trained on GPU, with minibatch training in the latter case. The α-CI and PVA metrics are missing for the PLMC are missing because predicted variance was not computed with this model: our current implementation requires the formation of a full p × p noise matrix for this, which causes a memory error in this setting.
These remarks being made, it can be seen that all numerical values are very close to these of Table 5: our models perform just as well in the many-HXS setting as in the standard one. Apart from this, the main observation is that the GPU training of our models – in particular the PLMC – is very fast: two minutes are enough to fit a PLMC to hundreds of thousands of HXS’s, making it fully ready for the big-data setting. Other minor remarks can be made: BPR performs less well on this dataset, which strengthens our intuition that GPs are more reliable than polynomial methods for modeling diverse HXS datasets; and both training-less models (Lazy-LMC and training-less SOGPs) still perform remarkably well on this test-case, although the gap between the Lazy-LMC and VLMC is larger than on the smaller dataset, as it was already observed in Figure 3b.
Appendix B
Optimal MOGP model size for neutronics calculations
The experiments of Appendix A showed that training a single model for very many HXS’s was feasible. However, it is also somewhat troublesome: it can be slow, even with GPU mini-batch training (more than 2 hours per assembly with the VLMC); the memory footprint can exceed the capacity of a single GPU; training is less stable, as illustrated in Figure 3b, etc. One should therefore wonder whether this methodology is really appropriate, or whether fitting smaller models to subsets of the data should be preferred. Eventual gains of the full-model option would be storage savings – by mutualizing latent processes across more outputs, the storage of the data defining them can be spared – and computational savings (likewise, the interpolation of latent processes would be mutualized). But in large-data and low-q settings, these gains can be seen to be negligible.
Consider for instance our finely-discretized dataset, which neutronics data is defined at the level of the fuel pin; assume that we split each assembly-wise dataset between its nc fuel cells, and define one PLMC or Lazy-LMC model per cell, keeping the same values of q and n for each. Each of these models now has p′=p/nc tasks. According to Section 4.1, the cost of making predictions with each one is 2p′q + n(ck + 2q), so the total cost for all of them is nc × (2p′q + n(ck + 2q)) = 2pq + nc n(ck + 2q). By contrast, the cost of predictions with a single assembly-wise model is 2pq + n(ck + 2q). The question is thus the relative size of 2pq and nc n(ck + 2q).
In the setting of our previous experiments, we have p = 3.5 ⋅ 105, q = 16, nc = 45 (one assembly with 17 × 17 cells and 1/8 symmetry), n = 201 and ck = 30. Therefore, 2pq ≃ 1.12 ⋅ 107, while nc n(ck + 2q)≃5.6 ⋅ 105: the lifting of the low-dimensional interpolated values to the “real” HXS space remains 20 times more costly than the interpolation of the latent processes, even if there are now 45 times more latent processes to interpolate (because the model was split into 45 smaller ones). This lifting operation doesn’t depend on the allocation of HXS’s into one or several models, only on the retained value of q. The exact same argument can be given about storage requirements: they remain dominated by the storage of the p × q matrix H, next to which the few parameters of each latent process and the training caches C of Section 4.1 are negligible.
Therefore, in the many-HXS setting, typically if pin-by-pin homogenization is performed, there is no clear benefit to gathering all HXS’s from a dataset into a single model: pin-wise models can be used instead without noticeable loss in computation speed or storage space. The question to be asked before deciding on a grouping of HXS’s is the relative size of the mixing matrix H and of the quantities defining the latent processes: as long as the former dominates, there is no harm in splitting a large model into smaller one37. Incidentally, there can be practical issues when gathering some categories of HXS’s into a single model. For instance, if pin-by-pin burnup values are used and HXS from all pins are modeled by the same MOGP, the tasks of this model will be trained with differing training inputs X, as HXS’s from different pins will be exposed to different burnups. This at least requires dedicated implementations, and may not be readily available (or available at all) for all models, as stated in Table 1.
Appendix C
Adequacy of leave-one-out cross-validation for model assessment
In Section 4.3, we gave closed-form expressions for the LOO-CV errors of our models, and indicated their possible uses in neutronics calculation chains. One of them is to assess the accuracy of trained models without resorting to an external test set. However, we would like to ensure that this methodology is trustworthy, that is, that the error metrics we obtain with LOO-CV are consistent with these obtained on test sets. This is the case if two conditions are met:
-
LOO-based error metrics are well-correlated with these measured on the test set, across various model instances. Indeed, it would be problematic if they designated a specific model declination (some combination of modeling choices) as “good”, but the predictions of this model turned out to be unsatisfactory.
-
LOO-based error metrics are well-calibrated with the ones measured on the test set. By this, we mean that the values of LOO-based metrics should always be close to these of their test-based counterparts: one must be able to expect some level of accuracy when applying the model to unseen data.
The only manner in which we can assess these criteria experimentally is by computing LOO estimates to test-set errors for various model declinations (models spanning the full range of the modeling choices discussed in this article: mean function, kernel function, number of latent processes, etc.). This was done in all grid searches of Section 2.5, and in the experiments of Section 3.3 on the number of latent processes. We thus dispose of a set of experiments with various model declinations, each yielding a series of error metrics ℳij computed on the test dataset – where i denotes the metric (RMSE, Errmean, PVA, etc.), and j the tested model – and their leave-one-out counterparts
computed on the training dataset. These metrics are RMSE, Errmean, Errmax and q99; q99 is the quantile of L1 errors at level 99%, added here for additional information about large errors. From this we compute the ratios
, and the correlations
. For each metric ℳi, we summarize its computed ratios rij (of which there are as many as models tested) by a few of their statistics: minimum, maximum, mean and standard deviation. Results of this analysis are displayed below in Tables C.1 for the PLMC and C.2 for the ICM; each one corresponds to the grid search previously performed on these models, which feature respectively 561 and 300 tested model declinations, ensuring the meaningfulness of the computed statistics. As stated above, these statistics were also computed on the experiments of Section 3.3 studying the variations of MOGP performance with q: results are fully consistent with these obtained on grid searches, but less relevant as the number of items is small (for each model, 18 declinations differing only by their value of q). We still display them for the Lazy-LMC in Table C.3, by lack of an alternative: the only modeling choice left in this model is indeed the number of latent processes.
Analysis of the agreement between test-based and LOO-based accuracy metrics for the PLMC. For any metric ℳi, ℳij denotes its value for the j-th model declination,
its LOO counterpart, and
.
Analysis of the agreement between test-based and LOO-based accuracy metrics for the ICM. For any metric ℳi, ℳij denotes its value for the j-th model declination,
its LOO counterpart, and
. A few models have undefined PVA values because of some near-zero variance estimate, which are not taken into account in the statistics.
Analysis of the agreement between test-based and LOO-based accuracy metrics for the Lazy-LMC (not measured on a grid search, but on the experiment of Fig. 3a). For any metric ℳi, ℳij denotes its value for the j-th model declination,
its LOO counterpart, and
.
Results are clear: LOO estimates are very good proxies of test set errors for PLMC and Lazy-LMC, but much less for ICM. For the first two, the ratios rij are close to 1 on average, and contained in the range [0.36; 2.48] meaning that any LOO metric is at worse 3 times smaller or larger than its test-data counterpart; moreover, the correlations between LOO and non-LOO metrics are always very close to 1. The situation of the ICM is much more troublesome: correlations between metrics and their LOO counterparts are small (≃20%), and ratios between these quantities can take values smaller than 1% – although their means are still rightfully close to 1. Taking a closer look at the experimental values, it appears that this discrepancy is due to a few very ill-performing model instances, which LOO estimates are nonetheless satisfactory. As in Section 3.5, this suggests some subpar behavior of linear algebra operations in our implementation of this model: while it usually yields consistent and satisfying results38, it sometimes breaks down on edge cases – computation of small variance terms in contexts of ill-conditioned, poorly-performing model instances. Recall indeed that our LOO expressions (Eqs. (16) and (18)) involve variance terms, which can be very small for little-noisy HXS’s and particularly sensitive to numerical issues (see Sect. 3.5).
Lastly, let us mention that the same tests as in Tables C.1–C.3 were performed on the metrics assessing the quality of variance assessment, α-CI and PVA. The results, given in the Section 5.5 of [24], are catastrophic: correlations are tiny, and ratios are erratic. This is another indication of the malfunction of MOGP predictive variance in our application when no specific attention is paid to it. The same experiments should now be undertaken after fixing predictive variances with the solutions suggested in Section 3.5, in order to definitely validate the reliability of these LOO-CV estimates.
Cite this article as: Olivier Truffinet, Karim Ammar, Jean-Philippe Argaud, Nicolas Gérard Castaing, Bertrand Bouriquet. Multi-output Gaussian processes for the reconstruction of homogenized cross-sections, EPJ Nuclear Sci. Technol. 11, 61 (2025). https://doi.org/10.1051/epjn/2025055
All Tables
Description of the Lazy-LMC model. For each parameter featured in the definition of the LMC (Eq. (5)), a procedure is given to set its value to an adequate one. See Section 2.3 for more information.
Description of all modeling choices tested with a grid search. See Section 2.4 for explanations of the notation ψ(r) for stationary kernels, the lengthscale parameters λj, and the notations of variable decomposition. The piecewise-polynomial kernel of order 2, taken from [35], is defined as such :
, where
and ( ⋅ )+ = max(⋅, 0) denotes the positive part.
Performance metrics on all assemblies of the standard-discretized dataset, for the studied MOGP models (all with q = 16 latent processes) and three baselines. The best and worst value for each metric are highlighted in green and red respectively. Definitions of the metrics are given in Section 3.2; they are all averaged across all HXS’s. The ErrΣ metrics were not computed for the polynomial method, which was tested at a time when concentration data was unavailable. The training time Ttrain is defined at the level of one assembly. n is the number of datapoints used for training.
Summary of the process of fixing predictive variance for VLMC and PLMC. Each row denotes a modification performed on them and the corresponding results. These modifications are incremental: each one is kept in the following rows.
Performance metrics on all assemblies of the finely-discretized dataset, for three of the studied MOGP models (all with q = 16 latent processes) and two baselines. The best and worst value for each metric are highlighted in green and red respectively. Definitions of the metrics are given in Section 3.2. The training time Ttrain is defined at the level of one assembly. PLMC and VLMC were trained on GPU, with minibatch training in the latter case. The α-CI and PVA metrics are missing for the PLMC are missing because predicted variance was not computed with this model: our current implementation requires the formation of a full p × p noise matrix for this, which causes a memory error in this setting.
Analysis of the agreement between test-based and LOO-based accuracy metrics for the PLMC. For any metric ℳi, ℳij denotes its value for the j-th model declination,
its LOO counterpart, and
.
Analysis of the agreement between test-based and LOO-based accuracy metrics for the ICM. For any metric ℳi, ℳij denotes its value for the j-th model declination,
its LOO counterpart, and
. A few models have undefined PVA values because of some near-zero variance estimate, which are not taken into account in the statistics.
Analysis of the agreement between test-based and LOO-based accuracy metrics for the Lazy-LMC (not measured on a grid search, but on the experiment of Fig. 3a). For any metric ℳi, ℳij denotes its value for the j-th model declination,
its LOO counterpart, and
.
All Figures
![]() |
Fig. 1. Typical aspect of the 2-group thermal absorption XHS of Xenon 135 (homogenized on a fuel cell). The wireframe mesh corresponds to a realistic multivariate grid for multilinear interpolation. (a) Projection on the Bu (burnup) and Tf (fuel temperature) variables. (b) Projection on the Tm (moderator temperature) and Cb (boron concentration) variables. |
| In the text | |
![]() |
Fig. 2. Schematized view of the three sets of input points used in the experiments, projected in the plane (Tf, Cb). Color fading represents the euclidian distance to the projection plane. |
| In the text | |
![]() |
Fig. 3. Variation with the number of latent processes q of the prediction error (RMSE measured on one assembly of the test dataset) for the studied MOGP models. (a) Standard-discretized dataset. (b) Finely-discretized dataset. |
| In the text | |
![]() |
Fig. 4. Compression loss 1 − EVR(r) of the training data (normalized HXS’s of one assembly) in function of the truncation rank r of the SVD, where EVR(r) is its cumulative explained variance ratio. The erratic behavior of the left part of subfigure b/ is due to the compression loss reaching single-float precision (while figure a/ was computed in double-float precision). (a) Standard-discretized dataset. (b) Finely-discretized dataset. |
| In the text | |
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.






![$$ \begin{aligned} y_{j}(\mathbf x ) = m_{j}(\mathbf x ) + \sum _{i=1}^{q} H_{j,i} \, u_{i}(\mathbf x ) + \epsilon _{j}(\mathbf x ) \quad \forall j \in [\![1 ; p]\!] \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq5.gif)






![$$ \begin{aligned} \hat{y}(\mathbf x _{*})&= \mathbf k_{*} ^{T} \mathbf K ^{-1} \mathbf y = \sum _{i=1}^{n} \left[ \mathbf k_{*} ^{T} \mathbf K ^{-1} \right]_{i} \, y_{i} = \sum _{i=1}^{n} \alpha _{i}^{y}(\mathbf x _{*}) \, y_{i} \\&= \sum _{i=1}^{n} \left[ \mathbf K ^{-1} \mathbf y \right]_{i} \, k(\mathbf x _{*}, \mathbf x_{i} ) = \sum _{i=1}^{n} C_{i} \, k(\mathbf x _{*}, \mathbf x_{i} ) \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq36.gif)

![$$ \begin{aligned} \hat{u}_{j}(\mathbf x _{*})&= \mathbf k_{j*} ^{T} \mathbf K _{j}^{-1} \mathbf U_{j} = \sum _{i=1}^{n} C_{ij} \, k_{j}(\mathbf x _{*}, \mathbf x_{i} ) \quad \forall j \in [1;q]\\ \hat{y}_{l}(\mathbf x _{*})&= \sum _{j=1}^{q} H_{lj} \, \hat{u}_{j}(\mathbf x _{*}) \quad \forall l \in [1;p], \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq40.gif)


![$$ \begin{aligned} y_{i} - y_{i}^\mathrm{LOO}&= \left[(\mathbf K + \sigma ^{2} \mathbf I_{n} )^{-1}\mathbf y \right]_{i}/v_{i}^\mathrm{LOO} \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq47.gif)
![$$ \begin{aligned} v_{i}^\mathrm{LOO}&= \left[(\mathbf K + \sigma ^{2} \mathbf I_{n} )^{-1}\right]_{ii} \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq48.gif)
![$$ \begin{aligned}[\mathbf U ]_{ij} - u_{ij}^\mathrm{LOO}&= \left[(\mathbf K_{j} + \sigma _{j}^{2} \mathbf I_{n} )^{-1}\mathbf U_{j}^{T} \right]_{i} / \vartheta _{ij}^\mathrm{LOO} \quad \forall j \in [1;q] \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq49.gif)
![$$ \begin{aligned} \vartheta _{ij}^\mathrm{LOO}&= \left[(\mathbf K_{j} + \sigma _{j}^{2} \mathbf I_{n} )^{-1}\right]_{ii} \quad \forall j \in [1;q] \end{aligned} $$](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq50.gif)


![$ =[k(\mathbf{x_{i}, x_{k}})]_{1\leq i \leq n}^{1\leq j \leq n} $](/articles/epjn/full_html/2025/01/epjn20250040/epjn20250040-eq54.gif)
