Istituto per le Applicazioni del Calcolo ”M. Picone”,
Consiglio Nazionale delle Ricerche,
Via dei Taurini, 19 - 00185 Roma - Italy
Regularization of Kriging interpolation on irregularly spaced data.
Abstract
Interpolation models are critical for a wide range of applications, from numerical optimization to artificial intelligence. The reliability of the provided interpolated value is of utmost importance, and it is crucial to avoid the insurgence of spurious noise. Noise sources can be prevented using proper countermeasures when the training set is designed, but the data sparsity is inevitable in some cases. A typical example is represented by the application of an optimization algorithm: the area where the minimum or maximum of the objective function is assumed to be present is where new data is abundantly added, but other areas of design variable space are significantly neglected. In these cases, a regularization of the interpolation model becomes absolutely crucial. In this paper we are presenting an approach for the regularization of an interpolator based on the control of its kernel function via the condition number of the self-correlation matrix.
1 Introduction
In these years, the attention to the use of the so-called digital twins has gained a great popularity, particularly in the field of artificial intelligence (AI). The possibility to apply numerical methods replacing a complex and expensive real system, testing a number of different possibilities and configurations before producing the real object, is a great help for the design team. In addition, the comparison of the observed performances of a system and its ideal values, obtained as the output of the digital twin, can be useful for scheduling maintenance activities, thus reducing management costs.
Generally speaking, in most of the cases ”digital twin” is a different way to call an interpolation model, since the response of the digital twin is commonly based on the manipulation of a number of observed values of the different quantities characterizing the system status: data interpolation represents therefore the core element. If we need high quality in the data interpolation, and of the digital twin consequently, a great attention must be dedicated to the design of the training set, uniformly distributing the sample points. Unfortunately, if the building of the interpolator is not static, and new points can be added dynamically, the resulting training set could lost its uniformity, and also the quality of the interpolation could become poor.
An example is represented by the use of the digital twin inside an optimization loop: classically, the optimization algorithm is using the digital twin instead of the true objective function (if computationally expensive) for the detection of the optimal solution. Once the current optimal solution is detected, the candidate solution is verified and a new point can be added to the training set. If this operation is repeated, it is common that all the new points will belong to the same area in the design variable space, creating an imbalance in the point distribution.
The digital twin can be also based on real (non simulated) data, collected when the system is operational, or can be obtained from measurements of a natural phenomenon. Irregularly-sampled time series occur in many domains including healthcare, geophysics and many more. They can be challenging to model because they do not naturally yield a fixed-dimensional representation as required by many standard machine learning model. Adding new points in these cases could enhance the difficulties, because new data are not completely under the control of the researchers, so they are often not improving t database under the standpoint of its uniformity and regularity.
To mitigate the effect of the lack of uniformity of the samples, a localization of the interpolation may be performed: the digital twin can divide the entire database into different parts, accessing the most appropriate one depending on the required configuration, but some irregularities at the borders of the splitting areas could appear. Viceversa, if we prefer to preserve the integrity of the database, producing a single interpolator for the complete set of possible system configurations, a kind of regularization technique is mandatory.
Some solutions have been already presented in literature, and their number is quite large. The proposed approaches can be categorized in two great families: the so called dynamic kriging, presented and reviewed in [Zhao et al.(2011)], and more generic regularization methods (an example is presented in [Zhang et al.(2019)]).
The dynamic kriging tunes the optimal values of the kernel parameters by optimizing a number sub-problems resulting from the cross-validation technique: it has been already preliminarily tested in [Peri(2015)]. The procedure could be extremely time consuming, since the number of subproblem to be solved could be very large, depending on the number of available sampling points, and the final parameters are obtained solving a suite of sub-problems all different than the final one. Anyway, this technique has been demonstrated to be pretty efficient.
In the regularization methods, the kernel function is modified to include the regularization terms: in this case, the solved problem is the same as the final one, but the use of regularization terms can transform kriging from interpolation to approximation, losing precision locally.
In this article we present a regularization method suitable for a generic kernel method, or in any case for all methods that require the inversion of a square matrix for the determination of the key elements of the interpolator (typically, the weights of some kernel functions). There is not a modification of the kernel function e.g. by using some penalties term, since we are seeking for the optimal value of the parameters ruling the kernel functions. The full training set is used, without any further manipulation or reduction, and all the available information are exploited. Results are presented for the kriging interpolation method, that is responding to the aforementioned requirements. Numerical evidences are absolutely satisfactory under the perspective of preserving a regular behavior of the interpolation, increasing the overall credibility of the produced output.
2 Basics about kriging
Since we are producing results using the kriging method, we are going here to recall some basic elements. Among the different techniques for data interpolation, kriging represents a very attractive option. It was firstly introduced in [Matheron(1963)], and since then a large number of studies and variants have been proposed. A rather complete review paper is [Kleijnen(2009)]. Following the notation reported in [Kleijnen(2009)], the estimated value can be obtained as
(1) |
where d indicates the position of the point where the estimate is required, D is the position of the training points and w(D) is the value of the function to be interpolated at the sample points.
To select the optimal values for the weights , a least-squares problem is solved minimizing the Sum of Squared Residuals, that is, minimizing the Mean Squared Error (MSE) of the predictor y(d)
where d may be any point in the design variable space. Moreover, this minimization must account for the condition that the predictor is unbiased:
where in deterministic simulation E(w(d)) may be replaced by w(d). It can be proven that the solution of this constrained minimization problem implies that the sum of the weights is 1. Finally, it can be also proven that the optimal value of the weights is
(2) |
where is the covariance matrix involving the training set data, and is the covariance vector including the training set data and the interpolation location. See [Kleijnen(2009)] for details.
What is still missing is the definition/selection of the covariance function. In the original formulation, the concept of semi-variogram is adopted instead, and it can be estimated experimentally using the training set[Matheron(1963)]. Unfortunately, the semi-variogram represents a statistical quantity, and the number of training points is rarely adequate in real applications for producing a significative output, so that some analytical function are used instead. Practically, an algebraic expression, typically an exponential law, is used to fit the (few points of the) sample points, obtaining a kind of experimental semi-variogram: In fact, the data do not clearly indicate a trend, and even less they are able to indicate any continuous curve to apply: consequently, for any algebraic model adopted, the difference between the experimental values and the semi-Approximate variogram is usually large. A commonly adopted function is
(3) |
denotes the importance of input : the higher is, the less effect input has. denotes the smoothness of the correlation function: = 2 implies an infinitely differentiable function. As a consequence, the correlation function is a function of these two parameters, to be selected.
The tuning of the free parameters, if the previous formulation for the covariance function is adopted, can be performed splitting the available samples in two different sets: training and validation, using the first one to derive and the second one to tune . An example is reported in [Peri(2015)]. But this technique is difficult to apply in practice, since the number of sampled points is generally not large, and the use of some of these points as validation points is further impoverish the training set. By the way, the exactness of the solution is guaranteed by kriging on the training point, so that the shift of a training point in the validation set is not intrinsically beneficial.
In the following, we are going to fix = 2, so that the only free parameters are .
3 Kriging regularization - numerical experiments
The basic idea behind the present scheme for the regularization of kriging is simple. The determination of the weights is obtained by solving a linear system: if we find a way to maximize the condition number of the matrix of the linear system, the solution could be much more stable, and possibly much more regular. To verify this hypothesis, using the aforementioned definitions, an optimization problem has been solved, minimizing the condition number of the self-correlation matrix acting on the parameters . The starting point for the optimization problem is obtained by investigating a number of alternative solutions for , producing a perturbation around an initial value. Starting from the so determined best solution (best condition number), a local minimization algorithm of the class of the Direct Search Methods [Powell(2006)] is applied for the fine tuning of the parameters .
Six different 2D test functions have been adopted. 2D functions have been selected in order to provide a clearer evidence of the obtained results: the method is obviously valid for an arbitrary space dimension. The selected six functions are:
-
1.
Griewank function
where -
2.
Sasena function
where -
3.
Franke’s function
where -
4.
G-function
where -
5.
Irregular TF
where -
6.
Cosin2
where
A first comment comes from the observation of figure 1, where four different examples, with different number of training points, are compared. The improvement of the condition number of the self-correlation matrix, defined as the ratio between the current value and the initial value, is reported as a function of the non-dimensional value of the iterations of the optimization problem (the ratio between the current number of iteration and the maximum value). It is clear how, for a moderate number of training points, the improvement of the condition number is not large, while it becomes high in case of a very small or very large number of training points. The reasons for this improvement are exactly opposite: while for 16 training points the absolute value of the condition number is already high, a great improvement achieved in relative terms translates into a small change of the model in practice. On the opposite, the condition number can be largely improved in absolute value when the training set is large and unevenly distributed, and we have both a great improvement in absolute and relative terms. Figure 1 is demonstrating that very good results that have been obtained in terms of condition number.

Due to the nature of the kriging method, the matrix is depending only from the correlation function, and not from the shape of the objective function. As a consequence, if the point distribution, in a normalized design variable space, is the same, also the results for different objective functions are exactly the same. For this reason, the figure 1 is here reported for the case of the Franke’s function, but in reality it is valid also for any other objective function. The results for all the algebraic functions listed above are shown below, but the use of one function instead of another is completely equivalent, unless the training set distribution is not changed.
Results about the regularization of the interpolating function are reported in figure 2: only the graphical results for last two training sets are reported, since for smaller values of the training points differences are not visible. On top of figure 2, the case with 64 training points is reported, and there is not a clear difference between the case with and without regularization. On the contrary, for the case with 121 training points, reported in the bottom part of figure 2, the difference is clear. To highlight irregularities, contour levels are provided at the bottom of the graph. In the case without regularization, the presence of high-frequency noise is clear, while for the regularized interpolation we have a global behavior comparable with the case with 64 training points, but with a sharper distinction between the basins of attraction. Summarizing, we can affirm that the regularization procedure produces beneficial effects.


The same behavior is observed, as anticipated earlier, for all the six investigated algebraic function reported in figure 3. The noise is clearly eliminated by the regularization procedure, improving the overall quality and consistency of the interpolation: this feature is of paramount importance for the use of kriging in combination with local optimization algorithms.






Figure 4 provides a local quantification of the differences between the interpolator and the true value of the function for the Griewank and Sasena functions. Two examples are given, even if not necessary, simply to allow to notice the great similarities: minor differences are linked only to the different absolute value of the functions, and the location of the areas where the discrepancies are greatest is absolutely the same. The dots at the base of the graph are reporting the position of the training points. We can see how regularization reduces the differences to almost zero in the largest part of the domain. On the contrary, in a corner, where training is poor, regularization produces a local increase of the differences: it is quite obvious that regularization cannot improve the content of information, but only reshape interpolation more regularly, so that regions where the point density is low are subject to these consequences.


4 Conclusions
A regularization method for a generic kernel interpolator, based on the observation of the condition number of the kernel matrix, has been here described and successfully tested. Implementation and results has been produced for the kriging interpolator only, but we can use this algorithm also for different kernel methods: this is desirable for a larger comparison, in the future.
Also a direct and systematic comparison with other regularization methods, e.g. the ones based on the split of the sampling points in two different sets, training and validation, as reported in [Zhang et al.(2019)] and [Peri(2015)], could be also desirable.
Acknowledgements
D.P. research was partly funded by Italian Minister of Instruction, University and Research (MIUR) to support this research with funds coming from PRIN Project 2017 (No. 2017KKJP4X entitled ”Innovative numerical methods for evolutionary partial differential equations and applications”).
Conflict of interest
The author declares no conflicts of interest.
Data avaliability
The data that support the findings of this study are available on request from the author.
References
- [Kleijnen(2009)] Kleijnen, J. P., 2009: Kriging metamodeling in simulation: A review. European Journal of Operational Research, 192 (3), 707–716.
- [Matheron(1963)] Matheron, G., 1963: Principles of geostatistics. Economic Geology, 58 (8), 1246–1266.
- [Peri(2015)] Peri, D., 2015: Improving predictive quality of kriging metamodel by variogram adaptation. MARINE VI : proceedings of the VI International Conference on Computational Methods in Marine Engineering, CIMNE, Ed., Vol. 1.
- [Powell(2006)] Powell, M. J. D., 2006: The NEWUOA software for unconstrained optimization without derivatives, 255–297. Springer US, Boston, MA.
- [Zhang et al.(2019)] Zhang, Y., W. Yai, and X. Chen, 2019: A regularization method for constructing trend function in kriging model. Structural and Multidisciplinary Optimization, 59, 1221–1239.
- [Zhao et al.(2011)] Zhao, L., K. K. Choi, and I. Lee, 2011: Metamodeling method using dynamic kriging for design optimization. AIAA Journal, 49, 2034–2046.