Global sensitivity analysis and Wasserstein spaces
Abstract
Sensitivity indices are commonly used to quantity the relative influence of any specific group of input variables on the output of a computer code. In this paper, we focus both on computer codes the output of which is a cumulative distribution function and on stochastic computer codes. We propose a way to perform a global sensitivity analysis for these kinds of computer codes. In the first setting, we define two indices: the first one is based on Wasserstein Fréchet means while the second one is based on the Hoeffding decomposition of the indicators of Wasserstein balls. Further, when dealing with the stochastic computer codes, we define an “ideal version” of the stochastic computer code thats fits into the frame of the first setting. Finally, we deduce a procedure to realize a second level global sensitivity analysis, namely when one is interested in the sensitivity related to the input distributions rather than in the sensitivity related to the inputs themselves. Several numerical studies are proposed as illustrations in the different settings.
Keywords: Global sensitivity indices, functional computer codes, stochastic computer codes, second level uncertainty, Fréchet means, Wasserstein spaces.
AMS subject classification 62G05, 62G20, 62G30, 65C60, 62E17.
1 Introduction
The use of complex computer models for the analysis of applications from sciences, engineering and other fields is by now routine. For instance, in the area of marine submersion, complex computer codes have been developed to simulate submersion events (see, e.g., [3, 34] for more details). In the context of aircraft design, sensitivity analysis and metamodelling are intensively used to optimize the design of an airplane (see, e.g., [51]). Several other concrete examples of stochastic computer codes can be found in [42].
Often, the models are expensive to run in terms of computational time. Thus it is crucial to understand the global influence of one or several inputs on the output of the system under study with a moderate number of runs afforded [53]. When these inputs are regarded as random elements, this problem is generally called (global) sensitivity analysis. We refer to [16, 52, 56] for an overview of the practical aspects of global sensitivity analysis.
A classical tool to perform global sensitivity analysis consists in computing the Sobol indices. These indices were first introduced in [50] and then considered by [55]. They are well tailored when the output space is . The Sobol indices compare, using the Hoeffding decomposition [33], the conditional variance of the output knowing some of the input variables to the total variance of the output. Many different estimation procedures of the Sobol indices have been proposed and studied in the literature. Some are based on Monte-Carlo or quasi Monte-Carlo design of experiments (see [38, 47] and references therein for more details). More recently a method based on nested Monte-Carlo [28] has been developed. In particular, an efficient estimation of the Sobol indices can be performed through the so-called Pick-Freeze method. For the description of this method and its theoretical study (consistency, central limit theorem, concentration inequalities and Berry- Esseen bounds), we refer to [36, 25] and references therein. Some other estimation procedures are based on different designs of experiments using for example polynomial chaos expansions (see [57] and the reference therein for more details).
Since Sobol indices are variance based, they only quantify the influence of the inputs on the mean behaviour of the code. Many authors proposed other criteria to compare the conditional distribution of the output knowing some of the inputs to the distribution of the output. In [47, 49, 48], the authors use higher moments to define new indices while, in [6, 7, 15], the use of divergences or distances between measures allows to define new indices. In [20], the authors use contrast functions to build indices that are goal oriented. Although these works define nice theoretical indices, the existence of a relevant statistical estimation procedure is still in most cases an open question. The case of vectorial-valued computer codes is considered in [26] where a sensitivity index based on the whole distribution utilizing the Cramér-von-Mises distance is defined. Within this framework, the authors show that the Pick-Freeze estimation procedure provides an asymptotically Gaussian estimator of the index. The definition of the Cramér-von-Mises indices has been extended to computer codes valued in general metric spaces in [21, 27].
Nowadays, the computer code output is often no longer a real-valued multidimensional variable but rather a function computed at various locations. In that sense, it can be considered as a functional output. Some other times, the computer code is stochastic in the sense that the same inputs can lead to different outputs. When the output of the computer code is a function (for instance, a cumulative distribution function) or when the computer code is stochastic, Sobol indices are no longer well tailored. It is then crucial to define indices adapted to the functional or random aspect of the output. When the output is vectorial or valued in an Hilbert space some generalizations of Sobol indices are available [39, 24]. Nevertheless, these indices are still based on the Hoeffding decomposition of the output; so that they only quantify the relative influence of an input through the variance. More recently, indices based on the whole distribution have been developed [15, 8, 6]. In particular, the method relying on Cramér-von-Mises distance [26] compares the conditionnal cumulative distribution function with the unconditional one by considering the Hoeffding decomposition of half-space indicators (rather than the Hoeffding decomposition of the output itself) and by integrating them. This method was then extend to codes taking values in a Riemannian manifold [21] and then in general metric spaces [27].
In this work, we focus on two kinds of computer codes: 1) computer codes the output of which is the cumulative distribution function of a real random variable and 2) real-valued stochastic computer codes. A first step will consist in performing global sensitivity analysis for these kinds of computer codes. Further, we will deduce how to perform second level sensitivity analysis using the tools developed in the first step. A code with cumulative distribution function as output can be seen as a code taking values in the space of all probability measures on . This space can be endowed with a metric (for example, the Wasserstein metric [59]). This point of view allows to define at least two different indices for this kind of codes, generalizing the framework of [27]. The first one is based on Wasserstein Fréchet means while the second one is based on the Hoeffding decomposition of the indicators of Wasserstein balls. Further, stochastic codes (see Section 5 for a bibliographical study) can be seen as a “discrete approximation” of codes having cumulative distribution functions as values. Then it is possible to define “natural” indices for such stochastic codes. Finally, second level sensitivity analysis aims at considering uncertainties on the type of the input distributions and/or on the parameters of the input distributions (see Section 6 for a bibliographical study). Actually, this kind of problem can be embedded in the framework of stochastic codes.
The article is organized as follows. In Section 2, we introduce and precisely define a general class of global sensitivity indices. We also present statistical methods to estimate these indices. In Section 3, we recall some basic facts on Wasserstein distances, Wasserstein costs and Fréchet means. In Section 4, we define and study the statistical properties of two new global sensitivity indices for computer codes valued in general Wasserstein spaces. Further, in Section 5, we study the case of stochastic computer codes. Finally, Section 6 is dedicated to the sensitivity analysis with respect to the distributions of the input variables.
2 Sensitivity indices for codes valued in general metric spaces
We consider a black-box code defined on a product of measurable spaces () taking its values in a metric space . The output denoted by is then given by
(1) |
We denote by the distribution of the output code .
The aim of this work is to give some partial answers to the following questions.
-
Question 1
How can we perform Global Sensitivity Analysis (GSA) when the output space is the space of probability distribution functions (p.d.f.) on or the space of cumulative distribution functions (c.d.f.)?
-
Question 2
How can we perform GSA for stochastic computer codes?
-
Question 3
How can we perform GSA with respect to the choice of the distributions of the input variables?
2.1 The general metric spaces sensitivity index
In [27], the authors performed GSA for codes taking values in general metric spaces. To do so, they consider a family of test functions parameterized by elements of and defined by
Let and . Assuming that the test functions are L2-functions with respect to the product measure (where is the product -times of the distribution of the output code ) on , they allow to defined the general metric space sensitivity index with respect to by
(2) |
Roughly speaking, there are two parts in the previous indices. First, for any value of , we consider the numerator and the denominator of the classical Sobol index of . We call this part the Sobol’ part. Second, we integrate each part with respect to the measure ; this will be called the integration part.
As explained in [27], by construction, the indices lie in and share the same properties as their Sobol counterparts:
-
1.
the different contributions sum to 1;
-
2.
they are invariant by translation, by any isometric and by any non-degenerated scaling of .
Estimation
Three different estimation procedures are available in this context. The two first methods are based on the so-called Pick-Freeze scheme. More precisely, the Pick-Freeze scheme, considered in [36], is a well tailored design of experiment. Namely, let be the random vector such that if and if where is an independent copy of . We then set
(3) |
Further, the procedure consists in rewriting the variances of the conditional expectation in terms of covariances as follows:
(4) |
Alternatively, the third estimation procedure that can be seen as an ingenious and effective approximation of the Pick-Freeze scheme is based on rank statistics. Until now, it is unfortunately only available to estimate first order indices in the case of real-valued inputs.
-
•
First method - Pick-Freeze. Introduced in [26], this procedure is based on a double Monte-Carlo scheme to estimate the Cramér-von-Mises indices . More precisely, to estimate in our context, one considers the following design of experiment consisting in:
-
1.
a classical Pick-Freeze -sample, that is two -samples of : , ;
-
2.
another -samples of independent of : , , .
The empirical estimator of the numerator of is then given by
while the one of the denominator is
For , , and given by , the index is nothing more than the index defined in [26] based on the Cramér-von-Mises distance and the whole distribution of the output. Its estimator defined as the ratio of and with has been proved to be asymptotically Gaussian [26, Theorem 3.8]. The proof relies on Donsker’s theorem and the functional delta method [58, Theorem 20.8]. Hence, in the general case of , the central limit theorem will be still valid as soon as the collection forms a Donsker’s class of functions.
-
1.
-
•
Second method - U-statistics. As done in [27], this method allows the practitioner to get rid of the additional random variables for and . The estimator is now based on U-statistics and deals simultaneously with the Sobol part and the integration part with respect to . It suffices to rewrite as
(5) where,
(6) denoting by the pair and, for ,
(7) with and . Finally, one considers the empirical version of (5) as estimator of :
(8) where, for ,
(9) and the function:
is the symmetrized version of . In [27, Theorem 2.3], the estimator has been proved to be consistent and asymptotically Gaussian.
Even if the Pick-Freeze procedure is quite general, it presents some drawbacks. First of all, the Pick-Freeze design of experiment is peculiar and may not be available in real applications. Moreover, it can be unfortunately very time consuming in practice. For instance, estimating all the first order Sobol indices requires calls to the computer code.
-
•
Third method - Rank-based. In [14], Chatterjee proposes an efficient way based on ranks to estimate a new coefficient of correlation. This estimation procedure can be seen as an approximation of the Pick-Freeze scheme and then has been exploited in [23] to perform a more efficient estimation of . Anyway, this method is only well tailored for estimating first order indices i.e. in the case of for some and when the input .
More precisely, an i.i.d. sample of pairs of real-valued random variables () is considered, assuming for simplicity that the laws of and are both diffuse (ties are excluded). The pairs are rearranged in such a way that
and, for any , is the output computed from . Let be the rank of , that is,
The new correlation coefficient is then given by
(10) In [14], it is proved that converges almost surely to a deterministic limit which is actually equal to when . Further, the author also proves a central limit theorem when and are independent, which is clearly not relevant in the context of sensitivity analysis (where and are assumed to be dependent through the computer code).
In our context, recall that and let . Let also be the rank of in the sample of and define
(11) Then the empirical estimator of only requires a -sample of and is given by the ratio between
and
It is worth mentioning that plays the same role as (the Pick-Freeze version of ) in the Pick-Freeze estimation procedure. Anyway, the strength of the rank-based estimation procedure lies in the fact that only one -sample of is required while samples of size are necessary in the Pick-Freeze estimation of a single index (worse, samples of size are required when one wants to estimates indices).
Comparison of the estimation procedures
First, the Pick-Freeze estimation procedure allows the estimation of several sensitivity indices: the classical Sobol indices for real-valued outputs, as well as their generalization for vectorial-valued codes, but also the indices based on higher moments [49] and the Cramér-von-Mises indices which take into account on the whole distribution [26, 21]. Moreover, the Pick-Freeze estimators have desirable statistical properties. More precisely, this estimation scheme has been proved to be consistent and asymptotically normal (i.e. the rate of convergence is ) in [36, 25, 27]. The limiting variances can be computed explicitly, allowing the practitioner to build confidence intervals. In addition, for a given sample size , exponential inequalities have been established. Last but not least, the sequence of estimators is asymptotically efficient from such a design of experiment (see, [58] for the definition of the asymptotic efficiency and [25] for more details on the result).
However, the Pick-Freeze estimators have two major drawbacks. First, they rely on a particular experimental design that may be unavailable in practice. Second, the number of model calls to estimate all first order Sobol indices grows linearly with the number of input parameters. For example, if we consider input parameters and only calls are allowed, then only a sample of size is available to estimate each single first order Sobol index.
Secondly, the estimation procedure based on U-statistics has the same kind of asymptotic guarantees as consistency and asymptotic normality. Furthermore, the estimation scheme is reduced to evaluations of the code. Last but not least, using the results of Hoeffding [33] on U-statistics, the asymptotic normality is proved straightforwardly.
Finally, embedding Chatterjee’s method in the GSA framework (called rank-based method in this framework) thereby eliminates the two drawbacks of the classical Pick-Freeze estimation. In addition, the rank-based method allows for the estimation of a large class of GSA indices which include the Sobol indices and the higher order moment indices proposed by Owen [47, 49, 48]. Using a single sample of size , it is now possible to estimate at the same time all the first order Sobol indices , first order Cramér-von-Mises indices, and other useful first order sensitivity indices as soon as all inputs are real valued.
2.2 The universal sensitivity index
Formula (2) can be generalized in the following ways.
-
1.
The point in the definition of the test functions is allowed to belong to another measurable space than .
-
2.
The probability measure in (2) can be replaced by any “admissible” probability measure.
Such generalizations lead to the definition of a universal sensitivity index and its procedures of estimation.
Definition 2.1.
Let belongs to some measurable space endowed with some probability measure . For any , we define the universal sensitivity index with respect to by
(12) |
Notice that the index is obtained by the integration over with respect to of the Hoeffding decomposition of . Hence, by construction, this index lies in and shares the same properties as its Sobol counterparts:
-
1.
the different contributions sum to 1;
-
2.
it is invariant by translation, by any isometric and by any non-degenerated scaling of .
The universality is twofold. First, it allows to consider more general relevant indices. Secondly, this definition encompasses, as particular cases, the classical sensitivity indices. Indeed,
-
•
the so-called Sobol index with respect to is ;
-
•
the Cramér-von-Mises index with respect to is where and ;
-
•
the general metric space sensitivity index with respect to is where .
An example where is different from will be considered in Section 4.
Estimation
Here, we assume that is different from and we follow the same tracks as for the estimation of in Section 2.1.
-
•
First method - Pick-Freeze. We use the same design of experiment as in the First method of Section 2.1 but instead of considering that the additional -samples for and are drawn with respect to the distribution of the output, they are now drawn with respect to the law . More precisely, one considers the following design of experiment consisting in:
-
1.
a classical Pick-Freeze sample, that is two -samples of : , ;
-
2.
-distributed -samples , and that are independent of for .
The empirical estimator of the numerator of is then given by
while the one of the denominator is
As previously, it is straightforward (as soon as the collection forms a Donsker’s class of functions) to adapt the proof of Theorem [26, Theorem 3.8] to prove the asymptotic normality of the estimator.
-
1.
-
•
Second method - U-statistics. This method is not relevant in this case since .
-
•
Third method - Rank-based. Here, the design of experiment reduces to:
-
1.
a -sample of : , ;
-
2.
a -sample of that is -distributed: , , independent of , .
The empirical estimator of is then given by the ratio between
and
-
1.
We recall that this last method only applies for first order sensitivity indices and real-valued input variables.
2.3 A sketch of answer to Questions 1 to 3
In the sequel, we discuss how pertinent choices of the metric, of the class of functions and of the probability measure can provide some partial answers to Questions 1 to 3 raised at the beginning of Section 2. For instance, in order to answer to Question 1, we can consider that is the space of probability measures on that are -functions and we endow this space with the Wasserstein metric (see Section 3.1 for some recalls on Wasserstein metrics). We will propose two possible approaches to define interesting sensitivity indices in this framework.
- •
- •
The case of stochastic computer computer codes raised in Question 2 will be addressed as follows. A computer code (to be defined) valued in will be seen as an ideal case of stochastic computer codes. Finally, it will be possible to treat Question 3 using the framework of Question 2.
3 Wasserstein spaces and random distributions
3.1 Definition
For any , we define the -Wasserstein distance between two probability distributions that are -integrable and characterized by their c.d.f.’s and on by:
where and mean that and are random variables with respective c.d.f.’s and . We define the Wasserstein space as the space of all -integrable measures defined on endowed with the -Wasserstein distance . In the sequel, any measure is identified to its c.d.f. or in some cases to its p.d.f. In the unidimensional case (), it is a well known fact that has an explicitly expression given by
(13) |
Here and are the generalized inverses of the increasing functions and and is a random variable uniformly distributed on . Of course, and have c.d.f.’s and . The representation (13) of the -Wasserstein distance when can be generalized to a wider class of “contrast functions”. For more details on Wasserstein spaces, one can refer to [59] and [5] and the references therein.
Definition 3.1.
We call contrast function any application from to satisfying the "measure property" defined by
meaning that defines a negative measure on .
For instance, satisfies . If satisfies then any function of the form also satisfies . If is a convex real function, satisfies . In particular, satisfy and actually so does as soon as .
Definition 3.2.
We define the Skorohod space of all distribution functions as the space of all non-decreasing functions from to that are càd-làg with limit (resp. ) in (resp. ) equipped with the supremum norm.
Definition 3.3.
For any , any , and any positive contrast function , we define the -Wasserstein cost by
Obviously, with . The following theorem can be found in ([11]).
Theorem 3.4 (Cambanis, Simon, Stout [11]).
Let be a contrast function. Then
where is a random variable uniformly distributed on .
3.2 Extension of the Fréchet mean to contrast functions
Definition 3.5.
We call a loss function any positive and measurable function . Then, we define a Fréchet feature of a random variable taking values in a measurable space as (whenever it exists):
(14) |
When is a metric space endowed with a distance , the Fréchet feature corresponds to the classical Fréchet mean (see [22]). In particular, minimizes which is an extension of the definition of the classical mean in that minimizes .
Now we consider and . Further, Equation (14) becomes
where is a measurable function from a measurable space to .
Theorem 3.6.
Let be a positive contrast function. Assume that the application defined by is measurable. In addition, assume that exists and is unique. Then there exists a unique Fréchet mean of denoted by and we have
Proof of Theorem 3.6.
Since satisfies , we have
by Fubini’s theorem. Now, for all , the quantity is minimum for .
and, in particular, for , one gets
Conversely, by the definition of , we have for all ,
and, in particular, for , one gets
The theorem then follows by the uniqueness of the minimizer. ∎
In the previous theorem, we propose a very genereal non parametric framework for the random element together with some assumptions on existence and uniqueness of the Fréchet feature and measurability of the map . Nevertheless, it is possible to construct explicit parametric models for for whom theses assumptions are satisfied. For instance, the authors of [4] ensures measurability for some parametric models on using results of [18]. Notice that in [19] a new sensitivity indice is defined for each feature associated to a contrast function. In section 4.2 we will restrict our analysis to Fréchet means, hence to Sobol indices.
3.3 Examples
The Fréchet mean in the space is the inverse of the function . Another example is the Fréchet median. Since the median in is related to the cost, the Fréchet median of a random c.d.f. is
More generally, we recall that, for , the -quantile in is the Fréchet mean associated to the contrast function , also called the pinball function. Then we can define an -quantile of a random c.d.f. as:
where is the set of the -quantiles of a random variable taking values in . Naturally, taking leads to the median.
Let us illustrate the previous definitions on an example. Let be a random variable with c.d.f. which is assumed to be increasing and continuous. Let also and two real random variables such that >0. Then we consider the random c.d.f. of :
Naturally, the Fréchet mean of is and its -quantile is given by
4 Sensitivity analysis in general Wasserstein spaces
In this section, we consider that our computer code is -valued; namely, the output of an experiment is the c.d.f. or the p.d.f. of a measure . For instance, in [9], [40] and [46], the authors deal with p.d.f.-valued computer codes (and stochastic computer codes). In other words, they define the following application:
where is the set of p.d.f.’s:
Here, we choose to identify any element of with its c.d.f. In this framework, the output of the cmoputer code is then a c.d.f. denoted by
(16) |
Here, denotes the law of the c.d.f. . In addition, we set . The case of a general can be handled analogously.
4.1 Sensitivity anlaysis using Equation (2) and Wasserstein balls
Consider , , and three elements of and, for , the family of test functions
(17) |
Then, for all , (2) becomes
(18) |
As explained in Section 2.1, is obtained by integration over of the Hoeffding decomposition of . Hence, by construction, this index lies in and shares the same properties as its Sobol counterparts:
-
1.
the different contributions sum to 1;
-
2.
it is invariant by translation, by any isometric and by any non-degenerated scaling of .
4.2 Sensitivity analysis using Equation (12) and Fréchet means
In the classical framework where the output is real, we recall that the Sobol index with respect to is defined by
(19) |
by the property of the conditional expectation. In one hand, one may extend this formula to the framework of this section where the output of interest is the c.d.f. :
where with the Fréchet mean of . From Theorem 3.6, we get
leading to
(20) |
In the other hand, one can consider (12), with ,
(21) |
and the uniform probability measure on . In that case,
Then
is exactly the same as in (20). Thus, as explained in Section 2.2, lies in and:
-
1.
the different contributions sum to 1;
-
2.
it is invariant by translation, by any isometric and by any non-degenerated scaling of .
4.3 Estimation procedure
As noticed in the previous section, both
with defined in (17) and
with defined in (21), are particular cases of indices of the form (12).
When belongs to the same space as the output and when is equal to , one may first use the Pick-Freeze estimations of the indices given in (4.1) and (20). To do so, it is convenient once again to use (4) leading to
(22) |
and
(23) |
where and are respectively the Pick-Freeze versions of and . Secondly, one may resort to the estimations based on U-statistics together on the Pick-Freeze design of experiment. Thirdly, it is also possible and easy to obtain rank-based estimations in the vein of (10).
4.4 Numerical comparison of both indices
Example 4.1 (Toy model).
Let be independent and positive random variables. We consider the c.d.f.-valued code , the output of which is given by
(24) |
so that
(25) |
In addition, one gets
and
and finally
For or , it remains to plug the previous formulas in (20) to get the explicit expressions of the indices .
Now, in order to get a closed formula for the indices defined in (4.1), we assume is Bernoulli distributed with parameter . In (4.1), the distributions and can be either , , , or with respective probabilities , , , and . In the sequel, we give, for all sixteen possibilities for the distribution of , the corresponding contributions for the numerator and for the denominator of (4.1).
With probability , and . Then , , and if and only if . Since , the contribution of this case to the denominator is thus
Moreover,
so that, the contribution to the numerator is here given by
Similarly, one gets
Moreover, regarding the indices with respect to and ,
and the contribution to the numerator is given by
The remaining fifteen cases can be treated similarly and are gathered (with the first case developed above) in the following table. Finally, for , one may compute the explicit expression of :
Some numerical values have not been explicited in the table but given below:
Case 2 | ||||
Case 6 | ||||
Case 11 | ||||
Case 15 |
Case 1 | , | Case 2 | , |
---|---|---|---|
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 3 | , | Case 4 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 5 | , | Case 6 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 7 | , | Case 8 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 9 | , | Case 10 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 11 | , | Case 12 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 13 | , | Case 14 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. | ||
Case 15 | , | Case 16 | , |
Prob. | Prob. | ||
Num. 1 | Num. 1 | ||
Num. 2 | Num. 2 | ||
Num. 3 | Num. 3 | ||
Num. 1,3 | Num. 1,3 | ||
Den. | Den. |
In Figure 1, we have represented the indices , , , and given by (20) with respect to the values of and varying from 0 to 1 for a fixed value of . We have considered three different values of : (first row), , (second row) and (third row). Analogously, the same kind of illustration for the indices , , , and given by (4.1) is provided in Figure 2 . In addition, the regions of predominance of each index are plotted in Figure 3. The values of and still vary from 0 to 1 and the fixed values of considered are: (first row), , (second row) and (third row). Finally, the same kind of illustration for the indices is given in Figure 4.




In order to compare the estimation accuracy of the Pick-Freeze method and the rank-based method at a fixed size, we assume that only calls of the computer code are allowed to estimate the indices and for , , and . We only focus on the first order indices since, as explained previously, the rank-based procedure has not been developed yet for higher order indices. We repeat the estimation procedure 500 times. The boxplots of the mean square errors for the estimation of the Fréchet indices and the Wasserstein indices have been plotted in Figure 5. We observe that, for a fixed sample size (corresponding to a Pick-Freeze sample size ), the rank-based estimation procedure performs much better than the Pick-Freeze method with significantly lower mean errors.
![]() |
5 Sensitivity analysis for stochastic computer codes
This section deals with stochastic computer codes in the sense that two evaluations of the code for the same input lead to different outputs.
5.1 State of the art
A first natural way to handle stochastic computer codes is definitely to consider the expectation of the output code. Indeed, as mentioned in [9], previous works dealing with stochastic simulators together with robust design or optimization and sensitivity analysis consist mainly in approximating the mean and variance of the stochastic output [17, 10, 37, 2] and then performing a global sensitivity analysis on the expectation of the output code [42].
As pointed out by [35], another approach is to consider that the stochastic code is of the form where the random element contains the classical input variables and the variable is an extra unobserved random input.
Such an idea was exploited in [36] to compare the estimation of the Sobol indices in an “exact” model to the estimation of the Sobol indices in an associated metamodel.
Analogously, the author of [43] assumes the existence of an extra random variable which is not chosen by the practitioner but rather generated at each computation of the output independently of . In this framework, he builds two different indices. The first index is obtained by substituting for in the classical definition of the first order Sobol index . In this case, is considered as another input, even though it is not observable. The second index is obtained by substituting for in the Sobol index. The noise is then smoothed out.
Similarly, the authors of [31] traduces the randomness of the computer code using such an extra random variable . In practice, their algorithm returns realizations of the first order Sobol indices for , denoted by for and . Then, for any , they approximate the statistical properties of by considering the sample -th moments given by
noticing that
5.2 The space as an ideal version of stochastic computer codes
When dealing with stochastic computer codes, the practitioner is generally interested in the distribution of the output for a given . As previously seen, one can translate this type of codes in terms of a deterministic code by considering an extra input which is not chosen by the practitioner itself but which is a latent variable generated randomly by the computer code and independently of the classical input. As usual in the framework of sensitivity analysis, one considers the input as a random variable. All the random variables (the one chosen by the practitioner and the one generated by the computer code) are built on the same probability space, leading to the function :
where is the extra random variable lying in . We naturally denote the output random variable by .
Hence, one may define another (deterministic) computer code associated with whose output is the probability measure:
The framework of (5.2) is exactly the one of Section 4.1 and has already been handled. Obviously, in practice, one does not assess the output of the code but one can only obtain an empirical approximation of the measure given by evaluations of at , namely,
Further, (5.2) can be seen as an ideal version of (5.2). Concretely, for a single random input , we will evaluate times the code defined by (5.2) (so that the code will generate independently hidden variables , …, ) and one may observe
leading to the measure approximating the distribution of . We emphasize on the fact that the random variables are not observed.
5.3 Sensitivity analysis
Let us now present the methodology we adopt. In order to study the sensitivity of the distribution , one can use the framework introduced in Section 4.1 and the index given by (4.1).
In an ideal scenario which corresponds to the framework of (5.2), one may asses to the probability measure for any . Then following the estimation procedure of Section 4.3, one gets an estimation of the sensitivity index with good asymptotic properties [27, Theorem 2.3].
In the more realistic framework presented above in (5.2), we only have access to the approximation of rendering more complex the estimation procedure and the study of the asymptotic properties. In this case, the general design of experiments is the following:
where is the total number of evaluations of the stochastic computer code (5.2). Then we construct the approximations of (standing for ) for any given by
(28) |
From there, one may use one of the three estimation procedures presented in Section 2.1.
-
•
First method - Pick-Freeze. It suffices to plug the empirical version of each measure under concern in (22).
-
•
Second method - U-statistics. For , let
(29) where as previously seen is the symmetrized version of defined in (• ‣ 2.1) and . Then we estimate by
(30) -
•
Third method - Rank-based. The rank-based estimation procedure may also easily extend to this context by using the empirical version of each measure under concern instead of the true one , as explained into more details in the numerical study developed in Section 5.1.
Actually, these estimators are easy to compute since for two discrete measures supported on a same number of points and given by
the Wasserstein distance between and simply writes
(31) |
where is the -th order statistics of .
5.4 Central limit theorem for the estimator based on U-statistics
Proposition 5.1.
In some particular frameworks, one may derive easily a suitable value of . Two examples are given in the following.
Example 5.2.
If the inverse of the random variable has a finite expectation, then, by Markov inequality,
and it suffices to choose so that as goes to infinity.
Example 5.3 (Uniform example).
Assume that is uniformly distributed on and that is a Gaussian distribution centered at with unit variance. Then the Wasserstein distance rewrites as so that the random variable is given by
Consequently,
Notice that is triangular distributed with parameter , , and leading to
In addition,
Now, and are two independent random variables uniformly distributed on . Then (see Figure 6), one has
whence
Thus it turns out that . Consequently, a suitable choice for is .
Analogously, one may derive suitable choices for in some particular cases. For instance, we refer the reader to [5] to get upper bounds on for several values of and several assumptions on the distribution on : general, uniform, Gaussian, beta, log concave… Here are some results.
-
•
In the general framework, the upper bound for relies on the functional
where is the c.d.f. associated to and its p.d.f. See Cf. [5, Theorems 3.2, 5.1 and 5.3].
-
•
Assume that is uniformly distributed on . Then by [5, Theorems 4.7, 4.8 and 4.9], for any ,
for any and for any ,
and for any ,
E.g. .
- •
Example 5.3 - continued. We consider that is uniformly distributed on and is a Gaussian distribution centered at with unit variance. Then, by [5, Corollary 6.14], we have for any ,
and for any and for any ,
where depends only on . Since we have already chosen , it remains to take so that to fulfill the condition .
5.5 Numerical study
Example 4.1 - continued. Here, we consider again the code given by (24). Having in mind the notation of Section 5.2, we consider the ideal version of the code:
where is the uniform distribution on , the c.d.f. of which is given by (24) and its stochastic counterpart:
where is a realization of .
Hence, we no longer assume that one may observe realizations of associated to the initial realizations of . Instead, for any of the initial realizations of , we assess realizations of a uniform random variable on .
In order to compare the estimation accuracy of the Pick-Freeze method and the rank-based method at a fixed size, we assume that only calls of the computer code are allowed to estimate the indices and for , , and . We only focus on the first order indices since, as explained previously, the rank-based procedure has not been developed yet for higher order indices. The empirical c.d.f. based on the empirical measures for in (28) are constructed with evaluations. We repeat the estimation procedure 500 times. The boxplots of the mean square errors for the estimation of the Fréchet indices and the Wasserstein indices have been plotted in Figure 7. We observe that, for a fixed sample size (corresponding to a Pick-Freeze sample size ), the rank-based estimation procedure performs much better than the Pick-Freeze method with significantly lower mean errors.
![]() |
6 Sensitivity analysis with respect to the law of the inputs
6.1 State of the art
The paper [44] is devoted to second level uncertainty which corresponds to the uncertainty on the type of the input distributions and/or on the parameters of the input distributions. As mentioned by the authors, such uncertainties can be handled in two different manners: (1) aggregating them with no distinction [13, 12] or (2) separating them [44]. In [13], e.g., the uncertainty concerns the parameters of the input distributions. The authors study the expectation with respect to the distribution of the parameters of the conditional output. In [12], the second level uncertainties are transformed into first level uncertainties considering the aggregated vector containing the input random variables vector together with the vector of uncertain parameters. Alternatively, in [44], the uncertainty brought by the lack of knowledge of the input distributions and the uncertainty of the random inputs are treated separately. A double Monte-Carlo algorithm is first considered. In the outer loop, a Monte-Carlo sample of input distribution is generated, while the inner loop proceeds to a global sensitivity analysis associated to each distribution. A more efficient algorithm is also proposed with a unique Monte-Carlo loop. The sensitivity analysis is then performed using the so-called Hilbert-Schmidt dependence measures (HSIC indices) on the input distributions rather than the input random variables themselves. See, e.g., [29] for the definition of the HSIC indices and more details on the algorithms.
In [45], a different approach is adopted. A failure probability is studied while the uncertainty concerns the parameters of the input distributions. An algorithm with low computational cost is proposed to handle such uncertainty together with the rare event setting. A single initial sample allows to compute the failure probabilities associated to different parameters of the input distributions. A similar idea is exploited in [41] in which the authors consider input perturbations. and Perturbed-Law based Indices that are used to quantify the impact of a perturbation of an input p.d.f. on a failure probability. Analogously, the authors of [30, 32] are interested in (marginal) p.d.f. perturbations and the aim is to study the “robustness of the Sobol indices to distributional uncertainty and to marginal distribution uncertainty” which correspond to second level uncertainty. For instance, the basic idea of the approach proposed in [30] is to view the total Sobol index as an operator which inputs the p.d.f. and returns the Sobol index. Then the analysis of robustness is done computing and studying the Fréchet derivative of this operator. The same principle is used in [32] to treat the robustness with respect to the marginal distribution uncertainty.
Last but not least, it is worth mentioning the classical approach of epistemic global sensitivity analysis of Dempster-Shafer theory (see, e.g., [54, 1]). This theory describes the random variables together with an epistemic uncertainty traduced in terms of an associated epistemic variable on a set , a mass function representing a probability measure on the set of all subsets . This lack of knowledge leads to an upper and lower bound of the c.d.f. and can be viewed as a second level uncertainty.
6.2 Link with stochastic computer codes
We propose a new procedure that stems from the the methodology in the context of stochastic computer codes described in Section 5. We still denote by () the distribution of the input () in the model given by (1). There are several ways to model the uncertainty with respect to the choice of each . Here we adopt the following framework. We assume that each belongs to some family of probability measures endowed with the probability measure . In general, there might be measurability issues and the question of how to define a field on some general spaces can be tricky. We will restrict our study to the simple case where the existence of the probability measure on is given by the construction of the set . More precisely, we proceed as follows.
-
•
First, for , let be an integer and let . Then consider the probability space where is the Borel field and is a probability measure on .
-
•
Second, for , we consider an identifiable parametric set of probability measure on : . Let us denote by the one-to-one mapping from to defined by and define the field on by
Then we endow this measurable space with the probability defined, for any , by
-
•
Third, in order to perform a second level sensitivity analysis on (1), we introduce the stochastic mapping from to defined by
(34) where are independently drawn according to the distribution . Hence is a stochastic computer code from to and once the probability measures on each are defined, we can perform sensitivity analysis using the framework of Section 5.
6.3 Numerical study
As in [32], let us consider the synthetic example defined on by
(35) |
We are interested in the uncertainty in the support of the random variables , and . To do so, we follow the notation and framework of [32]. For , 2, and 3, we assume that is uniformly distributed on the interval , where and are themselves uniformly distributed on and respectively. As remarked in [32], it seems natural that will vary more in the -direction when is close to 0 and less when is close to 1.
As mentioned in Section 6.1, the authors of [32] view the total Sobol index as an operator which inputs the p.d.f. and returns the total Sobol index. Then they study the Fréchet derivative of this operator and determine the most influential p.d.f., which depends on a parameter denoted by . Finally, they make vary this parameter .
Here, we adopt the methodology explained in the previous section (Section 6.2). Namely, we consider the stochastic computer code given by:
(36) |
where the ’s are independently drawn according to the uniform measure on with and themselves uniformly distributed on and respectively. Then to estimate the indices , for , 2, and 3, we proceed as follows.
-
1.
For , 2, and 3, we produce a -sample of intervals .
-
2.
For , 2, and 3, and, for , we generate a -sample of , where is uniformly distributed on .
-
3.
For , we compute the -sample of the output using
Thus we get a -sample of the empirical measures of the distribution of the output given by:
-
4.
For , 2, and 3, we order the intervals and get the Pick-Freeze versions of to treat the sensitivity analysis regarding the -th input.
- 5.
Notice that we only consider the estimators based on the Pick-Freeze method since we allow for both bounds of the interval to vary and, as explained previously, the rank-based procedure has not been developed yet neither for higher order indices nor in higher dimensions.
First, we compute the estimators of following the previous procedure with a sample size and an approximation size . We also perform another batch of simulations allowing for higher variability on the bounds: is now uniformly distributed on while is now uniformly distributed on . The results are displayed in Table 1.
u | |||||||
---|---|---|---|---|---|---|---|
0.07022 | 0.08791 | 0.09236 | 0.14467 | 0.21839 | 0.19066 | ||
0.11587 | 0.06542 | 0.169529 | 0.22647 | 0.40848 | 0.34913 |
Second, we run another simulations allowing for more variability on the upper bound related to the third input only: is uniformly distributed on (instead of ). The results are displayed in Table 2. We still use a sample size and an approximation size .
u | ||||||
---|---|---|---|---|---|---|
0.01196 | 0.06069 | 0.56176 | -0.01723 | 0.63830 | 0.59434 |
Third, we perform a classical GSA on the inputs rather than on the parameters of their distributions. Namely, we estimate the index with a sample size . The reader is referred to [26, Section 3] for the definition of the index and its Pick-Freeze estimator together with their properties. The results are displayed in Table 3.
u | ||||||
---|---|---|---|---|---|---|
0.13717 | 0.15317 | 0.33889 | 0.33405 | 0.468163 | 0.53536 |
In Table 3, we see that the Cramér-von-Mises index related to is more than twice as important as and (when considering only first order effects). Nevertheless, when one is interested in the choice of the input distributions of , , and , the first row in Table 1 shows that each choice is equally important. Now, if one give more freedom to the space where the distribution lives, the relative importance may change as one can see in Table 2 (second row) and in Table 3. More precisely, in Table 2, the variability of the third input distribution (namely, the variability of its upper bound) is five times bigger than the other variabilities. Not surprisingly, it results that the importance of the choice of the third input distribution is then much more important than the choices of the distributions of the two first inputs.
7 Conclusion
In this article, we present a very general way to perform sensitivity analysis when the output of a computer code lives in a metric space. The main idea is to consider real-valued squared integrable test functions parameterized by a finite number of elements of a probability space. Then Hoeffding decomposition of the test functions is computed and integrated with respect to the parameter . This very general and flexible definition allows, in one hand, to recover a lot of classical indices (namely, the Sobol indices and the Cramér-von-Mises indices) and, in the other hand, to perform a well tailored and interpretable sensitivity analysis. Furthermore, a sensitivity analysis is also made possible for computer codes the output of which is a c.d.f., for stochastic computer codes (that are seen as an approximation of c.d.f.-valued computer codes). Last but not least, it enables also to perform second level sensitivity analysis by embedding second level sensitivity analysis as a particular case of stochastic computer codes.
Acknowledgment
Appendix A Proofs
A.1 Notation
It is convenient to have short expressions for terms that converge in probability to zero. We follow [58]. The notation (respectively ) stands for a sequence of random variables that converges to zero in probability (resp. is bounded in probability) as . More generally, for a sequence of random variables ,
For deterministic sequences and , the stochastic notation reduce to the usual and . Finally, stands for a generic constant that may differ from one line to another.
A.2 Proof of Proposition 5.1
One has
By [27, Theorem 2.3], the second term in the right hand side of the previous equation is asymptotically Gaussian. If we prove that the first term in the right hand side is , then by Slutsky’s Lemma [58, Lemma 2.8], is asymptotically Gaussian.
Now we prove that . We write
Since , for and and converges almost surely respectively to 0 and , the denominator converges almost surely. Thus it suffices to prove that the numerator is which reduces to prove that for , where (respectively ) has been defined in (29) (resp. (9)). Let for example. The other terms can be treated analogously. Here, . We write
where the random variable in the expectation in the right hand side of the previous inequality is a Bernoulli random variable whose distribution does not depend on . Let be the following event
Obviously, we get , where stands for the complementary of in . Furthermore,
Finally, we introduce and we study:
It remains to choose first, so that and second, such that . Consequently, . Analogously, one gets for =2, 3 and 4.
References
- [1] D. A. Alvarez. Reduction of uncertainty using sensitivity analysis methods for infinite random sets of indexable type. International journal of approximate reasoning, 50(5):750–762, 2009.
- [2] B. Ankenman, B. L. Nelson, and J. Staum. Stochastic kriging for simulation metamodeling. In 2008 Winter Simulation Conference, pages 362–370. IEEE, 2008.
- [3] J. D. Betancourt, F. Bachoc, T. Klein, D. Idier, R. Pedreros, and J. Rohmer. Gaussian process metamodeling of functional-input code for coastal flood hazard assessment. Reliability Engineering and System Safety, 198, June 2020.
- [4] J. Bigot and T. Klein. Consistent estimation of a population barycenter in the Wasserstein space. ArXiv e-prints, Dec. 2012.
- [5] S. Bobkov and M. Ledoux. One-dimensional empirical measures, order statistics, and kantorovich transport distances. Memoirs of the American Mathematical Society, To appear, 2019.
- [6] E. Borgonovo. A new uncertainty importance measure. Reliability Engineering & System Safety, 92(6):771–784, 2007.
- [7] E. Borgonovo, W. Castaings, and S. Tarantola. Moment independent importance measures: New results and analytical test cases. Risk Analysis, 31(3):404–428, 2011.
- [8] E. Borgonovo and B. Iooss. Moment Independent Importance Measures and a Common Rationale. Manuscript.
- [9] T. Browne, B. Iooss, L. Le Gratiet, J. Lonchampt, and E. Remy. Stochastic simulators based optimization by gaussian process metamodels - application to maintenance investments planning issues. Quality and Reliability Engineering International, 32:2067–2080, 2016.
- [10] D. Bursztyn and D. M. Steinberg. Screening experiments for dispersion effects. In Screening, pages 21–47. Springer, 2006.
- [11] S. Cambanis, G. Simons, and W. Stout. Inequalities for when the marginals are fixed. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete, 36(4):285–294, 1976.
- [12] V. Chabridon. Reliability-oriented sensitivity analysis under probabilistic model uncertainty. PhD thesis, Université Clermont-Auvergne, 2018.
- [13] V. Chabridon, M. Balesdent, J.-M. Bourinet, J. Morio, and N. Gayton. Evaluation of failure probability under parameter epistemic uncertainty: application to aerospace system reliability assessment. Aerospace Science and Technology, 69:526–537, 2017.
- [14] S. Chatterjee. A new coefficient of correlation. arXiv e-prints, page arXiv:1909.10140, Sep 2019.
- [15] S. Da Veiga. Global sensitivity analysis with dependence measures. J. Stat. Comput. Simul., 85(7):1283–1305, 2015.
- [16] E. De Rocquigny, N. Devictor, and S. Tarantola. Uncertainty in industrial practice. Wiley Chisterter England, 2008.
- [17] G. Dellino and C. Meloni. Uncertainty management in simulation-optimization of complex systems. Springer, 2015.
- [18] J. Fontbona, H. Guérin, and S. Méléard. Measurability of optimal transportation and strong coupling of martingale measures. Electron. Commun. Probab., 15:124–133, 2010.
- [19] J.-C. Fort, T. Klein, and N. Rachdi. New sensitivity analysis subordinated to a contrast. May 2013.
- [20] J.-C. Fort, T. Klein, and N. Rachdi. New sensitivity analysis subordinated to a contrast. Comm. Statist. Theory Methods, 45(15):4349–4364, 2016.
- [21] R. Fraiman, F. Gamboa, and L. Moreno. Sensitivity indices for output on a Riemannian manifold. arXiv e-prints, page arXiv:1810.11591, Oct 2018.
- [22] M. Fréchet. Les éléments aléatoires de nature quelconque dans un espace distancié. Ann. Inst. H.Poincaré, Sect. B, Prob. et Stat., 10:235–310, 1948.
- [23] F. Gamboa, P. Gremaud, T. Klein, and A. Lagnoux. Global Sensitivity Analysis: a new generation of mighty estimators based on rank statistics. Working paper or preprint, Feb. 2020.
- [24] F. Gamboa, A. Janon, T. Klein, and A. Lagnoux. Sensitivity analysis for multidimensional and functional outputs. Electronic Journal of Statistics, 8:575–603, 2014.
- [25] F. Gamboa, A. Janon, T. Klein, A. Lagnoux, and C. Prieur. Statistical inference for Sobol pick-freeze Monte Carlo method. Statistics, 50(4):881–902, 2016.
- [26] F. Gamboa, T. Klein, and A. Lagnoux. Sensitivity analysis based on Cramér–von Mises distance. SIAM/ASA J. Uncertain. Quantif., 6(2):522–548, 2018.
- [27] F. Gamboa, T. Klein, A. Lagnoux, and L. Moreno. Sensitivity analysis in general metric spaces. Working paper or preprint, Feb. 2020.
- [28] T. Goda. Computing the variance of a conditional expectation via non-nested monte carlo. Operations Research Letters, 45(1):63 – 67, 2017.
- [29] A. Gretton, O. Bousquet, A. Smola, and B. Schölkopf. Measuring statistical dependence with Hilbert-Schmidt norms. In International conference on algorithmic learning theory, pages 63–77. Springer, 2005.
- [30] J. Hart and P. A. Gremaud. Robustness of the Sobol’indices to distributional uncertainty. International Journal for Uncertainty Quantification, 9(5), 2019.
- [31] J. L. Hart, A. Alexanderian, and P. A. Gremaud. Efficient computation of Sobol’indices for stochastic models. SIAM Journal on Scientific Computing, 39(4):A1514–A1530, 2017.
- [32] J. L. Hart and P. A. Gremaud. Robustness of the Sobol’indices to marginal distribution uncertainty. SIAM/ASA Journal on Uncertainty Quantification, 7(4):1224–1244, 2019.
- [33] W. Hoeffding. A class of statistics with asymptotically normal distribution. Ann. Math. Statistics, 19:293–325, 1948.
- [34] D. Idier, A. Aurouet, F. Bachoc, A. Baills, J. D. Betancourt, J. Durand, R. Mouche, J. Rohmer, F. Gamboa, T. Klein, J. Lambert, G. Le Cozannet, S. Leroy, J. Louisor, R. Pedreros, and A.-L. Véron. Toward a User-Based, Robust and Fast Running Method for Coastal Flooding Forecast, Early Warning, and Risk Prevention. Journal of Coastal Research, Special Issue, 95:11–15, 2020.
- [35] B. Ioss, T. Klein, and A. Lagnoux. Sobol’ sensitivity analysis for stochastic numerical codes. In Proceedings of the SAMO 2016 Conference, Reunion Island, France, pages 48–49, 2016.
- [36] A. Janon, T. Klein, A. Lagnoux, M. Nodet, and C. Prieur. Asymptotic normality and efficiency of two Sobol index estimators. ESAIM: Probability and Statistics, 18:342–364, 1 2014.
- [37] J. P. Kleijnen. Design and analysis of simulation experiments. In International Workshop on Simulation, pages 3–22. Springer, 2015.
- [38] S. Kucherenko and S. Song. Different numerical estimators for main effect global sensitivity indices. Reliability Engineering & System Safety, 165:222–238, 2017.
- [39] M. Lamboni, H. Monod, and D. Makowski. Multivariate sensitivity analysis to measure global contribution of input factors in dynamic models. Reliability Engineering & System Safety, 96(4):450–459, 2011.
- [40] L. Le Gratiet. Asymptotic normality of a Sobol index estimator in gaussian process regression framework. Preprint, 2013.
- [41] P. Lemaître, E. Sergienko, A. Arnaud, N. Bousquet, F. Gamboa, and B. Iooss. Density modification-based reliability sensitivity analysis. Journal of Statistical Computation and Simulation, 85(6):1200–1223, 2015.
- [42] A. Marrel, B. Iooss, S. Da Veiga, and M. Ribatet. Global sensitivity analysis of stochastic computer models with joint metamodels. Statistics and Computing, 22(3):833–847, 2012.
- [43] G. Mazo. An optimal tradeoff between explorations and repetitions in global sensitivity analysis for stochastic computer models. 2019.
- [44] A. Meynaoui, A. Marrel, and B. Laurent. New statistical methodology for second level global sensitivity analysis. working paper or preprint, Feb. 2019.
- [45] J. Morio. Influence of input pdf parameters of a model on a failure probability estimation. Simulation Modelling Practice and Theory, 19(10):2244–2255, 2011.
- [46] V. Moutoussamy, S. Nanty, and B. Pauwels. Emulators for stochastic simulation codes. In CEMRACS 2013—modelling and simulation of complex systems: stochastic and deterministic approaches, volume 48 of ESAIM Proc. Surveys, pages 116–155. EDP Sci., Les Ulis, 2015.
- [47] A. Owen. Better estimation of small Sobol’sensitivity indices. arXiv preprint arXiv:1204.4763, 2012.
- [48] A. Owen. Variance components and generalized Sobol’ indices. SIAM/ASA Journal on Uncertainty Quantification, 1(1):19–41, 2013.
- [49] A. Owen, J. Dick, and S. Chen. Higher order Sobol’ indices. Information and Inference, 3(1):59–81, 2014.
- [50] K. Pearson. On the partial correlation ratio. Proceedings of the Royal Society of London. Series A, Containing Papers of a Mathematical and Physical Character, 91(632):492–498, 1915.
- [51] N. Peteilh, T. Klein, T. Druot, N. Bartoli, and R. P. Liem. Challenging Top Level Aircraft Requirements based on operations analysis and data-driven models, application to take-off performance design requirements. In AIAA AVIATION 2020 FORUM, AIAA AVIATION 2020 FORUM, Reno, NV, United States, June 2020. American Institute of Aeronautics and Astronautics, American Institute of Aeronautics and Astronautics.
- [52] A. Saltelli, K. Chan, and E. Scott. Sensitivity analysis. Wiley Series in Probability and Statistics. John Wiley & Sons, Ltd., Chichester, 2000.
- [53] T. J. Santner, B. Williams, and W. Notz. The Design and Analysis of Computer Experiments. Springer-Verlag, 2003.
- [54] P. Smets et al. What is Dempster-Shafer’s model. Advances in the Dempster-Shafer theory of evidence, pages 5–34, 1994.
- [55] I. Sobol. Global sensitivity indices for nonlinear mathematical models and their Monte Carlo estimates. Mathematics and Computers in Simulation, 55(1-3):271–280, 2001.
- [56] I. M. Sobol. Sensitivity estimates for nonlinear mathematical models. Math. Modeling Comput. Experiment, 1(4):407–414, 1993.
- [57] B. Sudret. Global sensitivity analysis using polynomial chaos expansions. Reliability Engineering & System Safety, 93(7):964–979, 2008.
- [58] A. W. van der Vaart. Asymptotic statistics, volume 3 of Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, Cambridge, 1998.
- [59] C. Villani. Topics in Optimal Transportation. American Mathematical Society, 2003.