This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bayesian model inversion using stochastic spectral embedding

P.-R. Wagner, S. Marelli, B. Sudret
(14.05.2020)
Abstract

In this paper we propose a new sampling-free approach to solve Bayesian model inversion problems that is an extension of the previously proposed spectral likelihood expansions (SLE) method. Our approach, called stochastic spectral likelihood embedding (SSLE), uses the recently presented stochastic spectral embedding (SSE) method for local spectral expansion refinement to approximate the likelihood function at the core of Bayesian inversion problems.

We show that, similar to SLE, this approach results in analytical expressions for key statistics of the Bayesian posterior distribution, such as evidence, posterior moments and posterior marginals, by direct post-processing of the expansion coefficients. Because SSLE and SSE rely on the direct approximation of the likelihood function, they are in a way independent of the computational/mathematical complexity of the forward model. We further enhance the efficiency of SSLE by introducing a likelihood specific adaptive sample enrichment scheme.

To showcase the performance of the proposed SSLE, we solve three problems that exhibit different kinds of complexity in the likelihood function: multimodality, high posterior concentration and high nominal dimensionality. We demonstrate how SSLE significantly improves on SLE, and present it as a promising alternative to existing inversion frameworks.

Keywords: Bayesian model inversion, inverse problems, polynomial chaos expansions, spectral likelihood expansions, stochastic spectral likelihood embedding, sampling-free inversion.

1 Introduction

Computational models are an invaluable tool for decision making, scientific advances and engineering breakthroughs. They establish a connection between a set of input parameters and output quantities with wide-ranging applications. Model inversion uses available experimental observations of the output to determine the set of input parameters that maximize the predictive potential of a model. The importance of efficient and reliable model inversion frameworks can hardly be overstated, considering that they establish a direct connection between models and the real world. Without it, the most advanced model predictions might lack physical meaning and, consequently, be useless for their intended applications.

Bayesian model inversion is one way to formalize this problem (Jaynes, 2003; Gelman et al., 2014). It is based on Bayesian inference and poses the problem in a probabilistic setting by capitalizing on Bayes’ theorem. In this setting a so-called prior (i.e., before observations) probability distribution about the model parameters is updated to a so-called posterior (i.e., after observations) distribution. The posterior distribution is the probability distribution of the input parameters conditioned on the available observations, and the main outcome of the Bayesian inversion process.

In Bayesian model inversion, the connection between the model output and the observations is established through a probabilistic discrepancy model. This model, which is a function of the input parameters, leads to the so-called likelihood function. The specific form of the likelihood function depends on the problem at hand, but typically it has a global maximum for the input parameters with the model output that is closest to the available observations (w.r.t. some metric), and rapidly goes to zero with increasing distance to those parameters.

Analytical expressions for the posterior distribution can only be found in few academic examples (e.g., conjugate priors with a linear forward model, Bishop (2006); Gelman et al. (2014)). In general model inversion problems, such analytical solutions are not available though. Instead, it is common practice to resort to sampling methods to generate a sample distributed according to the posterior distribution. The family of Markov chain Monte Carlo (MCMC) algorithms are particularly suitable for generating such a posterior sample (Beck and Au, 2002; Robert and Casella, 2004).

While MCMC and its extensions are extensively used in model inversion, and new algorithms are continuously being developed (Haario et al., 2001; Ching and Chen, 2007; Goodman and Weare, 2010; Neal, 2011), it has a few notable shortcomings that hinder its application in many practical cases. It is well known that there are no robust convergence criteria for MCMC algorithms, and that their performance is particularly sensitive to their tuning parameters. Additionally, samples generated by MCMC algorithms are often highly correlated, thus requiring extensive heuristic post-processing and empirical rules (Gelman and Rubin, 1992; Brooks and Gelman, 1998). MCMC algorithms are also in general not well suited for sampling multimodal posterior distributions.

When considering complex engineering scenarios, the models subject to inversion are often computationally expensive. Because MCMC algorithms usually require a significant number of forward model evaluations, it has been proposed to accelerate the procedure by using surrogate models in lieu of the original models. These surrogate models are either constructed non-adaptively before sampling from the posterior distribution (Marzouk et al., 2007; Marzouk and Xiu, 2009) or adaptively during the sampling procedure (Li and Marzouk, 2014; Birolleau et al., 2014; Cui et al., 2014; Conrad et al., 2016; Conrad et al., 2018; Yan and Zhou, 2019). Adaptive techniques can be of great benefit with posterior distributions that are concentrated in a small subspace of the prior domain, as the surrogate only needs to be accurate near high density areas of the posterior distribution.

Polynomial chaos expansions (PCE) are a widely used surrogate modelling technique based on expanding the forward model onto a suitable polynomial basis (Ghanem and Spanos, 1991; Xiu and Karniadakis, 2002). In other words, it provides a spectral representation of the computational forward model. Thanks to the introduction of sparse regression (see, e.g. Blatman and Sudret (2011)), its computation has become feasible even in the presence of complex and computationally expensive engineering models. This technique has been successfully used in conjunction with MCMC to reduce the total computational costs associated with sampling from the posterior distribution (Marzouk et al., 2007; Wagner et al., 2020).

Alternative approaches to compute the posterior distribution or its statistics include the Laplace approximation at a posterior mode (Tierney and Kadane, 1986; Tierney et al., 1989b, a), approximate Bayesian computations (ABC) (Marin et al., 2012; Sisson et al., 2018), optimal transport approaches (El Moselhy and Marzouk, 2012; Parno, 2015; Marzouk et al., 2016) and embarrassingly parallel quasi-Monte Carlo sampling (Dick et al., 2017; Gantner and Peters, 2018).

Stochastic spectral embedding (SSE) is a metamodelling technique suitable for approximating functions with complex localized features recently developed in Marelli et al. (2021). In this paper we propose to extend this technique with an ad-hoc adaptive sample enrichment strategy that makes it suitable to efficiently approximate likelihood functions in Bayesian model inversion problems. This method can be seen as a generalization of the previously proposed spectral likelihood expansions (SLE) approach (Nagel and Sudret, 2016).

Due to its deep connection to SLE, we call the application of SSE to likelihood functions stochastic spectral likelihood embedding (SSLE). We show that, due to its local spectral characteristics, this approach allows us to analytically derive expressions for the posterior marginals and general posterior moments by post-processing its expansion coefficients.

The paper is organized as follows: In Section 2 we establish the basics of Bayesian inference and particularly Bayesian model inversion. We then give an introduction into spectral function decomposition with a focus on polynomial chaos expansions and their application to likelihood functions (SLE) in Section 3. In Section 4 we present the main contribution of the paper, namely the derivation of Bayesian posterior quantities of interest through SSLE and the extension of the SSE algorithm with an adaptive sampling strategy. Finally, in Section 5 we showcase the performance of our approach on three case studies of varying complexity.

2 Model inversion

The problem of model inversion occurs whenever the predictions of a model are to be brought into agreement with available observations or data. This is achieved by properly adjusting a set of input parameters of the model. The goal of inversion can be twofold: on the one hand the inferred input parameters might be used to predict new realizations of the model output. On the other hand, the inferred input parameters might be the main interest. Model inversion is a common problem in many engineering disciplines, that in some cases is still routinely solved manually, i.e. by simply changing the input parameters until some, often qualitative, goodness-of-fit criterion is met. More quantitative inversion approaches aim at automatizing this process, by establishing a metric (e.g., L2L^{2}-distance) between the data and the model response, which is then minimized through suitable optimization algorithms.

While such approaches can often be used in practical applications, they tend not to provide measures of the uncertainties associated with the inferred model input or predictions. These uncertainties are useful in judging the accuracy of the inversion, as well as indicating non-informative measurements. In fact, the lack of uncertainty quantification in the context of model inversion can lead to erroneous results that have far-reaching consequences in subsequent applications. One approach to consider uncertainties in inverse problems is the Bayesian framework for model inversion that will be presented hereinafter.

2.1 Bayesian inference

Consider some non-observable parameters 𝑿𝒟𝑿\bm{X}\in{\mathcal{D}}_{\bm{X}} and the observables 𝒀𝒟𝒀\bm{Y}\in{\mathcal{D}}_{\bm{Y}}. Furthermore, let 𝒴={𝒚(1),,𝒚(N)}{{\mathcal{Y}}=\{\bm{y}^{(1)},\dots,\bm{y}^{(N)}\}} be a set of NN measurements, i.e., noisy observations of a set of realizations of 𝒀\bm{Y}. Statistical inference consists in drawing conclusions about 𝑿\bm{X} using the information from 𝒴{\mathcal{Y}} (Gelman et al., 2014). These measurements can be direct observations of the parameters (𝒀=𝑿\bm{Y}=\bm{X}) or some quantities indirectly related to 𝑿\bm{X} through a function or model :𝒟𝑿𝒟𝒀{\mathcal{M}}:{\mathcal{D}}_{\bm{X}}\to{\mathcal{D}}_{\bm{Y}}. One way to conduct this inference is through Bayes’ theorem of conditional probabilities, a process known as Bayesian inference.

Denoting by π()\pi(\cdot) a probability density function (PDF) and by π(|𝒙)\pi(\cdot|\bm{x}) a PDF conditioned on 𝒙\bm{x}, Bayes’ theorem can be written as

π(𝒙|𝒴)=π(𝒴|𝒙)π(𝒙)π(𝒴),\pi(\bm{x}|{\mathcal{Y}})=\frac{\pi({\mathcal{Y}}|\bm{x})\pi(\bm{x})}{\pi({\mathcal{Y}})}, (1)

where π(𝒙)\pi(\bm{x}) is known as the prior distribution of the parameters, i.e., the distribution of 𝑿\bm{X} before observing the data 𝒴{\mathcal{Y}}. The conditional distribution π(𝒴|𝒙)\pi({\mathcal{Y}}|\bm{x}), known as likelihood, establishes a connection between the observations 𝒴{\mathcal{Y}} and a realization of the parameters 𝑿=𝒙\bm{X}=\bm{x}. For a given realization 𝒙\bm{x}, it returns the probability density of observing the data 𝒴{\mathcal{Y}}. Under the common assumption of independence between individual observations, {𝒚(i)}i=1N\{\bm{y}^{(i)}\}_{i=1}^{N}, the likelihood function takes the form:

:𝒙(𝒙;𝒴)=defπ(𝒴|𝒙)=i=1Nπ(𝒚(i)|𝒙).{\mathcal{L}}:\bm{x}\mapsto{\mathcal{L}}(\bm{x};\,{\mathcal{Y}})\stackrel{{\scriptstyle\text{def}}}{{=}}\pi({\mathcal{Y}}|\bm{x})=\prod_{i=1}^{N}\pi(\bm{y}^{(i)}|\bm{x}). (2)

The likelihood function is a map 𝒟𝑿+{\mathcal{D}}_{\bm{X}}\rightarrow\mathbb{R}_{+}, and it attains its maximum for the parameter set with the highest probability of yielding 𝒴{\mathcal{Y}}. With this, Bayes’ theorem from Eq. (1) can be rewritten as:

π(𝒙|𝒴)=(𝒙;𝒴)π(𝒙)Z,withZ=𝒟𝑿(𝒙;𝒴)π(𝒙)d𝒙,\pi(\bm{x}|{\mathcal{Y}})=\frac{{\mathcal{L}}(\bm{x};\,{\mathcal{Y}})\pi(\bm{x})}{Z},\quad\text{with}\quad Z=\int_{{\mathcal{D}}_{\bm{X}}}{\mathcal{L}}(\bm{x};\,{\mathcal{Y}})\pi(\bm{x})\,{\rm d}\bm{x}, (3)

where ZZ is a normalizing constant often called evidence or marginal likelihood. On the left-hand side, π(𝒙|𝒴)\pi(\bm{x}|{\mathcal{Y}}) is the posterior PDF, i.e., the distribution of 𝑿\bm{X} after observing data 𝒴{\mathcal{Y}}. In this sense, Bayes’ theorem establishes a general expression for updating the prior distribution using a likelihood function to incorporate information from the data.

2.2 Bayesian model inversion

Bayesian model inversion describes the application of the Bayesian inference framework to the problem of model inversion (Beck and Katafygiotis, 1998; Kennedy and O’Hagan, 2001; Jaynes, 2003; Tarantola, 2005). The two main ingredients needed to infer model parameters within the Bayesian framework are a prior distribution π(𝒙)\pi(\bm{x}) of the model parameters and a likelihood function {\mathcal{L}}. In practical applications, prior information about the model parameters is often readily available. Typical sources of such information are physical parameter constraints or expert knowledge. Additionally, prior inversion attempts can serve as guidelines to assign informative prior distributions. In cases where no prior information about the parameters is available, so-called non-informative or invariant prior distributions (Jeffreys, 1946; Harney, 2016) can also be assigned. The likelihood function serves instead as the link between model parameters 𝑿\bm{X} and observations of the model output 𝒴{\mathcal{Y}}. To connect these two quantities, it is necessary to choose a so-called discrepancy model that gives the relative probability that the model response to a realization of 𝑿=𝒙\bm{X}=\bm{x} describes the observations. One common assumption for this probabilistic model is that the measurements are perturbed by a Gaussian additive discrepancy term 𝑬𝒩(𝜺|𝟎,𝚺)\bm{E}\sim{\mathcal{N}}(\bm{\varepsilon}|\bm{0},\bm{\Sigma}), with covariance matrix 𝚺\bm{\Sigma}. For a single measurement 𝒚(i)\bm{y}^{(i)} it reads:

𝒚(i)=(𝒙)+𝜺.\bm{y}^{(i)}={\mathcal{M}}(\bm{x})+\bm{\varepsilon}. (4)

This discrepancy between the model output (𝑿){\mathcal{M}}(\bm{X}) and the observables 𝒀\bm{Y} can result from measurement error or model inadequacies. By using this additive discrepancy model, the distribution of the observables conditioned on the parameters 𝒀|𝒙\bm{Y}|\bm{x} is written as:

π(𝒚(i)|𝒙)=𝒩(𝒚(i)|(𝒙),𝚺),\pi(\bm{y}^{(i)}|\bm{x})={\mathcal{N}}(\bm{y}^{(i)}|{\mathcal{M}}(\bm{x}),\bm{\Sigma}), (5)

where 𝒩(|𝝁,𝚺){\mathcal{N}}(\cdot|\bm{\mu},\bm{\Sigma}) denotes the multivariate Gaussian PDF with mean value 𝝁\bm{\mu} and covariance matrix 𝚺\bm{\Sigma}. The likelihood function {\mathcal{L}} is then constructed using this probabilistic model π(𝒚(i)|𝒙)\pi(\bm{y}^{(i)}|\bm{x}) and Eq. (2). For a given set of measurements 𝒴{\mathcal{Y}} it thus reads:

(𝒙;𝒴)=defi=1N𝒩(𝒚(i)|(𝒙),𝚺).{\mathcal{L}}(\bm{x};\,{\mathcal{Y}})\stackrel{{\scriptstyle\text{def}}}{{=}}\prod_{i=1}^{N}{\mathcal{N}}(\bm{y}^{(i)}|{\mathcal{M}}(\bm{x}),\bm{\Sigma}). (6)

With the fully specified Bayesian model inversion problem, Eq. (3) directly gives the posterior distribution of the model parameters π(𝒙|𝒴)\pi(\bm{x}|{\mathcal{Y}}). In the setting of model inversion, the posterior distribution represents therefore the state of belief about the true data-generating model parameters, considering all available information: computational forward model, discrepancy model and measurement data (Beck and Katafygiotis, 1998; Jaynes, 2003).

Often, the ultimate goal of model inversion is to provide a set of inferred parameters, with associate confidence measures/intervals. This is often achieved by computing posterior statistics (e.g., moments, mode, etc.). Propagating the posterior through secondary models is also of interest. So-called quantites of interest (QoI) can be expressed by calculating the posterior expectation of suitable functions of the parameters h(𝒙):Mh(\bm{x}):\mathbb{R}^{M}\to\mathbb{R}, with 𝑿|𝒴π(𝒙|𝒴)\bm{X}|{\mathcal{Y}}\sim\pi(\bm{x}|{\mathcal{Y}}), as in:

𝔼[h(𝑿)|𝒴]=𝒟𝑿|𝒴h(𝒙)π(𝒙|𝒴)d𝒙.{\mathbb{E}}\left[h(\bm{X})|{\mathcal{Y}}\right]=\int_{{\mathcal{D}}_{\bm{X}|{\mathcal{Y}}}}h(\bm{x})\pi(\bm{x}|{\mathcal{Y}})\,{\rm d}\bm{x}. (7)

Depending on hh, this formulation encompasses posterior moments (h(𝒙)=xih(\bm{x})=x_{i} or h(𝒙)=(xi𝔼[Xi])2h(\bm{x})=(x_{i}-{\mathbb{E}}\left[X_{i}\right])^{2} for the first and second moments, respectively), posterior covariance (h(𝒙)=xixj𝔼[Xi]𝔼[Xj]h(\bm{x})=x_{i}x_{j}-{\mathbb{E}}\left[X_{i}\right]{\mathbb{E}}\left[X_{j}\right]) or expectations of secondary models (h(𝒙)=(𝒙)h(\bm{x})={\mathcal{M}}^{\star}(\bm{x})).

3 Spectral function decomposition

To pose a Bayesian inversion problem, the specification of a prior distribution and a likelihood function described in the previous section is sufficient. Its solution, however, is not available in closed form in the general case.

Spectral likelihood expansion (SLE) is a recently proposed method that aims at solving the Bayesian inversion problem by finding a polynomial chaos expansion (PCE) of the likelihood function in a basis orthogonal w.r.t. the prior distribution (Nagel and Sudret, 2016). This representation allows one to derive analytical expressions for the evidence ZZ, the posterior distribution, the posterior marginals, and many types of QoIs, including the posterior moments.

We offer here a brief introduction to regression-based, sparse PCE before introducing SLE, but refer the interested reader to more exhaustive resources on PCE (Ghanem and Spanos, 1991; Xiu and Karniadakis, 2002) and sparse PCE (Xiu, 2010; Blatman and Sudret, 2010, 2011).

Let us consider a random variable 𝑿\bm{X} with independent components {Xi,i=1,,M}\{X_{i},i=1,\dots,M\} and associated probability density functions πi(xi)\pi_{i}(x_{i}) so that π(𝒙)=i=1Mπi(xi).\pi(\bm{x})=\prod_{i=1}^{M}\pi_{i}(x_{i}). Assume further that :𝒟𝑿=i=1M𝒟XiM\mathcal{M}:{\mathcal{D}}_{\bm{X}}=\prod_{i=1}^{M}{\mathcal{D}}_{X_{i}}\subseteq\mathbb{R}^{M}\to\mathbb{R} is a scalar function of 𝑿\bm{X} which fulfills the finite variance condition (𝔼[(𝑿)2]<+{\mathbb{E}}\left[\mathcal{M}(\bm{X})^{2}\right]<+\infty). Then it is possible to find a so-called truncated polynomial chaos approximation of \mathcal{M} such that

(𝑿)PCE(𝑿)=def𝜶𝒜a𝜶Ψ𝜶(𝑿)\mathcal{M}(\bm{X})\approx\mathcal{M}_{\mathrm{PCE}}(\bm{X})\stackrel{{\scriptstyle\text{def}}}{{=}}\sum_{\bm{\alpha}\in{\mathcal{A}}}a_{\bm{\alpha}}\Psi_{\bm{\alpha}}(\bm{X}) (8)

where 𝜶\bm{\alpha} is an MM-tuple (α1,,αM)M(\alpha_{1},\dots,\alpha_{M})\in\mathbb{N}^{M} and 𝒜M{\mathcal{A}}\subset\mathbb{N}^{M}. For most parametric distributions, well-known classical orthonormal polynomials {Ψ𝜶}𝜶M\{\Psi_{\bm{\alpha}}\}_{\bm{\alpha}\in\mathbb{N}^{M}} satisfy the necessary orthonormality condition w.r.t. π(𝒙)\pi(\bm{x}) (Xiu and Karniadakis, 2002). For more general distributions, arbitrary orthonormal polynomials can be constructed numerically through the Stieltjes procedure (Gautschi, 2004). If additionally, 𝒜{\mathcal{A}} is a sparse subset of M\mathbb{N}^{M}, the truncated expansion in Eq. (8) is called a sparse PCE.

In this contribution, we restrict the discussion to independent inputs, because it is computationally challenging, albeit possible, to construct an orthogonal basis for dependent inputs (e.g. through Gram-Schmidt decomposition of ). Furthermore, Torre et al. (2019) demonstrated that for purely predictive purposes, ignoring input dependence can significantly improve the predictive performance of PCE, at the cost of losing basis orthonormality. The latter, however, is required to derive analytical post-processing quantities such as the moments of the posterior distribution (see Sections 3.1 and 4.1).

There exist different algorithms to produce a sparse PCE in practice, i.e. select a sparse basis 𝒜{\mathcal{A}} and compute the corresponding coefficients. A powerful class of methods are regression-based approaches that rely on an initial input sample 𝒳\mathcal{X}, called experimental design, and corresponding model evaluations (𝒳)\mathcal{M}(\mathcal{X}) (See, e.g. Lüthen et al. (2021) for a recent survey). Additionally, it is possible to design adaptive algorithms that choose the truncated basis size (Blatman and Sudret, 2011; Jakeman et al., 2015).

To assess the accuracy of PCE, the so-called generalization error 𝔼[((𝑿)PCE(𝑿))2]\mathbb{E}\left[(\mathcal{M}(\bm{X})-\mathcal{M}_{\mathrm{PCE}}(\bm{X}))^{2}\right] shall be evaluated. A robust generalization error estimator is given by the leave-one-out (LOO) cross validation technique. This estimator is obtained by

εLOO=1Ki=1K((𝒙(i))PCEi(𝒙(i)))2,\varepsilon_{\mathrm{LOO}}=\frac{1}{K}\sum_{i=1}^{K}\left(\mathcal{M}(\bm{x}^{(i)})-\mathcal{M}_{\mathrm{PCE}}^{\sim i}(\bm{x}^{(i)})\right)^{2}, (9)

where PCEi\mathcal{M}_{\mathrm{PCE}}^{\sim i} is constructed by leaving out the ii-th point from the experimental design 𝒳{\mathcal{X}}. For methods based on linear regression, it can be shown (Chapelle et al., 2002; Blatman and Sudret, 2010) that the LOO error is available analytically by post-processing the regressor matrix.

3.1 Spectral likelihood expansions

The idea of SLE is to use sparse PCE to find a spectral representation of the likelihood function {\mathcal{L}} occurring in Bayesian model inversion problems (see Eq. (2)). We present here a brief introduction to the method and the main results of Nagel and Sudret (2016).

Likelihood functions can be seen as scalar functions of the input random vector 𝑿π(𝒙)\bm{X}\sim\pi(\bm{x}). In this work we assume priors of the type π(𝒙)=i=1Mπi(xi)\pi(\bm{x})=\prod_{i=1}^{M}\pi_{i}(x_{i}), i.e. with independent marginals, their spectral expansion then reads:

(𝑿)SLE(𝑿)=def𝜶𝒜a𝜶Ψ𝜶(𝑿),{\mathcal{L}}(\bm{X})\approx{\mathcal{L}}_{\mathrm{SLE}}(\bm{X})\stackrel{{\scriptstyle\text{def}}}{{=}}\sum_{\bm{\alpha}\in{\mathcal{A}}}a_{\bm{\alpha}}\Psi_{\bm{\alpha}}(\bm{X}), (10)

where the explicit dependence on 𝒴{\mathcal{Y}} was dropped for notational simplicity.

Upon computing the basis and coefficients in Eq. (8), the solution to the inverse problem is converted to merely post-processing the coefficients a𝜶a_{\bm{\alpha}}. The following expressions can be derived for the individual quantities:

Evidence

The evidence emerges as the coefficient of the constant polynomial a𝟎a_{\bm{0}}

Z=𝒟𝑿(𝒙)π(𝒙)d𝒙SLE,1π=a𝟎.Z=\int_{{\mathcal{D}}_{\bm{X}}}{\mathcal{L}}(\bm{x})\pi(\bm{x})\,{\rm d}\bm{x}\approx\left<{\mathcal{L}}_{\mathrm{SLE}},1\right>_{\pi}=a_{\bm{0}}. (11)
Posterior

Upon computing the evidence ZZ, the posterior can be evaluated directly through

π(𝒙|𝒴)SLE(𝒙)π(𝒙)Z=π(𝒙)a𝟎𝜶𝒜a𝜶Ψ𝜶(𝒙).\pi(\bm{x}|{\mathcal{Y}})\approx\frac{{\mathcal{L}}_{\mathrm{SLE}}(\bm{x})\pi(\bm{x})}{Z}=\frac{\pi(\bm{x})}{a_{\bm{0}}}\sum_{\bm{\alpha}\in{\mathcal{A}}}a_{\bm{\alpha}}\Psi_{\bm{\alpha}}(\bm{x}). (12)
Posterior marginals

We split the random vector 𝑿\bm{X} into two vectors 𝑿𝘂\bm{X}_{{\bm{\mathsf{u}}}} with components {Xi}i𝘂𝒟𝑿𝘂\{X_{i}\}_{i\in{\bm{\mathsf{u}}}}\in\mathcal{D}_{\bm{X}_{{\bm{\mathsf{u}}}}} and 𝑿𝘃\bm{X}_{{\bm{\mathsf{v}}}} with components {Xi}i𝘃𝒟𝑿𝘃\{X_{i}\}_{i\in{\bm{\mathsf{v}}}}\in\mathcal{D}_{\bm{X}_{{\bm{\mathsf{v}}}}}, where 𝘂{\bm{\mathsf{u}}} and 𝘃{\bm{\mathsf{v}}} are two non-empty disjoint index sets such that 𝘂𝘃={1,,M}{\bm{\mathsf{u}}}\cup{\bm{\mathsf{v}}}=\{1,\dots,M\}. Denote further by π𝘂(𝒙𝘂)=defi𝘂πi(xi)\pi_{{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}})\stackrel{{\scriptstyle\text{def}}}{{=}}\prod_{i\in{\bm{\mathsf{u}}}}\pi_{i}(x_{i}) and π𝘃(𝒙𝘃)=defi𝘃πi(xi)\pi_{{\bm{\mathsf{v}}}}(\bm{x}_{{\bm{\mathsf{v}}}})\stackrel{{\scriptstyle\text{def}}}{{=}}\prod_{i\in{\bm{\mathsf{v}}}}\pi_{i}(x_{i}) the prior marginal density functions of 𝑿𝘂\bm{X}_{{\bm{\mathsf{u}}}} and 𝑿𝘃\bm{X}_{{\bm{\mathsf{v}}}} respectively. The posterior marginals then read:

π𝘂(𝒙𝘂|𝒴)=𝒟𝑿𝘃π(𝒙|𝒴)d𝒙𝘃π𝘂(𝒙𝘂)a𝟎𝜶𝒜𝘃=0a𝜶Ψ𝜶(𝒙𝘂),\pi_{{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}}|{\mathcal{Y}})=\int_{{\mathcal{D}}_{\bm{X}_{{\bm{\mathsf{v}}}}}}\pi(\bm{x}|{\mathcal{Y}})\,{\rm d}\bm{x}_{{\bm{\mathsf{v}}}}\approx\frac{\pi_{{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}})}{a_{\bm{0}}}\sum_{\bm{\alpha}\in{\mathcal{A}}_{{\bm{\mathsf{v}}}=0}}a_{\bm{\alpha}}\Psi_{\bm{\alpha}}(\bm{x}_{{\bm{\mathsf{u}}}}), (13)

where 𝒜𝘃=0={𝜶𝒜:αi=0i𝘃}{\mathcal{A}}_{{\bm{\mathsf{v}}}=0}=\{\bm{\alpha}\in\mathcal{A}:\alpha_{i}=0\Leftrightarrow i\in{\bm{\mathsf{v}}}\}. The series in the above equation constitutes a subexpansion that contains non-constant polynomials only in the directions i𝘂i\in{\bm{\mathsf{u}}}.

Quantities of interest

Finally, it is also possible to analytically compute posterior expectations of functions that admit a polynomial chaos expansion in the same basis of the form h(𝑿)𝜶𝒜b𝜶Ψ𝜶(𝑿)h(\bm{X})\approx\sum_{\bm{\alpha}\in{\mathcal{A}}}b_{\bm{\alpha}}\Psi_{\bm{\alpha}}(\bm{X}). Eq. (7) then reduces to the spectral product:

𝔼[h(𝑿)|𝒴]=1a𝟎𝜶𝒜a𝜶b𝜶.{\mathbb{E}}\left[h(\bm{X})|{\mathcal{Y}}\right]=\frac{1}{a_{\bm{0}}}\sum_{\bm{\alpha}\in{\mathcal{A}}}a_{\bm{\alpha}}b_{\bm{\alpha}}. (14)

The quality of these results depends only on the approximation error introduced in Eq. (10). The latter, in turn, depends mainly on the chosen PCE truncation strategy (Blatman and Sudret, 2011; Nagel and Sudret, 2016) and the number of points used to compute the coefficients (i.e., the experimental design). It is known that likelihood functions typically have quasi-compact supports (i.e., (𝑿)0{\mathcal{L}}(\bm{X})\approx 0 on a majority of 𝒟𝑿{\mathcal{D}}_{\bm{X}}). Such functions require a very high polynomial degree to be approximated accurately, which in turn can lead to the need for prohibitively large experimental designs.

4 Stochastic spectral embedding

Stochastic spectral embedding (SSE) is a multi-level approach to surrogate modeling originally proposed in Marelli et al. (2021). It attempts to approximate a given square-integrable function \mathcal{M} with independent inputs 𝑿\bm{X} through

SSE(𝑿)=k𝒦𝟏𝒟𝑿k(𝑿)^Sk(𝑿),\mathcal{M}\approx\mathcal{M}_{\text{SSE}}(\bm{X})=\sum\limits_{k\in\mathcal{K}}\bm{1}_{\mathcal{D}_{\bm{X}}^{k}}(\bm{X})\,\widehat{\mathcal{R}}_{S}^{k}(\bm{X}), (15)

where 𝒦2\mathcal{K}\subseteq\mathbb{N}^{2} is a set of multi-indices with elements k=(,p)k=(\ell,p) for which =0,,L\ell=0,\dots,L and p=1,,Pp=1,\dots,P_{\ell} where LL is the number of levels and PP_{\ell} is the number of subdomains at a specific level \ell. We call ^Sk(𝑿)\widehat{\mathcal{R}}_{S}^{k}(\bm{X}) a residual expansion given by

^Sk(𝑿)=𝜶𝒜ka𝜶kΨ𝜶k(𝑿).\widehat{\mathcal{R}}^{k}_{S}(\bm{X})=\sum\limits_{\bm{\alpha}\in\mathcal{A}^{k}}a_{\bm{\alpha}}^{k}\Psi_{\bm{\alpha}}^{k}(\bm{X}). (16)

In the present paper the term j𝒜kajkΨjk(𝑿)\sum_{j\in\mathcal{A}^{k}}a_{j}^{k}\Psi_{j}^{k}(\bm{X}) denotes a polynomial chaos expansion (see Eq. (8)) constructed in the subdomain 𝒟𝑿k{\mathcal{D}}_{\bm{X}}^{k}, but in principle it can refer to any spectral expansion (e.g., Fourier series). A schematic representation of the summation in Eq. (15) is given in Figure 2. The detailed notation and the algorithm to sequentially construct an SSE are given in the sequel.

4.1 Stochastic spectral likelihood embedding

Viewing the likelihood as a function of a random variable 𝑿\bm{X} with independent marginals, we can directly use Eq. (15) to write down its SSLE representation

(𝑿)SSLE(𝑿)=defk𝒦𝟏𝒟𝑿k(𝑿)^Sk(𝑿),{\mathcal{L}}(\bm{X})\approx{\mathcal{L}}_{\mathrm{SSLE}}(\bm{X})\stackrel{{\scriptstyle\text{def}}}{{=}}\sum\limits_{k\in\mathcal{K}}\bm{1}_{\mathcal{D}_{\bm{X}}^{k}}(\bm{X})\,\widehat{\mathcal{R}}_{S}^{k}(\bm{X}), (17)

where the variable 𝑿\bm{X} is distributed according to the prior distribution π(𝒙)\pi(\bm{x}) and, consequently, the local basis used to compute ^Sk(𝑿)\widehat{\mathcal{R}}_{S}^{k}(\bm{X}) is orthonormal w.r.t. that distribution.

Due to the local spectral properties of the residual expansions, the SSLE representation of the likelihood function retains all of the post-processing properties of SLE (Section 3.1):

Evidence

The normalization constant ZZ emerges as the sum of the constant polynomial coefficients weighted by the prior mass:

Z=k𝒦𝜶𝒜ka𝜶k𝒟𝑿kΨ𝜶k(𝒙)π(𝒙)d𝒙=k𝒦𝒱ka𝟎k,where𝒱k=𝒟𝑿kπ(𝒙)d𝒙.Z=\sum_{k\in{\mathcal{K}}}\sum_{\bm{\alpha}\in{\mathcal{A}}^{k}}a_{\bm{\alpha}}^{k}\int_{{\mathcal{D}}_{\bm{X}}^{k}}\Psi_{\bm{\alpha}}^{k}(\bm{x})\pi(\bm{x})\,{\rm d}\bm{x}=\sum_{k\in{\mathcal{K}}}\mathcal{V}^{k}a_{\bm{0}}^{k},\quad\text{where}\quad\mathcal{V}^{k}=\int_{{\mathcal{D}}_{\bm{X}}^{k}}\pi(\bm{x})\,{\rm d}\bm{x}. (18)
Posterior

This allows us to write the posterior as

π(𝒙|𝒴)SSLE(𝒙)π(𝒙)Z=π(𝒙)k𝒦𝒱ka𝟎kk𝒦𝟏𝒟𝑿k(𝒙)^Sk(𝒙).\pi(\bm{x}|{\mathcal{Y}})\approx\frac{{\mathcal{L}}_{\mathrm{SSLE}}(\bm{x})\pi(\bm{x})}{Z}=\frac{\pi(\bm{x})}{\sum_{k\in{\mathcal{K}}}\mathcal{V}^{k}a_{\bm{0}}^{k}}\sum_{k\in{\mathcal{K}}}\bm{1}_{\mathcal{D}_{\bm{X}}^{k}}(\bm{x})\widehat{\mathcal{R}}^{k}_{S}(\bm{x}). (19)
Posterior marginal

Utilizing again the disjoint sets 𝘂{\bm{\mathsf{u}}} and 𝘃{\bm{\mathsf{v}}} from Eq. (13) it is also possible to analytically derive posterior marginal PDFs as

π𝘂(𝒙𝘂|𝒴)=𝒟𝑿𝘃π(𝒙|𝒴)d𝒙𝘃π𝘂(𝒙𝘂)k𝒦𝒱ka𝟎kk𝒦𝟏𝒟𝑿𝘂k(𝒙𝘂)^S,𝘂k(𝒙𝘂)𝒱𝘃k\pi_{{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}}|{\mathcal{Y}})=\int_{{\mathcal{D}}_{\bm{X}_{{\bm{\mathsf{v}}}}}}\pi(\bm{x}|{\mathcal{Y}})\,{\rm d}\bm{x}_{{\bm{\mathsf{v}}}}\approx\frac{\pi_{{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}})}{\sum_{k\in{\mathcal{K}}}\mathcal{V}^{k}a_{\bm{0}}^{k}}\sum_{k\in{\mathcal{K}}}\bm{1}_{{\mathcal{D}}_{\bm{X}_{{\bm{\mathsf{u}}}}}^{k}}(\bm{x}_{{\bm{\mathsf{u}}}})\widehat{\mathcal{R}}^{k}_{S,{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}})\mathcal{V}^{k}_{{\bm{\mathsf{v}}}} (20)

where

^S,𝘂k(𝒙𝘂)=𝜶𝒜𝘃=0ka𝜶kΨ𝜶k(𝒙𝘂)and𝒱𝘃k=𝒟𝑿𝘃kπ𝘃(𝒙𝘃)d𝒙𝘃.\widehat{\mathcal{R}}^{k}_{S,{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}})=\sum_{\bm{\alpha}\in{\mathcal{A}}_{{\bm{\mathsf{v}}}=0}^{k}}a_{\bm{\alpha}}^{k}\Psi_{\bm{\alpha}}^{k}(\bm{x}_{{\bm{\mathsf{u}}}})\quad\text{and}\quad\mathcal{V}^{k}_{{\bm{\mathsf{v}}}}=\int_{{\mathcal{D}}_{\bm{X}_{{\bm{\mathsf{v}}}}}^{k}}\pi_{{\bm{\mathsf{v}}}}(\bm{x}_{{\bm{\mathsf{v}}}})\,{\rm d}\bm{x}_{{\bm{\mathsf{v}}}}. (21)

^S,𝘂k(𝒙𝘂)\widehat{\mathcal{R}}^{k}_{S,{\bm{\mathsf{u}}}}(\bm{x}_{{\bm{\mathsf{u}}}}) is a subexpansion of ^Sk(𝒙)\widehat{\mathcal{R}}^{k}_{S}(\bm{x}) that contains only non-constant polynomials in the directions i𝘂i\in{\bm{\mathsf{u}}}. Note that, as we assumed that the prior distribution has independent components, the constants 𝒱k\mathcal{V}^{k} and 𝒱𝘃k\mathcal{V}^{k}_{{\bm{\mathsf{v}}}} are obtained as products of univariate integrals which are available analytically from the prior marginal cumulative distribution functions (CDFs).

Quantities of interest

Expected values of function h(𝒙)=𝜶𝒜kb𝜶kΨ𝜶k(𝒙)h(\bm{x})=\sum_{\bm{\alpha}\in{\mathcal{A}}^{k}}b_{\bm{\alpha}}^{k}\Psi_{\bm{\alpha}}^{k}(\bm{x}) for k𝒦k\in{\mathcal{K}} under the posterior can be approximated by:

𝔼[h(𝑿)|𝒴]=𝒟𝑿h(𝒙)π(𝒙|𝒴)d𝒙=1Zk𝒦𝜶𝒜ka𝜶k𝒟𝑿kh(𝒙)Ψ𝜶k(𝒙)π(𝒙)d𝒙=1Zk𝒦𝜶𝒜ka𝜶kb𝜶k,\displaystyle\begin{split}\mathbb{E}\left[h(\bm{X})|{\mathcal{Y}}\right]&=\int_{{\mathcal{D}}_{\bm{X}}}h(\bm{x})\pi(\bm{x}|{\mathcal{Y}})\,{\rm d}\bm{x}\\ &=\frac{1}{Z}\sum_{k\in{\mathcal{K}}}\sum_{\bm{\alpha}\in{\mathcal{A}}^{k}}a_{\bm{\alpha}}^{k}\int_{{\mathcal{D}}_{\bm{X}}^{k}}h(\bm{x})\Psi_{\bm{\alpha}}^{k}(\bm{x})\pi(\bm{x})\,{\rm d}\bm{x}\\ &=\frac{1}{Z}\sum_{k\in{\mathcal{K}}}\sum_{\bm{\alpha}\in{\mathcal{A}}^{k}}a_{\bm{\alpha}}^{k}b_{\bm{\alpha}}^{k},\end{split} (22)

where b𝜶kb_{\bm{\alpha}}^{k} are the coefficients of the PCE of hh in the card(𝒦)\mathrm{card}({\mathcal{K}}) bases {Ψ𝜶k}𝜶𝒜k\{\Psi_{\bm{\alpha}}^{k}\}_{\bm{\alpha}\in{\mathcal{A}}^{k}}. This can also be used for computing posterior moments like mean, variance or covariance.

These expressions can be seen as a generalization of the ones for SLE detailed in Section 3.1. For a single-level global expansion (i.e., card(𝒦)=1\mathrm{card}({\mathcal{K}})=1) and consequently 𝒱(0,1)=1\mathcal{V}^{(0,1)}=1, they are identical.

4.2 Modifications to the original algorithm

The original algorithm for computing an SSE was presented in Marelli et al. (2021). It recursively partitions the input domain 𝒟𝑿\mathcal{D}_{\bm{X}} and constructs truncated expansions of the residual. We reproduce it below for reference but replace the model \mathcal{M} with the likelihood function \mathcal{L}. We further simplify the algorithm by choosing a partitioning strategy with NS=2N_{S}=2.

  1. 1.

    Initialization:

    1. (a)

      =0\ell=0, p=1p=1

    2. (b)

      𝒟𝑿,p=𝒟𝑿\mathcal{D}_{\bm{X}}^{\ell,p}=\mathcal{D}_{\bm{X}}

    3. (c)

      (𝑿)=(𝑿){\mathcal{R}}^{\ell}(\bm{X})=\mathcal{L}(\bm{X})

  2. 2.

    For each subdomain 𝒟X,p\mathcal{D}_{\bm{X}}^{\ell,p}, p=1,,Pp=1,\cdots,P_{\ell}:

    1. (a)

      Calculate the truncated expansion ^S,p(𝑿,p)\widehat{\mathcal{R}}_{S}^{\ell,p}(\bm{X}^{\ell,p}) of the residual (𝑿,p){\mathcal{R}}^{\ell}(\bm{X}^{\ell,p}) in the current subdomain

    2. (b)

      Update the residual in the current subdomain +1(𝑿,p)=(𝑿,p)^S,p(𝑿,p){\mathcal{R}}^{\ell+1}(\bm{X}^{\ell,p})={\mathcal{R}}^{\ell}(\bm{X}^{\ell,p})-\widehat{\mathcal{R}}_{S}^{\ell,p}(\bm{X}^{\ell,p})

    3. (c)

      Split the current subdomain 𝒟𝑿,p\mathcal{D}_{\bm{X}}^{\ell,p} in 22 subdomains 𝒟𝑿+1,{s1,s2}\mathcal{D}_{\bm{X}}^{\ell+1,\{s_{1},s_{2}\}} based on a partitioning strategy

    4. (d)

      If <L\ell<L, +1\ell\leftarrow\ell+1, go back to 2a, otherwise terminate the algorithm

  3. 3.

    Termination

    1. (a)

      Return the full sequence of 𝒟𝑿,p\mathcal{D}_{\bm{X}}^{\ell,p} and ^S,p(𝑿,p)\widehat{\mathcal{R}}_{S}^{\ell,p}(\bm{X}^{\ell,p}) needed to compute Eq. (15).

In practice, the residual expansions ^S,p(𝑿,p)\widehat{\mathcal{R}}_{S}^{\ell,p}(\bm{X}^{\ell,p}) are computed using a fixed experimental design 𝒳\mathcal{X} and corresponding model evaluations (𝒳)\mathcal{L}(\mathcal{X}). The algorithm then only requires the specification of a partitioning strategy and a termination criterion, as detailed in Marelli et al. (2021).

Likelihood functions are typically characterized by a localized behaviour: Close to the data-generating parameters they peak while quickly decaying to 0 in the remainder of the prior domain. This means that in a majority of the domain the likelihood evaluation is non-informative. Directly applying the original algorithm is then expected to waste many likelihood evaluations.

We therefore modify the original algorithm by adding an adaptive sampling scheme (Section 4.2.1) that includes the termination criterion and introducing an improved partitioning strategy (Section 4.2.2) that is especially suitable for finding compact support features. The rationale for these modifications is presented next.

4.2.1 Adaptive sampling scheme

The proposed algorithm has two parameters: the experimental design size for the residual expansions NrefN_{\mathrm{ref}} and the final experimental design size corresponding to the available computational budget NEDN_{\mathrm{ED}}. At the initialization of the algorithm NrefN_{\mathrm{ref}} points are sampled as a first experimental design. At every further iteration, additional points are then sampled from the prior distribution. These samples are generated in a space-filling way (e.g. through latin hypercube sampling) in the newly created subdomains 𝒟𝑿+1,s\mathcal{D}_{\bm{X}}^{\ell+1,s} to have always exactly NrefN_{\mathrm{ref}} points available for constructing the residual expansions. The algorithm is terminated, once the computational budget NEDN_{\mathrm{ED}} has been exhausted.

Adding new samples points requires the evaluation of the likelihood function. Because multiple points are added at every iteration of the algorithm, this step can be easily performed simultaneously on multiple computational nodes.

At every step, the proposed algorithm chooses a single refinement domain from the set of unsplit, i.e. terminal domains, creates two new subdomains by splitting the refinement domain and constructs residual expansions after enriching the experimental design. The selection of this refinement domain is based on the error estimator k\mathcal{E}^{k} that is defined by

+1,s={ELOO+1,s𝒱+1,s,if^S+1,s,ELOO,s𝒱+1,s,otherwise.\mathcal{E}^{\ell+1,s}=\begin{cases}E^{\ell+1,s}_{\mathrm{LOO}}\mathcal{V}^{\ell+1,s},\quad&\text{if}\quad\exists~{}\widehat{\mathcal{R}}_{S}^{\ell+1,s},\\ E^{\ell,s}_{\mathrm{LOO}}\mathcal{V}^{\ell+1,s},\quad&\text{otherwise}.\end{cases} (23)

This estimator incorporates the subdomain size through the prior mass 𝒱+1,s\mathcal{V}^{\ell+1,s}, and the approximation accuracy, through the leave-one-out estimator. The distinction is necessary to assign an error estimator also to domains that have too few points to construct a residual expansion, in which case the error estimator of the previous level ELOO,sE^{\ell,s}_{\mathrm{LOO}} is reused.

The algorithm sequentially splits and refines subdomains with large approximation errors. Because likelihood functions typically have the highest complexity close to their peak, these regions tend to have larger approximation errors and are therefore predominantly picked for refinement. The proposed way of adaptive sampling then ends up placing more points near the likelihood peak, thereby reducing the number of non-informative likelihood evaluations.

The choice of a constant NrefN_{\mathrm{ref}} is simple and could in principle be replaced by a more elaborate strategy (e.g., based on the approximation error of the current subdomain relative to the total approximation error). A benefit of this enrichment criterion is that all residual expansions are computed with experimental designs of the same size. Upon choosing the domain with the maximum approximation error among the terminal domains, the error estimators then have a more comparable estimation accuracy.

4.2.2 Partitioning strategy

The partitioning strategy determines how a selected refinement domain is split. As described in Marelli et al. (2021), it is easy to define the splitting in the uniformly distributed quantile space 𝑼\bm{U} and map the resulting split domains 𝒟𝑼,p\mathcal{D}_{\bm{U}}^{\ell,p} to the (possibly unbounded) real space 𝑿\bm{X} through an appropriate isoprobabilistic transform (e.g., the Rosenblatt transform (Rosenblatt, 1952)).

Similar to the original SSE algorithm presented in Marelli et al. (2021), we split the refinement domain in half w.r.t. its prior mass. The original algorithm chooses the splitting direction based on the partial variance in the refinement domain. This approach is well suited for generic function approximation problems. For the approximation of likelihood functions, however, we propose a partitioning strategy that is more apt for dealing with their compact support.

We propose to pick the split direction along which a split yields a maximum difference in the residual empirical variance between the two candidate subdomains created by the split. This can easily be visualized with an example given by the M=2M=2 dimensional domain 𝒟𝑿,p{\mathcal{D}}_{\bm{X}}^{\ell,p} in Figure LABEL:sub@fig:SSE:algo:splitting:1. Assume this subdomain was selected as the refinement domain. To decide along which dimension to split, we construct the MM candidate subdomain pairs {𝒟spliti,1,𝒟spliti,2}i=1,,M\{{\mathcal{D}}_{\mathrm{split}}^{i,1},{\mathcal{D}}_{\mathrm{split}}^{i,2}\}_{i=1,\dots,M} and estimate the corresponding {Espliti}i=1,,M\{E_{\mathrm{split}}^{i}\}_{i=1,\dots,M} in those subdomains defined by

Espliti=def|Var[+1(𝒳spliti,1)]Var[+1(𝒳spliti,2)]|.E_{\mathrm{split}}^{i}\stackrel{{\scriptstyle\text{def}}}{{=}}\left|{\rm Var}\left[\mathcal{R}^{\ell+1}(\mathcal{X}^{i,1}_{\mathrm{split}})\right]-{\rm Var}\left[\mathcal{R}^{\ell+1}(\mathcal{X}^{i,2}_{\mathrm{split}})\right]\right|. (24)

In this expression, 𝒳spliti,1\mathcal{X}^{i,1}_{\mathrm{split}} and 𝒳spliti,2\mathcal{X}^{i,2}_{\mathrm{split}} denote subsets of the experimental design 𝒳\mathcal{X} inside the subdomains 𝒟spliti,1{\mathcal{D}}_{\mathrm{split}}^{i,1} and 𝒟spliti,2{\mathcal{D}}_{\mathrm{split}}^{i,2} respectively. The occurring variances can be easily estimated with the empirical variance of the residuals in the respective candidate subdomains.

After computing the residual variance differences, the split is carried out along the dimension

d=argmaxi{1,,M}Espliti,d=\arg\max_{i\in\{1,\dots,M\}}E_{\mathrm{split}}^{i}, (25)

i.e., to keep the subdomains 𝒟splitd,1{\mathcal{D}}_{\mathrm{split}}^{d,1} and 𝒟splitd,2{\mathcal{D}}_{\mathrm{split}}^{d,2} that introduce the largest difference in variance. For d=1d=1, the resulting split can be seen in Figure LABEL:sub@fig:SSE:algo:splitting:4.

Refer to caption
(a) Refinement domain
Refer to caption
(b) Split along d=1d=1
Refer to caption
(c) Split along d=2d=2
Refer to caption
(d) Selected pair
Figure 1: Partitioning strategy for a 2D example visualized in the quantile space 𝑼\bm{U}. The refinement domain 𝒟𝑼,p{\mathcal{D}}_{\bm{U}}^{\ell,p} is split into two subdomains 𝒟𝑼+1,s1{\mathcal{D}}_{\bm{U}}^{\ell+1,s_{1}} and 𝒟𝑼+1,s2{\mathcal{D}}_{\bm{U}}^{\ell+1,s_{2}}.

The choice of this partitioning strategy can be justified heuristically with the goal of approximating compact support functions. Assume that the likelihood function has compact support, this criterion will avoid cutting through its support and instead identify a split direction that results in one subdomain with large variance (expected to contain the likelihood support) and a subdomain with small variance. In subsequent steps, the algorithm will proceed by cutting away low variance subdomains, until the likelihood support is isolated.

4.2.3 The adaptive SSLE algorithm

The algorithm is presented here with its two parameters NrefN_{\mathrm{ref}}, the minimum experimental design size needed to expand a residual, and NEDN_{\mathrm{ED}}, the final experimental design size. The sample 𝒳,p{\mathcal{X}}^{\ell,p} refers to 𝒳𝒟X,p{\mathcal{X}}\cap{\mathcal{D}}_{X}^{\ell,p}, i.e. the subset of 𝒳\mathcal{X} inside 𝒟X,p{\mathcal{D}}_{X}^{\ell,p}. Further, the multi-index set 𝒯2\mathcal{T}\in\mathbb{N}^{2} at each step of the algorithm gathers all indices (,p)(\ell,p) of unsplit subdomains. It thus denotes the terminal domains: 𝒟𝑿k,k𝒯\mathcal{D}_{\bm{X}}^{k},k\in\mathcal{T}. For visualization purposes we show the first iterations of the algorithm for a two-dimensional example in Figure 2.

  1. 1.

    Initialization:

    1. (a)

      𝒟𝑿0,1=𝒟𝑿\mathcal{D}_{\bm{X}}^{0,1}=\mathcal{D}_{\bm{X}}

    2. (b)

      Sample from prior distribution 𝒳={𝒙(1),,𝒙(Nref)}\mathcal{X}=\{\bm{x}^{(1)},\cdots,\bm{x}^{(N_{\mathrm{ref}})}\}

    3. (c)

      Calculate the truncated expansion ^S0,1(𝑿)\widehat{\mathcal{R}}_{S}^{0,1}(\bm{X}) of (𝑿)\mathcal{L}(\bm{X}) in the full domain 𝒳0,1\mathcal{X}^{0,1}, retrieve its approximation error 0,1\mathcal{E}^{0,1} and initialize 𝒯={(0,1)}\mathcal{T}=\{(0,1)\}

    4. (d)

      1(𝑿)=(𝑿)^S0,1(𝑿){\mathcal{R}}^{1}(\bm{X})=\mathcal{L}(\bm{X})-\widehat{\mathcal{R}}_{S}^{0,1}(\bm{X})

  2. 2.

    For (,p)=argmaxk𝒯k(\ell,p)=\operatorname*{arg\,max}_{k\in\mathcal{T}}\mathcal{E}^{k}:

    1. (a)

      Split the current subdomain 𝒟𝑿,p\mathcal{D}_{\bm{X}}^{\ell,p} in 22 sub-parts 𝒟𝑿+1,{s1,s2}\mathcal{D}_{\bm{X}}^{\ell+1,\{s_{1},s_{2}\}} and update 𝒯\mathcal{T}

    2. (b)

      For each split s={s1,s2}s=\{s_{1},s_{2}\}

      1. i.

        If |𝒳+1,s|<Nref|\mathcal{X}^{\ell+1,s}|<N_{\mathrm{ref}} and Nref|𝒳+1,s|<NED|𝒳|N_{\mathrm{ref}}-|\mathcal{X}^{\ell+1,s}|<N_{\mathrm{ED}}-|\mathcal{X}|

        1. A.

          Enrich sample 𝒳\mathcal{X} with Nref|𝒳+1,s|N_{\mathrm{ref}}-|\mathcal{X}^{\ell+1,s}| new points inside 𝒟𝑿+1,s\mathcal{D}_{\bm{X}}^{\ell+1,s}

      2. ii.

        If |𝒳+1,s|=Nref|\mathcal{X}^{\ell+1,s}|=N_{\mathrm{ref}}

        1. A.

          Create the truncated expansion ^S+1,s(𝑿+1,s)\widehat{\mathcal{R}}_{S}^{\ell+1,s}(\bm{X}^{\ell+1,s}) of the residual +1(𝑿+1,s){\mathcal{R}}^{\ell+1}(\bm{X}^{\ell+1,s}) in the current subdomain using 𝒳+1,s\mathcal{X}^{\ell+1,s}

        2. B.

          Update the residual in the current subdomain +2(𝑿+1,s)=+1(𝑿+1,s)^S+1,s(𝑿+1,s){\mathcal{R}}^{\ell+2}(\bm{X}^{\ell+1,s})={\mathcal{R}}^{\ell+1}(\bm{X}^{\ell+1,s})-\widehat{\mathcal{R}}_{S}^{\ell+1,s}(\bm{X}^{\ell+1,s})

      3. iii.

        Retrieve the approximation error +1,s\mathcal{E}^{\ell+1,s} from Eq. (23)

    3. (c)

      If no new expansions were created, terminate the algorithm, otherwise go back to 2

  3. 3.

    Termination

    1. (a)

      Return the full sequence of 𝒟𝑿,p\mathcal{D}_{\bm{X}}^{\ell,p} and ^S,p(𝑿,p)\widehat{\mathcal{R}}_{S}^{\ell,p}(\bm{X}^{\ell,p}) needed to compute Eq. (15)

The updating of the multi-index set in Step 2a refers to removing the current index (,p)(\ell,p) from the set and adding to it the newly created indices (+1,s1)(\ell+1,s_{1}) and (+1,s2)(\ell+1,s_{2}).

Refer to captionRefer to caption
(a) Initialization
Refer to captionRefer to caption
(b) First iteration
Refer to captionRefer to caption
(c) 33rd iteration
Figure 2: Graphical representation of the first steps of the adaptive SSLE algorithm described in Section 4.2.3 for a two-dimensional problem with independent prior distribution. Upper row: partitioning in the quantile space; Lower row: partitioning in the unbounded real space with π(𝒙)\pi(\bm{x}) contour lines in dashed blue. Red dots show the adaptive experimental design that has a constant size of Nref=5N_{\mathrm{ref}}=5 in each created subdomain. The terminal domains 𝒯\mathcal{T} are highlighted in orange. The splitting direction in each subdomain is determined randomly in this example.

4.2.4 Convergence of the adaptive SSLE algorithm

Convergence of the original SSE algorithm is guaranteed in the mean-square sense by the spectral convergence in each subdomain Marelli et al. (2021). This convergence property directly applies to the SSLE of any likelihood function as, per existence of a finite maximum likelihood value, they fulfil the finite variance condition Nagel and Sudret (2016).

This result cannot be directly extended to the present adaptive (greedy) setting, because a combination of parameters might in general lead to a whole subdomain not being explored further. This can only happen, however, if the error estimator k\mathcal{E}^{k} tremendously underestimates the actual error in a terminal domain. The choice of the leave-one-out cross-validation error estimator (see Eq. (23)), with its tendency to avoid overfitting, reduces significantly the probability of this scenario. Further investigations in this direction are ongoing.

5 Case studies

To showcase the effectiveness of the proposed SSLE approach, we present three case studies with different types of likelihood complexity: (i) a one-dimensional vibration problem with a bimodal posterior, (ii) a six-dimensional heat transfer problem that exhibits high posterior concentrations (i.e. highly informative likelihood) and (iii) a 6262-dimensional diffusion problem with low active dimensionality that models concentration-driven diffusion in a one-dimensional domain.

For all case studies, we adopt the adaptive sparse-PCE based on LARS approach developed in Blatman and Sudret (2011) through its numerical implementation in UQLab (Marelli and Sudret, 2014, 2019). Each ^Sk\widehat{\mathcal{R}}_{S}^{k} is therefore a degree- and qq-norm-adaptive polynomial chaos expansion. We further introduce a rank truncation of r=2r=2, i.e. we limit the maximum number of input interactions (Marelli and Sudret, 2019) to two variables at a time. The truncation set for each spectral expansion (Eq. (8)) thus reads:

𝒜M,p,q,r={𝜶M:𝜶qp,𝜶0r}.\mathcal{A}^{M,p,q,r}=\{\bm{\alpha}\in\mathbb{N}^{M}:||\bm{\alpha}||_{q}\leq p,||\bm{\alpha}||_{0}\leq r\}. (26)

where

𝜶q=(i=1Mαiq)1q,q(0,1];𝜶0=i=1M1{αi>0}.||\bm{\alpha}||_{q}=\left(\sum_{i=1}^{M}\alpha_{i}^{q}\right)^{\frac{1}{q}},q\in(0,1];\quad||\bm{\alpha}||_{0}=\sum_{i=1}^{M}1_{\{\alpha_{i}>0\}}. (27)

The qq-norm is adaptively increased between q={0.5,,0.8}q=\{0.5,\cdots,0.8\} while the maximum polynomial degree is adaptively increased in the interval p={0,1,,p}p=\{0,1,\cdots,p\}, where the maximum degree p=20p=20 for case study (i) and (ii) and p=3p=3 for case study (iii) due to its high dimensionality.

In case study (ii) and (iii), the performance of SLE (Nagel and Sudret, 2016), the original non-adaptive SSLE (Marelli et al., 2021) and the proposed adaptive SSLE approach presented in Section 4.2 is compared. The comparison was omitted for case study (i), because only adaptive SSLE succeeded in solving the problem. For clarity, we henceforth abbreviate the adaptive SSLE algorithm to adSSLE.

To simplify the comparison, the same partitioning strategy employed for adSSLE (Section 4.2.2) was employed for the non-adaptive SSLE approach. Also, the same experimental designs were used for the non-adaptive SSLE and the SLE approaches. Finally, the same parameter NrefN_{\mathrm{ref}} was used to define the enrichment samples in adSSLE and the termination criterion in non-adaptive SSLE.

Because the considered case studies do not admit a closed form solution, we validate the algorithm instead through MCMC reference solutions with long chain lengths to produce samples from the posterior distributions. In this respect, identifiability of the “true” underlying data-generating model, while clearly important in inverse problems in general, is not central in the present discussion. As a proxy for identifiability, however, we consider several different case studies with more or less informative (peaked) likelihood and posterior distributions.

To assess the performance of the three algorithms considered, we define an error measure that allows to quantitatively compare the similarity of the SSLE, adSSLE and SLE solution with the reference MCMC solution. This comparison is inherently difficult, as a sampling-based approach (MCMC) needs to be compared to a functional approximation (SSLE, adSSLE, SLE). We proceed to compare the univariate posterior marginals, available analytically in SSLE, adSSLE and SLE (See Eq. (13) and Eq. (20)), to the reference posterior marginals estimated with kernel density estimation (KDE, Wand and Jones (1995)) from the MCMC sample. Denoting by π^i(κi|𝒴)\hat{\pi}_{i}(\kappa_{i}|{\mathcal{Y}}) the SSLE, adSSLE or SLE approximations and by πi(κi|𝒴){\pi}_{i}(\kappa_{i}|{\mathcal{Y}}) the reference solution, we define the following error measure

η=def1Mi=1MJSD(π^i(κi|𝒴)||πi(κi|𝒴))\eta\stackrel{{\scriptstyle\text{def}}}{{=}}\frac{1}{M}\sum_{i=1}^{M}\operatorname*{JSD}\left(\hat{\pi}_{i}(\kappa_{i}|{\mathcal{Y}})||\pi_{i}(\kappa_{i}|{\mathcal{Y}})\right) (28)

where MM is the dimensionality of the problem and JSD\operatorname*{JSD} is the Jensen-Shannon divergence (Lin, 1991), a symmetric and bounded ([0,log(2)][0,\log(2)]) distance measure for probability distributions based on the Kullback-Leibler divergence.

The purpose of the error measure η\eta is to allow for a fair comparison between the different methods investigated. It is not a practical measure for engineering applications because it relies on the availability of a reference solution, and it its magnitude does not have a clear quantitative interpretation. However, it is considerably more comprehensive than a pure moment-based error measure. Because it is averaged over all MM marginals, it encapsulates the approximation accuracy of all univariate posterior marginals in a scalar value.

As all algorithms (SSLE, adSSLE, SLE) depend on a randomly chosen experimental designs, we produce 2020 replications for case study (ii) and 55 replications for case study (iii) by running them multiple times. The computational cost of all examples is given in terms of number of likelihood evaluations NEDN_{\mathrm{ED}}, as they dominate the total computational cost in any engineering setting. All computations shown were carried out on a standard laptop, with the longest single runs taking 1h\approx 1~{}h, which includes both forward model evaluations and the overall algorithmic overhead.

5.1 1-dimensional vibration problem

In this first case study, the goal is the inference of a single unknown parameter with a multimodal posterior distribution. This problem is difficult to solve with standard MCMC methods due to the presence of a probability valley, i.e. low probability region, between essentially disjoint posterior peaks. It also serves as an illustrative example of how the proposed adaptive algorithm constructs an adSSLE. The problem is fabricated and uses synthetic data, but is presented in the context of a relevant engineering problem.

Consider the oscillator displayed in Figure 3 subject to harmonic (i.e., sinusoidal) excitation. Assume the prior information about its stiffness X=defkX\stackrel{{\scriptstyle\text{def}}}{{=}}k is that it follows a lognormal distribution with μ=0.8N/m\mu=0.8~{}N/m and σ=0.1N/m\sigma=0.1~{}N/m. Its true value shall be determined using measurements of the oscillation amplitude at the location of the mass mm. The known properties of the oscillator system are the oscillator mass m=1kgm=1~{}kg, the excitation frequency ω=1rad/s\omega=1~{}rad/s and the viscous damping coefficient c=0.1Ns/mc=0.1~{}Ns/m. The oscillation amplitude is measured in five independent oscillation events and normalized by the forcing amplitude yielding the measured amplitude ratios 𝒴={9.01,8.67,8.84,9.22,8.54}{\mathcal{Y}}=\{9.01,8.67,8.84,9.22,8.54\}.

Refer to caption
Figure 3: 1-dimensional vibration problem: Sketch of the linear oscillator

This problem is well known in mechanics and in the linear case (i.e., assuming small deformations and linear material behavior) can be solved analytically with the amplitude of the frequency response function. This function returns the ratio between the steady state amplitude of a linear oscillator and the amplitude of its excitation. It is given by

(X)=mω2(Xmω2)2+(cω)2.{\mathcal{M}}(X)=\frac{m\omega^{2}}{\sqrt{(X-m\omega^{2})^{2}+(c\omega)^{2}}}. (29)

We assume a discrepancy model with known discrepancy standard deviation σ\sigma. In conjunction with the available measurements 𝒴{\mathcal{Y}}, this leads to the following likelihood function:

(𝒙;𝒴)=i=15𝒩(y(i)|(x),σ2).{\mathcal{L}}(\bm{x};\,{\mathcal{Y}})=\prod_{i=1}^{5}{\mathcal{N}}(y^{(i)}|{\mathcal{M}}(x),\sigma^{2}). (30)

We employ the adSSLE algorithm to approximate this likelihood function with Nref=10N_{\mathrm{ref}}=10. A few iterations from the solution process are shown in Figure 4. The top plots show the subdomains 𝒟X,p{\mathcal{D}}_{X}^{\ell,p} constructed at each refinement step, highlighting the terminal domains 𝒯{\mathcal{T}}. The middle plots display the residual between the true likelihood and the approximation at the current iteration, as well as the adaptively chosen experimental design 𝒳\mathcal{X}. The bottom plots display the target likelihood function and its current approximation.

The initial global approximation of the first iteration in Figure LABEL:sub@fig:cs1:SSEBehaviour:1 is a constant polynomial based on the initial experimental design. By the third iteration, the algorithm has identified the subdomain 𝒟X2,2{\mathcal{D}}_{X}^{2,2} as the one of interest and proceeds to refine it in subsequent steps. By the 88th iteration both likelihood peaks have been identified. Finally, by the 1010th iteration in Figure LABEL:sub@fig:cs1:SSEBehaviour:4, both likelihood peaks are approximated well by the adSSLE approach.

The last iteration shows how the algorithm splits domains and adds new sample points. There is a clear clustering of subdomains and sample points near the likelihood peaks at X=0.95X=0.95 and X=1.05X=1.05.

Refer to captionRefer to captionRefer to caption
(a) 11st iteration, NED=10N_{\mathrm{ED}}=10
Refer to captionRefer to captionRefer to caption
(b) 33rd iteration, NED=30N_{\mathrm{ED}}=30
Refer to captionRefer to captionRefer to caption
(c) 88th iteration, NED=80N_{\mathrm{ED}}=80
Refer to captionRefer to captionRefer to caption
(d) 1010th iteration, NED=100N_{\mathrm{ED}}=100
Figure 4: One-dimensional vibration problem: Illustration of the adSSLE algorithm approximating the likelihood function {\mathcal{L}}.

The results from Eq. (22) show that without further computations it would be possible to directly extract the posterior moments by post-processing the SSLE coefficients. In the present bimodal case, however, the posterior moments are not very meaningful. Instead, the available posterior approximation gives a full picture of the inferred parameter X|𝒴X|{\mathcal{Y}}. It is shown together with the true posterior and the original prior distribution in Figure 5.

Refer to caption
Figure 5: One-dimensional vibration problem: Comparison of the true multimodal posterior and its adSSLE based approximation.

For this case study, non-adaptive experimental design approaches like the standard SSLE (Marelli et al., 2021) and the original SLE algorithm (Nagel and Sudret, 2016) will almost surely fail for the considered experimental design of NED=100N_{\mathrm{ED}}=100. In numerous trial runs these approaches did not manage to accurately reconstruct the likelihood function due to a lack of informative samples near the likelihood peaks.

5.2 Moderate-dimensional heat transfer problem

This case study is a complex engineering problem that can be modified to exhibit high posterior concentration in the prior domain. It was originally presented in Nagel and Sudret (2016) and solved there using SLE. We again solve the same problem with SSLE and compare the performance of SLE (Nagel and Sudret, 2016), the original non-adaptive SSLE (Marelli et al., 2021) and the proposed adSSLE approach presented in Section 4.2. To investigate the performance of the algorithm in the case of high posterior concentrations (e.g. due to a large data-set), two instances of the problem with different discrepancy parameters are investigated.

Consider the diffusion-driven stationary heat transfer problem sketched in Figure LABEL:sub@fig:cs3:setup. It models a 2D plate with a background matrix of constant thermal conductivity κ0\kappa_{0} and 66 inclusions with conductivities 𝜿=def(κ1,,κ6)\bm{\kappa}\stackrel{{\scriptstyle\text{def}}}{{=}}(\kappa_{1},\dots,\kappa_{6})^{\intercal}. The diffusion driven steady state heat distribution is described by a heat equation in Euclidean coordinates 𝒓=def(r1,r2)\bm{r}\stackrel{{\scriptstyle\text{def}}}{{=}}(r_{1},r_{2})^{\intercal} of the form

(κ(𝒓)T~(𝒓))=0,\nabla\cdot(\kappa(\bm{r})\nabla\tilde{T}(\bm{r}))=0, (31)

where the thermal conductivity field is denoted by κ\kappa and the temperature field by T~\tilde{T}. The boundary conditions of the plate are given by a no-heat-flux Neumann boundary conditions on the left and right sides (T~/r1=0\partial\tilde{T}/r_{1}=0), a Neumann boundary condition on the bottom (κ0T~/r2=q2{\kappa_{0}\partial\tilde{T}/r_{2}=q_{2}}) and a temperature T~1\tilde{T}_{1} Dirichlet boundary condition on the top.

We employ a finite element (FE) solver to solve the weak form of Eq. (31) by discretizing the domain into approximately 10510^{5} triangular elements. A sample solution returned by the FE-solver is shown in Figure 6b.

Refer to caption
(a) Problem sketch
Refer to caption
(b) Steady state solution
Figure 6: Moderate-dimensional heat transfer problem: Model setup and exemplary solution.

In this example we intend to infer the thermal conductivities 𝜿\bm{\kappa} of the inclusions. We assume the same problem constants as in Nagel and Sudret (2016) (i.e., q2=2,000W/m2q_{2}=2{,}000~{}\text{W/m}^{2}, T~1=200K\tilde{T}_{1}=200~{}\text{K}, κ0=45W/m/K\kappa_{0}=45\text{W/m/K}). The forward model {\mathcal{M}} takes as an input the conductivities of the inclusions 𝜿\bm{\kappa}, solves the finite element problem and returns the steady state temperature 𝑻~=def(T~1,,T~20)\tilde{\bm{T}}\stackrel{{\scriptstyle\text{def}}}{{=}}(\tilde{T}_{1},\dots,\tilde{T}_{20})^{\intercal} at the measurement points, i.e., :𝜿𝑻~{\mathcal{M}}:\bm{\kappa}\mapsto\tilde{\bm{T}}.

To solve the inverse problem, we assume a multivariate lognormal prior distribution with independent marginals on the inclusion conductivities, i.e. π(𝜿)=i=16𝒩(κi|μ=30W/m/K,σ=6W/m/K)\pi(\bm{\kappa})=\prod_{i=1}^{6}\mathcal{LN}(\kappa_{i}|\mu=30~{}\text{W/m/K},\sigma=6~{}\text{W/m/K}). We further assume an additive Gaussian discrepancy model, which yields the likelihood function

(𝜿;𝒴)=12πσ2exp(12σ2(𝑻(𝜿))(𝑻(𝜿))),{\mathcal{L}}(\bm{\kappa};{\mathcal{Y}})=\frac{1}{2\pi\sigma^{2}}\exp\left(-\frac{1}{2\sigma^{2}}\left(\bm{T}-{\mathcal{M}}(\bm{\kappa})\right)^{\intercal}\left(\bm{T}-{\mathcal{M}}(\bm{\kappa})\right)\right), (32)

with a discrepancy standard deviation of σ\sigma.

As measurements, we generate one temperature field with 𝜿^=def(32,36,20,24,40,28)W/m/K\hat{\bm{\kappa}}\stackrel{{\scriptstyle\text{def}}}{{=}}(32,36,20,24,40,28)^{\intercal}~{}\text{W/m/K} and collect its values at 2020 points indicated by black dots in Figure LABEL:sub@fig:cs3:setup. We then perturb these temperature values with additive Gaussian noise and use them as the inversion data 𝒴=def𝑻=(T1,,T20){\mathcal{Y}}\stackrel{{\scriptstyle\text{def}}}{{=}}\bm{T}=(T_{1},\dots,T_{20})^{\intercal}.

We look at two instances of this problem that differ only by the discrepancy parameter σ\sigma from Eq. (32). The prior model response has a standard deviation of approximately 0.3K0.3~{}\text{K}, depending on the measurement point TiT_{i}. We therefore solve the problem first with a large value σ=0.25K\sigma=0.25~{}\text{K} and second with a small value σ=0.1K\sigma=0.1~{}\text{K}. As the discrepancy standard deviation determines how peaked the likelihood function is, the first problem has a likelihood function with a much wider support and is in turn significantly easier to solve than the second one. It is noted here that in practice, the peakedness of the likelihood function is either increased by a smaller discrepancy standard deviation, or the inclusion of additional experimental data.

To monitor the dependence of the algorithms on the number of likelihood evaluations, we solve both problems with a set of maximum likelihood evaluations NED={1,000;2,000,5,000;10,000;30,000}N_{\mathrm{ED}}=\{1{,}000;2{,}000,5{,}000;10{,}000;30{,}000\}. The number of refinement samples is set to Nref=1,000N_{\mathrm{ref}}=1{,}000.

As a benchmark, we use reference posterior samples generated by the affine-invariant ensemble sampler MCMC algorithm (Goodman and Weare, 2010) with 30,00030{,}000 steps and 5050 parallel chains, requiring a total of NED=1.5106N_{\mathrm{ED}}=1.5\cdot 10^{6} likelihood evaluations. Based on numerous heuristic convergence tests and due to the large number of MCMC steps, the resulting samples can be considered to accurately represent the true posterior distributions.

The results of the analyses are summarized in Figure 7, where the error measure η\eta is plotted against the number of likelihood evaluations for the large and small standard deviation case. For the large discrepancy standard deviation case, both SSLE approaches clearly outperform standard SLE w.r.t. the error measure η\eta. This is most significant at mid-range experimental designs (NED=5,000;10,000N_{\mathrm{ED}}={5{,}000;10{,}000}), where SLE does not reach the required high degrees and fails to accurately approximate the likelihood function. At larger experimental designs SLE catches up to non-adaptive SSLE but is still outperformed by the proposed adSSLE approach. The real strength of the adaptive algorithm shows for the case of a small discrepancy standard deviation, where the limitations of fixed experimental designs become obvious. When the likelihood function is nonzero in a small subdomain of the prior, the global SLE and non-adaptive SSLE approach will fail in practice because of the insufficient number of samples placed in the informative regions. The adSSLE approach, however, works very well in these types of problems. It manages to identify the regions of interest and produces a likelihood approximation that accurately reproduces the posterior marginals.

Refer to caption
(a) Large discrepancy, σ=0.25K\sigma=0.25~{}\text{K}
Refer to caption
(b) Small discrepancy, σ=0.1K\sigma=0.1~{}\text{K}
Figure 7: Moderate-dimensional heat transfer problem: Convergence of the η\eta error measure (Eq. (28)) as a function of the experimental design size NEDN_{\mathrm{ED}} in 2020 replications for SLE, SSLE with a static experimental design and the proposed adSSLE approach. We display the two discrepancy standard deviation cases σ={0.25,0.1}K\sigma=\{0.25,0.1\}~{}\text{K}.

Tables 1 and 2 show the convergence of the adSSLE method moment estimate (mean and variance) to the reference solution for a single run. In brackets next to the moment estimates ξ\xi, the relative error ϵ=def|ξMCMCξSSLE|/ξMCMC\epsilon\stackrel{{\scriptstyle\text{def}}}{{=}}|\xi_{\mathrm{MCMC}}-\xi_{\mathrm{SSLE}}|/\xi_{\mathrm{MCMC}} is also shown. Due to the non-strict positivity of the SSLE estimate, one variance estimate computed with Eq. (22) is negative and is therefore omitted from Table 2.

Table 1: Moderate-dimensional heat transfer problem: adSSLE results with large discrepancy standard deviation σ=0.25K\sigma=0.25~{}\text{K}. Relative errors w.r.t. MCMC reference solution are shown in brackets.
𝔼[κi|𝒴]{\mathbb{E}}\left[\kappa_{i}|{\mathcal{Y}}\right] κ1\kappa_{1} κ2\kappa_{2} κ3\kappa_{3} κ4\kappa_{4} κ5\kappa_{5} κ6\kappa_{6}
adSSLE NED=1,000N_{\mathrm{ED}}=1{,}000 30.00(0.7%)30.00(0.7\%) 30.00(7.1%)30.00(7.1\%) 19.51(5.8%)19.51(5.8\%) 30.01(7.1%)30.01(7.1\%) 35.39(2.9%)35.39(2.9\%) 30.00(14.6%)30.00(14.6\%)
NED=2,000N_{\mathrm{ED}}=2{,}000 31.74(6.5%)31.74(6.5\%) 29.80(7.7%)29.80(7.7\%) 20.89(0.9%)20.89(0.9\%) 30.00(7.1%)30.00(7.1\%) 35.56(2.4%)35.56(2.4\%) 28.30(8.1%)28.30(8.1\%)
NED=5,000N_{\mathrm{ED}}=5{,}000 29.70(0.4%)29.70(0.4\%) 31.74(1.7%)31.74(1.7\%) 20.21(2.4%)20.21(2.4\%) 30.21(6.5%)30.21(6.5\%) 36.45(0.0%)36.45(0.0\%) 27.58(5.4%)27.58(5.4\%)
NED=10,000N_{\mathrm{ED}}=10{,}000 29.96(0.5%)29.96(0.5\%) 32.41(0.3%)32.41(0.3\%) 19.90(3.9%)19.90(3.9\%) 31.55(2.3%)31.55(2.3\%) 36.57(0.3%)36.57(0.3\%) 26.20(0.1%)26.20(0.1\%)
NED=30,000N_{\mathrm{ED}}=30{,}000 29.92(0.4%)29.92(0.4\%) 32.40(0.3%)32.40(0.3\%) 19.94(3.7%)19.94(3.7\%) 31.96(1.1%)31.96(1.1\%) 36.53(0.2%)36.53(0.2\%) 26.15(0.1%)26.15(0.1\%)
MCMC 29.81\mathbf{29.81} 32.30\mathbf{32.30} 20.71\mathbf{20.71} 32.31\mathbf{32.31} 36.45\mathbf{36.45} 26.18\mathbf{26.18}
Var[κi|𝒴]\sqrt{{\rm Var}\left[\kappa_{i}|{\mathcal{Y}}\right]}
adSSLE NED=1,000N_{\mathrm{ED}}=1{,}000 6.00(86.5%)6.00(86.5\%) 6.01(41.7%)6.01(41.7\%) 6.00(142.0%)6.00(142.0\%) 5.99(18.1%)5.99(18.1\%) 4.44(15.9%)4.44(15.9\%) 5.99(97.1%)5.99(97.1\%)
NED=2,000N_{\mathrm{ED}}=2{,}000 6.09(89.3%)6.09(89.3\%) 4.78(12.9%)4.78(12.9\%) 3.44(38.6%)3.44(38.6\%) 6.00(18.1%)6.00(18.1\%) 4.34(13.5%)4.34(13.5\%) 5.38(76.9%)5.38(76.9\%)
NED=5,000N_{\mathrm{ED}}=5{,}000 2.19(32.0%)2.19(32.0\%) 5.61(32.4%)5.61(32.4\%) 3.09(24.4%)3.09(24.4\%) 6.06(19.4%)6.06(19.4\%) 3.62(5.5%)3.62(5.5\%) 4.78(57.2%)4.78(57.2\%)
NED=10,000N_{\mathrm{ED}}=10{,}000 2.62(18.7%)2.62(18.7\%) 4.40(3.7%)4.40(3.7\%) 2.82(13.6%)2.82(13.6\%) 5.15(1.5%)5.15(1.5\%) 3.96(3.6%)3.96(3.6\%) 3.14(3.4%)3.14(3.4\%)
NED=30,000N_{\mathrm{ED}}=30{,}000 2.40(25.5%)2.40(25.5\%) 4.38(3.4%)4.38(3.4\%) 2.70(8.8%)2.70(8.8\%) 5.23(3.1%)5.23(3.1\%) 3.87(1.2%)3.87(1.2\%) 3.19(5.0%)3.19(5.0\%)
MCMC 3.22\mathbf{3.22} 4.24\mathbf{4.24} 2.48\mathbf{2.48} 5.08\mathbf{5.08} 3.83\mathbf{3.83} 3.04\mathbf{3.04}
Table 2: Moderate-dimensional heat transfer problem: adSSLE results with small discrepancy standard deviation σ=0.1K\sigma=0.1~{}\text{K}. Relative errors w.r.t. MCMC reference solution are shown in brackets. Field with an asterisk (*) indicates negative variance estimate.
𝔼[κi|𝒴]{\mathbb{E}}\left[\kappa_{i}|{\mathcal{Y}}\right] κ1\kappa_{1} κ2\kappa_{2} κ3\kappa_{3} κ4\kappa_{4} κ5\kappa_{5} κ6\kappa_{6}
adSSLE NED=1,000N_{\mathrm{ED}}=1{,}000 30.00(3.8%)30.00(3.8\%) 30.00(10.7%)30.00(10.7\%) 30.00(62.2%)30.00(62.2\%) 30.00(5.7%)30.00(5.7\%) 30.00(22.8%)30.00(22.8\%) 30.00(17.6%)30.00(17.6\%)
NED=2,000N_{\mathrm{ED}}=2{,}000 30.00(3.8%)30.00(3.8\%) 34.58(2.9%)34.58(2.9\%) 30.00(62.2%)30.00(62.2\%) 30.00(5.7%)30.00(5.7\%) 30.00(22.8%)30.00(22.8\%) 30.00(17.6%)30.00(17.6\%)
NED=5,000N_{\mathrm{ED}}=5{,}000 30.00(3.8%)30.00(3.8\%) 34.70(3.2%)34.70(3.2\%) 25.30(36.8%)25.30(36.8\%) 30.00(5.7%)30.00(5.7\%) 37.28(4.0%)37.28(4.0\%) 30.00(17.6%)30.00(17.6\%)
NED=10,000N_{\mathrm{ED}}=10{,}000 30.00(3.8%)30.00(3.8\%) 34.71(3.3%)34.71(3.3\%) 18.92(2.3%)18.92(2.3\%) 30.00(5.7%)30.00(5.7\%) 37.87(2.5%)37.87(2.5\%) 25.44(0.3%)25.44(0.3\%)
NED=30,000N_{\mathrm{ED}}=30{,}000 31.16(0.0%)31.16(0.0\%) 34.28(2.0%)34.28(2.0\%) 18.57(0.4%)18.57(0.4\%) 31.37(1.4%)31.37(1.4\%) 38.68(0.4%)38.68(0.4\%) 25.56(0.2%)25.56(0.2\%)
MCMC 31.17\mathbf{31.17} 33.61\mathbf{33.61} 18.49\mathbf{18.49} 31.83\mathbf{31.83} 38.84\mathbf{38.84} 25.51\mathbf{25.51}
Var[κi|𝒴]\sqrt{{\rm Var}\left[\kappa_{i}|{\mathcal{Y}}\right]}
adSSLE NED=1,000N_{\mathrm{ED}}=1{,}000 6.00(268.3%)6.00(268.3\%) 6.00(149.7%)6.00(149.7\%) 6.00(362.1%)6.00(362.1\%) 6.00(69.4%)6.00(69.4\%) 5.99(202.8%)5.99(202.8\%) 5.99(264.7%)5.99(264.7\%)
NED=2,000N_{\mathrm{ED}}=2{,}000 6.00(268.5%)6.00(268.5\%) 4.53(88.9%)4.53(88.9\%) 6.00(362.2%)6.00(362.2\%) 6.00(69.5%)6.00(69.5\%) 6.00(203.1%)6.00(203.1\%) 6.00(265.2%)6.00(265.2\%)
NED=5,000N_{\mathrm{ED}}=5{,}000 6.00(268.5%)6.00(268.5\%) 4.43(84.3%)4.43(84.3\%) 2.87(120.8%)2.87(120.8\%) 6.00(69.5%)6.00(69.5\%) 4.29(116.8%)4.29(116.8\%) 6.00(265.2%)6.00(265.2\%)
NED=10,000N_{\mathrm{ED}}=10{,}000 4.49(176.0%)4.49(176.0\%) ()*(*) 1.84(42.1%)1.84(42.1\%) 2.58(27.2%)2.58(27.2\%) 3.89(96.6%)3.89(96.6\%) 2.66(61.8%)2.66(61.8\%)
NED=30,000N_{\mathrm{ED}}=30{,}000 1.20(26.6%)1.20(26.6\%) 3.60(49.8%)3.60(49.8\%) 1.62(25.2%)1.62(25.2\%) 2.66(24.7%)2.66(24.7\%) 3.06(54.6%)3.06(54.6\%) 1.74(5.9%)1.74(5.9\%)
MCMC 1.63\mathbf{1.63} 2.40\mathbf{2.40} 1.30\mathbf{1.30} 3.54\mathbf{3.54} 1.98\mathbf{1.98} 1.64\mathbf{1.64}

The full posterior marginals obtained from one run of adSSLE with NED=30,000N_{\mathrm{ED}}=30{,}000 are also compared to those of the reference MCMC and displayed in Figure 8. The individual plots show the univariate posterior marginals (i.e. π(xi|𝒴)\pi(x_{i}|\mathcal{Y})) on the main diagonal and the bivariate posterior marginals (i.e. π(𝒙ij|𝒴)\pi(\bm{x}_{ij}|\mathcal{Y})) in the ii-th row and jj-th column. It can be clearly seen that the posterior characteristics are very well captured. However, the adSSLE approach sometimes fails to accurately represent the tails of the distribution. This is especially obvious in the small discrepancy case in Figure LABEL:sub@fig:cs2:Posterior:SSESmall where the tail is sometimes cut off. We emphasize here that the SSLE marginals are obtained analytically as 1D and 2D surfaces for the univariate and bivariate marginals respectively. For the reference MCMC approach, on the other hand, they need to be approximated with histograms based on the available posterior sample.

Refer to caption
(a) Large discrepancy, σ=0.25K\sigma=0.25~{}\text{K}, adSSLE
Refer to caption
(b) Large discrepancy, σ=0.25K\sigma=0.25~{}\text{K}, MCMC
Refer to caption
(c) Small discrepancy, σ=0.1K\sigma=0.1~{}\text{K}, adSSLE
Refer to caption
(d) Small discrepancy, σ=0.1K\sigma=0.1~{}\text{K}, MCMC
Figure 8: Moderate-dimensional heat transfer problem: Comparison plots of the posterior distribution marginals computed from adSSLE (NED=30,000N_{\mathrm{ED}}=30{,}000) and MCMC (NED=1.5106N_{\mathrm{ED}}=1.5\cdot 10^{6}). The prior marginals are shown in red.

5.2.1 Convergence of the posterior moments

In practical inference applications, posterior moments are often one of the main quantities of interest. An estimator of these moments is readily available at every refinement step of adSSLE through Eq. (22).

Tracking the evolution of the posterior moments throughout the adSSLE iterations can be used as a heuristic estimator of the convergence of the adSSLE algorithm. However, only the stability of the solution can be assessed, without guarantees on the bias. As an example, we now consider the large discrepancy problem and plot the evolution of the posterior mean and standard deviation for every XiX_{i} as a function of the number of likelihood evaluations in Figure 9. It can be seen that after 10,000\sim 10{,}000 likelihood evaluations, most moment estimators achieve convergence to a value close to the reference solution. This plot also reveals a small bias of the 𝔼[X3|𝒴]{\mathbb{E}}\left[X_{3}|{\mathcal{Y}}\right] and Var[X3|𝒴]\sqrt{{\rm Var}\left[X_{3}|{\mathcal{Y}}\right]} estimators, that was previously highlighted in Table 1.

Refer to caption
(a) Mean
Refer to caption
(b) Standard deviation
Figure 9: Moderate-dimensional heat transfer problem: Evolution of the posterior moment estimation for a typical run of the adSSLE algorithm on the small discrepancy problem. The thin lines show the MCMC reference solution.

5.2.2 Influence of NrefN_{\mathrm{ref}}

The main hyperparameter of the proposed adSSLE algorithm is the number NrefN_{\mathrm{ref}}, which corresponds to the number of sample points that are required at each PCE construction step (see Section 4.2). In Figure 10 we display the effect of different NrefN_{\mathrm{ref}} values on the convergence in the small and large discrepancy problems.

Refer to caption
(a) Large discrepancy, σ=0.25K\sigma=0.25~{}\text{K}
Refer to caption
(b) Small discrepancy, σ=0.1K\sigma=0.1~{}\text{K}
Figure 10: Moderate-dimensional heat transfer problem: Convergence of the η\eta error measure (Eq. (28)) as a function of the experimental design size NEDN_{\mathrm{ED}} in 2020 replications for the proposed adSSLE approach with different NrefN_{\mathrm{ref}} parameters. We display the two discrepancy standard deviation cases σ={0.25,0.1}K\sigma=\{0.25,0.1\}~{}\text{K}.

NrefN_{\mathrm{ref}} influences the accuracy of the two error estimators used inside the adSSLE algorithm. They are: (i) the residual expansion accuracy ELOOE_{\mathrm{LOO}} in Eq. (23) and (ii) the splitting error EsplitE_{\mathrm{split}} in Eq. (24).

Small values of NrefN_{\mathrm{ref}} allow to quickly obtain a crude likelihood approximation with limited experimental design sizes NEDN_{\mathrm{ED}}, but this comes at the cost of lower convergence rates at larger NEDN_{\mathrm{ED}}. This behaviour can be partially attributed to the deterioration of residual expansion error ELOOE_{\mathrm{LOO}} in Eq. (23). At small experimental design sizes, the overall number of terminal domains is relatively small and this effect is not as pronounced. At larger experimental designs and higher numbers of subdomains, however, the error estimators high variances can lead to difficulties in identifying the true high error subdomains.

Large values of NrefN_{\mathrm{ref}} lead to slower initial convergence rates because of the smaller number of overall subdomains. The algorithm stability, however, is increased because both error estimators have lower variance and thereby allow the algorithm to more reliably identify the true high error subdomains and choose the split directions that maximize Eq. (25).

5.3 High-dimensional diffusion problem

The last cast study shows that adSSLE for Bayesian model inversion remains feasible in high dimensional problems with low effective dimensionality. The considered forward model is often used as a standard benchmark in UQ computations (Shin and Xiu, 2016; Fajraoui et al., 2017). It represents the 1-D diffusion along a domain with coordinate ξ[0,1]\xi\in[0,1] given by the following boundary value problem:

ξ[κ(ξ)uξ(ξ)]=1,with{u(0)=0,uξ(1)=1.-\frac{\partial}{\partial\xi}\left[\kappa(\xi)\frac{\partial u}{\partial\xi}(\xi)\right]=1,\quad\text{with}\quad\begin{cases}u(0)=0,\\ \frac{\partial u}{\partial\xi}(1)=1.\end{cases} (33)

The concentration field uu can be used to describe any steady-state diffusion driven process (e.g., heat diffusion, concentration diffusion, etc.). Assume that the diffusion coefficient κ\kappa is a log-normal random field given by κ(ξ,ω)=exp(10+3g(ξ))\kappa(\xi,\omega)=\exp{(10+3g(\xi))} where gg is a standard normal stationary Gaussian random field with exponential autocorrelation function ρ(ξ,ξ)=exp(3|ξξ|)\rho(\xi,\xi^{\prime})=\exp{(-3\left|\xi^{\prime}-\xi\right|)}. Let gg be approximated through a truncated Karhunen-Loève expansion

g(ξ)k=1MXkek(ξ),g(\xi)\approx\sum_{k=1}^{M}X_{k}e_{k}(\xi), (34)

with the pairwise uncorrelated random variables XkX_{k} denoting the field coefficients and the real valued function eke_{k} obtained from the solution of the Fredholm equation for ρ\rho (Ghanem and Spanos, 1991). The truncation variable is set to M=62M=62 to explain 99%99\% of the variance. Some realizations of the random field and resulting concentrations are shown in Figure 11.

In this example, the random vector of coefficients 𝑿=(X1,,X62)\bm{X}=(X_{1},\dots,X_{62}) shall be inferred using a single measurement of the diffusion field at u(ξ=1)u(\xi=1) given by 𝒴=0.16{\mathcal{Y}}=0.16. The considered model therefore takes as an input a realization of that random vector, and returns the diffusion field at ξ=1\xi=1, i.e., :𝒙u(1){\mathcal{M}}:\bm{x}\mapsto u(1). It is expected that due to the single measurement location at ξ=1\xi=1, only very little information about the parameters will be recovered in the inverse problem.

Refer to caption
(a) Diffusion coefficient
Refer to caption
(b) Concentration
Figure 11: High-dimensional diffusion problem: 55 independent realizations of 𝑿\bm{X} and the resulting κ\kappa and uu with the model output u(1)u(1) highlighted with a circle.

We impose a standard normal prior on the field coefficients such that π(𝒙)=i=162𝒩(xi|0,1)\pi(\bm{x})=\prod_{i=1}^{62}{\mathcal{N}}(x_{i}|0,1) and assume the standard additive discrepancy model with known discrepancy variance σ2=106\sigma^{2}=10^{-6}. This yields the likelihood function

(𝒙;𝒴)=12πσ2exp(12σ2(0.16(𝒙))2).{\mathcal{L}}(\bm{x};{\mathcal{Y}})=\frac{1}{2\pi\sigma^{2}}\exp\left(-\frac{1}{2\sigma^{2}}\left(0.16-{\mathcal{M}}(\bm{x})\right)^{2}\right). (35)

We proceed to compare the performance of standard SLE, non-adaptive SSLE and the proposed adSSLE approach on this example. We solve the problem with a set of maximum likelihood evaluations NED={2,000;4,000;6,000;8,000;10,000}N_{\mathrm{ED}}=\{2{,}000;4{,}000;6{,}000;8{,}000;10{,}000\}.

In the present high-dimensional case, it is necessary to set NrefN_{\mathrm{ref}} to a relatively large number (Nref=2,000N_{\mathrm{ref}}=2{,}000). At smaller NrefN_{\mathrm{ref}} numbers the variance of the estimator of EsplitE_{\mathrm{split}} in Eq. (24) makes it difficult for the algorithm to correctly identify the splitting direction that maximizes Eq. (25).

To compare the results of the algorithms, they are compared to a reference MCMC solution obtained with the affine-invariant ensemble sampler (Goodman and Weare, 2010) algorithm with 100,000100{,}000 steps and 100100 parallel chains at a total cost of 10710^{7} likelihood evaluations.

To allow a quantitative comparison, we again use the error measure from Eq. (28) with M=62M=62. It is plotted for a set of maximum likelihood evaluations in Figure 12. It is clear that both SSLE algorithms outperform SLE, while the adSSLE approach manages to improve the performance of SLE by an order of magnitude.

The overall small magnitude of the error η\eta in Figure 12 can be attributed to the low active dimensionality of this problem. Despite its high nominal dimensionality (M=62M=62), this problem in fact has only very few active dimensions, as the first few variables are significantly more important than the rest. In physical terms, very local fluctuations of the conductivity do not influence the output u(1)u(1) which results from an integration of these fluctuations. Therefore, the biggest change between prior and posterior distribution happens in the first few parameters (see also Table 3), while the other parameters remain unchanged by the updating procedure. This results in a small value of the Jensen-Shannon divergence for the inactive dimensions that lower the average value η\eta as defined in Eq. (28).

Refer to caption
Figure 12: High-dimensional diffusion problem: Convergence of the η\eta error measure (Eq. (28)) over five replications for SLE, SSLE with a static experimental design and the proposed adSSLE approach.

To highlight the results of one adSSLE algorithm instance with NED=10,000N_{\mathrm{ED}}=10{,}000, we display plots of the marginal posteriors in Figure 13. Due to the low active dimensionality of the problem, we focus on the first 33 parameters {X1,X2,X3}\{X_{1},X_{2},X_{3}\}. The remaining posterior parameters are not significantly influenced by the considered data. The comparative plots show a good agreement between the adSSLE and the reference solution, especially w.r.t. the interaction between X1X_{1} and {X2,X3}\{X_{2},X_{3}\}. As can be expected from the single measurement location, only the first parameter X1X_{1} is significantly influenced by the Bayesian updating procedure.

For the same instance, we also compute the first two posterior moments for all posterior marginals and compare them to the MCMC reference solution. The resulting values are presented in Table 3. Keeping in mind that the prior distribution is a multivariate standard normal distribution (𝔼[Xi]=0{\mathbb{E}}\left[X_{i}\right]=0 and Var[Xi|𝒴]=1\sqrt{{\rm Var}\left[X_{i}|{\mathcal{Y}}\right]}=1 for i=1,,62i=1,\dots,62) it is obvious from this table that the data most significantly affects the first three parameters.

The adSSLE approach manages to accurately recover the first two posterior moments at the relatively low cost of NED=10,000N_{\mathrm{ED}}=10{,}000 likelihood evaluations. The average absolute error for 𝔼[Xi]{\mathbb{E}}\left[X_{i}\right] and Var[Xi|𝒴]\sqrt{{\rm Var}\left[X_{i}|{\mathcal{Y}}\right]} is approximately 0.020.02.

Refer to caption
(a) adSSLE
Refer to caption
(b) MCMC
Figure 13: High-dimensional diffusion problem: Comparison plots of the posterior distribution marginals for the first 33 parameters {X1,X2,X3}\{X_{1},X_{2},X_{3}\} computed from adSSLE (NED=10,000N_{\mathrm{ED}}=10{,}000) and MCMC (NED=107N_{\mathrm{ED}}=10^{7}). The prior marginals are shown in red.
Table 3: High-dimensional diffusion problem: Posterior mean and standard deviation for selected marginals Xi|𝒴,i=1,,6,10,20,,50,62X_{i}|{\mathcal{Y}},i=1,\dots,6,10,20,\dots,50,62. The values in brackets are computed from the MCMC reference solution. The prior is a multivariate standard normal distribution (𝔼[Xi]=0{\mathbb{E}}\left[X_{i}\right]=0 and Var[Xi|𝒴]=1\sqrt{{\rm Var}\left[X_{i}|{\mathcal{Y}}\right]}=1).
X1X_{1} X2X_{2} X3X_{3} X4X_{4} X5X_{5} X6X_{6}
𝔼[Xi|𝒴]{\mathbb{E}}\left[X_{i}|{\mathcal{Y}}\right] 0.88(0.89)-0.88(-0.89) 0.14(0.14)0.14(0.14) 0.07(0.09)0.07(0.09) 0.04(0.03)0.04(-0.03) 0.04(0.00)-0.04(-0.00) 0.04(0.03)-0.04(0.03)
Var[Xi|𝒴]\sqrt{{\rm Var}\left[X_{i}|{\mathcal{Y}}\right]} 0.16(0.14)0.16(0.14) 1.07(1.04)1.07(1.04) 0.97(1.03)0.97(1.03) 1.02(1.00)1.02(1.00) 1.00(1.00)1.00(1.00) 0.98(1.03)0.98(1.03)
X10X_{10} X20X_{20} X30X_{30} X40X_{40} X50X_{50} X62X_{62}
𝔼[Xi|𝒴]{\mathbb{E}}\left[X_{i}|{\mathcal{Y}}\right] 0.00(0.03)0.00(0.03) 0.00(0.00)0.00(0.00) 0.02(0.01)-0.02(-0.01) 0.03(0.01)-0.03(-0.01) 0.00(0.01)-0.00(0.01) 0.00(0.03)0.00(-0.03)
Var[Xi|𝒴]\sqrt{{\rm Var}\left[X_{i}|{\mathcal{Y}}\right]} 1.03(1.01)1.03(1.01) 1.03(0.98)1.03(0.98) 1.00(1.00)1.00(1.00) 1.04(1.02)1.04(1.02) 1.00(0.99)1.00(0.99) 1.00(0.98)1.00(0.98)

6 Conclusions

Motivated by the recently proposed spectral likelihood expansions (SLE) framework for Bayesian model inversion presented in Nagel and Sudret (2016), we showed that the same analytical post-processing capabilities can be derived when the novel SSE approach from Marelli et al. (2021) is applied to likelihood functions, giving rise to the proposed SSLE approach. Because SSE is designed for models with complex local characteristics, it was expected to outperform SLE on practically relevant, highly localized likelihood functions. To further improve SSLE performance, we introduced a novel adaptive sampling procedure and modified partitioning strategy.

There are a few unsolved shortcomings of SSLE that will be addressed in future works. Namely, the discontinuities at the subdomain boundaries may be a source of error that should be addressed. Additionally, for the adSSLE algorithm, it is not possible at the moment to specify the optimal NrefN_{\mathrm{ref}} parameter a priori. In light of the considerable influence of that parameter as shown in Section 5.2.2, it might be necessary to adaptively adjust it, or decouple this parameter from the termination criterion.

Approximating likelihood functions through local PCEs prohibits the enforcement of strict positivity throughout the function domain. For visualization purposes this is not an issue, as negative predictions can simply be set to 0 in a post-processing step. When computing posterior expectations with Eq. (22), however, this can lead to erroneous results such as negative posterior variances. One way to enforce strict positivity is through an initial transformation of the likelihood function (e.g., log-likelihood logk𝒦fkPCE(𝑿)\log{{\mathcal{L}}}\approx\sum_{k\in{\mathcal{K}}}f_{k}^{\mathrm{PCE}}(\bm{X})). This is avoided in the present work because it comes at a loss of the desirable analytical post-processing properties.

Another class of problems that can cause instability when focusing directly on the likelihood function (both for SLE and SSLE), is that of problems with a very sharp likelihood function. This can happen, e.g. in the case of many available measurements and small discrepancy variances, that cause the likelihood function to reduce to the Dirac delta function, which has a very dense spectral representation.

In problems with few active dimensions (e.g. High-dimensional diffusion problem, Section 5.3), SSLE performs well because the likelihood is constant in many dimensions. The local sparse PCE construction can exploit this and could therefore handle up to hundreds of input dimensions. However, if the number of active dimensions is high as well, the present SSLE algorithm will require prohibitively large experimental designs, thereby rendering it unfeasible.

The biggest advantage of SSLE, however, is that it poses the challenging Bayesian computation in a function approximation setting. This yields an analytical expression of the posterior distribution and preserves the analytical post-processing capabilities of SLE while delivering highly improved likelihood function approximations. As opposed to many existing algorithms, SSLE can even efficiently solve Bayesian problems with multiple posterior modes. As shown in the case studies, the proposed adaptive algorithm further capitalizes on the compact support nature of likelihood functions and leads to significant performance gains, especially at larger experimental designs.

Acknowledgements

The PhD thesis of the first author is supported by ETH grant #44 17-1.

References

  • Beck and Au (2002) Beck, J. L. and S.-K. Au (2002). Bayesian updating of structural models and reliability using Markov chain Monte Carlo simulation. Journal of Engineering Mechanics 128(4), 380–391.
  • Beck and Katafygiotis (1998) Beck, J. L. and L. S. Katafygiotis (1998). Updating models and their uncertainties. I: Bayesian statistical framework. Journal of Engineering Mechanics 124(4), 455–461.
  • Birolleau et al. (2014) Birolleau, A., G. Poëtte, and D. Lucor (2014). Adaptive Bayesian inference for discontinuous inverse problems, application to hyperbolic conservation laws. Communications in Computational Physics 16(1), 1–34.
  • Bishop (2006) Bishop, C. M. (2006). Pattern recognition and machine learning. Information Science and Statistics. Springer.
  • Blatman and Sudret (2010) Blatman, G. and B. Sudret (2010). An adaptive algorithm to build up sparse polynomial chaos expansions for stochastic finite element analysis. Probabilistic Engineering Mechanics 25, 183–197.
  • Blatman and Sudret (2011) Blatman, G. and B. Sudret (2011). Adaptive sparse polynomial chaos expansion based on Least Angle Regression. Journal of Computational Physics 230, 2345–2367.
  • Brooks and Gelman (1998) Brooks, S. P. and A. Gelman (1998). General methods for monitoring convergence of iterative simulations. Journal of Computational and Graphical Statistics 7(4), 434–455.
  • Chapelle et al. (2002) Chapelle, O., V. Vapnik, and Y. Bengio (2002). Model selection for small sample regression. Journal of Machine Learning Research 48(1), 9–23.
  • Ching and Chen (2007) Ching, J. and Y. Chen (2007). Transitional Markov chain Monte Carlo method for Bayesian model updating, model class selection, and model averaging. Journal of Engineering Mechanics 133(7), 816–832.
  • Conrad et al. (2018) Conrad, P. R., A. D. Davis, Y. M. Marzouk, N. S. Pillai, and A. Smith (2018, jan). Parallel local approximation MCMC for expensive models. SIAM/ASA Journal on Uncertainty Quantification 6(1), 339–373.
  • Conrad et al. (2016) Conrad, P. R., Y. M. Marzouk, N. S. Pillai, and A. Smith (2016, oct). Accelerating asymptotically exact MCMC for computationally intensive models via local approximations. Journal of the American Statistical Association 111(516), 1591–1607.
  • Cui et al. (2014) Cui, T., Y. M. Marzouk, and K. E. Willcox (2014, aug). Data-driven model reduction for the Bayesian solution of inverse problems. International Journal for Numerical Methods in Engineering 102(5), 966–990.
  • Dick et al. (2017) Dick, J., R. N. Gantner, Q. T. Le Gia, and C. Schwab (2017). Multilevel higher-order quasi-Monte Carlo Bayesian estimation. Mathematical Models and Methods in Applied Sciences 27(5), 953–995.
  • El Moselhy and Marzouk (2012) El Moselhy, T. A. and Y. M. Marzouk (2012). Bayesian inference with optimal maps. Journal of Computational Physics 231(23), 7815–7850.
  • Fajraoui et al. (2017) Fajraoui, N., S. Marelli, and B. Sudret (2017). Sequential design of experiment for sparse polynomial chaos expansions. SIAM/ASA Journal on Uncertainty Quantification 5(1), 1061–1085.
  • Gantner and Peters (2018) Gantner, R. N. and M. D. Peters (2018). Higher-order quasi-Monte Carlo for Bayesian shape inversion. SIAM/ASA Journal on Uncertainty Quantification 6(2), 707–736.
  • Gautschi (2004) Gautschi, W. (2004). Orthogonal polynomials: Computation and approximation. Numerical Mathematics and Scientific Computation. Oxford University Press.
  • Gelman et al. (2014) Gelman, A., J. B. Carlin, H. S. Stern, D. B. Dunson, A. Vehtari, and D. B. Rubin (2014). Bayesian Data Analysis (3 ed.). Texts in Statistical Science. CRC Press.
  • Gelman and Rubin (1992) Gelman, A. and D. B. Rubin (1992). Inference from iterative simulation using multiple sequences. Statistical Science 7(4), 457–472.
  • Ghanem and Spanos (1991) Ghanem, R. G. and P. Spanos (1991). Stochastic finite elements – A spectral approach. Springer Verlag, New York. (Reedited by Dover Publications, Mineola, 2003).
  • Goodman and Weare (2010) Goodman, J. and J. Weare (2010). Ensemble samplers with affine invariance. Communications in Applied Mathematics and Computational Science 5(1), 65–80.
  • Haario et al. (2001) Haario, H., E. Saksman, and J. Tamminen (2001). An adaptive Metropolis algorithm. Bernoulli 7(2), 223–242.
  • Harney (2016) Harney, H. L. (2016). Bayesian Inference: Data Evaluation and Decisions (2 ed.). Springer International Publishing.
  • Jakeman et al. (2015) Jakeman, J., M. S. Eldred, and K. Sargsyan (2015). Enhancing 1\ell_{1}-minimization estimates of polynomial chaos expansions using basis selection. Journal of Computational Physics 289, 18–34.
  • Jaynes (2003) Jaynes, E. T. (2003). Probability theory: The logic of science. Cambridge University Press.
  • Jeffreys (1946) Jeffreys, H. (1946). An invariant form for the prior probability in estimation problems. Proceedings of the Royal Society of London A: Mathematical, Physical and Engineering Sciences 186(1007), 453–461.
  • Kennedy and O’Hagan (2001) Kennedy, M. C. and A. O’Hagan (2001). Bayesian calibration of computer models. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 63(3), 425–464.
  • Li and Marzouk (2014) Li, J. and Y. M. Marzouk (2014). Adaptive construction of surrogates for the Bayesian solution of inverse problems. SIAM Journal on Scientific Computing 36(3), A1163–A1186.
  • Lin (1991) Lin, J. (1991). Divergence measures based on the Shannon entropy. Transactions on Information Theory 37(1), 145–151.
  • Lüthen et al. (2021) Lüthen, N., S. Marelli, and B. Sudret (2021). Sparse polynomial chaos expansions: Review and benchmark. SIAM/ASA Journal on Uncertainty Quantification.
  • Marelli and Sudret (2014) Marelli, S. and B. Sudret (2014). UQLab: A framework for uncertainty quantification in Matlab. In Vulnerability, Uncertainty, and Risk (Proceedings of the 2nd International Conference on Vulnerability, Risk Analysis and Management (ICVRAM2014), Liverpool, United Kingdom), pp.  2554–2563.
  • Marelli and Sudret (2019) Marelli, S. and B. Sudret (2019). UQLab user manual – Polynomial chaos expansions. Technical report, Chair of Risk, Safety and Uncertainty Quantification, ETH Zurich, Switzerland. Report # UQLab-V1.3-104.
  • Marelli et al. (2021) Marelli, S., P.-R. Wagner, C. Lataniotis, and B. Sudret (2021). Stochastic spectral embedding. International Journal for Uncertainty Quantification 11(2), 25–47.
  • Marin et al. (2012) Marin, J.-M., P. Pudlo, C. P. Robert, and R. J. Ryder (2012). Approximate Bayesian computational methods. Statistics and Computing 22(6), 1167–1180.
  • Marzouk et al. (2016) Marzouk, Y., T. Moselhy, M. Parno, and A. Spantini (2016). Sampling via measure transport: An introduction. In R. Ghanem, D. Higdon, and H. Owhadi (Eds.), Handbook of Uncertainty Quantification. Springer International Publishing.
  • Marzouk et al. (2007) Marzouk, Y. M., H. N. Najm, and L. A. Rahn (2007). Stochastic spectral methods for efficient Bayesian solution of inverse problems. Journal of Computational Physics 224, 560–586.
  • Marzouk and Xiu (2009) Marzouk, Y. M. and D. Xiu (2009). A stochastic collocation approach to Bayesian inference in inverse problems. Communications in Computational Physics 6(4), 826–847.
  • Nagel and Sudret (2016) Nagel, J. and B. Sudret (2016). Spectral likelihood expansions for Bayesian inference. Journal of Computational Physics 309, 267–294.
  • Neal (2011) Neal, R. M. (2011). MCMC using Hamiltonian dynamics. In S. Brooks, A. Gelman, G. L. Jones, and X.-L. Meng (Eds.), Handbook of Markov Chain Monte Carlo, Handbooks of Modern Statistical Methods, Chapter 5, pp.  113–162. Chapman & Hall/CRC.
  • Parno (2015) Parno, M. D. (2015). Transport maps for accelerated Bayesian computation. PhD thesis, Massachusetts Institute of Technology (MIT).
  • Robert and Casella (2004) Robert, C. P. and G. Casella (2004). Monte Carlo statistical methods (2nd Ed.). Springer Series in Statistics. Springer Verlag.
  • Rosenblatt (1952) Rosenblatt, M. (1952). Remarks on a multivariate transformation. Annals of Mathematical Statistics 23(3), 470–472.
  • Shin and Xiu (2016) Shin, Y. and D. Xiu (2016). Nonadaptive quasi-optimal points selection for least squares linear regression. SIAM Journal on Scientific Computing 38(1), A385–A411.
  • Sisson et al. (2018) Sisson, S. A., Y. Fan, and M. A. Beaumont (Eds.) (2018). Handbook of approximate Bayesian computation. Chapman and Hall/CRC.
  • Tarantola (2005) Tarantola, A. (2005). Inverse Problem Theory and Methods for Model Parameter Estimation. Society for Industrial and Applied Mathematics (SIAM).
  • Tierney and Kadane (1986) Tierney, L. and J. B. Kadane (1986). Accurate approximations for posterior moments and marginal densities. Journal of the American Statistical Association 81(393), 82–86.
  • Tierney et al. (1989a) Tierney, L., R. E. Kass, and J. B. Kadane (1989a). Approximate marginal densities of nonlinear functions. Biometrika 76(3), 425–433.
  • Tierney et al. (1989b) Tierney, L., R. E. Kass, and J. B. Kadane (1989b). Fully exponential Laplace approximations to expectations and variances of nonpositive functions. Journal of the American Statistical Association 84(407), 710–716.
  • Torre et al. (2019) Torre, E., S. Marelli, P. Embrechts, and B. Sudret (2019, jul). Data-driven polynomial chaos expansion for machine learning regression. Journal of Computational Physics 388, 601–623.
  • Wagner et al. (2020) Wagner, P.-R., R. Fahrni, M. Klippel, A. Frangi, and B. Sudret (2020, feb). Bayesian calibration and sensitivity analysis of heat transfer models for fire insulation panels. Engineering Structures 205(15), 110063.
  • Wand and Jones (1995) Wand, M. and M. C. Jones (1995). Kernel smoothing. Chapman and Hall, Boca Raton.
  • Xiu (2010) Xiu, D. (2010). Numerical methods for stochastic computations – A spectral method approach. Princeton University press.
  • Xiu and Karniadakis (2002) Xiu, D. and G. E. Karniadakis (2002). The Wiener-Askey polynomial chaos for stochastic differential equations. SIAM Journal on Scientific Computing 24(2), 619–644.
  • Yan and Zhou (2019) Yan, L. and T. Zhou (2019, mar). Adaptive multi-fidelity polynomial chaos approach to Bayesian inference in inverse problems. Journal of Computational Physics 381, 110–128.