This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\newsiamremark

remarkRemark \newsiamremarkhypothesisHypothesis \newsiamremarkexampleExample \newsiamthmclaimClaim \headersUQ for Large Linear Inverse ProblemsC. Travelletti, D. Ginsbourger, and N. Linde

Uncertainty Quantification and Experimental Design for Large-Scale Linear Inverse Problems under Gaussian Process Priors thanks: \fundingThis work was funded by the Swiss National Science Fundation (SNF) through project no. 178858.

Cédric Travelletti Institute of Mathematical Statistics and Actuarial Science, University of Bern, Bern, Switzerland (). [email protected]    David Ginsbourger Institute of Mathematical Statistics and Actuarial Science, University of Bern, Bern, Switzerland () [email protected]    Niklas Linde Institute of Earth Sciences, University of Lausanne, Lausanne, Switzerland () [email protected]
Abstract

We consider the use of Gaussian process (GP) priors for solving inverse problems in a Bayesian framework. As is well known, the computational complexity of GPs scales cubically in the number of datapoints. We here show that in the context of inverse problems involving integral operators, one faces additional difficulties that hinder inversion on large grids. Furthermore, in that context, covariance matrices can become too large to be stored. By leveraging recent results about sequential disintegrations of Gaussian measures, we are able to introduce an implicit representation of posterior covariance matrices that reduces the memory footprint by only storing low rank intermediate matrices, while allowing individual elements to be accessed on-the-fly without needing to build full posterior covariance matrices. Moreover, it allows for fast sequential inclusion of new observations. These features are crucial when considering sequential experimental design tasks. We demonstrate our approach by computing sequential data collection plans for excursion set recovery for a gravimetric inverse problem, where the goal is to provide fine resolution estimates of high density regions inside the Stromboli volcano, Italy. Sequential data collection plans are computed by extending the weighted integrated variance reduction (wIVR) criterion to inverse problems. Our results show that this criterion is able to significantly reduce the uncertainty on the excursion volume, reaching close to minimal levels of residual uncertainty. Overall, our techniques allow the advantages of probabilistic models to be brought to bear on large-scale inverse problems arising in the natural sciences. Particularly, applying the latest developments in Bayesian sequential experimental design on realistic large-scale problems opens new venues of research at a crossroads between mathematical modelling of natural phenomena, statistical data science and active learning.

1 Introduction

Gaussian processes (GP) provide a powerful Bayesian approach to regression (Rasmussen and Williams,, 2006). While traditional regression considers pointwise evaluations of an unknown function, GPs can also include data in the form of linear operators (Solak et al.,, 2003; Särkkä,, 2011; Jidling et al.,, 2017; Mandelbaum,, 1984; Tarieladze and Vakhania,, 2007; Hairer et al.,, 2005; Owhadi and Scovel,, 2015; Klebanov et al.,, 2020). This allows GPs to provide a Bayesian framework to address inverse problems (Tarantola and Valette,, 1982; Stuart,, 2010; Dashti and Stuart,, 2016). Even though GP priors have been shown to perform well on smaller-scale inverse problems, difficulties arise when one tries to apply them to large inverse problems (be it high dimensional or high resolution) and these get worse when one considers sequential data assimilation settings such as in Chevalier et al., 2014a . The main goal of this work is to overcome these difficulties and to provide solutions for scaling GP priors to large-scale Bayesian inverse problems.

Specifically, we focus on a triple intersection of i) linear operator data, ii) large number of prediction points and iii) sequential data acquisition. There exists related works that focus on some of these items individually. For example, methods for extending Gaussian processes to large datasets (Hensman et al.,, 2013; Wang et al.,, 2019) or to a large number of prediction points (Wilson et al.,, 2020) gained a lot of attention over the last years. Also, much work has been devoted to extending GP regression to include linear constraints (Jidling et al.,, 2017) or integral observations (Hendriks et al.,, 2018; Jidling et al.,, 2019). On the sequential side, methods have been developed relying on infinite-dimensional state-space representations (Särkkä et al.,, 2013) and have also been extended to variational GPs (Hamelijnck et al.,, 2021). There are also works that focus on the three aspects at the same time (Solin et al.,, 2015). We note that all these approaches rely on approximations, while our method does not.

The topic of large-scale sequential assimilation of linear operator data has also been of central interest in the Kalman filter community. To the best of our knowledge, techniques employed in this framework usually rely on a low rank representation of the covariance matrix, obtained either via factorization (Kitanidis,, 2015) or from an ensemble estimate (Mandel,, 2006). Our goal in this work is to elaborate similar methods for Gaussian processes without relying on a particular factorization of the covariance matrix. As stated above, we focus on the case where: i) the number of prediction points is large, ii) the data has to be assimilated sequentially, and iii) it comes in the form of integral operators observations. Integral operators are harder to handle than pointwise observations since, when discretized on a grid (which is the usual inversion approach), they turn into a matrix with entries that are not predominantly null, in our case non-zero for most grid points. For the rest of this work, we will only consider settings that enjoy these three properties. This situation is typical of Bayesian large-scale inverse problems because those are often solved on a discrete grid, forcing one to consider a large number of prediction points when inverting at high resolution; besides, the linear operators found in inverse problems are often of integral form (e.g. gravity, magnetics).  
{example*} For the rest of this work, we will use as red thread a real-world inverse problem that enjoys the above properties. This problem is that of reconstructing the underground mass density inside the Stromboli volcano, Italy, from observations of the vertical component of the gravity field at different locations of the surface of the volcano (with respect to a reference station), see Fig. 1.

Refer to caption
(a)
Refer to caption
(b)
Figure 1: Example inverse problem: (a) Simulated underground mass density inside the Stromboli volcano (realisation from GP prior). (b) vertical intensity of the generated gravity field at locations where data has been gathered (Linde et al.,, 2014). Colorscales were chosen arbitrarily.

We will use a real dataset of gravimetric observations that was collected during a field campaign in 2012 (courtesy of Linde et al., (2014)). In Linde et al., (2014) the inversion domain is discretized at 50[m]50~{}[m] resolution, which, as we explain in Section 3 results in larger-than-memory covariance matrices, calling for the methods developed in this work.

Our main contribution to overcome the above difficulties is the introduction of an implicit representation of the posterior covariance matrix that only requires storage of low rank intermediate matrices and allows individual elements to be accessed on-the-fly, without ever storing the full matrix. Our method relies on an extension of the kriging update formulae (Chevalier et al., 2014c, ; Emery,, 2009; Gao et al.,, 1996; Barnes and Watson,, 1992) to linear operator observations. As a minor contribution, we also provide a technique for computing posterior means on fine discretizations using a chunking technique and explain how to perform posterior simulations in the considered setting. The developed implicit representation allows for fast updates of posterior covariances under linear operator observations on very large grids. This is particularly useful when computing sequential data acquisition plans for inverse problems, which we demonstrate by computing sequential experimental designs for excursion set learning in gravimetric inversion. We find that our method provides significant computational time savings over brute-force conditioning and scales to problem sizes that are too large to handle using state-of-the-art techniques.

Although the contribution of this work is mostly methodological, the techniques developed here can be formulated in a rigorous abstract framework by using the language of Gaussian measures an disintegrations. Though less well-known than GPs, Gaussian measures provide an equivalent approach (Rajput and Cambanis,, 1972) that offers several advantages. First, it is more natural for discretization-independent formulations; second, it provides a clear description (in terms of dual spaces) of the observation operators for which a posterior can be defined and third in integrates well with excursion set estimation. Readers interessted in the technical details are refered to the results in (Travelletti and Ginsbourger,, 2022) upon which we build our implicit representation framework.

To demonstrate our method, we apply it to the Stromboli inverse problem described above. In this context, we show how it allows computing the posterior at high resolution, how hyperparameters of the prior can be trained and how we can sample from the posterior on a large grid. Finally, we illustrate how our method may be applied to a state-of-the-art sequential experimental design task for excursion set recovery. Sequential design criterion for excursion sets have gained a lot of attention recently (Azzimonti et al.,, 2019), but, to the best of our knowledge, this is the first time that sequential experimental design for set estimation is considered in the setting of Bayesian inverse problems.

2 Computing the Posterior under Operator Observations: Bottlenecks and Solutions

In this section, we focus on the challenges that arise when using Gaussian process priors to solve Bayesian inverse problems with linear operators observations and propose solutions to overcome these. Here an inverse problem is the task of recovering some unknown function ρC(D)\rho\in C(D) from observation of a bounded linear operator G:C(D)pG:C(D)\rightarrow\mathbb{R}^{p} applied to ρ\rho. To solve the problem within a Bayesian framework, one puts a Gaussian process prior on DD and uses the conditional law of the process, conditional on the data, to approximate the unknown ρ\rho.

Even if one can formulate the problem in an infinite dimensional setting (Travelletti and Ginsbourger,, 2022) and discretization should take place as late as possible (Stuart,, 2010), when solving inverse problems in practice there is always some form of discretization involved, be it through quadrature methods (Hansen,, 2010) or through basis expansion (Wagner et al.,, 2021). It turns out that, regardless of the type of discretization used, one quickly encounters computational bottlenecks arising from memory limitations when trying to scale inversion to real-world problems. We here focus on inverse problems discretized on a grid, but stress that the computational difficulties described next also plague spectral approaches.

Let 𝑾=(w1,,wr)Dr\bm{W}=\left(w_{1},...,w_{r}\right)\in D^{r} be a given set of discretization points, we consider observation operators that (after discretization) may be written as linear combinations of Dirac delta functionals

(1) G:C(D)p,G=(j=1rgijδwj)i,i=1,,p,G:C(D)\rightarrow\mathbb{R}^{p},~{}G=\left(\sum_{j=1}^{r}g_{ij}\delta_{w_{j}}\right)_{i},~{}i=1,...,p,

with arbitrary coefficients gijg_{ij}\in\mathbb{R}. Using (Travelletti and Ginsbourger,, 2022, Corollary 5) we can compute the conditional law of a GP ZGp(m,k)Z\sim Gp(m,k) on DD conditionally on the data 𝒀=G¯Z𝑾+ϵ\bm{Y}=\underline{G}Z_{\bm{W}}+\bm{\epsilon}, where ϵ𝒩(𝟎,τ2𝑰p)\bm{\epsilon}\sim\mathcal{N}\left(\bm{0},\tau^{2}\bm{I}_{p}\right) is some observational noise. The conditional mean and covariance are then given by:

(2) m~𝑿\displaystyle\tilde{m}_{\bm{X}} =m𝑿+K𝑿𝑾G¯T(G¯K𝑾𝑾G¯T+τ2𝑰p)1(𝒚G¯m𝑾),\displaystyle=m_{\bm{X}}+K_{\bm{X}\bm{W}}\underline{G}^{T}\left(\underline{G}K_{\bm{W}\bm{W}}\underline{G}^{T}+\tau^{2}\bm{I}_{p}\right)^{-1}\left(\bm{y}-\underline{G}m_{\bm{W}}\right),
(3) K~𝑿𝑿\displaystyle\tilde{K}_{\bm{X}\bm{X}^{\prime}} =K𝑿𝑿K𝑿𝑾G¯T(G¯K𝑾𝑾G¯T+τ2𝑰p)1K𝑾𝑿.\displaystyle=K_{\bm{X}\bm{X}^{\prime}}-K_{\bm{X}\bm{W}}\underline{G}^{T}\left(\underline{G}K_{\bm{W}\bm{W}}\underline{G}^{T}+\tau^{2}\bm{I}_{p}\right)^{-1}K_{\bm{W}\bm{X}^{\prime}}.

Notation: When working with discrete operators as in Eq. 1 it is more convenient to use matrices. Hence we will use G¯\underline{G} to denote the p×rp\times r matrix with elements gijg_{ij}. Relation Eq. 1 will then be written compactly as ρG¯ρ𝑾\rho\mapsto\underline{G}\rho_{\bm{W}}. Similarly, given a Gaussian process ZGP(m,k)Z\sim GP(m,k) on DdD\subset\mathbb{R}^{d} and another set of points 𝑿=(x1,,xm)TDm\bm{X}=\left(x_{1},...,x_{m}\right)^{T}\in D^{m}, we will use K𝑿𝑾K_{\bm{X}\bm{W}} to denote the m×rm\times r matrix obtained by evaluating the covariance function at all couples of points Kij=k(xi,wj)K_{ij}=k(x_{i},w_{j}). In a similar fashion, let Z𝑿mZ_{\bm{X}}\in\mathbb{R}^{m} denote the vector obtained by concatenating the values of the field at the different points. From now on, boldface letters will be used to denote concatenated quantities (usually datasets). The identity matrix will be denoted by 𝑰\bm{I}, the dimension being infered from the context.

Even if Eqs. 2 and 3 only involve basic matrix operations, their computational cost depends heavily on the number of prediction points 𝑿=(x1,,xm)\bm{X}=\left(x_{1},...,x_{m}\right) and on the number of discretization points 𝑾=(w1,,wr)\bm{W}=\left(w_{1},...,w_{r}\right), making their application to real-world inverse problems a non-trivial task. Indeed, when both mm and rr are big, there are two main difficulties that hamper the computation of the conditional distribution:

  • the r×rr\times r matrix K𝑾𝑾K_{\bm{W}\bm{W}} may never be built in memory due to its size, and

  • the m×mm\times m posterior covariance K𝑿𝑿K_{\bm{X}\bm{X}} may be too large to store.

The first of these difficulties can be solved by performing the product K𝑾𝑾GTK_{\bm{W}\bm{W}}G^{T} in chunks, as described in Section 2.3. The second one only becomes of particular interest in sequential data assimilation settings as considered in Section 2.1. In Section 2.2 we will introduce an (almost) matrix-free implicit representation of the posterior covariance matrix enabling us to overcome both these memory bottlenecks.

2.1 Sequential Data Assimilation in Bayesian Inverse Problems and Update Formulae

We now consider a sequential data assimilation problem where data is made available in discrete stages. At each stage, a set of observations described by a (discretized) operator G¯i\underline{G}_{i} is made and one observes a realization 𝒚i\bm{y}_{i} of

(4) 𝒀i=G¯iZ𝑾i+ϵi,\bm{Y}_{i}=\underline{G}_{i}Z_{\bm{W}_{i}}+\bm{\epsilon}_{i},

where 𝑾i\bm{W}_{i} is some set of points in DD. Then, the posterior mean and covariance after each stage of observation may be obtained by performing a low rank update of their counterparts at the previous stage. Indeed, (Travelletti and Ginsbourger,, 2022, Corollary 6) provides an extension of Chevalier et al., 2014c ; Emery, (2009); Gao et al., (1996); Barnes and Watson, (1992) to linear operator observations, and gives:

Theorem 2.1.

Let ZGp(m,k)Z\sim Gp(m,k) and let m(n)m^{(n)} and K(n)K^{(n)} denote the conditional mean and covariance function conditional on the data {𝐘i=𝐲i:i=1,,n}\{\bm{Y}_{i}=\bm{y}_{i}:i=1,...,n\} with 𝐘i\bm{Y}_{i} defined as in Eq. 4, where n1n\geq 1 and m(0)m^{(0)} and K(0)K^{(0)} are used to denote the prior mean and covariance. Then:

m𝑿(n)\displaystyle m^{(n)}_{\bm{X}} =m𝑿(n1)+λn(𝑿)T(𝒚nG¯nm𝑾n(n1)),\displaystyle=m^{(n-1)}_{\bm{X}}+\lambda_{n}\left(\bm{X}\right)^{T}\left(\bm{y}_{n}-\underline{G}_{n}~{}m^{(n-1)}_{\bm{W}_{n}}\right),
K𝑿𝑿(n)\displaystyle K^{(n)}_{\bm{X}\bm{X}^{\prime}} =K𝑿𝑿(n1)λn(𝑿)TSnλn(𝑿),\displaystyle=K^{(n-1)}_{\bm{X}\bm{X}^{\prime}}-\lambda_{n}\left(\bm{X}\right)^{T}S_{n}\lambda_{n}\left(\bm{X}^{\prime}\right),

with λn(𝐗)\lambda_{n}\left(\bm{X}\right), SnS_{n} defined as:

λn(𝑿)\displaystyle\lambda_{n}\left(\bm{X}\right) =Sn1G¯nK𝑾n𝑿(n1),\displaystyle=S_{n}^{-1}\underline{G}_{n}K^{(n-1)}_{\bm{W}_{n}\bm{X}},
Sn\displaystyle S_{n} =G¯nK𝑾n𝑾n(n1)G¯nT+τ2𝑰.\displaystyle=\underline{G}_{n}K^{(n-1)}_{\bm{W}_{n}\bm{W}_{n}}\underline{G}_{n}^{T}+\tau^{2}\bm{I}.

At each stage ii, these formulae require computation of the pi×mp_{i}\times m matrix λi(𝑿)\lambda_{i}\left(\bm{X}\right), which involves a pi×pip_{i}\times p_{i} matrix inversion, where pip_{i} is the dimension of the operator G¯n\underline{G}_{n} describing the current dataset to be included. This allows computational savings by reusing already computed quantities, avoiding inverting the full dataset at each stage, which would require a ptot2p_{tot}^{2} matrix inversion, where ptot=i=1npip_{tot}=\sum_{i=1}^{n}p_{i}.

In order for these update equations to bring computational savings, one has to be able to store the past covariances K𝑾n𝑾n(n1)K_{\bm{W}_{n}\bm{W}_{n}}^{(n-1)} (Chevalier et al.,, 2015). This makes their application to large-scale sequential Bayesian inverse problems difficult, since the covariance matrix on the full discretization may become too large for storage above a certain number of discretization points. The next section presents our main contributions to overcome this limitation. They rely on an implicit representation of the posterior covariance that allows the computational savings offered by the kriging update formulae to be brought to bear on large scale inverse problems.

2.2 Implicit Representation of Posterior Covariances for Sequential Data Assimilation in Large-Scale Bayesian Inverse Problems

We consider the same sequential data assimilation setup as in the previous section, and for the sake of simplicity we assume that 𝑾1,,𝑾n=𝑿\bm{W}_{1},...,\bm{W}_{n}=\bm{X} and use the lighter notation m(i):=m𝑿(i)m^{(i)}:=m^{(i)}_{\bm{X}} and K(i):=K𝑿𝑿(i)K^{(i)}:=K^{(i)}_{\bm{X}\bm{X}}. The setting we are interested in here is the one where 𝑿\bm{X} is so large that the covariance matrix gets bigger than the available computer memory.

Our key insight is that instead of building the full posterior covariance K(n)K^{(n)} at each stage nn, one can just maintain a routine that computes the product of the current posterior covariance with any other low rank matrix. More precisely, at each stage nn, we provide a routine CovMuln\textrm{CovMul}_{n} (Algorithm 1), that allows to compute the product of the current covariance matrix with any thin matrix Am×q,qmA\in\mathbb{R}^{m\times q},~{}q\ll m:

CovMuln:AK(i)A,\textit{CovMul}_{n}:A\mapsto K^{(i)}A,

where thin is to be understood as small enough so that the result of the multiplication can fit in memory.

This representation of the posterior covariance was inspired by the covariance operator of Gaussian measures. Indeed, if we denote by Cμ(n)C_{\mu^{(n)}} the covariance operator of the Gaussian measure associated to the posterior distribution of the GP at stage nn, then

(K(n)A)ij=k=1mCμ(n)δxi,δxkAkj.\left(K^{(n)}A\right)_{ij}=\sum_{k=1}^{m}\left\langle C_{\mu^{(n)}}\delta_{x_{i}},\delta_{x_{k}}\right\rangle A_{kj}.

Hence, the procedure CovMuln\textrm{CovMul}_{n} may be thought of as computing the action of the covariance operator of the Gaussian measure associated to the posterior on the Dirac delta functionals at the discretization points.

This motivates us to think in terms of an updatable covariance object, where the inclusion of new observations (the updating) amounts to redefining a right-multiplication routine. It turns out that by grouping terms appropriately in Theorem 2.1 such a routine may be defined by only storing low rank matrices at each data acquisition stage.  

Lemma 2.2.

For any nn\in\mathbb{N} and any m×qm\times q matrix AA:

K(n)A=K(0)Ai=1nK¯iRi1K¯iTA,K^{(n)}A=K^{(0)}A-\sum_{i=1}^{n}\bar{K}_{i}R_{i}^{-1}\bar{K}_{i}^{T}A,

with intermediate matrices K¯i\bar{K}_{i} and Ri1R_{i}^{-1} defined as:

K¯i:\displaystyle\bar{K}_{i}: =K(i1)G¯iT,\displaystyle=K^{(i-1)}\underline{G}_{i}^{T},
Ri1:\displaystyle R_{i}^{-1}: =(G¯iK(i1)G¯iT+τ2𝑰)1.\displaystyle=\left(\underline{G}_{i}K^{(i-1)}\underline{G}_{i}^{T}+\tau^{2}\bm{I}\right)^{-1}.

Hence, in order to compute products with the posterior covariance at stage nn, one only has to store nn matrices K¯i\bar{K}_{i}, each of size m×pim\times p_{i} and nn matrices Ri1R_{i}^{-1} of size pi×pip_{i}\times p_{i}, where pip_{i} is the number of observations made at stage ii (i.e. the number of lines in G¯i\underline{G}_{i}). In turn, each of these objects is defined by multiplications with the covariance matrix at previous stages, so that one may recursively update the multiplication procedure CovMuln\textrm{CovMul}_{n}. Algorithms 1, LABEL:, 2, and 3 may be used for multiplication with the current covariance matrix, update of the representation and update of the posterior mean.

Algorithm 1 Covariance Right Multiplication Procedure CovMuln\textrm{CovMul}_{n}
Precomputed matrices K¯i,Ri1\bar{K}_{i},~{}R_{i}^{-1}, i=1,,ni=1,\dotsc,n.
Prior multiplication routine CovMul0\textrm{CovMul}_{0}.
Input matrix AA.
K(n)AK^{(n)}A.
procedure CovMuln\textrm{CovMul}_{n}(A)
     Compute K(0)A=CovMul0(A)K^{(0)}A=\textrm{CovMul}_{0}(A).
     Return K(0)Ai=1nK¯iRi1K¯iTAK^{(0)}A-\sum_{i=1}^{n}\bar{K}_{i}R_{i}^{-1}\bar{K}_{i}^{T}A.
Algorithm 2 Updating intermediate quantities at conditioning stage nn
Last multiplication routine CovMuln1\textrm{CovMul}_{n-1}.
Measurement matrix G¯n\underline{G}_{n}, noise variance τ2\tau^{2}.
Step nn intermediate matrices K¯n\bar{K}_{n} and RnR_{n}
procedure Updaten\textrm{Update}_{n}
     Compute K¯n=CovMuln1G¯nT\bar{K}_{n}=\textrm{CovMul}_{n-1}\underline{G}_{n}^{T}.
     Compute Rn1=(G¯nK¯n+τ2𝑰)1R_{n}^{-1}=\left(\underline{G}_{n}\bar{K}_{n}+\tau^{2}\bm{I}\right)^{-1}.
Algorithm 3 Computation of conditional mean at step nn
Previous conditional mean m(n1)m^{(n-1)}.
Current data 𝒚n\bm{y}_{n} and forward G¯n\underline{G}_{n}.
Intermediate matrices K¯n\bar{K}_{n} and Rn1R_{n}^{-1}.
Step nn conditional mean m(n)m^{(n)}.
procedure MeanUpdaten\textrm{MeanUpdate}_{n}
     Return m(n1)+K¯nRn1(𝒚nG¯nm(n1))m^{(n-1)}+\bar{K}_{n}R_{n}^{-1}\left(\bm{y}_{n}-\underline{G}_{n}m^{(n-1)}\right).

2.3 Prior Covariance Multiplication Routine and Chunking

To use Algorithm 1, one should be able to compute products with the prior covariance matrix K(0)K^{(0)}. This may be achieved by chunking the set of grid points into ncn_{c} subsets 𝑿=(𝑿1,,𝑿nc)\bm{X}=\left(\bm{X}_{1},...,\bm{X}_{n_{c}}\right), where each 𝑿i\bm{X}_{i} contains a subset of the points. Without loss of generality, we assume all subsets to have the same size mcm_{c}. We may then write the product as

K𝑿𝑿(0)A=(K𝑿1𝑿(0)A,,K𝑿nc𝑿(0)A)T.K^{(0)}_{\bm{X}\bm{X}}A=\left(K^{(0)}_{\bm{X}_{1}\bm{X}}A,\dotsc,K^{(0)}_{\bm{X}_{n_{c}}\bm{X}}A\right)^{T}.\\

Each of the subproducts may then be performed separately and the results gathered together at the end. The individual products then involve matrices of size mc×mm_{c}\times m and m×qm\times q. One can then choose the number of chunks so that these matrices can fit in memory. Each block K𝑿i𝑿(0)K^{(0)}_{\bm{X}_{i}\bm{X}} may be built on-demand provided K𝑿𝑿(0)K^{(0)}_{\bm{X}\bm{X}} is defined through a given covariance function.
 
This ability of the prior covariance to be built quickly on-demand is key to our method. The fact that the prior covariance matrix does not need to be stored allows us to handle larger-than-memory posterior covariances by expressing products with it as a multiplication with the prior and a sum of multiplications with lower rank matrices.
 

Remark 2.3 (Choice of Chunk Size).

Thanks to chunking, the product may be computed in parallel, allowing for significant performance improvements in the presence of multiple computing devices (CPUs, GPUs, …). In that case, the chunk size should be chosen as large as possible to limit data transfers, but small enough so that the subproducts may fit on the devices.

2.4 Computational Cost and Comparison to Non-Sequential Inversion

For the sake of comparison, assume that all n datasets have the same size pcp_{c} and let p=npcp=np_{c} denote the total data size. The cost of computing products with the current posterior covariance matrix at some intermediate stage is given by:

Lemma 2.4 (Multiplication Cost).

Let AA be an m×qm\times q matrix. Then, the cost of computing KnAK_{n}A at some stage nn using Algorithms 1 and 2 is 𝒪(m2q+n(mpcq+pc2q))\mathcal{O}\left(m^{2}q+n(mp_{c}q+p_{c}^{2}q)\right).

Using this recursively, we can then compute the cost of creating the implicit representation of the posterior covariance matrix at stage nn:

Lemma 2.5 (Implicit Representation Cost).

To leading order in mm and pp, the cost of defining CovMuln\textrm{CovMul}_{n} is 𝒪(m2p+mp2+p2pc)\mathcal{O}\left(m^{2}p+mp^{2}+p^{2}p_{c}\right). This is also the cost of computing m(n)m^{(n)}.

This can then be compared with a non-sequential approach where all datasets would be concatenated into a single dataset of dimension pp. More precisely, define the p×mp\times m matrix G¯\underline{G} and the pp-dimensional vector 𝒚\bm{y} as the concatenations of all the measurements and data vectors into a single operator, respectively vector. Then computing the posterior mean using Eq. 2 with those new observation operators and data vector the cost is, to leading order in pp and mm:

𝒪(m2p+mp2+p3).\mathcal{O}\left(m^{2}p+mp^{2}+p^{3}\right).

In this light, we can now sum up the two main advantages of the proposed sequential approach:

  • the cubic cost 𝒪(p3)\mathcal{O}\left(p^{3}\right) arising from the inversion of the data covariances is decreased to 𝒪(p2pc)\mathcal{O}\left(p^{2}p_{c}\right) in the sequential approach

  • if a new set of observations has to be included, then the direct approach will require the 𝒪(m2p)\mathcal{O}\left(m^{2}p\right) computation of the product KG¯TK\underline{G}^{T}, which can become prohibitively expensive when the number of prediction points is large, whereas the sequential approach will only require a marginal computation of 𝒪(m2pc)\mathcal{O}\left(m^{2}p_{c}\right).

Aside from the computational cost, our implicit representation also provides significant memory savings compared to an explicit approach where the full posterior covariance matrix would be stored. The storage requirement for the implicit-representation as a function of the number of discretization points mm is shown in Figure 2.

Refer to caption
Figure 2: Memory footprint of the posterior covariance matrix as a function of discretization size for explicit and implicit representation.

2.5 Toy Example: 2D Fourier Transform

To illustrate the various methods presented in Section 2.2, we here apply them to a toy two-dimensional example: we consider a GP discretized on a large square grid {0,,M}×{0,,M}\{0,...,M\}\times\{0,...,M\} and try to learn it through various types of linear operator data.
 
More precisely, we will here allow two types of observations: pointwise field values and Fourier coefficient data. Indeed, the field values at the nodes Zkl,k,l{1,,M}Z_{kl},~{}k,l\in\{1,...,M\} are entirely determined by its discrete Fourier transform (DFT):

(5) Fuv=k=1Ml=1MZkle2πi(ukM+vlM),u,v{1,,M}.\displaystyle F_{uv}=\sum_{k=1}^{M}\sum_{l=1}^{M}Z_{kl}e^{-2\pi i\left(\frac{uk}{M}+\frac{vl}{M}\right)},~{}u,v\in\{1,...,M\}.

Note that the above may be viewed as the discretized version of some operator and thus fits our framework for sequential conditioning. One may then answer questions such as how much uncertainty is left after the first 10 Fourier coefficient Fkl,k,l=1,,10F_{kl},~{}k,l=1,...,10 have been observed? Or which Fourier coefficient provide the most information about the GP?

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Figure 3: Posterior mean after observation of n=10,50,100n=10,50,100 Fourier coefficients (top) and after observation of n=10,50,100n=10,50,100 field values along a space-filling design (bottom). Ground truth is shown on the left. Correlation parameter (Matérn 5/2) λ=0.5\lambda=0.5.

For example, Fig. 3 compares the posterior mean after observation of various Fourier coefficients, compared to observing pointwise values along a space-filling sequence. The GP model is a constant mean Matérn 5/25/2 random field on a square domain [1,1]2\left[-1,1\right]^{2}. The domain is discretized onto a 400×400400\times 400 grid. Note that, by nature, each Fourier coefficient involves the field values at all points of the discretization grid and thus direct computation of the posterior mean requires the full 4002×4002400^{2}\times 400^{2} covariance matrix (which would translate to roughly 100 GB of storage). This makes this situation suitable to demonstrate the techniques presented in this section.

In this example, the Fourier coefficients are ordered by growing ll_{\infty} norm. One observes that Fourier coefficients provide very different information than pointwise observations and decrease uncertainty in a more spatially uniform way, as shown in Fig. 4.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Figure 4: Posterior standard deviation after observation of n=10,50,100n=10,50,100 Fourier coefficients (top) and after observation of n=10,50,100n=10,50,100 field values along a space-filling design (bottom). Ground truth is shown on the left. Correlation parameter (Matérn 5/2) λ=0.5\lambda=0.5.

The extent to which Fourier coefficients provide more or less information than pointwise observations (depending on ordering) depends on the correlation length of the field’s covariance function. Indeed, for long correlation lengths, the low frequency Fourier coefficients contain most of the information about the field’s global behaviour, whereas for small correlation length, high frequency coefficients are needed to learn the small-scale structure.

3 Application: Scaling Gaussian Processes to Large-Scale Inverse Problems

In this section, we demonstrate how the implicit representation of the posterior covariance introduced in Section 2.2 allows scaling Gaussian processes to situations that are too large to handle using more traditional techniques. Such situations are frequently encountered in large-scale inverse problems and we will thus focus our exposition on such a problem arising in geophysics. In this setting, we demonstrate how our implicit representation allows to train prior hyperparameters in Section 3.1, in Section 3.2 we demonstrate posterior sampling on large grids and finally in Section 3.3 address a state-of-the-art sequential experimental design problem for excursion set recovery.

Example Gravimetric Inverse Problem: We focus on the problem of reconstructing the mass density distribution ρ:D\rho:D\rightarrow\mathbb{R} in some given underground domain DD from observations of the vertical component of the gravitational field at points s1,,sps_{1},...,s_{p} on the surface of the domain. Such gravimetric data are extensively used on volcanoes (Montesinos et al.,, 2006; Represas et al.,, 2012; Linde et al.,, 2017) and can be acquired non-destrucively at comparatively low costs by 1-2 persons in rough terrain using gravimeters, without the need for road infrastructure or other installations (such as boreholes) that are often lacking on volcanoes. Gravimeters are sensitive to spatial variations in density, which makes them useful to understand geology, localize ancient volcano conduits and present magma chambers, and to identify regions of loose light-weight material that are prone to landslides that could in the case of volcanic islands generate tsunamis.

As an example dataset we use gravimetric data gathered on the surface of the Strombli volcano during a field campaign in 2012 (Linde et al.,, 2014). In Section 3.3 we will also consider the problem of recovering high (or low) density regions inside the volcano. Fig. 5 displays the main components of the problem.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 5: Problem overview: (a) underground mass density (realisation from GP prior), (b) vertical intensity of the generated gravity field at selected locations, (c) high density regions and (d) low density regions. Thresholds and colorscales were chosen arbitrarily.

The observation operator describing gravity measurements is an integral one (see Appendix A), which, after discretization, fits the Bayesian inversion framework of Section 2. Indeed, this kind of problems is usually discretized on a finite grid of points 𝑿=(x1,,xm)\bm{X}=\left(x_{1},\dotsc,x_{m}\right), hence the available data is of the form

(6) 𝒀=G¯ρ𝑿+ϵ,\bm{Y}=\underline{G}\rho_{\bm{X}}+\bm{\epsilon},

where the p×mp\times m matrix G¯\underline{G} represents the discretized version of the observation operator for the gravity field at s1,,sps_{1},\dotsc,s_{p} and we assume i.i.d Gaussian noise ϵ𝒩(0,τ2Ip)\bm{\epsilon}\sim\mathcal{N}(0,\tau^{2}I_{p}). The posterior may then be computed using Eqs. 2 and 3. Note that the three-dimensional nature of the problem quickly makes it intractable for traditional inversion methods as the resolution of the inversion grid is refined. For example, when discretizing the Stromboli inversion domain as Linde et al., (2014) into cubic cells of 50m50m side length, one is left with roughly 21052\cdot 10^{5} cells, which translates to posterior covariance matrices of size 160GB160~{}\mathrm{GB} (using 32 bits floating point numbers). This characteristic of gravimetric inverse problems make them well-suited to demonstrate the implicit representation framework which we introduced in Section 2.2. In Section 3.3 we will show how our technique allows state-of-the-art adaptive design techniques to be applied to large real-world problems.

3.1 Hyperparameter Optimization

When using Gaussian process priors to solve inverse problems, one has to select the hyperparameters of the prior. There exists different approaches for optimizing hyperparameters. We here only consider maximum likelihood estimation (MLE).

We restrict ourselves to GP priors that have a constant prior mean m0m_{0}\in\mathbb{R} and a covariance kernel kk that depends on a prior variance parameter σ02\sigma_{0}^{2} and other correlation parameters 𝜽0t\bm{\theta}_{0}\in\mathbb{R}^{t}:

(7) k(x,y)=σ02r(x,y;𝜽0),k(x,y)=\sigma_{0}^{2}r(x,y;\bm{\theta}_{0}),

where r(.,.;𝜽0)r(.,.;\bm{\theta}_{0}) is a correlation function, such that r(x,x;𝜽0)=1,xDr(x,x;\bm{\theta}_{0})=1,\forall x\in D. The maximum likelihood estimator for the hyperparameters may then be obtained by minimizing the negative marginal log likelihood (nmll) of the data, which in the discretized setting of Section 2 may be writen as (Rasmussen and Williams,, 2006):

(8) (m0,σ0,𝜽0;𝒚)\displaystyle\mathcal{L}\left(m_{0},\sigma_{0},\bm{\theta}_{0};\bm{y}\right) =12logdetR+12(𝒚G¯m𝑿)TR1(𝒚G¯m𝑿)+n2log2π,\displaystyle=\frac{1}{2}\log\det R+\frac{1}{2}\Big{(}\bm{y}-\underline{G}m_{\bm{X}}\Big{)}^{T}R^{-1}\Big{(}\bm{y}-\underline{G}m_{\bm{X}}\Big{)}+\frac{n}{2}\log 2\pi,
(9) R\displaystyle R :=(G¯K𝑿𝑿G¯T+τ2𝑰n).\displaystyle:=\left(\underline{G}K_{\bm{X}\bm{X}}\underline{G}^{T}+\tau^{2}\bm{I}_{n}\right).

Since only the quadratic term depends on m0m_{0}, we can adapt concentration identities (Park and Baek,, 2001) to write the optimal m0m_{0} as a function of the other hyperparameters:

(10) m^0MLE(σ0,𝜽0)=(𝟏mTG¯TR1G¯𝟏m)1yTR1G𝟏m,\hat{m}_{0}^{MLE}\left(\sigma_{0},\bm{\theta}_{0}\right)=\Big{(}\bm{1}_{m}^{T}\underline{G}^{T}R^{-1}\underline{G}\bm{1}_{m}\Big{)}^{-1}y^{T}R^{-1}G\bm{1}_{m},

where 𝟏m\bm{1}_{m} denotes the mm-dimensional column vector containing only 11’s. Here we always assume RR to be invertible. The remaining task is then to minimize the concentrated nmll:

(σ0,𝜽0)(m^0MLE(σ0,𝜽0),σ0,𝜽0).\left(\sigma_{0},\bm{\theta}_{0}\right)\mapsto\mathcal{L}\left(\hat{m}_{0}^{MLE}\left(\sigma_{0},\bm{\theta}_{0}\right),\sigma_{0},\bm{\theta}_{0}\right).

Note that the main computational challenge in the minimization of Eq. 8 comes from the presence of the m×mm\times m matrix K𝑿𝑿K_{\bm{X}\bm{X}}. In the following, we will only consider the case of kernels that depend on a single length scale parameter: 𝜽0=λ0\bm{\theta}_{0}=\lambda_{0}\in\mathbb{R}, though the procedure described below can in principle be adapted for multidimensional 𝜽0\bm{\theta}_{0}.

In practice, for kernels of the form Eq. 7 the prior variance σ02\sigma_{0}^{2} may be factored out of the covariance matrix (for known noise variance), so that only the prior length scale λ0\lambda_{0} appears in this large matrix. One then optimizes these parameters separately, using chunking (Section 2.3) to compute matrix products. Since σ0\sigma_{0} only appears in an p×pp\times p matrix which does not need to be chunked (the data size pp being moderate in real applications), one can use automatic differentiation libraries such as Paszke et al., (2019) to optimize it by gradient descent. On the other hand, there is no way to factor out λ0\lambda_{0} out of the large matrix K𝑿𝑿K_{\bm{X}\bm{X}}, so we resort to a brute force approach by specifying a finite search space for it. To summarize, we proceed here in the following way:

  1. (i)

    (brute force search) Discretize the search space for the length scale by only allowing λ0Λ0\lambda_{0}\in\Lambda_{0}, where Λ0\Lambda_{0} is a discrete set (usually equally spaced values on a reasonable search interval);

  2. (ii)

    (gradient descent) For each possible value of λ0\lambda_{0}, minimize the (concentrated) \mathcal{L} over the remaining free parameter σ0\sigma_{0} by gradient descent.

We ran the above approach on the Stromboli dataset with standard stationary kernels (Matérn 3/23/2, Matérn 5/25/2, exponential). In agreement with Linde et al., (2014), the observational noise standard deviation is 0.1[mGal]0.1~{}[mGal]. The optimization results for different values of the length scale parameter are shown in Fig. 6. The best estimates of the parameter values for each kernel are shown in Table 1. The table also shows the practical range λ¯\bar{\lambda} which is defined as the distance at which the covariance falls to 5%5\% of its original value.

Refer to caption
(a)
Refer to caption
(b)
Figure 6: (a) Concentrated negative marginal log-likelihood and (b) optimal hyperparameter values for different length scale parameters λ0\lambda_{0}.

We asses the robustness of each kernel by predicting a set of left out observations using the other remaining observations. Fig. 7 displays RMSE and negative log predictive density for different proportion of train/test splits.

Hyperparameters Metrics
Kernel λ\lambda λ¯\bar{\lambda} m0m_{0} σ0\sigma_{0} \mathcal{L} Train RMSE
Exponential 1925.0 5766.8 535.4 308.9 -804.4 0.060
Matérn 3/2 651.6 1952.0 2139.1 284.65 -1283.5 0.071
Matérn 5/2 441.1 1321..3 2120.9 349.5 -1247.6 0.073
Table 1: Optimal hyperparameters (Stromboli dataset) for different kernels.
Refer to caption
(a)
Refer to caption
(b)
Figure 7: (a) Root mean squared error and (b) negative log predictive density on test set for the different models (with optimal hyperparameters). The full dataset contains 501 observations.

Note the the above procedure is more of a quality assurance than a rigorous statistical evaluation of the model, since all datapoints were already used in the fitting of the hyperparameters. Due to known pathologies of the exponential kernel (MLE for length scale parameter going to infinity), we choose to use the Matérn 3/2 model for the experiments of Section 3.2 and Section 3.3. The maximum likelihood estimator of the prior hyperparameters for this model are m^0MLE=2139.1[kg/m3]\hat{m}_{0}^{MLE}=2139.1~{}[kg/m^{3}], σ^0MLE=284.65[kg/m3]\hat{\sigma}_{0}^{MLE}=284.65~{}[kg/m^{3}] and λ^0MLE=651.6[m]\hat{\lambda}_{0}^{MLE}=651.6~{}[m].

3.2 Posterior Sampling

Our implicit representation also allows for efficient sampling from the posterior by using the residual kriging algorithm (Chilès and Delfiner,, 2012; de Fouquet,, 1994), which we here adapt to linear operator observations. Note that in order to sample a Gaussian process at mm sampling points, one needs to generate mm correlated Gaussian random variables, which involves covariance matrices of size m2m^{2}, leading to the same computational bottlenecks as described in Section 2. On the other hand, the residual kriging algorithm generates realizations from the posterior by updating realizations of the prior, as we explain next.

As before, suppose we have a GP ZZ defined on some compact Euclidean domain DD and assume ZZ has continuous sample paths almost surely. Furthermore, say we have pp observations described by linear operators 1,,pC(D)\ell_{1},...,\ell_{p}\in C(D)^{*}. Then the conditional expectation of ZZ conditional on the σ\sigma-algebra Σ:=σ(1(Z),,p(Z))\Sigma:=\sigma\left(\ell_{1}\left(Z\right),...,\ell_{p}\left(Z\right)\right) is an orthogonal projection (in the L2L^{2}-sense (Williams,, 1991)) of ZZ onto Σ\Sigma. This orthogonality can be used to decompose the conditional law of ZZ conditional on Σ\Sigma into a conditional mean plus a residual. Indeed, if we let ZZ^{{}^{\prime}} be another GP with the same distribution as ZZ and let Σ:=σ(1(Z),,p(Z))\Sigma^{{}^{\prime}}:=\sigma\left(\ell_{1}\left(Z^{{}^{\prime}}\right),...,\ell_{p}\left(Z^{{}^{\prime}}\right)\right), then we have the following equality in distribution:

(11) ZxΣ\displaystyle Z_{x}\mid\Sigma =𝔼[ZxΣ]+(Zx𝔼[ZxΣ]),all xD.\displaystyle=\mathbb{E}\left[Z_{x}\mid\Sigma\right]+\left(Z^{{}^{\prime}}_{x}-\mathbb{E}\left[Z^{{}^{\prime}}_{x}\mid\Sigma^{{}^{\prime}}\right]\right),~{}\text{all }x\in D.

Compared to direct sampling of the posterior, the above approach involves two main operations: sampling from the prior and conditioning under operator data. When the covariance kernel is stationary and belongs to one of the usual families (Gaussian, Matérn), methods exist to sample from the prior on large grids (Mantoglou and Wilson,, 1982); whereas the conditioning part may be performed using our implicit representation.

Remark 3.1.

Note that in a sequential setting as in Section 2.1, the residual kriging algorithm may be used to maintain an ensemble of realizations from the posterior distribution by updating a fixed set of prior realizations at every step in the spirit of Chevalier et al., (2015).

3.3 Sequential Experimental Design for Excursion Set Recovery

As a last example of application where our implicit update method provides substantial savings, we consider a sequential data collection task involving an inverse problem. Though sequential design criterion for inverse problems have already been considered in the litterature (Attia et al.,, 2018), most of them only focus on selecting observations to improve the reconstruction of the unknown parameter field, or some linear functional thereof.

We here consider a different setting. In light of recent progress in excursion set estimation (Azzimonti et al.,, 2016; Chevalier et al.,, 2013), we instead focus on the task of recovering an excursion set of the unknown parameter field ρ\rho, that is, we want to learn the unknown set Γ:={xD:ρ(x)T}\Gamma^{*}:=\{x\in D:\rho\left(x\right)\geq T\}, where TT is some threshold. In the present context of Stromboli, high density areas are related to dykes (previous feeding conduits of the volcano), while low density values are related to deposits formed by paroxysmal explosive phreato-magmatic events (Linde et al.,, 2014). To the best of our knowledge such sequential experimental design problems for excursion set learning in inverse problems have not been considered elsewhere in the litterature.

Remark 3.2.

For the sake of simplicity, we focus only on excursion sets above some threshold, but all the techniques presented here may be readily generalized to generalized excursion sets of the form Γ:={xD:ρ(x)I}\Gamma^{*}:=\{x\in D:\rho\left(x\right)\in I\} where II is any finite union of intervals on the extended real line.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Refer to caption
(i)
Figure 8: Realizations from Matérn 3/23/2 GP prior (hyperparameters taken from Table 1) with corresponding excursion sets: (left to right) Underground mass density field (arbitrary colorscale), high density regions and low density regions, thresholds: 2600[kg/m3]2600~{}[kg/m^{3}] and 1700[kg/m3]1700~{}[kg/m^{3}].

We here consider a sequential setting, where observations are made one at a time and at each stage we have to select which obsevation to make next in order to optimally reduce the uncertainty on our estimate of Γ\Gamma^{*}. Building upon Picheny et al., (2010); Bect et al., 2012a ; Azzimonti et al., (2019); Chevalier et al., 2014a , there exists several families of criteria to select the next observations. Here, we restrict ourselves to a variant of the weighted IMSE criterion (Picheny et al.,, 2010). The investigation of other state-of-the-art criteria is left for future work. We note in passing that most Bayesian sequential design criteria involve posterior covariances and hence tend to become intractable for large-scale problems. Moreover in a sequential setting, fast updates of the posterior covariance are crucial. Those characteristics make the problem considered here particularly suited for the implicit update framework intoduced in Section 2.2.

The weighted IMSE criterion selects next observations by maximizing the variance reduction they will provide at each location, weighted by the probability for that location to belong to the excursion set Γ\Gamma^{*}. Assuming that nn data collection stages have already been performed and using the notation of Section 2.1, the variant that we are considering here selects the next observation location by maximizing the weighted integrated variance reduction (wIVR):

(12) wIVRn(s)=D(Kxx(n)Kxx(n+1)[Gs])pn(x)𝑑x,\textrm{wIVR}^{n}(s)=\int_{D}\left(K^{(n)}_{xx}-K^{(n+1)}_{xx}\left[G_{s}\right]\right)p_{n}(x)dx,

where ss is some potential observation location, K(n+1)K^{(n+1)} denotes the conditional covariance after including a gravimetric observation made at ss (this quantity is independent of the observed data) and GsG_{s} is the forward operator corresponding to this observation. Also, here pnp_{n} denotes the coverage function at stage nn (we refer the reader to Appendix B for more details on Bayesian set estimation). After discretization, applying Theorem 2.1 turns this criterion into:

(13) x𝑿Kx𝑿(n)GsT(GsK𝑿𝑿(n)GsT+τ2𝑰)1GsK𝑿x(n)pn(x),\sum_{x\in\bm{X}}K^{(n)}_{x\bm{X}}G_{s}^{T}\left(G_{s}K^{(n)}_{\bm{X}\bm{X}}G_{s}^{T}+\tau^{2}\bm{I}\right)^{-1}G_{s}K^{(n)}_{\bm{X}x}p_{n}(x),

where we have assumed that all measurements are affected by 𝒩(0,τ2)\mathcal{N}(0,\tau^{2}) distributed noise.

Note that for large-scale problems, the wIVR criterion in the form given in Eq. 13 becomes intractable for traditional methods because of the presence of the full posterior covariance matrix K𝑿𝑿(n)K^{(n)}_{\bm{X}\bm{X}} in the parenthesis. The implicit representation presented in Section 2.2 can be used to overcome this difficulty. Indeed, the criterion can be evaluated using the posterior covariance multiplication routine Lemma 2.2 (where the small dimension qq is now equal to the number of candidate observations considered at a time, here 11 but batch acquisition scenarios could also be tackled). New observations can be seamlessly integrated along the way by updating the representation using Algorithm 2.

Experiments and Results: We now study how the wIVR criterion can help to reduce the uncertainty on excursion sets within the Stromboli volcano. We here focus on recovering the volume of the excursion set instead of its precise location. To the best of our knowledge, in the existing literature such sequential design criteria for excursion set recovery have only been applied to small-scale inverse problems and have not been scaled to larger, more realistic problems where the dimensions at play prevent direct access to the posterior covariance matrix.

In the following experiments, we use the Stromboli volcano inverse problem and work with a discretization into cubic cells of 50[m]50~{}[m] side length. We use a Matérn 3/23/2 GP prior with hyperparameters trained on real data (Table 1) to generate semi-realistic ground truths for the experiments. We then simulate numerically the data collection process by computing the response that results from the considered ground truth and adding random observational noise. When computing sequential designs for excursion set estimation, the threshold that defines the excursion set can have a large impact on the accuracy of the estimate. Indeed, different thresholds will produce excursion sets of different sizes, which may be easier or harder to estimate depending on the set estimator used. For the present problem, Fig. 9 shows the distribution of the excursion volume under the considered prior for different excursion thresholds.

Refer to caption
Figure 9: Distribution of excursion set volume under the prior for different thresholds. Size is expressed as a percentage of the volume of the inversion domain.

It turns out that the estimator used in our experiments (Vorob’ev expectation) behaves differently depending on the size of the excursion set to estimate. Indeed, the Vorob’ev expectation tends to produce a smoothed version of the true excursion set, which in our situation results in a higher fraction of false positives for larger sets. Thus, we consider two scenarios: a large scenario where the generated excursion sets have a mean size of 10%10\% of the total inversion volume and a small scenario where the excursion sets have a mean size of 5%5\% of the total inversion volume. One should note that those percentages are in broad accordance with the usual size of excursion sets that are of interest in geology. The chosen thresholds are 2500[kg/m3]2500~{}[kg/m^{3}] for the large excursions and 2600[kg/m3]2600~{}[kg/m^{3}] for the small ones.

Refer to caption
(a) (large scenario) threshold: 2500 [kg/m3]{[}kg/m^{3}{]}
Refer to caption
(b) (small scenario) threshold: 2600[kg/m3]2600{[}kg/m^{3}{]}
Figure 10: Distribution of excursion volume (with kernel density estimate) under the prior for the two considered thresholds, together with excursion volumes for each ground truth.

The experiments are run on five different ground truths, which are samples from a Matérn 3/23/2 GP prior (see previous paragraphs). The samples were chosen such that their excursion set for the large scenario have volumes that correspond to the 5%5\%, 27.5%27.5\%, 50%50\%, 72.5%72.5\% and 95%95\% quantiles of the prior excursion volume distribution for the corresponding threshold. Fig. 10 shows the prior excursion volume distribution together with the volumes of the five different samples used for the experiments. Fig. 11 shows a profile of the excursion set (small scenario) for one of the five samples used in the experiments. The data collection location from the 2012 field campaign (Linde et al.,, 2014) are denoted by black dots. The island boundary is denoted by blue dots. Note that, for the sake of realism, in the experiments we only allow data collection at locations that are situated on the island (data acquired on a boat would have larger errors); meaning that parts of the excursion set that are outside of the island will be harder to recover.

Refer to caption
(a) xy projection
Refer to caption
(b) xz projection
Refer to caption
(c) yz projection
Figure 11: Projection of the excursion set (small scenario) for the first ground truth. Island boundary denoted in blue, observation location from previous field campaign denoted by black dots. Distances are displayed in [m] and density in [kg/m3m^{3}].

Experiments are run by starting at a fixed starting point on the volcano surface, and then sequentially choosing the next observation locations on the volcano surface according to the wIVR criterion. Datapoints are collected one at a time. We here only consider myopic optimization, that is, at each stage, we select the next observation site sn+1s_{n+1} according to:

sn+1=argmins𝑺cwIVRn(s),s_{n+1}=\operatorname*{arg\,min}_{s\in\bm{S}_{c}}\textrm{wIVR}^{n}(s),

where ties are broken arbitrarily. Here 𝑺c\bm{S}_{c} is a set of candidates among which to pick the next observation location. In our experiments, we fix 𝑺c\bm{S}_{c} to consist of all surface points within a ball of radius 150150 meters around the last observation location. Results are summarized in Figs. 12 and 13, which shows the evolution of the fraction of true positives and false positives as a function of the number of observations gathered.

Refer to caption
(a) True positives
Refer to caption
(b) False positives
Figure 12: Evolution of true and false positives for the large scenario as a function of the number of observations.

We see that in the large scenario (Fig. 12) the wIVR criterion is able to correctly detect 70 to 80% of the excursion set (in volume) for each ground truth after 450 observations. For the small scenario (Fig. 13) the amount of true positives reached after 450 observations is similar, though two ground truths are harder to detect.

Note that in Figs. 12 and 13 the fraction of false negatives is expressed as a percentage of the volume of the complementary of the true excursion set DΓD\setminus\Gamma^{*}. We see that the average percentage of false positives after 450450 observations tends to lie between 55 and 15%15\%, with smaller excursion sets yielding fewer false positives. While the Vorob’ev expectation is not designed to minimize the amount of false positives, there exists conservative set estimators (Azzimonti et al.,, 2019) that specialize on this task. We identify the extension of such estimators to inverse problems as a promising venue for new research.

Refer to caption
(a) True positives
Refer to caption
(b) False positives
Figure 13: Evolution of true and false positives for the small scenario as a function of the number of observations.

In both figures we also plot the fraction of true positives and false positives that result from the data collection plan that was used in Linde et al., (2014). Here only the situation at the end of the data collection process is shown. We see that for some of the ground truths the wIVR criterion is able to outperform static designs by around 10%. Note that there are ground truths where it performs similarly to a static design. We believe this is due to the fact that for certain ground truths most of the information about the excursion set can be gathered by spreading the observations across the volcano, which is the case for the static design that also considers where it is practical and safe to measure.

Refer to caption
(a) xy projection
Refer to caption
(b) xz projection
Refer to caption
(c) yz projection
Figure 14: Projection of the true excursion set (small scenario) and visited locations (wIVR strategy) for the first ground truth. Island boundary is shown in blue. Distances are displayed in [m] and density in [kg/m3m^{3}].

Limiting Distribution: The dashed horizontal lines in Figs. 12 and 13 show the detection percentage that can be achieved using the limiting distribution. We define the limiting distribution as the posterior distribution one were to obtain if one had gathered data at all allowed locations (everywhere on the volcano surface). This distribution may be approximated by gathering data at all points of a given (fine grained) discretization of the surface. In general, this is hard to compute since it requires ingestion of a very large amount of data, but thanks to our implicit representation (Section 2.2) we can get access to this object, thereby, allowing new forms of uncertainty quantification.

In a sense, the limiting distribution represents the best we can hope for when covering the volcano with this type of measurements (gravimetric). It gives a measure of the residual uncertainty inherent to the type of observations used (gravimetric). Indeed, it is known that a given density field is not identifiable from gravimetric data alone (see Blakely, (1995) for example). Even if gravity data will never allow for a perfect reconstruction of the excursion set, we can use the limiting distribution to compare the performance of different sequential design criteria and strategies. It also provides a mean of quantifying the remaining uncertainty under the chosen class of models. A sensible performance metric is then the number of observations that a given criterion needs to approach the minimal level of residual uncertainty which is given by the limiting distributions.

As a last remark, we stress that the above results and the corresponding reconstruction qualities are tied to an estimator, in our case the Vorob’ev expectation. If one were to use another estimator for the excursion set, those results could change significantly.

Posterior Volume Distribution: Thanks to our extension of the residual kriging algorithm to inverse problems (see Section 3.2), we are able to sample from the posterior at the end of the data collection process. This opens new venues for uncertainty quantification in inverse problems. For example, we can use sampling to estimate the posterior distribution of the excursion volume and estimate the residual uncertainty on the size of the excursion set.

Refer to caption
(a) (large scenario) threshold: 2500[kg/m3]2500~{}{[}kg/m^{3}{]}
Refer to caption
(b) (small scenario) threshold: 2600[kg/m3]2600~{}{[}kg/m^{3}{]}
Figure 15: Empirical posterior distribution (after 450 observations) of the excursion volume for each ground truth. True volumes are denoted by vertical lines.

Fig. 15 shows the empirical posterior distribution of the excursion volume for each of the ground truths considered in the preceding experiments. When compared to the prior distribution, Fig. 10, one sees that the wIVR criterion is capable of significantly reducing the uncertainty on the excursion volume. This shows that though the location of the excursion set can only be recovered with limited accuracy, as shown in Figs. 12 and 13, the excursion volume can be estimated quite well. This is surprising given that the criterion used (wIVR) is a very crude one and was not designed for that task. On the other hand, there exist more refined criterion, like the so-called SUR strategies (sequential uncertainty reduction) (Chevalier et al., 2014b, ; Bect et al.,, 2019), among which some were specifically engineered to reduce the uncertainty on the excursion volume (Bect et al., 2012b, ). Even though those criterion are more computationally challenging than the wIVR one, expecially in the considered framework, we identify their application to large Bayesian inverse problem as a promising avenue for future research.

4 Conclusion and Perspectives

Leveraging new results about sequential disintegrations of Gaussian measures (Travelletti and Ginsbourger,, 2022), we have introduced an implicit almost matrix free representation of the posterior covariance of a GP and have demonstrated fast update of the posterior covariance on large grids under general linear functional observations. Our method allows streamline updating and fast extraction of posterior covariance information even when the matrices are larger than the available computing memory. Using our novel implicit representation, we have shown how targeted design criteria for excursion set recovery may be extended to inverse problems discretized on large grids. We also demonstrated UQ on such problems using posterior sampling via residual kriging. Our results suggest that using the considered design criteria allows reaching close-to-minimal levels of residual uncertainty using a moderate number of observations and also exhibit significant reduction of uncertainty on the excursion volume. The GP priors used in this work are meant as a proof of concept and future work should address the pitfalls of such priors, such as lack of positiveness of the realisations and lack of expressivity. Other promising research avenues include extension to multivariate excursions Fossum et al., (2021) and inclusion of more sophisticated estimators such as conservative estimates Azzimonti et al., (2019). On the dynamic programming side, extending the myopic optimization of the criterion to finite horizon optimization in order to provide optimized data collection trajectories is an obvious next step which could have significant impact on the geophysics community. Also, including location dependent observation costs such as accessibility in the design criterion could help provide more realistic observation plans.

Acknowledgements

This work was funded by the Swiss National Science Foundation (SNF) through project no. 178858.

References

  • Attia et al., (2018) Attia, A., Alexanderian, A., and Saibaba, A. (2018). Goal-oriented optimal design of experiments for large-scale Bayesian linear inverse problems. Inverse Problems, 34.
  • Azzimonti et al., (2016) Azzimonti, D., Bect, J., Chevalier, C., and Ginsbourger, D. (2016). Quantifying uncertainties on excursion sets under a Gaussian random field prior. SIAM/ASA Journal on Uncertainty Quantification, 4(1):850–874.
  • Azzimonti et al., (2019) Azzimonti, D., Ginsbourger, D., Chevalier, C., Bect, J., and Richet, Y. (2019). Adaptive design of experiments for conservative estimation of excursion sets. Technometrics, 0(0):1–14.
  • Banerjee and Das Gupta, (1977) Banerjee, B. and Das Gupta, S. (1977). Gravitational attraction of a rectangular parallelepiped. Geophysics, 42(5):1053–1055.
  • Barnes and Watson, (1992) Barnes, R. J. and Watson, A. (1992). Efficient updating of kriging estimates and variances. Mathematical Geology, 24(1):129–133.
  • Bect et al., (2019) Bect, J., Bachoc, F., and Ginsbourger, D. (2019). A supermartingale approach to gaussian process based sequential design of experiments. Bernoulli, 25(4A):2883–2919.
  • (7) Bect, J., Ginsbourger, D., Li, L., Picheny, V., and Vazquez, E. (2012a). Sequential design of computer experiments for the estimation of a probability of failure. Statistics and Computing, 22(3):773–793.
  • (8) Bect, J., Ginsbourger, D., Li, L., Picheny, V., and Vazquez, E. (2012b). Sequential design of computer experiments for the estimation of a probability of failure. Statistics and Computing, 22(3):773–793.
  • Blakely, (1995) Blakely, R. J. (1995). Potential Theory in Gravity and Magnetic Applications. Cambridge University Press.
  • (10) Chevalier, C., Bect, J., Ginsbourger, D., Vazquez, E., Picheny, V., and Richet, Y. (2014a). Fast parallel kriging-based stepwise uncertainty reduction with application to the identification of an excursion set. Technometrics, 56(4):455–465.
  • (11) Chevalier, C., Bect, J., Ginsbourger, D., Vazquez, E., Picheny, V., and Richet, Y. (2014b). Fast parallel kriging-based stepwise uncertainty reduction with application to the identification of an excursion set. Technometrics, 56(4):455–465.
  • Chevalier et al., (2015) Chevalier, C., David, G., and Emery, X. (2015). Fast update of conditional simulation ensembles. Mathematical Geosciences, 47:771–789.
  • Chevalier et al., (2013) Chevalier, C., Ginsbourger, D., Bect, J., and Molchanov, I. (2013). Estimating and quantifying uncertainties on level sets using the Vorob’ev expectation and deviation with Gaussian process models. In mODa 10–Advances in Model-Oriented Design and Analysis, pages 35–43. Springer.
  • (14) Chevalier, C., Ginsbourger, D., and Emery, X. (2014c). Corrected kriging update formulae for batch-sequential data assimilation. In Pardo-Igúzquiza, E., Guardiola-Albert, C., Heredia, J., Moreno-Merino, L., Durán, J., and Vargas-Guzmán, J., editors, Mathematics of Planet Earth. Lecture Notes in Earth System Sciences. Springer, Berlin, Heidelberg.
  • Chilès and Delfiner, (2012) Chilès, J.-P. and Delfiner, P. (2012). Geostatistics: Modeling Spatial Uncertainty. Wiley, 2 edition.
  • Dashti and Stuart, (2016) Dashti, M. and Stuart, A. M. (2016). The Bayesian approach to inverse problems. Handbook of Uncertainty Quantification, pages 1–118.
  • de Fouquet, (1994) de Fouquet, C. (1994). Reminders on the conditioning kriging. In Armstrong, M. and Dowd, P. A., editors, Geostatistical Simulations, pages 131–145, Dordrecht. Springer Netherlands.
  • Emery, (2009) Emery, X. (2009). The kriging update equations and their application to the selection of neighboring data. Computational Geosciences, 13(3):269–280.
  • Fossum et al., (2021) Fossum, T. O., Travelletti, C., Eidsvik, J., Ginsbourger, D., and Rajan, K. (2021). Learning excursion sets of vector-valued Gaussian random fields for autonomous ocean sampling. The Annals of Applied Statistics, 15(2):597 – 618.
  • Gao et al., (1996) Gao, H., Wang, J., and Zhao, P. (1996). The updated kriging variance and optimal sample design. Mathematical Geology, 28(3):295–313.
  • Hairer et al., (2005) Hairer, M., Stuart, A. M., Voss, J., and Wiberg, P. (2005). Analysis of SPDEs arising in path sampling. Part I: The Gaussian case. Communications in Mathematical Sciences, 3(4):587 – 603.
  • Hamelijnck et al., (2021) Hamelijnck, O., Wilkinson, W. J., Loppi, N. A., Solin, A., and Damoulas, T. (2021). Spatio-temporal variational gaussian processes. In Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W., editors, Advances in Neural Information Processing Systems, volume 34. Curran Associates, Inc.
  • Hansen, (2010) Hansen, P. C. (2010). Discrete Inverse Problems. Society for Industrial and Applied Mathematics.
  • Hendriks et al., (2018) Hendriks, J. N., Jidling, C., Wills, A., and Schön, T. B. (2018). Evaluating the squared-exponential covariance function in gaussian processes with integral observations.
  • Hensman et al., (2013) Hensman, J., Fusi, N., and Lawrence, N. D. (2013). Gaussian processes for big data. In Nicholson, A. E. and Smyth, P., editors, Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence, UAI 2013, Bellevue, WA, USA, August 11-15, 2013. AUAI Press.
  • Jidling et al., (2019) Jidling, C., Hendriks, J., Schön, T. B., and Wills, A. (2019). Deep kernel learning for integral measurements.
  • Jidling et al., (2017) Jidling, C., Wahlström, N., Wills, A., and Schön, T. B. (2017). Linearly constrained gaussian processes. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  • Kitanidis, (2015) Kitanidis, P. K. (2015). Compressed state Kalman filter for large systems. Advances in Water Resources, 76:120 – 126.
  • Klebanov et al., (2020) Klebanov, I., Sprungk, B., and Sullivan, T. J. (2020). The linear conditional expectation in hilbert space.
  • Linde et al., (2014) Linde, N., Baron, L., Ricci, T., Finizola, A., Revil, A., Muccini, F., Cocchi, L., and Carmisciano, C. (2014). 3-D density structure and geological evolution of Stromboli volcano (Aeolian Islands, Italy) inferred from land-based and sea-surface gravity data. Journal of Volcanology and Geothermal Research, 273:58–69.
  • Linde et al., (2017) Linde, N., Ricci, t., Baron, L., A., S., and G., B. (2017). The 3-D structure of the Somma-Vesuvius volcanic complex (Italy) inferred from new and historic gravimetric data. Scientific Reports, 7(1):8434.
  • Mandel, (2006) Mandel, J. (2006). Efficient implementation of the ensemble Kalman filter. Technical Report 231, University of Colorado at Denver and Health Sciences Center.
  • Mandelbaum, (1984) Mandelbaum, A. (1984). Linear estimators and measurable linear transformations on a hilbert space. Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete, 65(3):385–397.
  • Mantoglou and Wilson, (1982) Mantoglou, A. and Wilson, J. L. (1982). The turning bands method for simulation of random fields using line generation by a spectral method. Water Resources Research, 18(5):1379–1394.
  • Molchanov, (2005) Molchanov, I. (2005). Theory of random sets. Springer.
  • Montesinos et al., (2006) Montesinos, F. G., Arnoso, J., Benavent, M., and Vieira, R. (2006). The crustal structure of El Hierro (Canary Islands) from 3-D gravity inversion. Journal of Volcanology and Geothermal Research, 150(1-3):283–299.
  • Owhadi and Scovel, (2015) Owhadi, H. and Scovel, C. (2015). Conditioning gaussian measure on hilbert space.
  • Park and Baek, (2001) Park, J.-S. and Baek, J. (2001). Efficient computation of maximum likelihood estimators in a spatial linear model with power exponential covariogram. Computers & Geosciences, 27(1):1–7.
  • Paszke et al., (2019) Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., and Chintala, S. (2019). Pytorch: An imperative style, high-performance deep learning library. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems 32, pages 8024–8035. Curran Associates, Inc.
  • Picheny et al., (2010) Picheny, V., Ginsbourger, D., Roustant, O., Haftka, R., and Kim, N. (2010). Adaptive designs of experiments for accurate approximation of a target region. Journal of Mechanical Design, 132:071008.
  • Rajput and Cambanis, (1972) Rajput, B. S. and Cambanis, S. (1972). Gaussian processes and Gaussian measures. Ann. Math. Statist., 43(6):1944–1952.
  • Rasmussen and Williams, (2006) Rasmussen, C. E. and Williams, C. K. I. (2006). Gaussian Processes for Machine Learning. The MIT Press.
  • Represas et al., (2012) Represas, P., Catalão, J. a., Montesinos, F. G., Madeira, J., Mata, J. a., Antunes, C., and Moreira, M. (2012). Constraints on the structure of Maio Island (Cape Verde) by a three-dimensional gravity model: imaging partially exhumed magma chambers. Geophysical Journal International, 190(2):931–940.
  • Robbins, (1944) Robbins, H. E. (1944). On the Measure of a Random Set. The Annals of Mathematical Statistics, 15(1):70 – 74.
  • Särkkä, (2011) Särkkä, S. (2011). Linear operators and stochastic partial differential equations in Gaussian process regression. In International Conference on Artificial Neural Networks, pages 151–158. Springer.
  • Särkkä et al., (2013) Särkkä, S., Solin, A., and Hartikainen, J. (2013). Spatiotemporal learning via infinite-dimensional bayesian filtering and smoothing: A look at gaussian process regression through kalman filtering. IEEE Signal Process. Mag., 30(4):51–61.
  • Solak et al., (2003) Solak, E., Murray-Smith, R., Leithead, W. E., Leith, D. J., and Rasmussen, C. E. (2003). Derivative observations in Gaussian process models of dynamic systems. In Advances in neural information processing systems, pages 1057–1064.
  • Solin et al., (2015) Solin, A., Kok, M., Wahlström, N., Schön, T., and Särkkä, S. (2015). Modeling and interpolation of the ambient magnetic field by gaussian processes. IEEE Transactions on Robotics, PP.
  • Stuart, (2010) Stuart, A. M. (2010). Inverse problems: A Bayesian perspective. Acta Numerica, 19:451–559.
  • Tarantola and Valette, (1982) Tarantola, A. and Valette, B. (1982). Generalized nonlinear inverse problems solved using the least squares criterion. Reviews of Geophysics, 20(2):219–232.
  • Tarieladze and Vakhania, (2007) Tarieladze, V. and Vakhania, N. (2007). Disintegration of Gaussian measures and average-case optimal algorithms. Journal of Complexity, 23(4):851 – 866. Festschrift for the 60th Birthday of Henryk Woźniakowski.
  • Travelletti and Ginsbourger, (2022) Travelletti, C. and Ginsbourger, D. (2022). Disintegration of gaussian measures for sequential assimilation of linear operator data. arXiv, abs/2207.13581.
  • Wagner et al., (2021) Wagner, P.-R., Marelli, S., and Sudret, B. (2021). Bayesian model inversion using stochastic spectral embedding. Journal of Computational Physics, 436:110141.
  • Wang et al., (2019) Wang, K., Pleiss, G., Gardner, J., Tyree, S., Weinberger, K. Q., and Wilson, A. G. (2019). Exact Gaussian processes on a million data points. In Wallach, H., Larochelle, H., Beygelzimer, A., d’ Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems 32, pages 14622–14632. Curran Associates, Inc.
  • Williams, (1991) Williams, D. (1991). Probability with Martingales. Cambridge University Press.
  • Wilson et al., (2020) Wilson, J. T., Borovitskiy, V., Terenin, A., Mostowsky, P., and Deisenroth, M. (2020). Efficiently sampling functions from Gaussian process posteriors. ArXiv, abs/2002.09309.

Appendix A Forward operator for Gravimetric Inversion

Given some subsurface density ρ:D\rho:D\rightarrow\mathbb{R} inside a domain D3D\subset\mathbb{R}^{3} and some location ss outside the domain, the vertical component of the gravitational field at ss is given by:

(14) 𝔊s[ρ\displaystyle\mathfrak{G}_{s}\left[\rho =Dρ(x)g(x,s)𝑑x,\displaystyle=\int_{D}\rho(x)g(x,s)dx,

with Green kernel

(15) g(x,s)=x(3)s(3)xs3,g(x,s)=\frac{x^{(3)}-s^{(3)}}{\|x-s\|^{3}},

where x(3)x^{(3)} denotes the vertical component of xx.

We discretize the domain DD into mm identic cubic cells D=i=1mDiD=\cup_{i=1}^{m}D_{i} with centroids 𝑿=(X1,,Xm)\bm{X}=\left(X_{1},\dotsc,X_{m}\right) and assume the mass density to be constant over each cell, so the field ρ\rho may be approximated by the vector ρ𝑿\rho_{\bm{X}}. The vertical component of the gravitational field at ss is then given by:

i=1mDig(x,s)ρ(x)𝑑xi=1m(Dig(x,s)𝑑x)ρXi:=Gsρ𝑿.\int_{\cup_{i=1}^{m}D_{i}}g(x,s)\rho(x)dx\approx\sum_{i=1}^{m}\Bigg{(}\int_{D_{i}}g(x,s)dx\Bigg{)}\rho_{X_{i}}:=G_{s}\rho_{\bm{X}}.

Integrals of Green kernels over cuboids may be computed using the Banerjee formula (Banerjee and Das Gupta,, 1977).

Theorem A.1 (Banerjee).

The vertical gravity field at point (x0,y0,z0)(x_{0},y_{0},z_{0}) generated by a prism with corners (xh,xl(x_{h},x_{l}, yh,yl,)y_{h},y_{l},...) of uniform mass density ρ\rho is given by:

gz\displaystyle g_{z} =12γNρ[xlog(x2+y2+z2+yx2+y2+z2y)\displaystyle=\frac{1}{2}\gamma_{N}\rho\Bigg{[}x~{}\log\Big{(}\frac{\sqrt{x^{2}+y^{2}+z^{2}}+y}{\sqrt{x^{2}+y^{2}+z^{2}}-y}\Big{)}
+ylog(x2+y2+z2+xx2+y2+z2x)\displaystyle\phantom{\frac{1}{2}G\rho\Bigg{[}}+y~{}\log\Big{(}\frac{\sqrt{x^{2}+y^{2}+z^{2}}+x}{\sqrt{x^{2}+y^{2}+z^{2}}-x}\Big{)}
2zarctan(xyzx2+y2+z2)]|xlx0xhx0|yly0yhy0|zlz0zhz0\displaystyle\phantom{\frac{1}{2}G\rho\Bigg{[}}-2z~{}\arctan\Big{(}\frac{xy}{z\sqrt{x^{2}+y^{2}+z^{2}}}\Big{)}\Bigg{]}\Bigg{\rvert}_{x_{l}-x_{0}}^{x_{h}-x_{0}}\Bigg{\rvert}_{y_{l}-y_{0}}^{y_{h}-y_{0}}\Bigg{\rvert}_{z_{l}-z_{0}}^{z_{h}-z_{0}}

Appendix B Bayesian Inversion and Bayesian Set Estimation

Given a generic Bayesian linear inverse problem with unknown function ρ:D\rho:D\rightarrow\mathbb{R} and prior ZZ, there exists several approaches to approximate the excursion set Γ={xD:ρ(x)T}\Gamma^{*}=\{x\in D:\rho\left(x\right)\geq T\} using the posterior. For example, a naive estimate for Γ\Gamma^{*} may be obtained using the plug-in estimator:

Γ^plug-in:={xD:m~xT},\hat{\Gamma}_{\text{plug-in}}:=\{x\in D:~{}\tilde{m}_{x}\geq T\},

where m~x\tilde{m}_{x} denotes the posterior mean function of the GP prior. In this work, we will focus on recently developed more sophisticated approaches to Bayesian excursion set estimation (Azzimonti et al.,, 2016; Chevalier et al.,, 2013) based on the theory of random sets (Molchanov,, 2005). We here briefly recall some theory taken from the aforementioned source.

In the following, let Z~\tilde{Z} denote a random field on DD that is distributed according to the posterior distribution. Then, the posterior distribution of the field gives rise to a random closed set (RACS):

(16) Γ:={xD:Z~xT}.\Gamma:=\{x\in D:~{}\tilde{Z}_{x}\geq T\}.

One can then consider the probability for any point in the domain to belong to that random set. This is captured by the coverage function:

pΓ:\displaystyle p_{\Gamma}: D[0,1]\displaystyle D\rightarrow[0,1]
x[xΓ].\displaystyle x\mapsto\mathbb{P}\left[x\in\Gamma\right].

The coverage function allows us to define a parametric family of set estimates for Γ\Gamma, the Vorob’ev quantiles:

(17) Qα:={xD:pΓ(x)α}.Q_{\alpha}:=\{x\in D:p_{\Gamma}(x)\geq\alpha\}.

The family of quantiles QαQ_{\alpha} gives us a way to estimate Γ\Gamma by controlling the (pointwise) probability α\alpha that the members of our estimate lie in Γ\Gamma. There exists several approaches for choosing α\alpha. One could for example choose it so as to produce conservative estimates of the excursion set (Azzimonti et al.,, 2019). Another approach is to choose it such that the volume of the resulting quantile is equal to the expected volume of the excursion set. This gives rise to the Vorob’ev expectation.

Definition B.1.

(Vorob’ev Expectation) The Vorob’ev expectation is the quantile QαVQ_{\alpha_{V}} with threshold αV\alpha_{V} chosen such that

μ(Qα)𝔼[V(Γ)]μ(QαV),α>αV,\mu(Q_{\alpha})\leq\mathbb{E}[V(\Gamma)]\leq\mu(Q_{\alpha_{V}}),~{}\forall\alpha>\alpha_{V},

where V()V(\cdot) denotes the volume under the Lebesgue measure on d\mathbb{R}^{d}.

Note that the expected excursion volume may be computed using Robbins’s theorem, which states that under suitable conditions:

V¯Γ:=𝔼[λ(Γ)]=DpΓ(x)𝑑x.\bar{V}_{\Gamma}:=\mathbb{E}[\lambda(\Gamma)]=\int_{D}p_{\Gamma}(x)dx.

We refer the reader to Robbins, (1944) and Molchanov, (2005) for more details.

To illustrate the various Bayesian set estimation concepts introduced here, we apply them to a simple one-dimensional inverse problem, where one wants to estimate the excursion set above 1.01.0 of a function f:[1,1]f:[-1,1]\rightarrow\mathbb{R} after 3 pointwise evaluations of the function have been observed. Results are shown in Appendix B

Refer to caption
(a)
Refer to caption
(c)
Refer to caption
(d)
Figure 16: Bayesian set estimation for one-dimensional example. Excursion threshold in red. Posterior after 3 pointwise observations is considered. For left to right, top to bottom: i) True function (black), posterior and 2σ\sigma confidence region (blue). True excursion region is highlighted in black and plug-in estimate is highlighted in blue. ii) Posterior excursion probability / coverage function (dark red). ii) Estimated excursion regions using Vorob’ev quantiles at level α=0.5\alpha=0.5 (dark blue) and α=0.75\alpha=0.75 (light blue). iv) Estimated excursion region using Vorob’ev expectation, Vorob’ev threshold is αV=0.4\alpha_{V}=0.4.

Appendix C Proofs

Proof C.1.

(Lemma 2.2) We proceed by induction. The case n=1n=1 follows from Eq. 3. The induction step is directly given by Theorem 2.1.

Proof C.2.

(Lemma 2.4) The product is computed using Algorithm 1. It involves multiplication of AA with the prior covariance, which costs 𝒪(m2q)\mathcal{O}\left(m^{2}q\right) and multiplication with all the previous intermediate matrices, which contribute 𝒪(mpcq)\mathcal{O}\left(mp_{c}q\right) and 𝒪(pc2q)\mathcal{O}\left(p_{c}^{2}q\right) respectively, at each stage.

Proof C.3.

(Lemma 2.5) The cost of computing the ii-th pushforward K¯i\bar{K}_{i} is 𝒪(m2pc+i(mpc2+pc3))\mathcal{O}\left(m^{2}p_{c}+i(mp_{c}^{2}+p_{c}^{3})\right). Summing this cost for all stages i=1,,ni=1,\dotsc,n then gives 𝒪(m2P+mP2+p2pc)\mathcal{O}\left(m^{2}P+mP^{2}+p^{2}p_{c}\right). To that cost, one should add the cost of computing Ri1R_{i}^{-1}, which costs 𝒪(pc3)\mathcal{O}\left(p_{c}^{3}\right) at each stage, yielding a 𝒪(Ppc2)\mathcal{O}\left(Pp_{c}^{2}\right) contribution to the total cost, which is dominated by P2pcP^{2}p_{c} since pc<Pp_{c}<P.

Appendix D Supplementary Experimental Results

We here include more detailed analysis of the results of Section 3.3 that do not fit in the main text.

Figs. 12 and 13 showed that there are differences in detection performance for the different ground truths. These can be better understood by plotting the actual location of the excursion set for each of the ground truths as well as the observation locations chosen by the wIVR criterion, as done in Appendix D. One sees that the (comparatively) poor performance shown by Fig. 13 for Sample 2 in the small scenario may be explained by the fact that, for this ground truth, the excursion set is located mostly outside of the accessible data collection zone (island surface), so that the strategy is never able to collect data directly above the excursion.

Refer to caption
(a)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Refer to caption
(i)
Refer to caption
(j)
Figure 17: True excursion set and visited locations (wIVR strategy). Island boundary is shown in blue.