This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Low-rank approximation for multiscale PDEs

Ke Chen Ke Chen is a postdoctoral researcher in the Department of Mathematics at University of Texas at Austin. His email address is [email protected].    Shi Chen Shi Chen is a graduate student in the Department of Mathematics at University of Wisconsin, Madison. His email address is [email protected].    Qin Li Qin Li is an associate professor in the Department of Mathematics and the Wisconsin Institutes for Discovery at University of Wisconsin, Madison. Her email address is [email protected].    Jianfeng Lu Jianfeng Lu is a professor in the Department of Mathematics, the Department of Physics and the Department of Chemistry at Duke University. His email address is [email protected].    Stephen J. Wright Stephen J. Wright is a professor in the Department of Computer Sciences at University of Wisconsin, Madison. His email address is [email protected].

1 Introduction

Multiscale phenomena are ubiquitous, with applications in many physical sciences and engineering fields: aerospace, material sciences, geological structure analysis, and many others. The different scales often have different physics, which entangle to produce complicated nonlinearities. Partial differential equations (PDEs) are often used to model these problems, with different scales captured in the coefficients and functions that define the PDE. These PDE models are challenging to compute directly, so analysis and algorithms specifically targeted to multiscale problems have been developed and investigated. Following convention, we focus in this review on problems with two distinct scales, with a small positive parameter ϵ\epsilon capturing the ratio between the small and large scale.

Though modern multiscale analysis dates back to asymptotic PDE analysis that was seen already in Hilbert and Poincaré expansions early last century (see review in [MR2382139]), the impetus for computations involving multiscale PDEs came largely from the US Department of Energy (DOE) National Labs within the ASCI (Advanced Strategic Computing Initiative) [Ho:2009multiscale]. Since that time, analysis and computation in multiscale PDEs have taken different paths. Analysis has tended to follow a single “universal” strategy, passed down from tradition. The equation is decomposed into several levels according to asymptotic expansions involving the scale parameters, with the subequation at each level representing physics at a single scale, and subequations at the finer level feeding information to those at the coarser level. This analytical machinery has been used to treat multiscale PDEs arising from such varied backgrounds as kinetic theory, semi-classical quantum systems, and homogenization of composite materials, among others [MR2382139, MR2830582].

On the computational side, strategies for handling multiscale PDEs are more varied. Problems are usually handled by specifically designed solvers. One class of solvers called asymptotic-preserving schemes [MR3645390] are designed to preserve asymptotic limits of kinetic equations. These schemes usually contain some component of macro-solvers and micro-solvers, integrated in a clever way to reveal different structures in different regimes. Another class of solvers called numerical homogenization methods [MR2830582, MR3971243, MR1979846] usually target elliptic and parabolic equations in which the coefficients that represent media have oscillatory elements. These methods usually consist of offline and online stages, with either the homogenized media or the representative basis functions being prepared in the offline stage.

Why are most numerical schemes for multiscale PDEs equation-specific despite the analytical tools being largely unified? This intriguing question has motivated our investigations into devising a universal numerical strategy for solving multiscale PDEs. While the approach is yet to be developed fully, we believe that our progress on this issue is of wide interest, and this article surveys our progress to date. Crucially, our approach exploits the low-rank structure present in discretizations of multiscale PDEs.

To demonstrate the fundamental idea, we consider the following problem:

ϵuϵ=f,\mathcal{L}^{\epsilon}u^{\epsilon}=f\,, (1)

where ϵ\mathcal{L}^{\epsilon} is a linear partial differential operator that depends explicitly on the small parameter ϵ\epsilon, while ff represents the boundary condition or the source term, which is assumed to have no dependence on ϵ\epsilon. Multiscale problems that can be formulated in this way include elliptic equations with highly oscillating media and the neutron transport equation with small Knudsen number. Due to the ϵ\epsilon-dependence of the operator, the solution uϵu^{\epsilon} inherits structures at both fine and coarse scales.

An asymptotic limit is revealed by multiscale analysis using asymptotic expansions as ϵ0\epsilon\to 0. In this limit, the oscillation at the fine scale is fast and the detailed oscillation pattern no longer matters — only macroscopic quantities are relevant. Formally, writing the homogenization limit as

u=f,\mathcal{L}^{\ast}u^{\ast}=f\,, (2)

we have

uϵu0 as ϵ0.\|u^{\epsilon}-u^{\ast}\|\to 0\textrm{ as }\epsilon\to 0\,. (3)

The norm of the approximation error depends heavily on the particular equation at hand.

The numerical challenge in solving (1) is that many degrees of freedom may be needed. Naïve finite element or finite difference methods would require mesh size hϵh\ll\epsilon to resolve fine-scale structure of the solution at the ϵ\epsilon level. For a problem on d\mathbb{R}^{d}, the discretized system would therefore have O(ϵd)O(\epsilon^{-d}) degrees of freedom, leading to prohibitive computational and memory cost for small ϵ\epsilon. From an application perspective, it often suffices to characterize the solutions on the macroscopic level, where oscillations at the ϵ\epsilon scale are largely absent. This property raises the question of whether we can obtain an approximate solution of this type using only O(1)O(1) degrees of freedom. If we know how to derive (2), we can simply solve for uu^{\ast}, which has the required macroscopic properties, and typically requires a discretization with O(1)O(1) degrees of freedom. Often, though, the limiting equation (2) and its solution uu^{\ast} are difficult to find explicitly, even when it is possible to establish their existence. These difficulties have led researchers to propose problem-specific solutions.

We believe that a universal approach for finding the large-scale solution can be devised, and that exploitation of the low-rank structure of the solution space is the key to developing such an approach. As suggested above, the Green’s matrix 𝖦ϵ\mathsf{G}^{\epsilon} (the discretized Green’s function on fine grids) for the multiscale system (1) requires dimension O(ϵd)O(\epsilon^{-d}) to represent the underlying Green’s function accurately. However, if a limiting system such as (2) exists, this limiting system can be well-represented numerically by 𝖦\mathsf{G}^{\ast}, a Green’s matrix with dimension only O(1)O(1). This phenomenon suggests the system can largely be “compressed” and hence is of low rank; see illustration in Figure 1. In the language of numerical linear algebra, this transition amounts to performing a truncated singular value decomposition (SVD) of 𝖦ϵ\mathsf{G}^{\epsilon} to obtain 𝖦\mathsf{G}^{\ast}.

Refer to caption
Figure 1: PDEs with small parameters have homogenized limits, meaning the solutions to the original PDEs can be well-approximated by the solutions to the limiting equations. While analytically the two solution spaces are “close”, the original equation requires many more degrees of freedom to solve numerically than its limiting counterpart. The numerical Green’s matrix is intrinsically low rank.

If we obtain the truncated SVD of the matrix 𝖦ϵ\mathsf{G}^{\epsilon} by starting with a full SVD, the resulting algorithms would be impractical because of the large dimension of the matrix and the expense of preparing and storing the full matrix 𝖦ϵ\mathsf{G}^{\epsilon} and computing its SVD. Several new linear algebra solvers take a quite different approach. Instead of accessing the full matrix, these new solvers merely require computation of matrix-vector products, involving the target matrix and several randomly selected vectors (typically vectors with Gaussian i.i.d. entries). Translated to the PDE solver setting, these matrix-vector multiplications amount to computing numerical solutions to PDEs with some random source terms, a task that may be practical if the number of such operations required is modest. The randomized SVD (rSVD) approach is one method of this type. It is equipped with a thorough analysis and achieves optimality in terms of computational efficiency. We make use of this method in the techniques described in the remainder of this article.

The main theme of our article, then, is the use of randomized SVD solvers to exploit the low-rank features of multiscale PDEs. We will describe two strategies both of which are divided into “offline” and “online” stages. The offline stage sees the preparation of either the solution space or the boundary-to-boundary map used in the domain decomposition, while the online stage singles out the specific solution for the given source ff. The two strategies are described in Section 4.1 and 4.2, respectively. In Section 5, we present the nonlinear extension utilizing manifold learning algorithms for reconstructing the low-rank features of the solution manifold. Prior to these discussions, we describe in Section 2 two algorithm classes — asymptotic preserving and numerical homogenization — for identifying the asymptotic limits of multiscale problems. As examples, we use the multiscale radiative transfer equation (RTE) and the elliptic equation with rough media. Section 3 explores the two main elements of our approaches: the numerical low-rank feature of multiscale PDEs and the randomized SVD solver for efficient reconstruction of low-rank operator/spaces. We conclude with a discussion of future work in Section 6.

2 Examples

Kinetic equations and elliptic equations with oscillating media are two examples of multiscale PDEs, for which computational schemes were developed separately. The specific features of these problems were incorporated into the design of asymptotic preserving schemes and numerical homogenization methods, respectively. We review these techniques and highlight the shared low-rank property of these two problems.

2.1 Kinetic equations and asymptotic preserving methods

Kinetic equations, which originate from statistical mechanics, describe the evolution of probability density for identical particles in phase space. A model equation, the radiative transfer equation (RTE), characterizes the evolution of photon density. In the steady state, this equation is

vxuϵ+𝖲ϵ[uϵ]=f(x,v),(x,v)𝒦×𝕍,-v\cdot\nabla_{x}u^{\epsilon}+\mathsf{S}^{\epsilon}[u^{\epsilon}]=f(x,v)\,,\quad(x,v)\in\mathcal{K}\times\mathbb{V}\,, (4)

where f(x,v)f(x,v) is the light source, and the linear collision operator 𝖲ϵ\mathsf{S}^{\epsilon} describes the interaction of photons with the optical media. The small parameter ϵ\epsilon is encoded in this operator.

The operator 𝖲ϵ\mathsf{S}^{\epsilon} defines several distinct regimes. In the optically thick regime, it is defined by

𝖲ϵu(x,v)=1ϵ𝕍\displaystyle\mathsf{S}^{\epsilon}u(x,v)=\frac{1}{\epsilon}\int_{\mathbb{V}} k(x,v,v)u(x,v)dv\displaystyle k(x,v,v^{\prime})u(x,v^{\prime})\mathrm{d}v^{\prime} (5)
1ϵ𝕍k(x,v,v)u(x,v)dv.\displaystyle-\frac{1}{\epsilon}\int_{\mathbb{V}}k(x,v^{\prime},v)u(x,v)\mathrm{d}v^{\prime}\,.

In this case, k(x,v,v)k(x,v,v^{\prime}) is the scattering coefficients that describes the possibility of a photon located at xx changing its velocity from vv to vv^{\prime}, and the parameter ϵ\epsilon is called the Knudsen number, standing for the ratio of the mean free path to the typical domain length. When the medium is optically thick, the mean free path is small, with ϵ1\epsilon\ll 1. This means the photon particles are scattered fairly often, and the system statistically achieves the equilibrium state, which can itself be characterized mathematically. One example is to observe light in atmosphere, where the average mean free path is about 1010 m, and the observation is conducted at the scale of 1010 km, leading to ϵ103\epsilon\sim 10^{-3}. By performing asymptotic expansion in terms of ϵ\epsilon, the inhomogeneity in the velocity domain vanishes, and one can show that uε(x,v)u^{\varepsilon}(x,v) asymptotically approximates u(x)u^{\ast}(x), a function without dependence on vv that solves a diffusion equation. We have the following result from [MR2839402].

Theorem 1.

Suppose that uϵu^{\epsilon} solves (4) with collision term 𝖲\mathsf{S} being isotropic, that is, k(x,v,v)=σ(x)k(x,v,v^{\prime})=\sigma(x) for some function σ\sigma. Let 𝒦d\mathcal{K}\subset\mathbb{R}^{d} be bounded with smooth boundary, and 𝕍=𝕊d1\mathbb{V}=\mathbb{S}^{d-1}. Assume that the boundary condition is

uϵ(x,v)=ϕ(x,v)onx𝒦,vnx<0.u^{\epsilon}(x,v)=\phi(x,v)\quad\text{on}\quad x\in\partial\mathcal{K}\,,\;\;v\cdot n_{x}<0\,. (6)

Then

uϵuL2(dxdv)0,\|u^{\epsilon}-u^{\ast}\|_{L_{2}(\mathrm{d}x\mathrm{d}v)}\to 0\,, (7)

where u=u(x)u^{\ast}=u^{\ast}(x) solves

x(1σ(x)xu(x))=g(x),x𝒦,\nabla_{x}\cdot\left(\frac{1}{\sigma(x)}\nabla_{x}u^{\ast}(x)\right)=g(x)\,,\quad x\in\mathcal{K}\,, (8)

with the boundary condition

u(x)=ξϕ(x),onx𝒦,u^{\ast}(x)=\xi_{\phi}(x)\,,\quad\text{on}\quad x\in\partial\mathcal{K}\,,

where ξϕ(x)\xi_{\phi}(x) solves a proper boundary layer equation and gg can be obtained from f(x,v)dv\int f(x,v)\mathrm{d}{v}.

This result indicates that the homogenized operator as ϵ0\epsilon\to 0 is x((1/σ)x)\mathcal{L}^{\ast}\propto\nabla_{x}\cdot\left(({1}/{\sigma})\nabla_{x}\right). Similar results, when k(x,v,v)k(x,v,v^{\prime}) fails to have the form of σ(x)\sigma(x) in the anisotropic optical media, are still available, but the explicit form of \mathcal{L}^{\ast} is no longer available.

A second regime of interest for 𝖲ϵ\mathsf{S}^{\epsilon} is one in which the media is highly heterogeneous [MR1760042]:

𝖲ϵu(x,v)=𝕍\displaystyle\mathsf{S}^{\epsilon}u(x,v)=\int_{\mathbb{V}} k(xϵ,v,v)u(x,v)dv\displaystyle k\left(\frac{x}{\epsilon},v,v^{\prime}\right)u(x,v^{\prime})\mathrm{d}v^{\prime} (9)
𝕍k(xϵ,v,v)u(x,v)dv.\displaystyle-\int_{\mathbb{V}}k\left(\frac{x}{\epsilon},v^{\prime},v\right)u(x,v)\mathrm{d}v^{\prime}\,.

In this case, the photons go through the media that oscillates at a small scale: For example, sunlight passing through heavy cloud with a large number of small droplets or laser beam passing through crystals. The amplitude of kk determines the photon scattering frequency. Since kk oscillates rapidly, photons also change rapidly between the high- and low-scattering regimes. On a large scale, the photons can be viewed approximately as scattering with an averaged frequency. A mathematical result is as follows [MR1760042].

Theorem 2.

Let the conditions from Theorem 1 hold, and suppose that the collision term 𝖲ϵ\mathsf{S}^{\epsilon} is defined in (9). Then

uϵuL2(dxdv)0,\|u^{\epsilon}-u^{\ast}\|_{L_{2}(\mathrm{d}x\mathrm{d}v)}\to 0\,, (10)

where u(x,v)u^{\ast}(x,v) solves

vxu+𝖲[u]=f(x,v),(x,v)𝒦×𝕍,-v\cdot\nabla_{x}u^{\ast}+\mathsf{S}^{\ast}[u^{\ast}]=f(x,v)\,,\quad(x,v)\in\mathcal{K}\times\mathbb{V}\,, (11)

where 𝖲u(x,v)=σ(x)𝕍u(x,v)u(x,v)dv\mathsf{S}^{\ast}u(x,v)=\sigma^{\ast}(x)\int_{\mathbb{V}}u(x,v^{\prime})-u(x,v)\mathrm{d}v^{\prime} for some σ(x)\sigma^{\ast}(x). Furthermore, if k(x,v,v)=σ(x)k(x,v,v^{\prime})=\sigma(x) is periodic in xx with period [0,1]d[0,1]^{d}, then σ=[0,1]dσ(x)dx\sigma^{\ast}=\int_{[0,1]^{d}}\sigma(x)\mathrm{d}x.

In special cases, such as under periodic or random ergodic conditions, the function σ\sigma^{\ast} can be computed explicitly. (There are also works that investigate the asymptotic limit of the RTE when the system is both highly oscillatory and in diffusion regime; see [MR1878799].)

In both limiting regimes, the limiting equations (8) and (11) can be solved much more efficiently than the original equation (4). The discretization of (4) is constrained strongly by ϵ\epsilon, due either to stability (as in (5)) or accuracy (as in (9)). By contrast, the solution uu^{\ast} varies smoothly, containing no ϵ\epsilon-scale effects, so can be obtained accurately by applying a discretization with mesh width O(1)O(1) to the asymptotic limiting equation. If the latter equation is available, computation of uu^{\ast} by this means is the recommended methodology.

Methods for kinetic equations are termed “asymptotic preserving” (AP) if they can relax the requirement that the discretization width hh satisfies h=o(ϵ)h=o(\epsilon) yet still capture the asymptotic limits. Many different AP approaches have been proposed. For linear equations, existing AP methods rely on even-odd or micro-macro decomposition. For nonlinear equations, knowledge of the specific forms of the limits is usually required, and this knowledge is built into the solvers [MR3645390]. As mentioned above, these specific forms are often not available, so many AP methods cannot be applied to a large set of multiscale kinetic equations. This observation begs the question: Knowing the existence of the limit, but not its particular form, can we still devise efficient methods for solving kinetic equations?

2.2 Elliptic equations and numerical homogenization

Another class of multiscale equations that has been investigated deeply is elliptic equations with highly oscillatory coefficient. These problems have the form

x(aϵ(x)xuϵ)=f,-\nabla_{x}\cdot\left(a^{\epsilon}\left(x\right)\nabla_{x}u^{\epsilon}\right)=f\,, (12)

where ϵ1\epsilon\ll 1 is the scale on which the media oscillates. (The source term ff has no small-scale contribution.)

This equation is a model problem from petroleum engineering where it is crucial to precompute the underground flow before expensive construction of infrastructure takes place [MR2801210]. The problem is typically solved on kilometer-scale domains, but the heterogeneities in the media can scale at centimeters. Certain forms of this equation can be approximated effectively by an equation that can be solved efficiently. Suppose the media coefficient aϵ(x)a^{\epsilon}(x) has the form a(x,x/ϵ)a(x,x/\epsilon), that is, it varies on two scales (11 and ϵ\epsilon), and moreover is periodic with respect to the fast variable (the second argument in a(x,x/ϵ)a(x,x/\epsilon)). Then in the limiting regime as ϵ0\epsilon\to 0, the solution uϵu^{\epsilon} converges to that of a homogenized equation, with the media “smoothed-out,” as described in the following result [MR1185639].

Theorem 3.

Let uϵu^{\epsilon} solve (12) in the domain x𝒦x\in\mathcal{K} with zero boundary condition. Suppose a(x,x/ϵ)a(x,x/\epsilon) is periodic with respect to the second argument. Then

uϵuL2ϵuH2,\|u^{\epsilon}-u^{\ast}\|_{L^{2}}\lesssim\epsilon\left\|u^{\ast}\right\|_{H^{2}}\,, (13)

where uu^{\ast} solves the following effective equation with zero boundary condition:

x(a(x)xu)=f,x𝒦,-\nabla_{x}\cdot(a^{\ast}(x)\nabla_{x}u^{\ast})=f\,,\quad x\in\mathcal{K}\,, (14)

where aa^{\ast}, the effective media, can be computed from a cell problem (See Definition 2.1 in [MR1185639]).

As in the previous section, when a limiting equation can be derived explicitly, the best course for obtaining a useful solution it to solve this equation directly, as the mesh width in the discretization scheme can be much larger than ϵ\epsilon. See [MR2477579] for a discussion of a reduced number of basis functions and [MR2830582] for computation of the effective media.

However, the validity and the specific form of the effective limit are known only in special cases like the one described in Theorem 3. In other cases, we seek a solver that relies on as little analytical knowledge as possible. An approach known as numerical homogenization has been investigated extensively. This approach is founded on two principles: a discretization scheme independent of ϵ\epsilon, and a numerical solution scheme that captures the true limiting behavior of the solution on the discrete level. Variants of numerical homogenization include application of the \mathcal{H}-matrix, a purely algebraic technique [MR3445676]; and a Bayesian approach that views the source ff, and hence the solution uϵu^{\epsilon}, as Gaussian fields [MR3369060], which further translates to game theory [MR3971243]. All these methods are successful, but they all implicitly rely on properties of the underlying elliptic equation. Can we devise an approach that applies to general problems with oscillatory media that exploits the low-rank property in the solution space, without using analytical structure explicitly?

3 A unified framework for multiscale PDEs based on random sampling

We have given several examples of multiscale models that arise in applications, and mentioned several algorithmic approaches that make use of the limiting equations, when available. We describe next the foundations of a unified scheme that captures asymptotic limiting behavior automatically, even when the asymptotic limits are unavailable. Our method exploits low-rank structure and uses random sampling to discover this structure. We describe the low-rank property in Section 3.1 and the randomized SVD method for revealing this structure in Section 3.2.

3.1 Numerical rank

We consider a bounded linear operator 𝒜\mathcal{A}, which maps f𝒳f\in\mathcal{X} to a space 𝒴\mathcal{Y}, that is

𝒜:\displaystyle\mathcal{A}: 𝒳\displaystyle\quad\mathcal{X} \displaystyle\rightarrow 𝒴\displaystyle\quad\mathcal{Y}
f\displaystyle\quad f \displaystyle\mapsto u.\displaystyle\quad u.

In the PDE setting, 𝒜\mathcal{A} is the solution operator that maps the boundary conditions and/or source term ff to the solution uu. The numerical rank of such an operator is defined as follows.

Definition 1 (Numerical rank).

The numerical τ\tau-rank of 𝒜\mathcal{A} is the rank of the lowest-rank operator within the τ\tau-neighborhood of 𝒜\mathcal{A}, that is,

kτ(𝒜):=min{dimran𝒜~:𝒜~(𝒳,𝒴),𝒜~𝒜τ}.k_{\tau}(\mathcal{A}):=\min\{\dim\mathrm{ran}\tilde{\mathcal{A}}:\tilde{\mathcal{A}}\in\mathcal{L}(\mathcal{X},\mathcal{Y}),\|\tilde{\mathcal{A}}-\mathcal{A}\|\leq\tau\}\,.

In other words, kτ(𝒜)k_{\tau}(\mathcal{A}) is this smallest dimension of the range among all the operators within distance τ\tau of 𝒜\mathcal{A}.

When 𝒜\mathcal{A} is the PDE solution map, then 𝒜~\tilde{\mathcal{A}} with low rank is also a linear map with a finite dimensionality. It can be viewed as the discrete version (or a matrix of dimension kτ(𝒜)k_{\tau}(\mathcal{A})) that approximates 𝒜\mathcal{A} within τ\tau accuracy. The definition suggests that if 𝒜~\tilde{\mathcal{A}} can be found, it is optimal in the sense of numerical efficiency. The concept is rather similar to the Kolmogorov NN-width, defined as follows.

Definition 2 (Kolmogorov NN-width).

Given the linear operator 𝒜:𝒳𝒴\mathcal{A}:\mathcal{X}\to\mathcal{Y}, the Kolmogorov NN-width dN(𝒜)d_{N}(\mathcal{A}) is the shortest distance from its range to all NN-dimensional space, that is,

dN(𝒜):\displaystyle d_{N}(\mathcal{A}): =minS:dimS=Nd(𝒜,S)\displaystyle=\min_{S:\dim S=N}d(\mathcal{A},S) (15)
=minS:dimS=NsupfminvS𝒜fv𝒴f𝒳.\displaystyle=\min_{S:\dim S=N}\sup_{f}\min_{v\in S}\frac{\|\mathcal{A}f-v\|_{\mathcal{Y}}}{\|f\|_{\mathcal{X}}}\,.

Indeed, the Kolmogorov NN-width and numerical rank are related by the following result [MR4155236].

Proposition 1.

For any linear operator 𝒜:𝒳𝒴\mathcal{A}:\mathcal{X}\to\mathcal{Y}, we have the following.

  1. (a)

    If the numerical τ\tau-rank is NN, then dN(𝒜)τd_{N}(\mathcal{A})\leq\tau.

  2. (b)

    If dN(𝒜)τ<dN1(𝒜)d_{N}(\mathcal{A})\leq\tau<d_{N-1}(\mathcal{A}), then the numerical τ\tau-rank is NN.

For the three examples presented in Section 2, the numerical ranks can be calculated from their limiting equations. For one-dimensional RTE in the diffusion regime, if we denote by 𝒜ϵ\mathcal{A}^{\epsilon} and 𝒜\mathcal{A}^{\ast} the solution operators of (4) and (8), respectively, then noting that 𝒜\mathcal{A}^{\ast} can be approximated using 1/τ1/\sqrt{\tau} grid points to achieve τ\tau accuracy, when ϵ<τ\epsilon<\tau, the numerical rank is naturally kτ(𝒜ϵ)1/τϵk_{\tau}(\mathcal{A}^{\epsilon})\lesssim 1/\sqrt{\tau-\epsilon}. Without employing the knowledge of the existence of the limit, however, a brute-force discretization naturally requires O(1/ϵτα+1)O(1/\epsilon\tau^{\alpha+1}) degrees of freedom: O(1/ϵτ)O(1/\epsilon\tau) for the upwind discretization in xx and O(1/τα)O(1/\tau^{\alpha}) for the discretization in vv, where α\alpha depends on the particular numerical integral accuracy. Translating into Green’s-matrix language, this observation means that 𝖦ϵ\mathsf{G}^{\epsilon} is represented by O(1/ϵτα+1)O(1/\epsilon\tau^{\alpha+1}) degrees of freedom but its range can be captured by a compressed Green’s matrix 𝖦\mathsf{G}^{\ast} with just O(1)O(1) column vectors.

The same argument applies to the elliptic equation on a two-dimension domain with high oscillations. When second-order linear finite elements are used, with no knowledge of the limiting system, O(1/ϵ2τ)O(1/\epsilon^{2}\tau) degrees of freedom are required, dropping to O(1/τ)O(1/\tau) when the the limiting system is known. In other words, the full Green’s matrix 𝖦ϵ\mathsf{G}^{\epsilon} requiring O(1/ϵ2)O(1/\epsilon^{2}) degrees of freedom can be well-represented using just O(1)O(1) column vectors.

In all these cases, the degrees of freedom for a given numerical method are substantially larger than the numerical rank of the problem. Thus, much of the information in these full-blown representations is redundant and compressible. A low rank representation exists and yields a much more economical representation.

3.2 Random sampling in numerical linear algebra

Knowing the existence of the low rank structure and finding such a structure are very different goals. The Kolmogorov NN-width is a concept developed in but has made little impact in numerical PDEs for a simple reason: Traditional PDE solvers require a predetermined set of basis functions, while the Kolmogorov NN-width looks for “optimal” basis functions. How can an optimal basis be found without first forming the full basis? Translated to linear algebra, this question is about finding the dominant singular vectors in a matrix without forming the whole matrix. Specifically, if 𝖠m×n\mathsf{A}\in\mathbb{R}^{m\times n} is known to be approximately low rank, meaning that there exists 𝖴r\mathsf{U}_{r}, a m×rm\times r matrix with orthonormal columns with rmin(m,n)r\ll\min(m,n) and

𝖠𝖠r=𝖠𝖴r𝖴r𝖠𝖠,\|\mathsf{A}-\mathsf{A}_{r}\|=\|\mathsf{A}-\mathsf{U}_{r}\mathsf{U}_{r}^{\top}\mathsf{A}\|\ll\|\mathsf{A}\|\,,

can we find 𝖴r\mathsf{U}_{r} without forming the full matrix 𝖠\mathsf{A}?

In linear algebra, it is well-known that 𝖴r\mathsf{U}_{r} is simply the collection of the first rr singular vectors of 𝖠\mathsf{A}. Writing

𝖠=𝖴Σ𝖵=i=1nσiuivi,\mathsf{A}=\mathsf{U}\Sigma\mathsf{V}^{\top}=\sum_{i=1}^{n}\sigma_{i}u_{i}v^{\top}_{i}\,, (16)

where 𝖴=[u1,u2,,un]m×n\mathsf{U}=\left[u_{1}\,,u_{2}\,,\dots,u_{n}\right]\in\mathbb{R}^{m\times n} and 𝖵=[v1,v2,,vn]n×n\mathsf{V}=\left[v_{1}\,,v_{2}\,,\dots,v_{n}\right]\in\mathbb{R}^{n\times n} contain the left/right singular vectors and Σ=diag(σ1,σ2,,σn)\Sigma=\mathrm{diag}(\sigma_{1},\sigma_{2},\dotsc,\sigma_{n}) contains the singular values, then 𝖴r\mathsf{U}_{r} is the first rr columns in 𝖴\mathsf{U}.

The standard method for computing the SVD requires 𝖠\mathsf{A} to be stored and computed with. But the celebrated randomized SVD (rSVD) method [MR2806637] captures the range of a given matrix by means of random sampling of its column space, which requires only computation of matrix-vector products involving 𝖠\mathsf{A} and random vectors — operations that can be performed without full storage or knowledge of 𝖠\mathsf{A}. Implementation of rSVD is easy and its performance is robust.

The idea behind the algorithm is simple. If matrix 𝖠m×n\mathsf{A}\in\mathbb{R}^{m\times n} has approximate low rank rmin{m,n}r\ll\min\{m,n\}, the matrix maps an nn-dimensional sphere to an mm-dimensional ellipsoid that is “thin:” rr of its axes are significantly larger than the rest. With high probability, vectors that are randomly sampled on the nn-dimensional sphere are mapped to vectors that lie mostly in a rr-dimensional subspace of m\mathbb{R}^{m} — the range of 𝖠\mathsf{A}. An approximation to 𝖠r\mathsf{A}_{r} can be obtained by projecting onto this subspace.

A precise statement of the performance of randomized SVD is as follows [MR2806637].

Theorem 4.

Let 𝖠\mathsf{A} be an m×nm\times n matrix. Define

𝖸=𝖠Ω,\mathsf{Y}=\mathsf{A}\Omega\,, (17)

where Ω=[ω1,,ωr+p]\Omega=\left[\omega_{1}\,,\dotsc,\omega_{r+p}\right] is a matrix of size n×(r+p)n\times(r+p) with its entries randomly drawn from an i.i.d. normal distribution, where pp is an oversampling parameter. If σr+1σ1=O(1)\sigma_{r+1}\ll\sigma_{1}=O(1), then the projection of 𝖠\mathsf{A} onto the space spanned by 𝖸\mathsf{Y}, defined by

𝖯𝖸(𝖠)=𝖸(𝖸𝖸)1𝖸𝖠,\mathsf{P}_{\mathsf{Y}}(\mathsf{A})=\mathsf{Y}(\mathsf{Y}^{\top}\mathsf{Y})^{-1}\mathsf{Y}^{\top}\mathsf{A}\,,

yields that 𝖠𝖯𝖸(𝖠)σ1\|\mathsf{A}-\mathsf{P}_{\mathsf{Y}}(\mathsf{A})\|\ll\sigma_{1} with high probability, and

𝔼𝖠𝖯𝖸(𝖠)rp1σr+1σ1.\mathbb{E}\,\|\mathsf{A}-\mathsf{P}_{\mathsf{Y}}(\mathsf{A})\|\lesssim\frac{r}{p-1}\sigma_{r+1}\ll\sigma_{1}\,.

The result reconstructs the range of 𝖠\mathsf{A} in a nearly optimal way. It is optimal in efficiency because to capture a rank-rr matrix, only r+pr+p matrix-vector products involving 𝖠\mathsf{A} are required for the calculation of 𝖸\mathsf{Y}, and the oversampling parameter pp is typically quite modest. (p=5p=5 is a typical value.) The result is nearly optimal in accuracy as well. The error bound relies only on σr+1\sigma_{r+1}, which is expected to be smaller than σ1\sigma_{1}. The decay profile of singular values do not affect the approximation accuracy.

If a low-rank approximation to the matrix 𝖠\mathsf{A} is required, and not just an approximation of its range, another step involving multiplications with its transpose is needed. The full method is shown in Algorithm 1.

Algorithm 1 Randomized SVD
1:Given an m×nm\times n matrix 𝖠\mathsf{A}, target rank rr and oversampling parameter pp;
2:Set k=r+pk=r+p;
3:Stage A:
4:     Generate an n×kn\times k Gaussian test matrix Ω\Omega;
5:     Form 𝖸=𝖠Ω\mathsf{Y}=\mathsf{A}\Omega;
6:     Perform the QR-decomposition of 𝖸\mathsf{Y}: 𝖸=𝖰𝖱\mathsf{Y}=\mathsf{Q}\mathsf{R}
7:Stage B:
8:     Form 𝖡=𝖠𝖰\mathsf{B}=\mathsf{A}^{\top}\mathsf{Q};
9:     Compute the SVD of the k×nk\times n matrix 𝖡=𝖴~Σ𝖵\mathsf{B}^{\top}=\widetilde{\mathsf{U}}\Sigma\mathsf{V}^{\top};
10:     Set 𝖴=𝖰𝖴~\mathsf{U}=\mathsf{Q}\widetilde{\mathsf{U}};
11:Return: 𝖴,Σ,𝖵\mathsf{U},\Sigma,\mathsf{V}.

4 Random sampling for multiscale computation

Here we describe how rSVD can be incorporated into multiscale PDE solvers to exploit the low-rank structure of these equations. Our procedure is composed of both offline and online stages. Low rank structure is learned in the offline stage, while in the online stage, the solution for the given source / boundary term ff in (1) is extracted.

We consider in particular the following boundary value problem:

{(ϵuϵ)(x)=0,x𝒦,u(x)=ϕ(x),x𝒦,\begin{cases}(\mathcal{L}^{\epsilon}u^{\epsilon})(x)=0\,,\quad x\in\mathcal{K}\,,\\ \mathcal{B}u(x)=\phi(x)\,,\quad x\in\partial\mathcal{K}\,,\end{cases} (18)

where \mathcal{B} is the boundary condition operator, 𝒦\partial\mathcal{K} the boundary associated with domain 𝒦\mathcal{K}, and we now denote the source term (the boundary data) by ϕ\phi. Our fundamental goal is to construct the low-rank approximation to the Green’s operator 𝖦ϵ\mathsf{G}^{\epsilon} for (18). With this operator in hand, the solution can be computed for any value of the boundary conditions ϕ\phi at a relatively small incremental cost.

If we apply rSVD to approximate 𝖦ϵ\mathsf{G}^{\epsilon} directly, we need to compute products of this operator with random vectors. For the problem (18), this operation corresponds to solving the problem with ϕ\phi replaced by random boundary conditions. Even to solve one such problem efficiently is a computationally challenging task. We use the domain decomposition framework.

We start by partitioning the domain 𝒦\mathcal{K} into subdomains as follows:

𝒦=m=1M𝒦m,\mathcal{K}=\bigcup_{m=1}^{M}\mathcal{K}_{m}\,, (19)

where the patches 𝒦m\mathcal{K}_{m} overlap, in general. We denote by 𝒦m\partial\mathcal{K}_{m} the boundary associated with 𝒦m\mathcal{K}_{m}. Furthermore, we identify the subregions that intersect with 𝒦m\mathcal{K}_{m} as follows:

m={n:1nM,𝒦m𝒦n},\mathcal{I}_{m}=\{n\in\mathbb{N}:1\leq n\leq M\,,\mathcal{K}_{m}\cap\mathcal{K}_{n}\neq\varnothing\}\,,

and define the interior of the patch to be

𝒦~m=𝒦m\(nm𝒦n).\widetilde{\mathcal{K}}_{m}=\mathcal{K}_{m}\backslash\left(\bigcup_{n\in\mathcal{I}_{m}}\mathcal{K}_{n}\right)\,.

For this particular partition of the domain, we define the partition-of-unity functions χm\chi_{m}, m=1,2,,Mm=1,2,\dotsc,M to have the following properties:

m=1Mχm(x)=1,x𝒦,\displaystyle\sum_{m=1}^{M}\chi_{m}(x)=1\,,\quad\forall x\in\mathcal{K}\,, (20)
with{0χm(x)1,x𝒦mχm(x)=0,x𝒦\𝒦m.\displaystyle\text{with}\quad\begin{cases}0\leq\chi_{m}(x)\leq 1\,,&x\in\mathcal{K}_{m}\\ \chi_{m}(x)=0\,,&x\in\mathcal{K}\backslash\mathcal{K}_{m}.\end{cases}

We choose a discretization that resolves the small scales in the solution, defining a mesh width hϵh\ll\epsilon. (The number of subdomains MM is independent of ϵ\epsilon.) A typical decomposition is illustrated in Figure 2.

Refer to caption
Figure 2: Illustration of domain decomposition of a rectangular domain 𝒦\mathcal{K} in 2D with overlap.

How do we design the offline stage to “learn” the low-rank approximation? We propose two approaches that lead to two different kinds of algorithms. In the first approach we learn the optimal basis functions within each subdomain, while the second algorithm employs Schwarz iteration, preparing the boundary-to-boundary map in the offline stage. Other PDE solvers that utilize randomness can also be found in [MR3477310, MR3824169, MR4050504]. In particular in [MR3477310] the authors studied, specifically for elliptic type equations, the generalized eigenvalue problem of the stiffness and mass matrices, and give an error bound using the largest eigenvalue obtained offline.

4.1 Learning basis functions

In standard domain decomposition, the local discretized Green’s matrix 𝖦m\mathsf{G}_{m} is assembled from a “full” collection of basis functions in the patch 𝒦m\mathcal{K}_{m}. The global solution to (18), confined to each 𝒦m\mathcal{K}_{m}, is a linear combination of the columns 𝖦m\mathsf{G}_{m}. The coefficients of these combinations are chosen so that that the continuity conditions across patches and the exterior boundary condition are all satisfied. The complete process can be outlined as follows.

  1. (1)

    Offline stage: For m=1,2,,Mm=1,2,\dots,M, find

    𝖦m=[bm,1,bm,2],\mathsf{G}_{m}=\left[b_{m,1}\,,b_{m,2}\dots\right]\,,

    where each local function bm,nb_{m,n} is a solution to (18) restricted to the subdomain 𝒦m\mathcal{K}_{m}, with fine grid hϵh\ll\epsilon and delta-function boundary conditions. That is,

    {ϵbm,n=0,x𝒦mbm,n=δm,n,x𝒦m,\begin{cases}\mathcal{L}^{\epsilon}b_{m,n}=0\,,\quad&x\in\mathcal{K}_{m}\\ b_{m,n}=\delta_{m,n}\,,\quad&x\in\partial\mathcal{K}_{m}\,,\end{cases} (21)

    where δm,n\delta_{m,n} is the Kronecker delta that singles out the nn-th grid point on the boundary 𝒦m\partial\mathcal{K}_{m}.

  2. (2)

    Online stage: The global solution is

    u=m=1Mumχm=m=1Mχm𝖦mcm,u=\sum_{m=1}^{M}u_{m}\chi_{m}=\sum_{m=1}^{M}\chi_{m}\mathsf{G}_{m}{c}_{m},

    with the support of each um=𝖦mcmu_{m}=\mathsf{G}_{m}{c}_{m} confined to 𝒦m\mathcal{K}_{m}, where cmc_{m} is a vector of coefficients determined by the boundary conditions ϕ\phi and continuity conditions across the patches.

The complete basis represented by 𝖦m\mathsf{G}_{m} has a low-rank structure that can be revealed using randomized SVD. Instead of using delta functions as the boundary conditions, we propose to obtain basis functions by setting random values on 𝒦m\partial\mathcal{K}_{m}, as follows:

{ϵrm,n=0,x𝒦m,rm,n=ωm,n,x𝒦m,\begin{cases}\mathcal{L}^{\epsilon}r_{m,n}=0\,,\quad&x\in\mathcal{K}_{m},\\ r_{m,n}=\omega_{m,n}\,,\quad&x\in\partial\mathcal{K}_{m},\end{cases} (22)

where ωm,n\omega_{m,n} is defined to have a random value drawn i.i.d. from a normal distribution at each grid point in 𝒦m\partial\mathcal{K}_{m}. Denoting 𝖦mr={rm,1,rm,2,}\mathsf{G}_{m}^{\mathrm{r}}=\{r_{m,1}\,,r_{m,2}\,,\cdots\}, we have from linearity of the equation that

𝖦mr=𝖦mΩ,\mathsf{G}_{m}^{\mathrm{r}}=\mathsf{G}_{m}\Omega\,,

where Ω\Omega is a random i.i.d. matrix with entries ωm,n\omega_{m,n}. This 𝖦mr\mathsf{G}_{m}^{\mathrm{r}} is used in the online stage, as an accurate surrogate of 𝖦m\mathsf{G}_{m}, see Algorithm 2.

Algorithm 2 A general framework for multiscale PDE ϵuϵ=0\mathcal{L}^{\epsilon}u^{\epsilon}=0 over 𝒦\mathcal{K} with uϵ=f\mathcal{B}u^{\epsilon}=f on 𝒦\partial\mathcal{K}
1:Domain Decomposition
2:     Partition domain according to (19).
3:Offline Stage:
4:     Prepare i.i.d. Gaussian vectors ωm,i,i=1,,km\omega_{m,i},i=1,\ldots,k_{m} on each 𝒦m\partial\mathcal{K}_{m}.
5:     Solve the basis function rm,ir_{m,i} in (22) on each Ωm\Omega_{m}, and collect the local basis in 𝖦m\mathsf{G}_{m}.
6:Online Stage:
7:     Use continuity condition and global boundary data ϕ\phi to determine coefficient vectors c1,c2,,cMc_{1},c_{2},\dotsc,c_{M}, and set
u=m=1Mχm𝖦mcmu=\sum_{m=1}^{M}\chi_{m}\mathsf{G}_{m}c_{m} (23)
8:Return: approximate global solution uu.

Although we do not apply full-blown rSVD here, the homogenizable and low-rank property of the local solution space implies that 𝖦mr\mathsf{G}_{m}^{\mathrm{r}} and 𝖦m\mathsf{G}_{m} share similar range with the number of basis functions kmk_{m} in 𝖦mr\mathsf{G}_{m}^{\mathrm{r}} being much smaller than nmn_{m}, the number of grid points on 𝒦m\partial\mathcal{K}_{m}, and independent of ϵ\epsilon. In Figure 3 we plot the angles between 𝖦mr\mathsf{G}_{m}^{\mathrm{r}} and 𝖦m\mathsf{G}_{m} for two of the equations discussed in Section 2. In both cases, and for small ϵ\epsilon, the approximated Green’s matrix quickly recovers the true Green’s matrix as the number of samples kmk_{m} increases, and thus captures the local solution space. We should note that if 𝖦m\mathsf{G}_{m} does not have low-rank structure, in the sense that kmnmk_{m}\sim n_{m}, then solving (22) would be equally expensive as solving (21), hence the random sampling technique does not gain any computational efficiency when the system is not homogenizable.

In Figure 4 we showcase the basis functions on a patch for elliptic equation with media coefficient a(x1,x2)=1+1000 1S(x1,x2)a(x_{1},x_{2})=1+1000\,\mathbf{1}_{S}(x_{1},x_{2}), with S={(x1,x2)[0,1]2:(x1cos(100(x10.5)2+(x20.5)2))x20.5}S=\{(x_{1},x_{2})\in[0,1]^{2}:(x_{1}\cos(100\sqrt{(x_{1}-0.5)^{2}+(x_{2}-0.5)^{2}}))\leq x_{2}-0.5\}. For this non-conventional media without any periodic structure, the traditional multiscale methods are no longer valid, but our method still quickly captures the optimal basis.

Refer to caption
Refer to caption
Figure 3: Angle between the true Green’s matrix 𝖦m\mathsf{G}_{m} and the approximate version 𝖦mr\mathsf{G}_{m}^{\mathrm{r}}, confined on the interior of a patch 𝒦~m\widetilde{\mathcal{K}}_{m} for some mm, as the number of random samples increases. Left plot: Angle for 1D RTE (4) with diffusive kernel (5) and various values of ϵ\epsilon. Right plot: Angle for elliptic equation (12) on a rectangular local patch with ϵ=24\epsilon=2^{-4}.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 4: Optimal basis functions and their projections onto the approximate spaces. First row: First two singular vectors of the Green’s matrix on a local patch. Second row: Projections onto the space spanned by 6 random sampled basis functions. Note that the small random sample captures well the leading eigenvectors of the true Green’s operator.

For particular boundary conditions ϕ\phi, the global solution is assembled from the local basis functions in the online stage. Two numerical examples are shown in Figure 5. In both examples, there is little visible difference between the reference solution and the approximated one computed from the reduced random basis. Only 8.3% and 62.5%, respectively, of the degrees of freedom required by the full basis are needed to represent these solutions using a random basis; see details in [MR4155236].

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 5: First row: Solutions for 1D RTE (4) with Henyey-Greenstein kernel in (5) with ϵ=26\epsilon=2^{-6}. Second row: Solutions for the elliptic equation (12) with Dirichlet boundary condition and highly oscillatory medium a(x,x/ϵ)=2+sin(2πx1)cos(2πx2)+2+1.8sin(2πx1/ϵ)2+1.8cos(2πx2/ϵ)+2+sin(2πx2/ϵ)2+1.8cos(2πx1/ϵ)a(x,x/\epsilon)=2+\sin(2\pi x_{1})\cos(2\pi x_{2})+\frac{2+1.8\sin(2\pi x_{1}/\epsilon)}{2+1.8\cos(2\pi x_{2}/\epsilon)}+\frac{2+\sin(2\pi x_{2}/\epsilon)}{2+1.8\cos(2\pi x_{1}/\epsilon)} with ϵ=24\epsilon=2^{-4}. The left column shows reference solutions while the right column is obtained from randomized reduced bases.

Since we do not have access to the full set of basis functions, the condition that uu defined in (23) is continuous across subdomains can be satisfied only in a least-squares sense; see [MR4155236] for details.

4.2 A low-rank Schwarz method

Our second approach for exploiting the low-rank property in multiscale computations is based on Schwarz iteration. The Schwarz method is a standard iteration algorithm within the domain decomposition framework, in which boundary-value problems are solved on the patches, with neighboring patches subsequently exchanging information and re-solving until consistency is attained. The exchange of boundary information between neighboring patches is known as the boundary-to-boundary (BtB) map. The map has an exploitable low-rank property.

To develop the approach, we write the solution of (18) as

uϵ(x)=m=1Nχm(x)umϵ(x),u^{\epsilon}(x)=\sum_{m=1}^{N}\chi_{m}(x)u^{\epsilon}_{m}(x)\,, (24)

where the partition-of-unity functions χm\chi_{m} are defined in (20). The solution umϵu^{\epsilon}_{m} on patch mm is uniquely determined by fmf_{m}, its local boundary condition, according to the equation

{ϵumϵ=0,x𝒦mumϵ(x)=fm(x),x𝒦m.\begin{cases}\mathcal{L}^{\epsilon}u^{\epsilon}_{m}=0\,,&\quad x\in\mathcal{K}_{m}\\ u_{m}^{\epsilon}(x)=f_{m}(x)\,,&\quad x\in\partial\mathcal{K}_{m}\,.\end{cases} (25)

The Schwarz method starts by initial guesses to the local boundary conditions fm=fm0f_{m}=f_{m}^{0}, then on iteration tt, it solves the subproblems (25) with fm=fmtf_{m}=f_{m}^{t} to obtain all local solutions umϵu_{m}^{\epsilon}. By confining umϵu_{m}^{\epsilon} on the boundaries of adjacent patches, one updates the boundary conditions for surrounding patches:

fmt(x)𝒮mumϵ(x)|𝒦~m𝒫mfnt+1(x),nm.f^{t}_{m}(x)\xrightarrow{\mathcal{S}_{m}}u_{m}^{\epsilon}(x)|_{\widetilde{\mathcal{K}}_{m}}\xrightarrow{\mathcal{P}_{m}}f^{t+1}_{n}(x)\,,\forall n\in\mathcal{I}_{m}\,. (26)

Here 𝒮m\mathcal{S}_{m} denote the solution to (25) confined in the interior of 𝒦m\mathcal{K}_{m}, and 𝒫m\mathcal{P}_{m} takes the trace of the solution on the neighboring boundaries 𝒦m𝒦n\mathcal{K}_{m}\cap\partial\mathcal{K}_{n} for nmn\in\mathcal{I}_{m}, for the updated boundary condition.

Define the BtB map by 𝒜m:=𝒫m𝒮m\mathcal{A}_{m}:=\mathcal{P}_{m}\circ\mathcal{S}_{m}, and define 𝒜\mathcal{A} and ftf^{t} to be the aggregation of 𝒜m\mathcal{A}_{m} and fmtf_{m}^{t}, respectively, over i=1,2,,Mi=1,2,\dotsc,M. We can then write the updating procedure as

ft+1=𝒜ft.f^{t+1}=\mathcal{A}f^{t}\,.

The overall method is summarized in Algorithm 3.

Algorithm 3 Schwarz method for multiscale PDE ϵuϵ=0\mathcal{L}^{\epsilon}u^{\epsilon}=0 over 𝒦\mathcal{K} with uϵ=f\mathcal{B}u^{\epsilon}=f on 𝒦\partial\mathcal{K}
1:Given total iterations TT;
2:Domain Decomposition
3:     Partition domain according to (19).
4:Schwarz Iteration:
5:     Initialize fm0f_{m}^{0} for each 𝒦m\partial\mathcal{K}_{m} and set t=0t=0 .
6:     While |fntfnt1|>TOL|f_{n}^{t}-f_{n}^{t-1}|>\text{TOL}      
7:         Solve (25) for umtu_{m}^{t} using fmtf_{m}^{t} for each mm .
8:         Update fnt+1=umtf_{n}^{t+1}=u_{m}^{t} on 𝒦m𝒦n\mathcal{K}_{m}\cap\partial\mathcal{K}_{n}, nmn\in\mathcal{I}_{m} .
9:         tt+1t\to t+1.      
10:     End
11:     Solve (25) for umtu_{m}^{t} using fmtf_{m}^{t} for each mm .
12:     Assemble global solution u=m=1Nχmumtu=\sum_{m=1}^{N}\chi_{m}u_{m}^{t} .
13:Return: approximated global solution uTu^{T}.

Most of the computation in the Schwarz method during the iteration comes from solving the boundary-value PDEs on the patches, to implement the map 𝒜\mathcal{A}. Since the PDE is homogenizable, the solution space on each patch is approximately low rank, and the map 𝒜\mathcal{A} can be expected to inherit this property. If we can “learn” this operator in an offline stage, and simply apply a low-rank approximation repeatedly in the online stage, the online part of Algorithm 3 can be made much more efficient. In our approach, Algorithm 1 is used to compress the map 𝒜\mathcal{A}.

This approach is quite different from the one described in Section 4.1, in the sense that it is not only the range of the solution space, but the whole operator that is being approximated. To apply Algorithm 1, we need to define the “adjoint operator” for 𝒜\mathcal{A} on the PDE level. This operator is composed of the adjoints 𝒮m\mathcal{S}_{m}^{\ast} for the local solution operators 𝒮m\mathcal{S}_{m} of (25) on each domain 𝒦m\mathcal{K}_{m}. The form of 𝒮m\mathcal{S}_{m}^{\ast} is specific to the PDE; we use the elliptic equation as an example. Defining ϵ=(a(x,x/ϵ))\mathcal{L}^{\epsilon}=\nabla\cdot\left(a(x,x/\epsilon)\nabla\right), 𝒮m\mathcal{S}_{m}^{\ast} is defined in the following result.

Theorem 5.

Let 𝒮m\mathcal{S}_{m} be the confined solution operator for the elliptic equation with Dirichlet boundary condition on patch 𝒦m\mathcal{K}_{m}. Given any function gg supported on 𝒦~m\widetilde{\mathcal{K}}_{m}, the adjoint operator 𝒮m\mathcal{S}_{m}^{\ast} acting on gg is given by:

𝒮mg=ahn|𝒦m,{\mathcal{S}}_{m}^{\ast}g=a\frac{\partial h}{\partial n}\Big{|}_{\partial\mathcal{K}_{m}}\,, (27)

where hn\frac{\partial h}{\partial n} is the outer normal derivative on 𝒦m\partial\mathcal{K}_{m} and hh solves the following sourced elliptic equation:

{(a(x,xϵ)h(x))=g,x𝒦mh(x)=0,x𝒦m.\begin{cases}\nabla\cdot\left(a\left(x,\frac{x}{\epsilon}\right)\nabla h(x)\right)={g}\,,&x\in\mathcal{K}_{m}\\ h(x)=0\,,&x\in\partial\mathcal{K}_{m}\,.\end{cases} (28)

We also describe calculation of the adjoint operator 𝒮m\mathcal{S}_{m}^{\ast} for the RTE (4).

Theorem 6.

Let 𝒮m\mathcal{S}_{m} be the confined solution operator for RTE (4) and the conditions in Theorem 1 hold. Given any function gg supported on 𝒦~m×𝕍\widetilde{\mathcal{K}}_{m}\times\mathbb{V}, the adjoint operator 𝒮m\mathcal{S}_{m}^{\ast} is defined as follows

𝒮mg(x,v)=h(x,v),x𝒦m,vnx<0,\mathcal{S}_{m}^{\ast}g(x,v)=h(x,v)\,,\quad x\in\partial\mathcal{K}_{m}\,,\quad v\cdot n_{x}<0\,, (29)

where hh solves the adjoint RTE over 𝒦m\mathcal{K}_{m}, which is

vxh𝒮ϵ[h]=g(x,v),(x,v)𝒦m×𝕍,-v\cdot\nabla_{x}h-\mathcal{S}^{\epsilon}[h]=g(x,v)\,,\quad(x,v)\in\mathcal{K}_{m}\times\mathbb{V}\,, (30)

with outgoing boundary condition h(x,v)=0h(x,v)=0 on x𝒦mx\in\partial\mathcal{K}_{m} and vnx>0v\cdot n_{x}>0.

The specific form of the adjoint operator 𝒮m\mathcal{S}_{m}^{\ast} allows us to adapt the randomized SVD algorithm to compress the confined solution map 𝒮m\mathcal{S}_{m}; see Algorithm 4. This method requires only kk solves of local PDE (25) and sourced adjoint PDE (28) (or (30)), together with a QR factorization and SVD of relatively small matrices. The overall low-rank Schwarz iteration is then summarized in Algorithm 5.

Algorithm 4 Randomized SVD for 𝒮m\mathcal{S}_{m}
1:Given target rank rr and oversampling parameter pp;
2:Set k=r+pk=r+p;
3:Stage A:
4:     Generate kk random boundary conditions ξj\xi_{j} on 𝒦m\partial\mathcal{K}_{m}.
5:     Solve (25) using ξj\xi_{j} as boundary conditions and restrict the solution over 𝒦~m\widetilde{\mathcal{K}}_{m} to obtain ujϵu_{j}^{\epsilon}.
6:     Find orthonormal basis 𝖰=[q1,,qk]\mathsf{Q}=[q_{1},\dots,q_{k}] of U~=[u1,,uk]\widetilde{U}=[u_{1},\dots,u_{k}].
7:Stage B:
8:     Construct zero extension of qkq_{k} over 𝒦m\mathcal{K}_{m}, denoted by q~k\widetilde{q}_{k}.
9:     Solve (28) or (30) for hjh_{j} using q~k\widetilde{q}_{k} as source.
10:     Compute bjb_{j} using hjh_{j} by flux (27) or restriction on incoming boundary (29).
11:     Assemble all fluxes B=[b1,,bk]B=[b_{1},\dots,b_{k}].
12:     Compute SVD of B=𝖴~kΣk𝖵kB^{\ast}=\widetilde{\mathsf{U}}_{k}\Sigma_{k}\mathsf{V}_{k}^{\ast}.
13:     Compute 𝖴k=𝖰𝖴~k\mathsf{U}_{k}=\mathsf{Q}\widetilde{\mathsf{U}}_{k}.
14:Return 𝖴k,Σk,𝖵k\mathsf{U}_{k},\Sigma_{k},\mathsf{V}_{k}.

In Figure 6 we present the confined solution operator 𝒮m\mathcal{S}_{m} and its low-rank approximation 𝒮mr\mathcal{S}_{m}^{\mathrm{r}}. For the radiative transfer equation, the map is a 2880-by-40 matrix, with the size of each patch being 0.2×\times[-1,1], with Δx=1/360\Delta x=1/360 and Δv=1/40\Delta v=1/40. The random sampling procedure reconstructs it with just 6 samples. For the elliptic equation, the map is a 1600-by-160 matrix, and the size of each patch is 1×\times[0,1] with Δx=1/40\Delta x=1/40. The random sampling approximates it well with 60 samples. The compression rates for these examples are thus 6.7 and 2.7, respectively. See details in [MR4252068].

In Figure 7, we show numerical examples for the global solutions of two problems obtained from the approach of this section. The reference solution (obtained with a fine mesh) is well captured by the approximation that uses the low-rank BtB map as the surrogate in the Schwarz iteration. These two cases use just 15%15\% and 43%43\%, respectively, of the number of local solves needed to capture the BtB map at fine scale. While the relative error of the reduced Schwarz method decays as fast as the standard Schwarz iteration, as shown in Figure 8, the cost is much reduced. See Table 1 for a comparison of computation times.

Algorithm 5 Reduced Schwarz method for multiscale PDE ϵuϵ=0\mathcal{L}^{\epsilon}u^{\epsilon}=0 over 𝒦\mathcal{K} with uϵ=ϕ\mathcal{B}u^{\epsilon}=\phi on 𝒦\partial\mathcal{K}
1:Given rank kk, total iterations TT.
2:Domain Decomposition
3:     Partition domain according to (19).
4:Offline Stage:
5:     For all mm, use Algorithm 4 to find the rank-kk approximation to 𝒮m\mathcal{S}_{m}, denoted by 𝖴kmΣkm𝖵km,\mathsf{U}_{k}^{m}\Sigma_{k}^{m}\mathsf{V}_{k}^{m,\ast}.
6:Online:
7:     Initiate fm0(x)f_{m}^{0}(x) for each 𝒦m\partial\mathcal{K}_{m}, and set t=0t=0.
8:     While |fntfnt1|>TOL|f_{n}^{t}-f_{n}^{t-1}|>\text{TOL}      
9:         Evaluate umt=𝖴kmΣkm𝖵km,fmtu_{m}^{t}=\mathsf{U}_{k}^{m}\Sigma_{k}^{m}\mathsf{V}_{k}^{m,\ast}f_{m}^{t} for each mm.
10:         Update fnt+1=umtf_{n}^{t+1}=u_{m}^{t} on 𝒦m𝒦n\mathcal{K}_{m}\cap\partial\mathcal{K}_{n}, nmn\in\mathcal{I}_{m}.
11:         tt+1t\to t+1.      
12:     End
13:     Solve (25) for umtu_{m}^{t} using fmtf_{m}^{t} for each mm .
14:     Assemble global solution u=n=1Nχmumtu=\sum_{n=1}^{N}\chi_{m}u^{t}_{m}.
15:Return u(x)u(x).
Refer to caption
Refer to caption
Figure 6: The singular decay of the restricted local solution operator 𝒮m\mathcal{S}_{m} and its low rank approximation 𝒮mr\mathcal{S}_{m}^{\mathrm{r}} for the RTE (left) and the elliptic equation (right). In RTE (4) we use heterogeneous collision kernel (9) k(x/ϵ,v,v)=σϵ(x)=1811.1+cos(4πx)1.1+sin(2πx/ϵ)k(x/\epsilon,v,v^{\prime})=\sigma^{\epsilon}(x)=\frac{1}{81}\frac{1.1+\cos(4\pi x)}{1.1+\sin(2\pi x/\epsilon)}, and in elliptic equation we use a(x,x/ϵ)=2+1.8sin(πx1/ϵ)2+1.8cos(πx2/ϵ)+2+sin(πx2/ϵ)2+1.8sin(πx1)a(x,x/\epsilon)=\frac{2+1.8\sin(\pi x_{1}/\epsilon)}{2+1.8\cos(\pi x_{2}/\epsilon)}+\frac{2+\sin(\pi x_{2}/\epsilon)}{2+1.8\sin(\pi x_{1})} with ϵ=24\epsilon=2^{-4}.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 7: The comparison between the reference solutions (left column) and the approximation using reduced Schwarz method (right column). The first and second rows are for the RTE and the elliptic equation, respectively. In RTE (4) we use heterogeneous collision kernel (9) k(x/ϵ,v,v)=σϵ(x)=1ϵ11.1+cos(4πx)1.1+sin(2πx/ϵ2)k(x/\epsilon,v,v^{\prime})=\sigma^{\epsilon}(x)=\frac{1}{\epsilon_{1}}\frac{1.1+\cos(4\pi x)}{1.1+\sin(2\pi x/\epsilon_{2})}, with (ϵ1,ϵ2)=(1/81,1/9)(\epsilon_{1},\epsilon_{2})=(1/81,1/9), and in elliptic equation we use a(x,x/ϵ)=2+1.8sin(πx1/ϵ)2+1.8cos(πx2/ϵ)+2+sin(πx2/ϵ)2+1.8sin(πx1)a(x,x/\epsilon)=\frac{2+1.8\sin(\pi x_{1}/\epsilon)}{2+1.8\cos(\pi x_{2}/\epsilon)}+\frac{2+\sin(\pi x_{2}/\epsilon)}{2+1.8\sin(\pi x_{1})} with ϵ=24\epsilon=2^{-4}.
Refer to caption
Figure 8: Relative error for reduced Schwarz methods for various ranks kk for elliptic equation with oscillatory medium that is the same as Figure 7.
RTE offline (s) online (s)
Rank = 5 227.99 0.0029
Rank = 6 268.46 0.0162
Full rank 706.99 0.0148
Schwarz 1027.40
elliptic offline (s) online (s)
Rank = 40 49.7 0.049
Rank = 70 87.3 0.061
Schwarz 31.4
Table 1: Run time comparison between vanilla Schwarz method and the reduced Schwarz method. k=5,6k=5,6 for the RTE and k=40,70k=40,70 for the elliptic equation with Dirichlet boundary condition. The configuration of the media is the same as those in Figure 7.

5 Manifold learning and nonlinear multiscale problems

It is not straightforward to extend the techniques of the previous section to nonlinear PDEs. Despite low-rank properties still holding due to the existence of the limiting equation, the argument based on compressing the Green’s matrix no longer holds. The collection of solutions for different source / boundary terms is not longer a linear subspace, but a solution manifold.

We consider the general nonlinear multiscale problem in the following form

𝒩ϵuϵ=f,\mathcal{N}^{\epsilon}u^{\epsilon}=f\,, (31)

where 𝒩ϵ\mathcal{N}^{\epsilon} is a nonlinear differential operator that depends explicitly on the small parameter ϵ\epsilon. The term ff can be the source term, boundary conditions or initial conditions. Assume further that the equation has an asymptotic limit

𝒩u=f\mathcal{N}^{\ast}u^{\ast}=f (32)

as ϵ0\epsilon\to 0, that is, uϵu0\|u^{\epsilon}-u^{\ast}\|\to 0 as ϵ0\epsilon\to 0. The argument for the linear problem is still applicable: The degrees of freedom required by the classical numerical method for solving (31) grows rapidly as ϵ0\epsilon\to 0, while the existence of the homogenized equation (32) indicates that only O(1)O(1) degrees of freedom should be needed to resolve macro-scale features. From a manifold perspective, the solutions to (31) vary in a high dimensional space as ff changes, but this manifold is approximated to within distance O(ϵ)O(\epsilon) by another manifold whose dimension is O(1)O(1).

Suppose a manifold in a high dimensional space is approximately low-dimensional, can we quickly learn it without paying the high dimensional cost? We turn to manifold learning for answers to this question. We are particularly interested in adopting the ideas from the local linear embedding and multi-scale SVD approaches that learn the manifold from observed point clouds, and interpolate the local solution manifold using multiple tangent-space patches, see references in [chen2020manifold].

We denote the nonlinear solution map of (31) by 𝒮ϵ:f𝒳uϵ𝒴\mathcal{S}^{\epsilon}:f\in\mathcal{X}\to u^{\epsilon}\in\mathcal{Y}, which maps the source term or initial / boundary conditions f(x)f(x) to the solution of the equation. In the offline stage, we randomly sample a large number of configurations fif_{i} in 𝒳\mathcal{X}, and compute the associated solutions uiϵ=𝒮ϵfi𝒴u^{\epsilon}_{i}=\mathcal{S}^{\epsilon}f_{i}\in\mathcal{Y} on fine grids. These solutions form a point cloud in a high dimensional space 𝒴\mathcal{Y}. The {fi}\{f_{i}\}’s are then subdivided into a number of small neighborhoods, and we construct tangential approximations to the mapping 𝒮ϵ\mathcal{S}^{\epsilon} on each neighborhood. In the online stage, given a new configuration ff, we identify the small neighborhood to which it belongs and find the corresponding solution by performing linear interpolation. The overall offline-online strategy in summarized in Algorithm 6. We stress that some modifications are needed to reduce the cost of implementation. For example, the algorithm should be combined with domain decomposition (for example, Schwarz iteration) to further confine the computation to local domains, to save computational cost.

In Figure 9 we plot the local low-dimensional solution manifold for a nonlinear RTE (specifically, a linear RTE nonlinearly coupled with a temperature term). The solution manifold appears to have a local two-dimensional structure; the point clouds lie near on a two-dimensional plane. We refer to [chen2020manifold] for more details of the implementation and numerical results.

Algorithm 6 Manifold learning algorithm for solving 𝒩ϵuϵ=f\mathcal{N}^{\epsilon}u^{\epsilon}=f
1:Offline
2:     Randomly sample fi(x)f_{i}(x), i=1,,Ni=1,\dots,N, and find solutions uiϵ=𝒮ϵfiu^{\epsilon}_{i}=\mathcal{S}^{\epsilon}f_{i}.
3:Online: Given f(x)f(x):
4:     Step 1: Identify the kk-nearest neighbors of f(x)f(x), call them fij,j=1,2,,kf_{i_{j}},j=1,2,\dots,k, with fi1f_{i_{1}} being the nearest neighbor;
5:     Step 2: Compute 𝒮ϵϕui1ϵ+𝖴c\mathcal{S}^{\epsilon}\phi\approx u^{\epsilon}_{i_{1}}+\mathsf{U}\cdot{c} with
𝖴=[||ui2ϵui1ϵuikϵui1ϵ||],\mathsf{U}=\begin{bmatrix}|&&|\\ u^{\epsilon}_{i_{2}}-u^{\epsilon}_{i_{1}}&\dots&u^{\epsilon}_{i_{k}}-u^{\epsilon}_{i_{1}}\\ |&&|\end{bmatrix},
where cc is a set of coefficient that fits ffi1f-f_{i_{1}} with a linear combination of fijfi1f_{i_{j}}-f_{i_{1}}, for j=2,3,,kj=2,3,\dotsc,k.
6:Return uϵ=𝒮ϵϕu^{\epsilon}=\mathcal{S}^{\epsilon}\phi.
Refer to caption
Figure 9: Point cloud and its fitting plane for a 1D nonlinear RTE with Knudsen number ϵ=26\epsilon=2^{-6}, when confined in a small patch containing an interval [0.625,1.375][0.625\,,1.375]. The solution profile is approximately determined by the temperature TT at two grid points x=0.625x=0.625 and x=1.375x=1.375. The zz-axis shows the value of uϵ(x=1)u^{\epsilon}(x=1). The dependence is clearly linear and two-dimensional.

6 Looking Forward

We have seen a vast literature addressing all aspects of the computation of multiscale problems. Over the years, the research has been drifting gradually away from its origin, where solvers were influenced by analytical understanding, specifically of the limiting behavior of the specific PDE. Machine learning algorithms have shown more and more power in sketching the solution profile with a much reduced numerical cost. In particular, the existence of the homogenized limit suggests there are low-rank features in the discrete system, and that random linear algebra techniques and manifold learning methods, when utilized properly, can identify these features for a compressed representation of the PDE solutions.

We have reviewed two methods, both of which make use of the domain decomposition framework. They compress either basis functions or the boundary-to-boundary map in an offline learning stage. This review article serves as a showcase of the power of random solvers in numerical PDEs. For time-dependent problems, and homogenization problems that have weak-convergence (instead of strong) such as quantum systems in the semi-classical regime, further development of the approaches is needed. Incorporation of time and weak-limit in the algorithm-design lies at the core of future challenges.

7 Acknowledgements

The authors acknowledge generous support from NSF, ONR, DOE, and AFOSR. Due to restrictions on the allowed number of references, many important contributions are omitted. The reader is invited to consult the bibliographies in the cited references for a more extensive view of the literature.

References