This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Ritz method for the spectral fractional Laplacian equation using the Caffarelli-Silvestre extension

Yiqi Gu Department of Mathematics, The University of Hong Kong, Pokfulam, Hong Kong ([email protected], [email protected]). This work is supported by HKRGC GRF 12300218, 12300519, 17201020 and 17300021.    Micheal K. Ng11footnotemark: 1
Abstract

In this paper, we propose a novel method for solving high-dimensional spectral fractional Laplacian equations. Using the Caffarelli-Silvestre extension, the dd-dimensional spectral fractional equation is reformulated as a regular partial differential equation of dimension d+1d+1. We transform the extended equation as a minimal Ritz energy functional problem and search for its minimizer in a special class of deep neural networks. Moreover, based on the approximation property of networks, we establish estimates on the error made by the deep Ritz method. Numerical results are reported to demonstrate the effectiveness of the proposed method for solving fractional Laplacian equations up to ten dimensions. Technically, in this method, we design a special network-based structure to adapt to the singularity and exponential decaying of the true solution. Also, A hybrid integration technique combining Monte Carlo method and sinc quadrature is developed to compute the loss function with higher accuracy.

keywords:
Ritz method; Deep learning; Fractional Laplacian; Caffarelli-Silvestre extension; Singularity;
AMS:
65N15; 65N30; 68T07; 41A25

1 Introduction

As a nonlocal generalization of the Laplacian Δ-\Delta, the spectral fractional Laplacian (Δ)s(-\Delta)^{s} with a fraction ss arises in many areas of applications, such as anomalous diffusion [10, 30], turbulent flows [26], Lévy processes [16], quantum mechanics [18], finance [29, 12] and pollutant transport [31]. In this paper, we develop a network-based Ritz method for solving fractional Laplacian equations using the Caffarelli-Silvestre extension.

Let d+d\in\mathbb{N}^{+} be the dimension of the problem and Ω\Omega be a bounded Lipschitz domain in d\mathbb{R}^{d}. Also, suppose 0<s<10<s<1 and ff is a function defined in Ω\Omega, we consider the following spectral fractional Laplacian equation with homogeneous Dirichlet condition,

(1) {(Δ)sU(x)=f(x),xΩ,U(x)=0,xΩ,\begin{cases}(-\Delta)^{s}U(x)=f(x),\quad\forall x\in\Omega,\\ U(x)=0,\quad\forall x\in\partial\Omega,\end{cases}

where (Δ)s(-\Delta)^{s} is defined by the spectral decomposition of Δ-\Delta with the same boundary conditions. More precisely, we suppose the countable set {(λn,ϕn)}n=1\{(\lambda_{n},\phi_{n})\}_{n=1}^{\infty} are all the eigenvalues and orthonormal eigenfunctions of the following problem,

(2) {Δϕ=λϕ,inΩ,ϕ=0,onΩ,(ϕ,ϕ)=1,\begin{cases}-\Delta\phi=\lambda\phi,\quad\text{in}~{}\Omega,\\ \phi=0,\quad\text{on}~{}\partial\Omega,\\ (\phi,\phi)=1,\end{cases}

where (,)(\cdot,\cdot) is the standard inner product in L2(Ω)L^{2}(\Omega). Then for any UL2(Ω)U\in L^{2}(\Omega),

(3) (Δ)sU=n=1U~nλnsϕn,withU~n:=(U,ϕn).(-\Delta)^{s}U=\sum_{n=1}^{\infty}\tilde{U}_{n}\lambda_{n}^{s}\phi_{n},~{}\text{with}~{}\tilde{U}_{n}:=(U,\phi_{n}).

We remark that another definition of the fractional Laplacian is formulated by integrals with non-local structures, and these two definitions do not coincide. It is difficult to solve fractional Laplacian equations of either definition directly using numerical methods for regular differential equations (e.g., finite difference method or finite element method) due to the non-local property of the fractional operator [5, 1]. Instead, one effective approach is to use the Caffarelli-Silvestre extension [7]. Specifically, let us introduce a scalar variable yy and consider the following d+1d+1-dimensional problem

(4) {(yαu(x,y))=0,(x,y)𝒟:=Ω×(0,),limy0yαuy(x,y)=dsf(x),xΩ,u(x,y)=0,(x,y)L𝒟:=Ω×[0,),\begin{cases}\nabla\cdot\left(y^{\alpha}\nabla u(x,y)\right)=0,\quad\forall(x,y)\in\mathcal{D}:=\Omega\times(0,\infty),\\ -\underset{y\rightarrow 0}{\lim}~{}y^{\alpha}u_{y}(x,y)=d_{s}f(x),\quad\forall x\in\Omega,\\ u(x,y)=0,\quad\forall(x,y)\in\partial_{L}\mathcal{D}:=\partial\Omega\times[0,\infty),\end{cases}

where α=12s\alpha=1-2s and ds=212sΓ(1s)/Γ(s)d_{s}=2^{1-2s}\Gamma(1-s)/\Gamma(s). Suppose u(x,y)u(x,y) solves (4), then U(x):=u(x,0)U(x):=u(x,0) is a solution of (1) [24]. Consequently, one can solve the extended problem (4) with regular derivatives to avoid addressing spectral fractional differential operators, with the extra cost that (i) the dimension is increased from dd to d+1d+1; (ii) the domain is extended from the bounded one Ω\Omega to an unbounded cylinder L𝒟\partial_{L}\mathcal{D}. Several methods have been proposed for (4), such as the tensor product finite elements [24] and the enriched spectral method using Laguerre functions [11]. We remark that the Caffarelli-Silvestre extension is exclusively for the spectral fractional Laplacian and not for the integral fractional Laplacian. Moreover, the extension technique can be extended to more general fractional symmetric elliptic operators of the form (A(x)U(x))+C(x)U(x)-\nabla\cdot(A(x)\nabla U(x))+C(x)U(x) with A(x)A(x) being symmetric and positive definite and C(x)C(x) being positive.

However, conventional linear structures such as finite elements and polynomials are usually incapable of high-dimensional approximation in practice. For example, suppose a tensor product linear structure has N~\tilde{N} basis functions in each dimension, then the total degree of freedom is O(N~d)O(\tilde{N}^{d}), which is a huge number if dd is moderately large. Such a curse of dimensionality prevents one from using linear algebra structures in high-dimensional problems with d>3d>3. For the spectral fractional Laplacian, most existing methods based on the Caffarelli-Silvestre extension could solve numerical examples of dimension at most two, mainly due to the limitation of storage. Our primary target is to solve many physically relevant problems that appear in three or higher dimensional situations.

In recent years, deep neural networks (DNNs) are widely studied and utilized in applied problems. As a composition of simple neural functions, the DNN has parameters nonlinearly arrayed in the network structure. For a fully-connected DNN with depth LL, width MM and dimension of inputs dd, the total number of parameters is of O(LM2+Md)O(LM^{2}+Md). Therefore, the degree of freedom increases linearly with the dimension and DNNs are capable of dealing with high-dimensional approximation in practice. Theoretically, it is shown that DNNs have decent approximation properties for particular function spaces (e.g., Barron space). The seminal work of Barron [2, 3] deduces L2L^{2}-norm and LL^{\infty}-norm approximations of two-layer sigmoid networks. Recent work [17, 14, 15, 13, 27, 28, 9, 19] considers more variants of the network-based approximation for Barron-type spaces. Generally, given a Barron function g:Ωg:\Omega\rightarrow\mathbb{R}, there exists a two-layer neural network gMg_{M} with width MM and common-used activations such that

(5) ggMΩO(gM),\|g-g_{M}\|_{\Omega}\leq O\left(\frac{\|g\|_{\mathcal{B}}}{\sqrt{M}}\right),

where g\|g\|_{\mathcal{B}} is the Barron norm of gg, and Ω\|\cdot\|_{\Omega} can be L2(Ω)L^{2}(\Omega), H1(Ω)H^{1}(\Omega) or L(Ω)L^{\infty}(\Omega) norm under different hypothesis. It is worth mentioning that the above error bound is independent of the dimension of the input variable; hence the network-based approximation can overcome the curse of dimensionality.

In this paper, we solve (4) by Ritz method in which DNNs are employed to approximate the solution. More precisely, we reformulate (4) as a minimal Ritz energy functional problem and characterize the Sobolev space of the weak solution. Next, as a subset of the solution space, a class of DNNs is taken as the hypothesis space of the minimization. We design a special network structure for the DNN class such that (i) it satisfies the homogeneous boundary condition on L𝒟\partial_{L}\mathcal{D} in (4); (ii) it decays to zero exponentially as yy\rightarrow\infty; and (iii) it has a singularity behaves as O(yk+1α)O(y^{k+1-\alpha}) for integers kk at y=0y=0. Note that the second and third properties mentioned above are also satisfied by the true solution. Consequently, the special DNNs have better approximation performance than generic DNNs, which is also observed in our numerical experiments.

Theoretically, under a Barron-type framework, we investigate the approximation error between the special DNN class and the solution space under the Sobolev norm, which has a similar form to (5). Based on that, we estimate the solution error of the proposed Ritz method using the special DNN class, assuming that the true solution has components in the Barron space. The final solution error is of O(M1/2)O(M^{-1/2}), which is consistent with the approximation error. We remark that the error bound of our method is advantageous over the existing methods [24, 11] using finite element or Laguerre functions if the dimension is moderately large, since the error order 1/2-1/2 is independent of the dimension. Also, the error order is consistent with the deep Ritz method for regular Laplacian equations [23]

In the implementation, a combination of stochastic and deterministic methods is employed to compute the integrals in the energy functional. Specifically, we utilize the quasi-Monte Carlo method and the sinc quadrature rule [21] to evaluate the integrals in terms of xx and yy, respectively. For the former, due to the potentially high dimension of xx, Monte Carlo-type methods are effective and easy to implement. For the latter, although the integrand in terms of yy is one-dimensional, it has a singular term yαy^{\alpha} when α0\alpha\neq 0. While sinc quadrature is highly accurate for integrals with fractional powers and therefore preferred here. By numerical experiments, we demonstrate that our method can solve model problems up to d=10d=10 with desired accuracy. To the best of our knowledge, this is the first attempt to solve high-dimensional fractional Laplacian equations by deep learning methods.

Overall, the highlights of our work can be summarized as follows:

  • Development of a special approximation structure based on generic DNNs according to the special properties of the true solution;

  • Combination of stochastic Monte Carlo method for high dimensions and deterministic sinc quadrature for high accuracy in the learning process;

  • Simulation of 10-D fractional Laplacian equations with relative error O(102)O(10^{-2}).

The rest of the paper is organized as follows. In Section 2, we reformulate the problem (4) as the minimization of an energy functional and show their equivalence. In Section 3, the fully connected neural networks are introduced. We characterize the special structures of the hypothesis space and discuss its approximation property. In Section 4, we derive the error estimate for the proposed method. Numerical experiments are presented to show the effectiveness of our method in Section 5. Finally, some concluding remarks are given in Section 6.

2 Minimization of Energy Functional

We solve the regular partial differential equation (4) under the framework of Ritz method. The equation can be transformed to an equivalent minimal functional, and we look for Sobolev weak solutions instead of classical solutions. Similarly, one can also solve (4) using Galerkin method by introducing appropriate test spaces such as in [24, 11]. Since learning-based methods aim to find solutions via optimization, the use of Ritz method can be achieved for building such formulation.

2.1 The space of weak solutions

Let 𝒵\mathcal{Z} be any region and ω\omega be a positive weight function. We define the weighted L2L^{2} space as

(1) Lω2(𝒵):={v|𝒵|v(z)|2ω(z)dz<},L^{2}_{\omega}(\mathcal{Z}):=\left\{v~{}\Big{|}~{}\int_{\mathcal{Z}}|v(z)|^{2}\omega(z){\rm d}z<\infty\right\},

equipped with the inner product

(2) (v1,v2)ω,𝒵:=𝒵v1(z)v2(z)ω(z)dz,v1,v2Lω2(𝒵),\left(v_{1},v_{2}\right)_{\omega,\mathcal{Z}}:=\int_{\mathcal{Z}}v_{1}(z)v_{2}(z)\omega(z){\rm d}z,\quad\forall v_{1},v_{2}\in L^{2}_{\omega}(\mathcal{Z}),

and the induced norm

(3) vω,𝒵:=(v,v)ω,𝒵12.\|v\|_{\omega,\mathcal{Z}}:=\left(v,v\right)_{\omega,\mathcal{Z}}^{\frac{1}{2}}.

The weight ω\omega is omitted from the notations if ω1\omega\equiv 1.

We also define the weighted Sobolev space as

(4) Hω1(𝒵):={v|vLω2(𝒵),vLω2(𝒵)},H^{1}_{\omega}(\mathcal{Z}):=\left\{v~{}|~{}v\in L^{2}_{\omega}(\mathcal{Z}),\nabla v\in L^{2}_{\omega}(\mathcal{Z})\right\},

equipped with the norm

(5) v1,ω,𝒵:=(vω,𝒵2+vω,𝒵2)12.\|v\|_{1,\omega,\mathcal{Z}}:=\left(\|v\|_{\omega,\mathcal{Z}}^{2}+\|\nabla v\|_{\omega,\mathcal{Z}}^{2}\right)^{\frac{1}{2}}.

It is shown in [22] the solution of the extended problem (4) has a desirable property that it converges exponentially to zero as yy\rightarrow\infty. Therefore we can define the solution space as

(6) Hyα1,b(𝒟):={vHyα1(𝒟)|limyv(x,y)=0,v(x,y)|L𝒟=0},H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}):=\left\{v\in H^{1}_{y^{\alpha}}(\mathcal{D})~{}\Big{|}~{}\underset{y\rightarrow\infty}{\lim}v(x,y)=0,~{}v(x,y)|_{\partial_{L}\mathcal{D}}=0\right\},

with the norm

(7) vHyα1,b(𝒟)=vyα,𝒟.\|v\|_{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}=\|\nabla v\|_{y^{\alpha},\mathcal{D}}.

Denote the trace for all vHyα1,b(𝒟)v\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) by

(8) tr{v}(x)=v(x,0).{\rm tr}\{v\}(x)=v(x,0).

Moreover, for column vectors or vector-valued functions, we use |||\cdot| to denote their Euclidean norm.

2.2 Minimal energy functional

We aim to characterize the solution of (4) as a minimizer of an corresponding energy functional. For this, we define the functional

(9) [w]:=12(w,w)yα,𝒟ds(f,tr{w})Ω,wHyα1,b(𝒟).\mathcal{I}[w]:=\frac{1}{2}(\nabla w,\nabla w)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{w\}\right)_{\Omega},\quad\forall w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}).

We have the following result.

Theorem 1.

Assume uHyα1,b(𝒟)u\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) solves (4). Then

(10) [u]=minwHyα1,b(𝒟)[w].\mathcal{I}[u]=\underset{w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}{\min}\mathcal{I}[w].

Conversely, if uu satisfies (10), then uu solves the problem (4).

Proof.

To prove (10), for all wHyα1,b(𝒟)w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}), uwHyα1,b(𝒟)u-w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}). Then using the fact that (uw)|L𝒟=(uw)|Ω×{+}=0(u-w)|_{\partial_{L}\mathcal{D}}=(u-w)|_{\Omega\times\{+\infty\}}=0 and integration by parts we have

(11) ((yαu),uw)𝒟=(yαu,(uw))𝒟+(yαuy,uw)Ω×{0}.\left(\nabla\cdot\left(y^{\alpha}\nabla u\right),u-w\right)_{\mathcal{D}}=-\left(y^{\alpha}\nabla u,\nabla(u-w)\right)_{\mathcal{D}}+\left(-y^{\alpha}u_{y},u-w\right)_{\Omega\times\{0\}}.

Note the left hand side is equal to zero since (yαu)0\nabla\cdot\left(y^{\alpha}\nabla u\right)\equiv 0. And the second term on the left is

(12) (yαuy,uw)Ω×{0}=(yαuy|y=0,u(x,0)w(x,0))Ω=(limy0(yαuy),u(x,0)w(x,0))Ω=(dsf(x),tr{u}tr{w})Ω.\left(-y^{\alpha}u_{y},u-w\right)_{\Omega\times\{0\}}=\left(-y^{\alpha}u_{y}|_{y=0},u(x,0)-w(x,0)\right)_{\Omega}\\ =\left(\lim_{y\rightarrow 0}(-y^{\alpha}u_{y}),u(x,0)-w(x,0)\right)_{\Omega}=\left(d_{s}f(x),\text{tr}\{u\}-\text{tr}\{w\}\right)_{\Omega}.

Therefore, it follows (11)

(13) (yαu,(uw))𝒟=(dsf(x),tr{u}tr{w})Ω,\left(y^{\alpha}\nabla u,\nabla(u-w)\right)_{\mathcal{D}}=\left(d_{s}f(x),{\rm tr}\{u\}-{\rm tr}\{w\}\right)_{\Omega},

which implies

(14) (u,u)yα,𝒟ds(f,tr{u})Ω=(u,w)yα,𝒟ds(f,tr{w})Ω.\left(\nabla u,\nabla u\right)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{u\}\right)_{\Omega}=\left(\nabla u,\nabla w\right)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{w\}\right)_{\Omega}.

Using the inequality uw12|u|2+12|w|2\nabla u\cdot\nabla w\leq\frac{1}{2}|\nabla u|^{2}+\frac{1}{2}|\nabla w|^{2}, it leads to

(15) [u][w].\mathcal{I}[u]\leq\mathcal{I}[w].

On the other hand, suppose (10) holds. Fix any wHyα1,b(𝒟)w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) and write

(16) i(τ):=[u+τw],τ.i(\tau):=\mathcal{I}[u+\tau w],\quad\forall\tau\in\mathbb{R}.

Note

(17) i(τ)=12((u+τw),(u+τw))yα,𝒟ds(f,tr{u+τw})Ω=12(u,u)yα,𝒟ds(f,tr{u})Ω+τ((u,w)yα,𝒟ds(f,tr{w})Ω)+12τ2(w,w)yα,𝒟.i(\tau)=\frac{1}{2}\left(\nabla(u+\tau w),\nabla(u+\tau w)\right)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{u+\tau w\}\right)_{\Omega}\\ =\frac{1}{2}\left(\nabla u,\nabla u\right)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{u\}\right)_{\Omega}+\tau\left(\left(\nabla u,\nabla w\right)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{w\}\right)_{\Omega}\right)+\frac{1}{2}\tau^{2}\left(\nabla w,\nabla w\right)_{y^{\alpha},\mathcal{D}}.

Since u+τwHyα1,b(𝒟)u+\tau w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) for each τ\tau, i(τ)i(\tau) takes its minimum at τ=0\tau=0, and thus

(18) 0=i(0)=(u,w)yα,𝒟ds(f,tr{w})Ω.0=i^{\prime}(0)=\left(\nabla u,\nabla w\right)_{y^{\alpha},\mathcal{D}}-d_{s}\left(f,{\rm tr}\{w\}\right)_{\Omega}.

Using integration by parts we have

(19) ((yαu),w)𝒟+(dsf+limy0yαuy,tr{w})Ω=0.\left(\nabla\cdot\left(y^{\alpha}\nabla u\right),w\right)_{\mathcal{D}}+\left(d_{s}f+\underset{y\rightarrow 0}{\lim}~{}y^{\alpha}u_{y},{\rm tr}\{w\}\right)_{\Omega}=0.

Especially, (19) holds for all wCc(𝒟)w\in C^{\infty}_{c}(\mathcal{D}), which implies

(20) 𝒟((yαu))w=0,wCc(𝒟),\int_{\mathcal{D}}\left(\nabla\cdot\left(y^{\alpha}\nabla u\right)\right)w=0,\quad\forall w\in C^{\infty}_{c}(\mathcal{D}),

leading to (yαu)=0\nabla\cdot\left(y^{\alpha}\nabla u\right)=0 in 𝒟\mathcal{D}. And thus by (19)

(21) (dsf+limy0yαuy,tr{w})Ω=0.\left(d_{s}f+\underset{y\rightarrow 0}{\lim}~{}y^{\alpha}u_{y},{\rm tr}\{w\}\right)_{\Omega}=0.

Especially, tr{w}{\rm tr}\{w\} takes over all functions in Cc(Ω)C^{\infty}_{c}(\Omega), which leads to dsf+limy0yαuy=0d_{s}f+\underset{y\rightarrow 0}{\lim}~{}y^{\alpha}u_{y}=0 in Ω\Omega. ∎

By virtue of Theorem 1, it suffices to solve the following optimization

(22) minwHyα1,b(𝒟)[w],\underset{w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}{\min}\mathcal{I}[w],

whose solution is exactly a weak solution of the extended problem (4).

3 Neural Network Approximation

In the numerical computation, one aims to introduce a function set with a finite degree of freedom to approximate the solution space Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}), and minimize []\mathcal{I}[\cdot] in this appropriate set of functions. In many physical relevant problems, it is required to address d3d\geq 3, causing the dimension of 𝒟\mathcal{D} no less than 4. Potentially high dimensions impede the usage of conventional linear structures. However, as a nonlinear structure, DNNs can approximate high-dimensional functions by moderately less degree of freedom. This inspires us to use classes of DNNs to approximate Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) especially when dd is large.

3.1 Fully connected neural network

In our method, we employ the fully connected neural network (FNN) which is one of the most common neural networks in deep learning. Mathematically speaking, let σ(t)\sigma(t) be an activation function which is applied entry-wise to a vector xx to obtain another vector of the same size. Let L+L\in\mathbb{N}^{+} and M+M_{\ell}\in\mathbb{N}^{+} for =1,,L\ell=1,\dots,L, an FNN ϕ^\hat{\phi} is the composition of LL simple nonlinear functions as follows

(1) ϕ^(z;θ):=aThLhL1h1(z)for zd,\hat{\phi}(z;\theta):=a^{T}h_{L}\circ h_{L-1}\circ\cdots\circ h_{1}(z)\quad\text{for }z\in\mathbb{R}^{d},

where aMLa\in\mathbb{R}^{M_{L}}; h(z):=σ(Wz+b)h_{\ell}(z_{\ell}):=\sigma\left(W_{\ell}z_{\ell}+b_{\ell}\right) with WM×M1W_{\ell}\in\mathbb{R}^{M_{\ell}\times M_{\ell-1}} (M0:=dM_{0}:=d) and bMb_{\ell}\in\mathbb{R}^{M_{\ell}} for =1,,L\ell=1,\dots,L. Here MM_{\ell} is called the width of the \ell-th layer and LL is called the depth of the FNN. θ:={a,W,b:1L}\theta:=\{a,\,W_{\ell},\,b_{\ell}:1\leq\ell\leq L\} is the set of all parameters in ϕ^\hat{\phi} to determine the underlying neural network. Common types of activation functions include the sigmoid function (1+et)1(1+e^{-t})^{-1} and the rectified linear unit (ReLU) max(0,t)\max(0,t). We remark that, when solving kk-th order differential equations, many existing network-based methods use the ReLUk+1 activation function max{0,tk+1}\max\{0,t^{k+1}\}, so that their networks are CkC^{k} functions and can be applied by the differential operators. While in the minimization (22), only H1H^{1} regularity is required and therefore ReLU networks suffice.

Denote M:=max{M,1L}M:=\max~{}\{M_{\ell},~{}1\leq\ell\leq L\}, then it is clear that |θ|=O(M2L+Md)|\theta|=O(M^{2}L+Md). Comparatively, the degree of freedom of linear structures such as finite elements and tensor product polynomials increases exponentially with dd. Hence FNNs are more practicable in high-dimensional approximations. For simplicity, we consider the architecture M=MM=M_{\ell} for all \ell and denote L,M,σ\mathcal{F}_{L,M,\sigma} as the set consisting of all FNNs with depth LL, width MM and activation function σ\sigma. In the following passage, all functions involving an FNN will be denoted with the superscript ^\hat{}.

3.2 Special structures of the approximate class

Recent work [20, 25] indicates that deep FNNs can approximate smooth functions in \infty-norm within any desired accuracy as long as the depth or width is large enough. The approximation will be more accurate if the target function has higher regularity. However, it is shown in [8, 11] that the solution of (4) has a singularity at y=0y=0 which behaves as yk+1αy^{k+1-\alpha} for k+k\in\mathbb{N}^{+}. Therefore, it is not appropriate to naively use the class of generic FNNs. Instead, We aim to develop a special structure based on FNNs for the approximate class.

In the enriched spectral method [11], the solution of (19) is approximated by a structure consisting of two parts. One part is a linear combination of smooth basis functions, which approximates the regular component of the solution. The other part is a linear tensor product combination of smooth basis functions and a sequence of artificial terms {ykα}k=1,2,\{y^{k-\alpha}\}_{k=1,2,\cdots}, which is for the singular component of the solution. Following this idea, we use ϕ^(x,y)\hat{\phi}(x,y) to denote any function in the approximate class, and build its structure as the combination of two parts,

(2) ϕ^(x,y)=ϕ^1(x,y)+y1αϕ^2(x,y),\hat{\phi}(x,y)=\hat{\phi}_{1}(x,y)+y^{1-\alpha}\hat{\phi}_{2}(x,y),

where ϕ^1\hat{\phi}_{1}, ϕ^2\hat{\phi}_{2} are FNN-based smooth functions and the term y1αy^{1-\alpha} is introduced to adapt to the singularity at y=0y=0.

Moreover, since the true solution of the extended problem (4) converges to zero as yy\rightarrow\infty, the functions in the approximate class should also preserve this property. To realize it, we can introduce exponential terms concerning yy in the structure of ϕ^(x,y)\hat{\phi}(x,y). Specifically, we let

(3) ϕ^1(x,y)=ϕ^3(x,y)eγy,ϕ^2(x,y)=ϕ^4(x,y)eγ′′y,\hat{\phi}_{1}(x,y)=\hat{\phi}_{3}(x,y)e^{-\gamma^{\prime}y},\quad\hat{\phi}_{2}(x,y)=\hat{\phi}_{4}(x,y)e^{-\gamma^{\prime\prime}y},

where {ϕ^3,ϕ^4}\{\hat{\phi}_{3},\hat{\phi}_{4}\} are FNN-based functions and {γ,γ′′}>0\{\gamma^{\prime},\gamma^{\prime\prime}\}>0 are two auxiliary scalar parameters ensuring that ϕ^1\hat{\phi}_{1} and ϕ^2\hat{\phi}_{2} converges to zero as yy\rightarrow\infty exponentially.

In the end, as a subset of Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}), the approximate class should also consist of functions satisfying the boundary condition; namely, ϕ^|L𝒟=0\hat{\phi}|_{\partial_{L}\mathcal{D}}=0. We achieve this by setting

(4) ϕ^3(x,y)=ϕ^(x,y)h(x),ϕ^4(x,y)=ϕ^′′(x,y)h(x),\hat{\phi}_{3}(x,y)=\hat{\phi}^{\prime}(x,y)h(x),\quad\hat{\phi}_{4}(x,y)=\hat{\phi}^{\prime\prime}(x,y)h(x),

where {ϕ^,ϕ^′′}\{\hat{\phi}^{\prime},\hat{\phi}^{\prime\prime}\} are generic FNNs and h(x)h(x) is a smooth function constructed particularly to satisfy h(x)=0h(x)=0 on Ω\partial\Omega. For example, if Ω\Omega is a hypercube (a1,b1)××(ad,bd)(a_{1},b_{1})\times\cdots\times(a_{d},b_{d}), then h(x)h(x) can be chosen as h(x)=i=1d(xiai)(xibi)h(x)=\prod_{i=1}^{d}(x_{i}-a_{i})(x_{i}-b_{i}); if Ω\Omega has a boundary characterized by a level set, say Ω={x|ρ(x)=0}\partial\Omega=\{x~{}|~{}\rho(x)=0\}, for some continuous function ρ\rho, then h(x)h(x) can be chosen as h(x)=ρ(x)h(x)=\rho(x). Generally, we can set h(x)=F(dist(x,Ω))h(x)=F(\text{dist}(x,\partial\Omega)), where FF is a particular analytic function satisfying F(0)=0F(0)=0 and dist(x,Ω)\text{dist}(x,\partial\Omega) represents the distant between xx and Ω\partial\Omega. In practice, we expect to set h(x)h(x) as smooth as possible so that ϕ^3\hat{\phi}_{3} and ϕ^4\hat{\phi}_{4} preserve the same regularity as ϕ^\hat{\phi}^{\prime} and ϕ^′′\hat{\phi}^{\prime\prime}, respectively.

To sum up, we build the following special structure for the approximate class,

(5) ϕ^((x,y);θ)=ϕ^((x,y);θ)h(x)eγy+y1αϕ^′′((x,y);θ′′)h(x)eγ′′y,\hat{\phi}((x,y);\theta)=\hat{\phi}^{\prime}((x,y);\theta^{\prime})h(x)e^{-\gamma^{\prime}y}+y^{1-\alpha}\hat{\phi}^{\prime\prime}((x,y);\theta^{\prime\prime})h(x)e^{-\gamma^{\prime\prime}y},

where θ:={θ,θ′′,γ,γ′′}\theta:=\{\theta^{\prime},\theta^{\prime\prime},\gamma^{\prime},\gamma^{\prime\prime}\} is the set of all trainable parameters. We denote 𝒩L,M,σ,h\mathcal{N}_{L,M,\sigma,h} as the class of all neural networks having the structure in (5), namely,

(6) 𝒩L,M,σ,h={ϕ^:𝒟|ϕ^(x,y)=ϕ^(x,y)h(x)eγy+y1αϕ^′′(x,y)h(x)eγ′′y,ϕ^,ϕ^′′L,M,σ,γ,γ′′+}.\mathcal{N}_{L,M,\sigma,h}=\Big{\{}\hat{\phi}:\mathcal{D}\rightarrow\mathbb{R}~{}\Big{|}~{}\hat{\phi}(x,y)=\hat{\phi}^{\prime}(x,y)h(x)e^{-\gamma^{\prime}y}+y^{1-\alpha}\hat{\phi}^{\prime\prime}(x,y)h(x)e^{-\gamma^{\prime\prime}y},\\ \forall\hat{\phi}^{\prime},\hat{\phi}^{\prime\prime}\in\mathcal{F}_{L,M,\sigma},~{}\forall\gamma^{\prime},\gamma^{\prime\prime}\in\mathbb{R}^{+}\Big{\}}.

It is clear that 𝒩L,M,σ,h\mathcal{N}_{L,M,\sigma,h} is a subset of Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) as long as σ\sigma is Lipschitz continuous and has a polynomial growth bound; namely, |σ(t)|C(1+|t|p)|\sigma(t)|\leq C(1+|t|^{p}) for all tt\in\mathbb{R} with a constant C>0C>0 and an integer p>0p>0. In our Ritz method, 𝒩L,M,σ,h\mathcal{N}_{L,M,\sigma,h} is taken as the approximate class of Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}).

3.3 Approximation property

We will show that functions in Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) can be approximated by the special networks in 𝒩L,M,σ,h\mathcal{N}_{L,M,\sigma,h} as MM\rightarrow\infty. To illustrate the approximation property, we first introduce the Barron space and then derive the error bounds for neural-network approximation, assuming that the target function has components in the Barron space. In this section, we specifically focus on the FNNs with ReLU activation and specify σ(x)=max(x,0)\sigma(x)=\max(x,0). For simplicity, we concatenate variables xdx\in\mathbb{R}^{d} and yy\in\mathbb{R} by writing z=(x,y)d+1z=(x,y)\in\mathbb{R}^{d+1}.

3.3.1 Barron space

Let us first quickly review the Barron space and norm. We will focus on the definition discussed in [15] which represents infinitely wide two-layer ReLU FNNs. Following Section 3.1, recall the set of two-layer ReLU FNNs without output bias is given by

(7) 2,M,ReLU={ϕ^|ϕ^(z)=1Mi=1Maiσ(biTz+ci),(ai,bi,ci)×d+1×}.\mathcal{F}_{2,M,\text{ReLU}}=\left\{\hat{\phi}~{}\Big{|}~{}\hat{\phi}(z)=\frac{1}{M}\sum_{i=1}^{M}a_{i}\sigma(b_{i}^{T}z+c_{i}),\quad\forall(a_{i},b_{i},c_{i})\in\mathbb{R}\times\mathbb{R}^{d+1}\times\mathbb{R}\right\}.

For a probability measure π\pi on ×d+1×\mathbb{R}\times\mathbb{R}^{d+1}\times\mathbb{R}, we set the function

(8) fπ(z)=𝒟aσ(bTz+c)π(da,db,dc)=𝔼π[aσ(bTz+c)],zd+1,f_{\pi}(z)=\int_{\mathcal{D}}a\sigma(b^{T}z+c)\pi({\rm d}a,{\rm d}b,{\rm d}c)=\mathbb{E}_{\pi}[a\sigma(b^{T}z+c)],\quad\forall z\in\mathbb{R}^{d+1},

given this expression exists. For a function u:d+1u:\mathbb{R}^{d+1}\rightarrow\mathbb{R}, we use Πu\Pi_{u} to denote the set of all probability measures π\pi such that fπ(z)=u(z)f_{\pi}(z)=u(z) almost everywhere. Then the Barron norm is defined as

(9) u2:=infπΠu×d+1×a2(|b|+|c|)2π(da,db,dc)=infπΠu𝔼π[a2(|b|+|c|)2].\|u\|_{\mathcal{B}}^{2}:=\underset{\pi\in\Pi_{u}}{\inf}\int_{\mathbb{R}\times\mathbb{R}^{d+1}\times\mathbb{R}}a^{2}(|b|+|c|)^{2}\pi({\rm d}a,{\rm d}b,{\rm d}c)=\underset{\pi\in\Pi_{u}}{\inf}~{}\mathbb{E}_{\pi}[a^{2}(|b|+|c|)^{2}].

The infimum of the empty set is considered as ++\infty. The set of all functions with finite Barron norm is denoted by \mathcal{B}. It is shown in [15] that \mathcal{B} equipped with the Barron norm is a Banach space which is called Barron space.

3.4 Error estimation

Let uu be a function in Hyα1,b(𝒟)\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}), and further assume that uu converges to zero exponentially as yy\rightarrow\infty. In the error analysis, we make the following assumption that uu can be factorized explicitly by components vanishing on L𝒟\partial_{L}\mathcal{D} and decaying to zero exponentially as yy\rightarrow\infty.

Assumption 3.1.

There exist functions hH01(Ω)h\in H_{0}^{1}(\Omega), vv\in\mathcal{B} and some number η>0\eta>0 such that

(10) u(z)=v(z)h(x)eηy.u(z)=v(z)h(x)e^{-\eta y}.

Especially, if Assumption 3.1 holds, we can normalize hh such that |h|1|h|\leq 1 and |h|1|\nabla h|\leq 1. Assumption 3.1 is indeed satisfied in some situations. For example, in the case Ω=[1,1]2\Omega=[-1,1]^{2}, s=0.5s=0.5 and f=2πsin(πx1)sin(πx2)f=\sqrt{2}\pi\sin(\pi x_{1})\sin(\pi x_{2}), the solution uu of (4) is given by u=sin(πx1)sin(πx2)e2πyu=\sin(\pi x_{1})\sin(\pi x_{2})e^{-\sqrt{2}\pi y}. Note that by Taylor series sin(πx)=π2(x21)+π8(x21)2+O((x21)3)\sin(\pi x)=-\frac{\pi}{2}(x^{2}-1)+\frac{\pi}{8}(x^{2}-1)^{2}+O\left((x^{2}-1)^{3}\right), uu can be written as (10) with

v=(π+π4(x121)+O((x111)2))(π+π4(x221)+O((x221)2)),\displaystyle v=\left(-\pi+\frac{\pi}{4}(x_{1}^{2}-1)+O\left((x_{1}^{1}-1)^{2}\right)\right)\left(-\pi+\frac{\pi}{4}(x_{2}^{2}-1)+O\left((x_{2}^{2}-1)^{2}\right)\right),
h=(x121)(x221)4,η=2π.\displaystyle h=\frac{(x_{1}^{2}-1)(x_{2}^{2}-1)}{4},\quad\eta=\sqrt{2}\pi.

Now we investigate the approximation error between 𝒩2,M,ReLU,h\mathcal{N}_{2,M,\text{ReLU},h} and Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}). It suffices to consider a special subset 𝒮2,M,ReLU,h\mathcal{S}_{2,M,\text{ReLU},h} given by

(11) 𝒮2,M,ReLU,h={u^|u^(z)=v^(z)h(x)eηy,v^2,M,ReLU}.\mathcal{S}_{2,M,\text{ReLU},h}=\left\{\hat{u}~{}|~{}\hat{u}(z)=\hat{v}(z)h(x)e^{-\eta y},\quad\forall\hat{v}\in\mathcal{F}_{2,M,\text{ReLU}}\right\}.

Note in 𝒮2,M,ReLU,h\mathcal{S}_{2,M,\text{ReLU},h} only the parameters of v^\hat{v} are free and trainable, while η\eta is fixed. Clearly, 𝒮2,M,ReLU,h\mathcal{S}_{2,M,\text{ReLU},h} is a subset of 𝒩2,M,ReLU,h\mathcal{N}_{2,M,\text{ReLU},h}. The following theorem and proof are referred to the result in [23] for deep Ritz method in a bounded domain.

Theorem 2.

Let uHyα1,b(𝒟)u\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}). If Assumption 3.1 is true with |h|1|h|\leq 1 and |h|1|\nabla h|\leq 1, then there exists some u^𝒮2,M,ReLU,h\hat{u}\in\mathcal{S}_{2,M,\text{ReLU},h} such that

(12) u^uHyα1,b(𝒟)C(Ω,η)M12v,\|\hat{u}-u\|_{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}\leq C(\Omega,\eta)M^{-\frac{1}{2}}\|v\|_{\mathcal{B}},

with

(13) C(Ω,η)=232|Ω|12[RΩ2+1α+1+1α+3+(4η3+2η2)(RΩ2+1)+4η3+6η2+6η+38η4e2η]12,C(\Omega,\eta)=2^{\frac{3}{2}}|\Omega|^{\frac{1}{2}}\left[\frac{R_{\Omega}^{2}+1}{\alpha+1}+\frac{1}{\alpha+3}+\frac{(4\eta^{3}+2\eta^{2})(R_{\Omega}^{2}+1)+4\eta^{3}+6\eta^{2}+6\eta+3}{8\eta^{4}}e^{-2\eta}\right]^{\frac{1}{2}},

where RΩ:=max(supxΩ|x|,1)R_{\Omega}:=\max\left(\sup_{x\in\Omega}|x|,1\right).

Proof.

By the definition of the Barron norm, there exsits some probability measure π\pi such that fπ=vf_{\pi}=v a.e. and 𝔼π[a2(|b|+|c|)2]2v2\mathbb{E}_{\pi}\left[a^{2}(|b|+|c|)^{2}\right]\leq 2\|v\|_{\mathcal{B}}^{2}. For all (a,b,c)×d+1×(a,b,c)\in\mathbb{R}\times\mathbb{R}^{d+1}\times\mathbb{R}, using the facts |σ(t)||t||\sigma(t)|\leq|t| and |σ(t)|=χt0|\sigma^{\prime}(t)|=\chi_{t\geq 0} we have

(14) aσ(bTz+c)1,e2ηyyα,𝒟2\displaystyle\|a\sigma(b^{T}z+c)\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2} =\displaystyle= 𝒟[|aσ(bTz+c)|2+|(aσ(bTz+c))|2]e2ηyyαdz\displaystyle\int_{\mathcal{D}}\left[|a\sigma(b^{T}z+c)|^{2}+|\nabla(a\sigma(b^{T}z+c))|^{2}\right]e^{-2\eta y}y^{\alpha}{\rm d}z
\displaystyle\leq 𝒟[a2(bTz+c)2+a2χbTz+c0|(bTz+c)|2]e2ηyyαdz\displaystyle\int_{\mathcal{D}}\left[a^{2}(b^{T}z+c)^{2}+a^{2}\chi_{b^{T}z+c\geq 0}|\nabla(b^{T}z+c)|^{2}\right]e^{-2\eta y}y^{\alpha}{\rm d}z
\displaystyle\leq 𝒟[a2(|b||z|+|c|)2+a2|b|2]e2ηyyαdz.\displaystyle\int_{\mathcal{D}}\left[a^{2}(|b||z|+|c|)^{2}+a^{2}|b|^{2}\right]e^{-2\eta y}y^{\alpha}{\rm d}z.

Since |z|=(|x|2+y2)12(RΩ2+y2)12|z|=(|x|^{2}+y^{2})^{\frac{1}{2}}\leq(R_{\Omega}^{2}+y^{2})^{\frac{1}{2}}, we have

(15) 𝒟[a2(|b||z|+|c|)2+a2|b|2]e2ηyyαdz\displaystyle\int_{\mathcal{D}}\left[a^{2}(|b||z|+|c|)^{2}+a^{2}|b|^{2}\right]e^{-2\eta y}y^{\alpha}{\rm d}z
\displaystyle\leq 𝒟[(RΩ2+y2)a2(|b|+|c|)2+a2|b|2]e2ηyyαdz\displaystyle\int_{\mathcal{D}}\left[(R_{\Omega}^{2}+y^{2})a^{2}(|b|+|c|)^{2}+a^{2}|b|^{2}\right]e^{-2\eta y}y^{\alpha}{\rm d}z
\displaystyle\leq a2(|b|+|c|)2𝒟(RΩ2+y2+1)e2ηyyαdz\displaystyle a^{2}(|b|+|c|)^{2}\int_{\mathcal{D}}(R_{\Omega}^{2}+y^{2}+1)e^{-2\eta y}y^{\alpha}{\rm d}z
\displaystyle\leq a2(|b|+|c|)2|Ω|0(RΩ2+y2+1)e2ηyyαdy:=a2(|b|+|c|)2|Ω|I1.\displaystyle a^{2}(|b|+|c|)^{2}|\Omega|\int_{0}^{\infty}(R_{\Omega}^{2}+y^{2}+1)e^{-2\eta y}y^{\alpha}{\rm d}y:=a^{2}(|b|+|c|)^{2}|\Omega|\cdot I_{1}.

While I1I_{1} is bounded above since

(16) I101(RΩ2+y2+1)yαdy+1(RΩ2+y2+1)e2ηyydy=(RΩ2+1α+1+1α+3)+(4η3+2η2)(RΩ2+1)+4η3+6η2+6η+38η4e2η:=I2.I_{1}\leq\int_{0}^{1}(R_{\Omega}^{2}+y^{2}+1)y^{\alpha}{\rm d}y+\int_{1}^{\infty}(R_{\Omega}^{2}+y^{2}+1)e^{-2\eta y}y{\rm d}y\\ =\left(\frac{R_{\Omega}^{2}+1}{\alpha+1}+\frac{1}{\alpha+3}\right)+\frac{(4\eta^{3}+2\eta^{2})(R_{\Omega}^{2}+1)+4\eta^{3}+6\eta^{2}+6\eta+3}{8\eta^{4}}e^{-2\eta}:=I_{2}.

Combining (14),(15),(16) we have

(17) 𝔼π[aσ(bTz+c)1,e2ηyyα,𝒟2]𝔼π[a2(|b|+|c|)2]|Ω|I22I2|Ω|v2.\mathbb{E}_{\pi}\left[\|a\sigma(b^{T}z+c)\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\right]\leq\mathbb{E}_{\pi}\left[a^{2}(|b|+|c|)^{2}\right]|\Omega|\cdot I_{2}\leq 2I_{2}|\Omega|\cdot\|v\|_{\mathcal{B}}^{2}.

On the other hand, the mapping

×d+1×,(a,b,c)aσ(bTz+c)\mathbb{R}\times\mathbb{R}^{d+1}\times\mathbb{R},\quad(a,b,c)\rightarrow a\sigma(b^{T}z+c)

is continuous and hence Bochner measurable. Also, (17) leads to

𝔼π[aσ(bTz+c)1,e2ηyyα,𝒟](𝔼π[aσ(bTz+c)1,e2ηyyα,𝒟2])12<,\mathbb{E}_{\pi}\left[\|a\sigma(b^{T}z+c)\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}\right]\leq\left(\mathbb{E}_{\pi}\left[\|a\sigma(b^{T}z+c)\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\right]\right)^{\frac{1}{2}}<\infty,

which implies the integral 𝒟aσ(bTz+c)π(da,db,dc)\int_{\mathcal{D}}a\sigma(b^{T}z+c)\pi({\rm d}a,{\rm d}b,{\rm d}c) is a Bochner integral.

We note the fact that if ξ1,,ξM\xi_{1},\cdots,\xi_{M} are independent samples from a random variable ξ\xi, then

𝔼(1Mi=1Mξi𝔼ξ)2=𝔼(1Mi=1M(ξi𝔼ξ))2\displaystyle\mathbb{E}\left(\frac{1}{M}\sum_{i=1}^{M}\xi_{i}-\mathbb{E}\xi\right)^{2}=\mathbb{E}\left(\frac{1}{M}\sum_{i=1}^{M}(\xi_{i}-\mathbb{E}\xi)\right)^{2}
=\displaystyle= 1M2(i=1M𝔼(ξi𝔼ξ)2+1i<jM𝔼(ξi𝔼ξ)𝔼(ξj𝔼ξ))=1Mi=1M𝔼(ξ𝔼ξ)2\displaystyle\frac{1}{M^{2}}\left(\sum_{i=1}^{M}\mathbb{E}(\xi_{i}-\mathbb{E}\xi)^{2}+\sum_{1\leq i<j\leq M}\mathbb{E}(\xi_{i}-\mathbb{E}\xi)\cdot\mathbb{E}(\xi_{j}-\mathbb{E}\xi)\right)=\frac{1}{M}\sum_{i=1}^{M}\mathbb{E}(\xi-\mathbb{E}\xi)^{2}
=\displaystyle= 1M𝔼ξ21M(𝔼ξ)21M𝔼ξ2.\displaystyle\frac{1}{M}\mathbb{E}\xi^{2}-\frac{1}{M}(\mathbb{E}\xi)^{2}\leq\frac{1}{M}\mathbb{E}\xi^{2}.

By similar argument, for independent samples {(ai,bi,ci)}\left\{(a_{i},b_{i},c_{i})\right\} from π\pi, we have

𝔼πM[1Mi=1Maiσ(biTz+ci)fπ(z)1,e2ηyyα,𝒟2]1M𝔼π[aσ(bTz+c)1,e2ηyyα,𝒟2].\mathbb{E}_{\pi^{M}}\left[\left\|\frac{1}{M}\sum_{i=1}^{M}a_{i}\sigma(b_{i}^{T}z+c_{i})-f_{\pi}(z)\right\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\right]\leq\frac{1}{M}\mathbb{E}_{\pi}\left[\|a\sigma(b^{T}z+c)\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\right].

In particular, there exists {(ai,bi,ci)}\left\{(a_{i},b_{i},c_{i})\right\} such that

(18) 1Mi=1Maiσ(biTz+ci)fπ(z)1,e2ηyyα,𝒟21M𝔼π[aσ(bTz+c)1,e2ηyyα,𝒟2].\left\|\frac{1}{M}\sum_{i=1}^{M}a_{i}\sigma(b_{i}^{T}z+c_{i})-f_{\pi}(z)\right\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\leq\frac{1}{M}\mathbb{E}_{\pi}\left[\|a\sigma(b^{T}z+c)\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\right].

Let v^=1Mi=1Maiσ(biTz+ci)\hat{v}=\frac{1}{M}\sum_{i=1}^{M}a_{i}\sigma(b_{i}^{T}z+c_{i}), combining (17) and (18) have obtain

(19) v^v1,e2ηyyα,𝒟22I2|Ω|M1v2\left\|\hat{v}-v\right\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}\leq 2I_{2}|\Omega|M^{-1}\cdot\|v\|_{\mathcal{B}}^{2}

Finally, let u^=v^(z)h(x)eηy\hat{u}=\hat{v}(z)h(x)e^{-\eta y} and denote v¯:=v^v\bar{v}:=\hat{v}-v, then

(20) u^uHyα1,b(𝒟)2\displaystyle\|\hat{u}-u\|_{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}^{2}
=\displaystyle= (v¯(z)h(x)eηy)yα,𝒟2\displaystyle\|\nabla(\bar{v}(z)h(x)e^{-\eta y})\|_{y^{\alpha},\mathcal{D}}^{2}
=\displaystyle= 𝒟(|h(x)xv¯(z)+v¯(z)xh(x)|2+|yv¯(z)v¯(z)|2h2(x))e2ηyyαdz\displaystyle\int_{\mathcal{D}}\left(|h(x)\nabla_{x}\bar{v}(z)+\bar{v}(z)\nabla_{x}h(x)|^{2}+|\partial_{y}\bar{v}(z)-\bar{v}(z)|^{2}h^{2}(x)\right)e^{-2\eta y}y^{\alpha}{\rm d}z
\displaystyle\leq 4𝒟(|v¯|2+|v¯|2)e2ηyyαdz=4v^v1,e2ηyyα,𝒟2.\displaystyle 4\int_{\mathcal{D}}\left(|\nabla\bar{v}|^{2}+|\bar{v}|^{2}\right)e^{-2\eta y}y^{\alpha}{\rm d}z=4\|\hat{v}-v\|_{1,e^{-2\eta y}y^{\alpha},\mathcal{D}}^{2}.

The inequality (12) directly follows (19) and (20). ∎

Theorem 2 provides a fractional equation dimension-independent approximation error bound for neural networks in 𝒮2,M,ReLU,h\mathcal{S}_{2,M,\text{ReLU},h} given that the target function uu has a Barron component. This is a desired property since it avoids the curse of dimensionality. Since 𝒮2,M,ReLU,h𝒩2,M,ReLU,h\mathcal{S}_{2,M,\text{ReLU},h}\subset\mathcal{N}_{2,M,\text{ReLU},h}, Theorem 2 also holds for 𝒩2,M,ReLU,h\mathcal{N}_{2,M,\text{ReLU},h}.

4 Ritz method and Error Estimation

The extended problem (4) can be solved practically by Ritz method. Thanks to the approximation property of the neural network class discussed in Section 3.3, we can replace the hypothesis space Hyα1,b(𝒟)H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) with 𝒩L,M,σ,h\mathcal{N}_{L,M,\sigma,h} in the minimization (22), obtaining

(1) minw^𝒩L,M,σ,h[w^].\underset{\hat{w}\in\mathcal{N}_{L,M,\sigma,h}}{\min}\mathcal{I}[\hat{w}].

Then the solution of (1) will be an approximation to the solution of (22). Same as in Section 3.3, we will estimate the solution error for the two-layer ReLU networks; namely, we consider the case that L=2L=2 and σ(t)=max(0,t)\sigma(t)=\max(0,t). The final error of the original problem (1) will thereafter be presented.

4.1 Error estimation

Let uHyα1,b(𝒟)u^{*}\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}) be a minimizer of (22); namely,

(2) [u]=minwHyα1,b(𝒟)[w]\mathcal{I}[u^{*}]=\min_{w\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}\mathcal{I}[w]

Given that Assumption 3.1 holds for uu^{*} with a factorization u(z)=v(z)h(x)eηyu^{*}(z)=v^{*}(z)h^{*}(x)e^{-\eta^{*}y}, let u^𝒩2,M,ReLU,h\hat{u}^{*}\in\mathcal{N}_{2,M,\text{ReLU},h^{*}} be a minimizer of (1); namely,

(3) [u^]=minw^𝒩2,M,ReLU,h[w^].\mathcal{I}[\hat{u}^{*}]=\min_{\hat{w}\in\mathcal{N}_{2,M,\text{ReLU},h^{*}}}\mathcal{I}[\hat{w}].

We first introduce the following Céa Lemma [22].

Proposition 3.

Let XX be a Hilbert space, VXV\subset X any subset and a:X×Xa:X\times X\rightarrow\mathbb{R} a symmetric, continuous and α\alpha-coercive bilinear form. For fXf\in X^{\prime} define the quadratic energy Ef(u):=12a(u,u)f(u)E_{f}(u):=\frac{1}{2}a(u,u)-f(u) and denote its unique minimizer by ufu_{f}. Then for every vVv\in V it holds that

(4) vufXα1(2(Ef(v)infv~VEf(v~))+infv~V(v~uf,v~uf)).\|v-u_{f}\|_{X}\leq\sqrt{\alpha^{-1}\left(2\left(E_{f}(v)-\inf_{\tilde{v}\in V}E_{f}(\tilde{v})\right)+\inf_{\tilde{v}\in V}(\tilde{v}-u_{f},\tilde{v}-u_{f})\right)}.

It is trivial to show that (,)yα,𝒟(\nabla\cdot,\nabla\cdot)_{y^{\alpha},\mathcal{D}} is a symmetric, continuous and 11-coercive bilinear form. Therefore by Lemma 3 and Theorem 2 we have

(5) u^uHyα1,b(𝒟)infu^𝒩2,M,ReLU,h((u^u),(u^u))yα,𝒟infu^𝒩2,M,ReLU,hu^uHyα1,b(𝒟)C(Ω,η)M12v.\|\hat{u}^{*}-u^{*}\|_{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}\leq\sqrt{\inf_{\hat{u}\in\mathcal{N}_{2,M,\text{ReLU},h^{*}}}(\nabla(\hat{u}-u^{*}),\nabla(\hat{u}-u^{*}))_{y^{\alpha},\mathcal{D}}}\\ \leq\inf_{\hat{u}\in\mathcal{N}_{2,M,\text{ReLU},h^{*}}}\|\hat{u}-u^{*}\|_{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})}\leq C(\Omega,\eta^{*})M^{-\frac{1}{2}}\|v^{*}\|_{\mathcal{B}}.

To derive the error for the original problem (1), we introduce the following trace theorem [24],

Proposition 4.

The trace operator tr{\rm tr} defined in (8) satisfies tr{Hyα1,b(𝒟)}=s(Ω){\rm tr}\{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})\}=\mathbb{H}^{s}(\Omega), and

(6) tr{u}s(Ω)CuHyα1,b(𝒟),uHyα1,b(𝒟),\|{\rm tr}\{u\}\|_{\mathbb{H}^{s}(\Omega)}\leq C\|u\|_{H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D})},\quad\forall u\in H^{1,{\rm b}}_{y^{\alpha}}(\mathcal{D}),

where CC is a constant only depending on Ω\Omega and ss, and the space s(Ω)\mathbb{H}^{s}(\Omega) is defined as

(7) s(Ω)={u=n=1u~nϕnL2(Ω):us=n=1λns|u~n|2<},\mathbb{H}^{s}(\Omega)=\left\{u=\sum_{n=1}^{\infty}\tilde{u}_{n}\phi_{n}\in L^{2}(\Omega):\|u\|_{\mathbb{H}^{s}}=\sum_{n=1}^{\infty}\lambda_{n}^{s}|\tilde{u}_{n}|^{2}<\infty\right\},

with (λn,ϕn)(\lambda_{n},\phi_{n}) being all the eigenvalues and orthonormal eigenfunctions of (2). Especially, s(Ω)\mathbb{H}^{s}(\Omega) can characterized by

(8) s(Ω)u={Hs(Ω),s(0,12),H0012(Ω),s=12,H0s(Ω),s(12,1),\mathbb{H}^{s}(\Omega)u=\begin{cases}H^{s}(\Omega),\quad s\in(0,\frac{1}{2}),\\ H_{00}^{\frac{1}{2}}(\Omega),\quad s=\frac{1}{2},\\ H_{0}^{s}(\Omega),\quad s\in(\frac{1}{2},1),\end{cases}

where

(9) H0012(Ω)={uH12(Ω):Ω|u(x)|2dist(x,Ω)dx<}.H_{00}^{\frac{1}{2}}(\Omega)=\left\{u\in H^{\frac{1}{2}}(\Omega):\int_{\Omega}\frac{|u(x)|^{2}}{\text{\rm dist}(x,\partial\Omega)}{\rm d}x<\infty\right\}.

Recall that the solution of original problem (1) is exactly the trace of the solution of the extended problem (4). Also, Theorem 1 shows the equivalence between the extended problem (4) and the minimization problem (22). Therefore, combining (5) and (6) leads to the following error estimate for the original solution.

Theorem 5.

Given that the solution uu^{*} of (22) satisfies Assumption 3.1, say u(z)=v(z)h(x)eηyu^{*}(z)=v^{*}(z)h^{*}(x)e^{-\eta^{*}y}. Let U(x)U^{*}(x) be the solution of (1) and u^(x,y)\hat{u}^{*}(x,y) be the solution of (1) with L=2L=2, h=hh=h^{*} and σ()=max(0,)\sigma(\cdot)=\max(0,\cdot). Then

(10) u^(x,0)U(x)s(Ω)CM12v,\|\hat{u}^{*}(x,0)-U^{*}(x)\|_{\mathbb{H}^{s}(\Omega)}\leq CM^{-\frac{1}{2}}\|v^{*}\|_{\mathcal{B}},

where CC is a constant only depending on Ω\Omega, ss and η\eta^{*}.

For our network-based method, we obtain the solution error O(M1/2)O(M^{-1/2}) provided that the true solution has a Barron component. This order 1/2-1/2 is consistent with that of the deep Ritz method solving regular Laplacian equations [23] proved under the similar Barron framework. Moreover, the proposed method can be compared with existing methods solving fractional Laplacian [24, 11]. In the early work, finite elements or Laguerre functions are used to approximate the solution, having error orders of O(Npν/d)O(N_{\text{p}}^{-\nu/d}), where ν\nu characterizes the regularity of the true solution and NpN_{\text{p}} is the number of free parameters. If dd is relatively low, say d<2νd<2\nu, then ν/d<1/2-\nu/d<-1/2 and hence the early methods converges faster. Otherwise, the network-based methods outperform the existing methods in the error order. This fact implies that neural networks are advantageous over other classical structures in the high-dimensional approximation.

4.2 Implementation

In the proposed method, we solve the optimization (1), finding a minimizer of []\mathcal{I}[\cdot] in the hypothesis space 𝒩L,M,σ,h\mathcal{N}_{L,M,\sigma,h}. Note that

(11) [ϕ^(x,y)]=120yα(Ω|ϕ^(x,y)|2dx)dydsΩf(x)ϕ^(x,0)dx.\mathcal{I}[\hat{\phi}(x,y)]=\frac{1}{2}\int_{0}^{\infty}y^{\alpha}\left(\int_{\Omega}|\nabla\hat{\phi}(x,y)|^{2}{\rm d}x\right){\rm d}y-d_{s}\int_{\Omega}f(x)\hat{\phi}(x,0){\rm d}x.

In practice, numerical quadrature is required for the integrals in (11). Since dd might be moderately large, for the integral in terms of xx over Ω\Omega, one choice is using Monte Carlo-type quadrature. Specifically, we prescribe a set of NN quadrature nodes 𝒯={xn}n=1N\mathcal{T}=\{x_{n}\}_{n=1}^{N} which are uniformly distributed in Ω\Omega, then for each y(0,)y\in(0,\infty),

(12) Ω|ϕ^(x,y)|2dx|Ω|Nn=1N|ϕ^(xn,y)|2,Ωf(x)ϕ^(x,0)dx|Ω|Nn=1Nf(xn)ϕ^(xn,0).\int_{\Omega}|\nabla\hat{\phi}(x,y)|^{2}{\rm d}x\approx\frac{|\Omega|}{N}\sum_{n=1}^{N}|\nabla\hat{\phi}(x_{n},y)|^{2},\quad\int_{\Omega}f(x)\hat{\phi}(x,0){\rm d}x\approx\frac{|\Omega|}{N}\sum_{n=1}^{N}f(x_{n})\hat{\phi}(x_{n},0).

For the integral in terms of yy over (0,)(0,\infty), specific quadrature rule is needed due to the singularity at y=0y=0. It is shown in [4, 6] that the infinite integral involving fractional operators can be effectively computed by sinc quadrature. More precisely, we use the change of variable y=eμy=e^{\mu} so that

(13) 0yαg(y)dy=e(α+1)μg(eμ)dμ\int_{0}^{\infty}y^{\alpha}g(y){\rm d}y=\int_{-\infty}^{\infty}e^{(\alpha+1)\mu}g(e^{\mu}){\rm d}\mu

for all gLyα1(0,)g\in L^{1}_{y\alpha}(0,\infty). Given h¯>0\bar{h}>0, let

(14) N+:=π24sh¯2,N:=π24(1s)h¯2,μm:=mh¯,N_{+}:=\lceil\frac{\pi^{2}}{4s\bar{h}^{2}}\rceil,\quad N_{-}:=\lceil\frac{\pi^{2}}{4(1-s)\bar{h}^{2}}\rceil,\quad\mu_{m}:=m\bar{h},

then (13) is evaluated by quadrature nodes {μm}m=N,,N+\{\mu_{m}\}_{m=-N_{-},\cdots,N_{+}} and uniform weights h¯\bar{h}, namely,

(15) e(α+1)μg(eμ)dμh¯m=NN+e(α+1)μmg(eμm).\int_{-\infty}^{\infty}e^{(\alpha+1)\mu}g(e^{\mu}){\rm d}\mu\approx\bar{h}\sum_{m=-N_{-}}^{N_{+}}e^{(\alpha+1)\mu_{m}}g(e^{\mu_{m}}).

Combining (12) and (15) we have an approximate functional of \mathcal{I} given by

(16) 𝒯,h¯[ϕ^(x,y)]=|Ω|N(h¯m=NN+e(α+1)μmn=1N|ϕ^(xn,eμm)|2dsn=1Nf(xn)ϕ^(xn,0)).\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{\phi}(x,y)]=\frac{|\Omega|}{N}\left(\bar{h}\sum_{m=-N_{-}}^{N_{+}}e^{(\alpha+1)\mu_{m}}\sum_{n=1}^{N}|\nabla\hat{\phi}(x_{n},e^{\mu_{m}})|^{2}-d_{s}\sum_{n=1}^{N}f(x_{n})\hat{\phi}(x_{n},0)\right).

Practically, we solve the following optimization,

(17) minϕ^𝒩L,M,σ,h𝒯,h¯[ϕ^],\underset{\hat{\phi}\in\mathcal{N}_{L,M,\sigma,h}}{\min}\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{\phi}],

whose solution can be regarded as a practical numerical solution of (22).

5 Numerical Experiments

5.1 The setting

The proposed method is tested by numerical experiments in this section. Deep learning techniques are utilized to solve the minimization (17). The overall setting is summarized as follows.

  • Environment. The experiment is performed in Python environment. PyTorch library with CUDA toolkit is utilized for neural network implementation and GPU-based parallel computing. The codes can be simply implemented on a desktop.

  • Optimizer and hyper-parameters. The network-based optimization (17) is solved by the stochastic gradient descent (SGD) from PyTorch library. The SGD is implemented for totally 5000 epochs, with adaptively decaying learning rates.

  • Network setting. The special network structure (5) is adopted. We choose σ\sigma as the ReLU function. The parameters in θ\theta^{\prime} and θ′′\theta^{\prime\prime} are initialized by

    (1) a,Wl,blU(M,M),l=1,,L.a,W_{l},b_{l}\sim U(-\sqrt{M},\sqrt{M}),\quad l=1,\cdots,L.

    And γ,γ′′\gamma^{\prime},\gamma^{\prime\prime} are initialized as 0.5. All these parameters are trained in the learning process.

  • Numerical quadrature. For quadrature (12) over Ω\Omega, we adopt the quasi-Monte Carlo with Halton sequence. For the cases of d=3d=3 and 1010, totally N=105N=10^{5} and 5×1055\times 10^{5} sample points are used in the Monte Carlo quadrature, separated as 4 and 10 batches in the SGD, respectively. The sinc quadrature parameter h¯=\bar{h}= is set as 1/31/3.

  • Testing set and error evaluation. We generate a testing set 𝒳\mathcal{X} consisting of 10410^{4} random points uniformly distributed in Ω\Omega for error evaluation. Suppose ϕ^(x,y)\hat{\phi}(x,y) is the neural network obtained by our method and U(x)U(x) is the true solution of the original fractional Laplacian problem (1), then the following relative 2\ell^{2} error will be computed,

    (2) e2(𝒳):=(x𝒳|ϕ^(x,0)U(x)|2/x𝒳|U(x)|2)12.e_{\ell^{2}}(\mathcal{X}):=\left(\sum_{x\in\mathcal{X}}|\hat{\phi}(x,0)-U(x)|^{2}/\sum_{x\in\mathcal{X}}|U(x)|^{2}\right)^{\frac{1}{2}}.

5.2 A model problem

In this example, we solve the following problem with an explicit solution to test the accuracy of the proposed method,

(3) {(Δ)su(x)=(dπ2)si=1dsin(πxi),x:=[x1,,xd]Ω:=(1,1)d,u(x)=0,xΩ,\begin{cases}(-\Delta)^{s}u(x)=(d\pi^{2})^{s}\prod_{i=1}^{d}\sin(\pi x_{i}),\quad\forall x:=[x_{1},\cdots,x_{d}]\in\Omega:=(-1,1)^{d},\\ u(x)=0,\quad\forall x\in\partial\Omega,\end{cases}

whose true solution is given by u(x)=i=1dsin(πxi)u(x)=\prod_{i=1}^{d}\sin(\pi x_{i}). We specify the boundary function h(x)=i=1d(1xi2)h(x)=\prod_{i=1}^{d}(1-x_{i}^{2}) in the network architecture (5).

5.2.1 Network size test

First, we solve the problem (3) with s=0.5s=0.5 using special network (5) of various network sizes. The proposed method is implemented with network depth L=2,3L=2,3 and width M=25,50,100,200M=25,50,100,200. The errors of the numerical solutions are listed in Table 1 as well as the total running time. We also computed the numerical error orders with respect to MM. Note that most of the running time is occupied by training the networks, and very little is for the testing process (computing the errors).

From the table, it is observed that using larger sizes improves the accuracy at the expense of extra running time. For a fixed width MM, the obtained error is reduced by more than half from L=2L=2 to L=3L=3. For a fixed depth LL, the numerical order with respect to MM is roughly around the theoretical order 1/2-1/2 proved in Theorem 5, and the deviation is probably due to the stochasticity and capability of the learning algorithm. Overall, the size pair (L,M)=(3,200)(L,M)=(3,200) obtains the most accurate solution; hence we continue using it in the following tests.

L=2L=2 L=3L=3
MM Error Order Running Time Error Order Running Time
2525 8.252×1028.252\times 10^{-2} N.A. 7.8×1037.8\times 10^{3} 2.410×1022.410\times 10^{-2} N.A. 8.0×1038.0\times 10^{3}
5050 4.337×1024.337\times 10^{-2} 0.93-0.93 7.7×1037.7\times 10^{3} 2.071×1022.071\times 10^{-2} 0.22-0.22 8.0×1038.0\times 10^{3}
100100 3.522×1023.522\times 10^{-2} 0.3-0.3 7.9×1037.9\times 10^{3} 1.371×1021.371\times 10^{-2} 0.59-0.59 8.6×1038.6\times 10^{3}
200200 2.529×1022.529\times 10^{-2} 0.48-0.48 8.6×1038.6\times 10^{3} 8.424×1038.424\times 10^{-3} 0.7-0.7 1.3×1041.3\times 10^{4}
Table 1: Errors e2(𝒳)e_{\ell^{2}}(\mathcal{X}), error orders with respect to MM and running time (seconds) for various LL and MM when d=3d=3.

5.2.2 Comparison of structures

Next, this problem is solved for d=3d=3 with s=0.1s=0.1, 0.30.3, 0.50.5, 0.70.7, 0.90.9. We test both the proposed special structure (5) and the following simple FNN structure for a comparison,

(4) ϕ^(x,y)=ϕ^(x,y)h(x)e12y,\hat{\phi}(x,y)=\hat{\phi}^{\prime}(x,y)h(x)e^{-\frac{1}{2}y},

where ϕ^\hat{\phi}^{\prime} is a generic FNN. Note in (4), the power of the exponent is fixed with 1/2-1/2 instead of a trainable parameter, and no singular term is introduced. By this design, the comparison can reveal the advantages of using trainable exponential decaying powers and the special singular term. For both structures, we set L=3L=3 and M=200M=200 for the involved FNN.

The error curves versus epochs of the SGD are shown in Figure 1, and the errors of the finally obtained solutions are listed in Table 2. It is observed that for each structure, the smallest error is obtained when s=0.5s=0.5, while larger errors are obtained when ss is close to 0 or 1. This is natural since the true solution has no singularity if s=0.5s=0.5 and has higher singularity if ss approaches to 0 or 1. Moreover, from the comparison, it is clear that the special structure outperforms the simple one in obtaining smaller errors. We also remark that the time cost of the simple structure is only slightly less than the special structure.

5.2.3 High-dimensional simulation

Finally, we conduct a high-dimensional test by solving the model problem (3) of d=10d=10 and s=0.2,0.3,0.4,0.5,0.6,0.7,0.8s=0.2,0.3,0.4,0.5,0.6,0.7,0.8. We continue using the special structure of L=3L=3 and M=200M=200 for approximation. The error curves and final errors are shown in Figure 2 and Table 3. We observe that the proposed method is still effective for high dimensions, obtaining the errors O(102)O(10^{-2}). While the smallest error is obtained between s=0.4s=0.4 and 0.50.5.

To the best of our knowledge, our method is the first successful attempt at 10-D spectral fractional Laplacian equation. The existing methods [24, 11] using finite elements or Laguerre functions show their effectiveness in low-dimensional problems (d=1,2d=1,2), while they are incapable of high-dimensional cases due to the limitation of storage. Therefore, we do not conduct a numerical comparison in this work.

Refer to caption
(a) s=0.1s=0.1
Refer to caption
(b) s=0.3s=0.3
Refer to caption
(c) s=0.5s=0.5
Refer to caption
(d) s=0.7s=0.7
Refer to caption
(e) s=0.9s=0.9
Fig. 1: Errors e2(𝒳)e_{\ell^{2}}(\mathcal{X}) versus epochs for various ss using the special or simple structure when d=3d=3.
Special structure in (5) Simple structure in (4)
ss Error Running Time Error Running Time
0.10.1 4.619×1024.619\times 10^{-2} 8.7×1038.7\times 10^{3} 5.402×1015.402\times 10^{-1} 8.5×1038.5\times 10^{3}
0.30.3 7.576×1037.576\times 10^{-3} 9.9×1039.9\times 10^{3} 1.800×1021.800\times 10^{-2} 9.6×1039.6\times 10^{3}
0.50.5 8.424×1038.424\times 10^{-3} 1.3×1041.3\times 10^{4} 1.257×1021.257\times 10^{-2} 1.1×1041.1\times 10^{4}
0.70.7 8.476×1038.476\times 10^{-3} 1.8×1041.8\times 10^{4} 8.498×1038.498\times 10^{-3} 1.5×1041.5\times 10^{4}
0.90.9 1.691×1021.691\times 10^{-2} 2.5×1042.5\times 10^{4} 2.134×1022.134\times 10^{-2} 2.3×1042.3\times 10^{4}
Table 2: Errors e2(𝒳)e_{\ell^{2}}(\mathcal{X}) and running time (seconds) for various ss using the special or simple structure when d=3d=3.
Refer to caption
Fig. 2: Errors e2(𝒳)e_{\ell^{2}}(\mathcal{X}) versus epochs for various ss when d=10d=10.
ss Error Running Time
0.20.2 7.134×1027.134\times 10^{-2} 2.4×1042.4\times 10^{4}
0.30.3 6.029×1026.029\times 10^{-2} 2.5×1042.5\times 10^{4}
0.40.4 5.012×1025.012\times 10^{-2} 2.7×1042.7\times 10^{4}
0.50.5 5.158×1025.158\times 10^{-2} 2.9×1042.9\times 10^{4}
0.60.6 6.603×1026.603\times 10^{-2} 3.4×1043.4\times 10^{4}
0.70.7 6.824×1026.824\times 10^{-2} 4.1×1044.1\times 10^{4}
0.80.8 8.215×1028.215\times 10^{-2} 5.6×1045.6\times 10^{4}
Table 3: Errors e2(𝒳)e_{\ell^{2}}(\mathcal{X}) and running time (seconds) for various ss when d=10d=10.

6 Conclusion

In summary, we develop a novel deep learning-based method to solve the spectral fractional Laplacian equation numerically. First, we reformulate the fractional equation as a regular partial differential equation of one more dimension by the Caffarelli-Silvestre extension. Next, we transform the extended problem to an equivalent minimal functional problem, and characterize the space of the weak solutions. To deal with the possible high dimensions, we employ FNNs to construct a special approximate class of the solution space, by which the approximate solution has consistent properties of the true solution. We then solve the minimization in the approximate class. In theory, we studied the approximation error of the special class under Barron-type hypothesis, and thereafter derive the solution error of this method. Finally, the effectiveness of the proposed deep Ritz method is illustrated in high-dimensional simulations.

In this work, we consider the error between the minimizer of the energy functional \mathcal{I} in the approximation class and the true solution of the original problem. However, practically one needs to use numerical quadrature to compute \mathcal{I}, which brings extra errors. Consequently, future work may include the error analysis for the minimizer of the empirical loss function 𝒯,h¯\mathcal{I}_{\mathcal{T},\bar{h}} (16) discretized from \mathcal{I} by the Monte Carlo method and sinc quadrature. More precisely, let uu^{*} be the true solution of (4) and

(1) u^=argminw𝒩[w],u^𝒯,h¯=argminw𝒩𝒯,h¯[w]\hat{u}=\text{argmin}_{w\in\mathcal{N}}\mathcal{I}[w],\quad\hat{u}_{\mathcal{T},\bar{h}}=\text{argmin}_{w\in\mathcal{N}}\mathcal{I}_{\mathcal{T},\bar{h}}[w]

for some DNN-based hypothesis space 𝒩\mathcal{N}, then

(2) uu^𝒯,h¯uu^+u^u^𝒯,h¯.\|u^{*}-\hat{u}_{\mathcal{T},\bar{h}}\|\leq\|u^{*}-\hat{u}\|+\|\hat{u}-\hat{u}_{\mathcal{T},\bar{h}}\|.

Note that uu^\|u^{*}-\hat{u}\| has been investigated, it suffices to consider u^u^𝒯,h¯\|\hat{u}-\hat{u}_{\mathcal{T},\bar{h}}\|. Suppose \mathcal{I} has a bounded inverse, then

(3) u^u^𝒯,h¯1|[u^][u^𝒯,h¯]|1([u^𝒯,h¯][u^])1([u^𝒯,h¯][u^]+𝒯,h¯[u^𝒯,h¯]𝒯,h¯[u^𝒯,h¯])1[([u^𝒯,h¯]𝒯,h¯[u^𝒯,h¯])+(𝒯,h¯[u^][u^])],\|\hat{u}-\hat{u}_{\mathcal{T},\bar{h}}\|\leq\|\mathcal{I}^{-1}\|\left|\mathcal{I}[\hat{u}]-\mathcal{I}[\hat{u}_{\mathcal{T},\bar{h}}]\right|\leq\|\mathcal{I}^{-1}\|\left(\mathcal{I}[\hat{u}_{\mathcal{T},\bar{h}}]-\mathcal{I}[\hat{u}]\right)\\ \leq\|\mathcal{I}^{-1}\|\left(\mathcal{I}[\hat{u}_{\mathcal{T},\bar{h}}]-\mathcal{I}[\hat{u}]+\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{u}_{\mathcal{T},\bar{h}}]-\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{u}_{\mathcal{T},\bar{h}}]\right)\\ \leq\|\mathcal{I}^{-1}\|\left[\left(\mathcal{I}[\hat{u}_{\mathcal{T},\bar{h}}]-\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{u}_{\mathcal{T},\bar{h}}]\right)+\left(\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{u}]-\mathcal{I}[\hat{u}]\right)\right],

where we use the facts [u^][u^𝒯,h¯]\mathcal{I}[\hat{u}]\leq\mathcal{I}[\hat{u}_{\mathcal{T},\bar{h}}] and 𝒯,h¯[u^𝒯,h¯]𝒯,h¯[u^]\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{u}_{\mathcal{T},\bar{h}}]\leq\mathcal{I}_{\mathcal{T},\bar{h}}[\hat{u}]. On the right hand side of (3), the first term describes the generalization error of the empirical loss minimization over the hypothesis space, and the second term characterizes the bias coming from the numerical quadrature of the integrals. In [19], related analysis is conducted for Poisson equation and Schrödinger equation using Rademacher complexity, but it is only for the Monte-Carlo quadrature. It is promising and challenging to do similar analysis for our method on fractional Laplacian equations, which involves both the stochastic Monte-Carlo method and the deterministic sinc quadrature for approximation.

References

  • [1] G. Acosta and J. P. Borthagaray. A fractional Laplace equation: regularity of solutions and finite element approximations. SIAM J Numer Anal, 55:472–495, 2017.
  • [2] A. R. Barron. Neural net approximation. In Proceedings of the 7th Yale Workshop on Adaptive and Learning Systems, 1992.
  • [3] A. R. Barron. Universal approximation bounds for superpositions of a sigmoidal function. IEEE Trans. Inform. Theory, 39(3):930–945, 1993.
  • [4] A. Bonito, J. P. Borthagaray, R. H. Nochetto, E. Otarola, and A. J. Salgado. Numerical methods for fractional diffusion. arXiv e-prints, arXiv:1707.01566, 2017.
  • [5] A. Bonito, W. Lei, and J. E. Pasciak. Numerical approximation of the integral fractional Laplacian. Numer Math, 142:235–278, 2019.
  • [6] A. Bonito, W. Lei, and J. E. Pasciak. On sinc quadrature approximations of fractional powers of regularly accretive operators. J Numer Math, 27(2):57–68, 2019.
  • [7] L. Caffarelli and L. Silvestre. An extension problem related to the fractional Laplacian. Comm Partial Differential Equations, 32(8):1245–1260, 2007.
  • [8] A. Capella, J. Dávila, L. Dupaigne, and Y. Sire. Regularity of radial extremal solutions for some non-local semilinear equations. Comm Partial Differential Equations, 36(8):1353–1384, 2011.
  • [9] A. Caragea, P. Petersen, and F. Voigtlaender. Neural network approximation and estimation of classifiers with classification boundary in a Barron class. arXiv e-prints, arXiv:2011.09363, 2020.
  • [10] B. A. Carreras, V. E. Lynch, and G. M. Zaslavsky. Anomalous diffusion and exit time distribution of particle tracers in plasma turbulence model. Phys Plasmas, 8:5096–5103, 2001.
  • [11] S. Chen and J. Shen. An efficient and accurate numerical method for the spectral fractional Laplacian equation. J Sci Comput, 82(17), 2020.
  • [12] R. Cont and E. Voltchkova. A finite difference scheme for option pricing in jump diffusion and exponential lévy models. SIAM J Numer Anal, 43:1596–1626, 2005.
  • [13] W. E, C. Ma, S. Wojtowytsch, and L. Wu. Towards a mathematical understanding of neural network-based machine learning: what we know and what we don’t. arXiv e-prints, arXiv:2009.10713, 2020.
  • [14] W. E, C. Ma, and L. Wu. The Barron space and the flow-induced function spaces for neural network models. arXiv e-prints, arXiv:1906.08039, 2019.
  • [15] W. E and S. Wojtowytsch. Representation formulas and pointwise properties for Barron functions. arXiv e-prints, arXiv:2006.05982, 2020.
  • [16] B. Jourdain, S. Méléard, and W. A. Woyczynski. A probabilistic approach for nonlinear equations involving the fractional Laplacian and a singular operator. Potential Anal, 23(1):55–81, 2005.
  • [17] J. M. Klusowski and A. R. Barron. Approximation by combinations of ReLU and squared ReLU ridge functions with 1\ell^{1} and 0\ell^{0} controls. IEEE Trans. Inform. Theory, 64(12):7649–7656, 2018.
  • [18] N. Laskin. Fractional quantum mechanics and lévy path integrals. Phys Lett A, 268:298–305, 2000.
  • [19] J. Lu, Y. Lu, and M. Wang. A priori generalization analysis of the deep Ritz method for solving high dimensional elliptic equations. arXiv e-prints, arXiv:2101.01708, 2021.
  • [20] J. Lu, Z. Shen, H. Yang, and S. Zhang. Deep network approximation for smooth functions. arXiv e-prints, arXiv:2001.03040, 2020.
  • [21] J. Lund and K. L. Bowers. Sinc methods for quadrature and differential equations. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 1992.
  • [22] K. S. Miller and S. G. Samko. Completely monotonic functions. Integr Transf Spec F, 12(4):389–402, 2001.
  • [23] Johannes Müller and Marius Zeinhofer. Error estimates for the variational training of neural networks with boundary penalty. arXiv e-prints, arXiv:2103.01007, 2021.
  • [24] R. H. Nochetto, E. Otárola, and A. J. Salgado. A PDE approach to fractional diffusion in general domains: a priori error analysis. Found Comput Math, 15:733–791, 2015.
  • [25] Z. Shen, H. Yang, and S. Zhang. Deep network with approximation error being reciprocal of width to power of square root of depth. Neural Comput, 33(4):1005–1036, 2021.
  • [26] M. F. Shlesinger, B. J. West, and J. Klafter. Lévy dynamics of enhanced diffusion: application to turbulence. Phys Rev Lett, 58:1100–1103, 1987.
  • [27] J. W. Siegel and J. Xu. Approximation rates for neural networks with general activation functions. Neural Networks, 128:313–321, 2020.
  • [28] J. W. Siegel and J. Xu. High-order approximation rates for neural networks with ReLUk activation functions. arXiv e-prints, arXiv:2012.07205, 2020.
  • [29] P. Tankov. Financial modelling with jump processes, volume 2. CRC press, 2003.
  • [30] V. V. Uchaikin. Nonlocal models of cosmic ray transport in the galaxy. J Appl Math Phys, 3:187–200, 2015.
  • [31] J. L. Vázquez. Recent progress in the theory of nonlinear diffusion with fractional Laplacian operators. Discrete Contin Dyn, 7:857–885, 2014.