This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Carleman-based numerical method for quasilinear elliptic equations with over-determined boundary data and applications

Thuy T. Le Department of Mathematics and Statistics, University of North Carolina at Charlotte, Charlotte, NC, 28223, USA, [email protected].    Loc H. Nguyen Department of Mathematics and Statistics, University of North Carolina at Charlotte, Charlotte, NC, 28223, USA, [email protected].    Hung V. Tran Department of Mathematics, University of Wisconsin Madison, Madison, WI, 53706, USA, [email protected] (corresponding author).
Abstract

We propose a new iterative scheme to compute the numerical solution to an over-determined boundary value problem for a general quasilinear elliptic PDE. The main idea is to repeatedly solve its linearization by using the quasi-reversibility method with a suitable Carleman weight function. The presence of the Carleman weight function allows us to employ a Carleman estimate to prove the convergence of the sequence generated by the iterative scheme above to the desired solution. The convergence of the iteration is fast at an exponential rate without the need of an initial good guess. We apply this method to compute solutions to some general quasilinear elliptic equations and a large class of first-order Hamilton-Jacobi equations. Numerical results are presented.

Key words: numerical methods; Carleman estimate; linearization; boundary value problems; quasilinear elliptic equations; Hamilton-Jacobi equations; viscosity solutions; vanishing viscosity process.

AMS subject classification: 35D40, 35F30, 35J62, 35N25, 65N12, 78A46.

1 Introduction

The main aim of this paper is to develop a numerical method based on Carleman estimates to solve quasilinear elliptic PDEs with over-determined boundary data. We consider this new method as the second generation of Carleman-based numerical methods while the first generation is called the convexification, which will be mentioned in detail later. Let Ω\Omega be an open and bounded domain in d\mathbb{R}^{d}, d2d\geq 2, with smooth boundary Ω\partial\Omega. Let ff and gg be two smooth functions defined on Ω\partial\Omega. Let F:Ω¯××dF:\overline{\Omega}\times\mathbb{R}\times\mathbb{R}^{d}\to\mathbb{R} be a function in the class C2.C^{2}. Let A=(aij)i,j=1d:Ω¯d×dA=(a_{ij})_{i,j=1}^{d}:\overline{\Omega}\to\mathbb{R}^{d\times d} be C2C^{2}, symmetric, and positive definite, that is,

γ|ξ|2aij(𝐱)ξiξjγ1|ξ|2 for all 𝐱Ω¯,ξd,\gamma|\xi|^{2}\leq a_{ij}({\bf x})\xi_{i}\xi_{j}\leq\gamma^{-1}|\xi|^{2}\quad\mbox{ for all }{\bf x}\in\overline{\Omega},\ \xi\in\mathbb{R}^{d},

for some fixed γ(0,1)\gamma\in(0,1). The following problem is of our interests.

Problem 1.1.

Assume that the over-determined boundary value problem

{div(A(𝐱)u(𝐱))+F(𝐱,u(𝐱),u(𝐱))=0𝐱Ω,u(𝐱)=f(𝐱)𝐱Ω,νu(𝐱)=g(𝐱)𝐱Ω\left\{\begin{array}[]{ll}-{\rm div}(A({\bf x})\nabla u({\bf x}))+F({\bf x},u({\bf x}),\nabla u({\bf x}))=0&{\bf x}\in\Omega,\\ u({\bf x})=f({\bf x})&{\bf x}\in\partial\Omega,\\ \partial_{\nu}u({\bf x})=g({\bf x})&{\bf x}\in\partial\Omega\end{array}\right. (1.1)

has a solution uu^{*} in C2(Ω¯)C^{2}(\overline{\Omega}). Compute the function uu^{*}.

Problem 1.1 is motivated by a class of nonlinear inverse problems in PDEs, in which ff and gg are the data that can be measured. One important goal of inverse problems is to reconstruct the internal structure of a domain from boundary measurements, which allow us to impose both Dirichlet and Neumann data of the unknown in (1.1). Recently, a unified framework to solve such inverse problems was developed by the research group of the first and second authors, which has two main steps. In the first step, by introducing a change of variables, one derives a PDE of the form (1.1) from the given inverse problem, in which ff and gg can be computed directly by the given boundary data. In the second step, one numerically solves (1.1) to find uu^{*}. The knowledge of uu^{*} directly yields that of the solution to the corresponding inverse problem under consideration. See [15, 30] and the references therein for some works in this framework. Moreover, this unified framework was successfully tested with experimental data in [13, 14, 22]. Another motivation to study Problem 1.1 is to seek solutions to Hamilton-Jacobi equations under the circumstance that the Neumann data of the unknown can be computed by its Dirichlet data and the given form of the Hamiltonian, see e.g., [25, Assumption 1.1 and Remark 1.1]. Since inverse problems are out of the scope of this paper, we only focus on the applications in solving quasilinear elliptic PDEs and first-order Hamilton-Jacobi equations.

A natural approach to solve (1.1) is based on optimization. That means one sets the computed solution to (1.1) as a minimizer of a mismatch functional, e.g.,

vJ(v):=Ω|div(A(𝐱)v(𝐱))+F(𝐱,v(𝐱),v(𝐱))|2𝑑𝐱+a regularization termv\mapsto J(v):=\int_{\Omega}\big{|}-{\rm div}(A({\bf x})\nabla v({\bf x}))+F({\bf x},v({\bf x}),\nabla v({\bf x}))\big{|}^{2}\,d{\bf x}+\mbox{a regularization term}

subject to the Cauchy boundary conditions v|Ω=fv|_{\partial\Omega}=f and νv|Ω=g.\partial_{\nu}v|_{\partial\Omega}=g. The methods based on optimization are widely used in the scientific community, especially in computational mathematics, physics and engineering. Although effective and popular, the optimization-based approaches have some drawbacks:

  1. 1.

    In general, it is not clear that the obtained minimizer approximates the true solution to (1.1).

  2. 2.

    The mismatch functional JJ is not convex, and it might have multiple minima and ravines (see an example in [43] for illustration). To deliver reliable numerical solutions, one must know some good initial guesses of the true solutions.

  3. 3.

    The computation is expensive and time consuming.

Drawbacks # 1 and #2 can be treated by the convexification method, which is designed to globalize the optimization methods. The main idea of the convexification method is to employ some suitable Carleman weight functions to convexify the mismatch functionals. The convexity of weighted mismatch functionals is rigorously proved by Carleman estimates. Several versions of the convexification method have been developed in [1, 15, 16, 17, 19, 22] since it was first introduced in [21]. Moreover, we recently discovered that the convexification method can be used to solve a large class of first-order Hamilton-Jacobi equations [25].

In this paper, we introduce a new method to solve (1.1) based on linearization and Carleman estimates. Like the convexification method, our method delivers a reliable solution to (1.1) without requiring a good initial guess. This fact is rigorously proved. Unlike the convexification method which is time consuming, our new method quickly provides the desired solutions. Its converge rate is O(θn)O(\theta^{n}) as nn\to\infty for some θ(0,1).\theta\in(0,1).

We find the numerical solution to (1.1) by repeatedly solving the linearization of (1.1) by a new “Carleman weighted” quasi-reversibility method. The classical quasi-reversibility method was first proposed in [28], and it has been studied intensively since then (see [18] for a survey). By a Carleman weighted quasi-reversibility method, we mean that we let a suitable Carleman weight function involve in the cost functional suggested by the classical quasi-reversibility method. The presence of the Carleman weight function is the key for us to prove our convergence theorem. Our process to solve Problem 1.1 is as follows. We first choose any initial solution that might be far away from the true one. Denote this initial solution by the function u0u_{0}. Linearizing (1.1) about u0u_{0}, we obtain a linear PDE. We then solve this linear PDE by the Carleman weighted quasi-reversibility method to obtain an updated solution u1u_{1}. Using the Carleman weighted quasi-reversibility method rather than the classical one in this step is the key to our success. By iteration, we repeat this step to construct a sequence {un}n0\{u_{n}\}_{n\geq 0}. The convergence of this sequence to the true solution to (1.1) is proved by using Carleman estimates. We then apply this method to numerically solve some quasilinear elliptic equations. It is important to note that our approach works well for systems of quasilinear elliptic PDEs too.

Remark 1.1.

In general, (1.1) is over-determined and it might have no solution, especially when the boundary data contains some noise. Our iteration and linearization method using the Carleman weighted quasi-reversibility method in each step still delivers a function that “most fits” (1.1).

On the other hand, (1.1) with only Dirichlet boundary condition, that is, the equation

{div(A(𝐱)u(𝐱))+F(𝐱,u(𝐱),u(𝐱))=0𝐱Ω,u(𝐱)=f(𝐱)𝐱Ω\left\{\begin{array}[]{ll}-{\rm div}(A({\bf x})\nabla u({\bf x}))+F({\bf x},u({\bf x}),\nabla u({\bf x}))=0&{\bf x}\in\Omega,\\ u({\bf x})=f({\bf x})&{\bf x}\in\partial\Omega\end{array}\right. (1.2)

might have many solutions since we do not impose structural conditions on FF. For example, in case F(𝐱,u(𝐱),u(𝐱))=λuF({\bf x},u({\bf x}),\nabla u({\bf x}))=\lambda u for some λ\lambda\in\mathbb{R} and f=0f=0, (1.2) becomes an eigenvalue problem with possibly many solutions as eigenfunctions. In such cases, requiring the additional Neumann boundary condition is then natural, and (1.1) is not over-determined.

Next, we use our method to solve some first-order Hamilton-Jacobi equations. More precisely, to find viscosity solutions to the first-order equation

F(𝐱,u(𝐱),u(𝐱))=0𝐱Ω,F({\bf x},u({\bf x}),\nabla u({\bf x}))=0\quad{\bf x}\in\Omega,

we use the vanishing viscosity procedure and consider, for 0<ϵ010<\epsilon_{0}\ll 1,

ϵ0Δu(𝐱)+F(𝐱,u(𝐱),u(𝐱))=0𝐱Ω,-\epsilon_{0}\Delta u({\bf x})+F({\bf x},u({\bf x}),\nabla u({\bf x}))=0\quad{\bf x}\in\Omega,\\

with given Cauchy boundary data. Our new method is robust in the sense that it works for general nonlinearity F(𝐱,u,u)F({\bf x},u,\nabla u) that might not be convex in u.\nabla u. We refer the readers to [9, 8, 47] and the references therein for the theory of viscosity solutions. A weakness of our new approach in computing the viscosity solutions to Hamilton-Jacobi equations is that we need to require both Dirichlet and Neumann data of the unknown uu. See [25, Remark 1.1] for some circumstances that this requirement is fulfilled. There have been many important methods to solve Hamilton-Jacobi equations in the literature. For finite difference monotone and consistent schemes of first-order equations and applications, see [2, 10, 40, 44, 45] for details and recent developments. If F=F(𝐱,u,u)F=F({\bf x},u,\nabla u) is convex in u\nabla u and satisfies appropriate conditions, it is possible to construct some semi-Lagrangian approximations by the discretization of the Dynamical Programming Principle associated to the problem, see [11, 12] and the references therein.

The paper is organized as follows. In Section 2, we recall a Carleman estimate and two examples about inverse problems in which Problem 1.1 appears. In Section 3, we introduce the new iterative method based on linearization and Carleman estimates. In Section 4, we prove the convergence of our method. Some numerical results for quasilinear elliptic equations and first-order Hamilton-Jacobi equations are presented in Section 5. Concluding remarks are given in Section 6.

2 Preliminaries

In this section, we recall a Carleman estimate, which plays a key role for the proof of the convergence theorem in this paper. We then present an inverse scattering problem in which Problem 1.1 appears.

2.1 A Carleman estimate

In this section, we present a simple form of Carleman estimates. Carleman estimates were first employed to prove the unique continuation principle, see e.g., [6, 42], and they quickly became a powerful tool in many areas of PDEs afterwards. Let 𝐱0{\bf x}_{0} be a point in dΩ¯\mathbb{R}^{d}\setminus\overline{\Omega} such that r(𝐱)=|𝐱𝐱0|>1r({\bf x})=|{\bf x}-{\bf x}_{0}|>1 for all 𝐱Ω.{\bf x}\in\Omega. For each β>0\beta>0, define

μβ(𝐱)=rβ(𝐱)=|𝐱𝐱0|βfor all 𝐱Ω¯.\mu_{\beta}({\bf x})=r^{-\beta}({\bf x})=|{\bf x}-{\bf x}_{0}|^{-\beta}\quad\mbox{for all }{\bf x}\in\overline{\Omega}. (2.1)

We have the following lemma.

Lemma 2.1 (Carleman estimate).

There exist positive constants β0,λ0\beta_{0},\lambda_{0} depending only on 𝐱0{\bf x}_{0}, Ω\Omega, γ\gamma, and dd such that for all function vC2(Ω¯)v\in C^{2}(\overline{\Omega}) satisfying

v(𝐱)=νv(𝐱)=0for all 𝐱Ω,v({\bf x})=\partial_{\nu}v({\bf x})=0\quad\mbox{for all }{\bf x}\in\partial\Omega, (2.2)

the following estimate holds true

Ωe2λμβ(𝐱)|div(Av)|2𝑑𝐱CλΩe2λμβ(𝐱)|v(𝐱)|2𝑑𝐱+Cλ3Ωe2λμβ(𝐱)|v(𝐱)|2𝑑𝐱\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|{\rm div}(A\nabla v)|^{2}\,d{\bf x}\geq C\lambda\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|\nabla v({\bf x})|^{2}\,d{\bf x}+C\lambda^{3}\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|v({\bf x})|^{2}\,d{\bf x} (2.3)

for all ββ0\beta\geq\beta_{0} and λλ0\lambda\geq\lambda_{0}. Here, C=C(𝐱0,Ω,γ,d,β)>0C=C({\bf x}_{0},\Omega,\gamma,d,\beta)>0 depends only on the listed parameters.

Proof.

Lemma 2.1 is a direct consequence of [36, Lemma 5]. Let 0<R1<10<R_{1}<1 and R31R_{3}\gg 1 such that ΩB(𝐱0,R3)B(𝐱0,R1)¯\Omega\Subset B({\bf x}_{0},R_{3})\setminus\overline{B({\bf x}_{0},R_{1})}. Here, B(𝐱0,s)={𝐲d:|𝐲𝐱0|<s}B({\bf x}_{0},s)=\{{\bf y}\in\mathbb{R}^{d}:|{\bf y}-{\bf x}_{0}|<s\} for s>0.s>0. Extend vv to the whole d\mathbb{R}^{d} such that v(𝐱)=0v({\bf x})=0 for all 𝐱dΩ{\bf x}\in\mathbb{R}^{d}\setminus\Omega. Using a change of variable 𝐱𝐱𝐱0{\bf x}\mapsto{\bf x}-{\bf x}_{0} and [36, Lemma 5], there exists a number β01\beta_{0}\geq 1 depending on γ\gamma, R1R_{1} and R3R_{3} such that for all ββ0\beta\geq\beta_{0} and |λ|2R3β|\lambda|\geq 2R_{3}^{-\beta}, we have

B(𝐱0,R3)B(𝐱0,R1)¯r(𝐱)β+2e2λr(𝐱)β|div(Av)|2𝑑𝐱CB(𝐱0,R3)B(𝐱0,R1)¯e2λrβ(𝐱)|λ|β(β3λ2r2β2(𝐱)|v|2+|v|2)𝑑𝐱C(B(𝐱0,R3)B(𝐱0,R1)¯)|λ|βe2λrβ(𝐱)(r(𝐱)|v(𝐱)|2+β2λ2r2β1(𝐱)|v(𝐱)|2)dσ(𝐱)\int_{B({\bf x}_{0},R_{3})\setminus\overline{B({\bf x}_{0},R_{1})}}r({\bf x})^{\beta+2}e^{2\lambda r({\bf x})^{-\beta}}|{\rm div}(A\nabla v)|^{2}\,d{\bf x}\\ \geq C\int_{B({\bf x}_{0},R_{3})\setminus\overline{B({\bf x}_{0},R_{1})}}e^{2\lambda r^{-\beta}({\bf x})}|\lambda|\beta\big{(}\beta^{3}\lambda^{2}r^{-2\beta-2}({\bf x})|v|^{2}+|\nabla v|^{2}\big{)}\,d{\bf x}\\ -C\int_{\partial(B({\bf x}_{0},R_{3})\setminus\overline{B({\bf x}_{0},R_{1})})}|\lambda|\beta e^{2\lambda r^{-\beta}({\bf x})}(r({\bf x})|\nabla v({\bf x})|^{2}\\ +\beta^{2}\lambda^{2}r^{-2\beta-1}({\bf x})|v({\bf x})|^{2})\,d\sigma({\bf x}) (2.4)

for some constant CC depending only on dd and γ\gamma. Since v=0v=0 on (B(𝐱0,R3)B(𝐱0,R1)¯)Ω\big{(}B({\bf x}_{0},R_{3})\setminus\overline{B({\bf x}_{0},R_{1})}\big{)}\setminus\Omega and since ΩB(𝐱0,R3)B(𝐱0,R1)¯\Omega\Subset B({\bf x}_{0},R_{3})\setminus\overline{B({\bf x}_{0},R_{1})}, allowing CC to depend on β\beta and Ω\Omega, we deduce the Carleman estimate (2.3) from (2.4). ∎

An alternative way to obtain (2.3) is to apply the Carleman estimate in [29, Chapter 4, §1, Lemma 3] for general parabolic operators. The arguments to obtain (2.3) using [29, Chapter 4, §1, Lemma 3] are similar to that in [32, Section 3] with the Laplacian replaced by the operator div(A){\rm div}(A\nabla\cdot).

Remark 2.1.

We specially draw the reader’s attention to different forms of Carleman estimates for all three main kinds of differential operators (elliptic, parabolic and hyperbolic) and their applications in inverse problems and computational mathematics [3, 4, 23]. It is worth mentioning that some Carleman estimates hold true for all functions vv satisfying v|Ω=0v|_{\partial\Omega}=0 and νv|Γ=0\partial_{\nu}v|_{\Gamma}=0 where Γ\Gamma is a part of Ω\partial\Omega, see e.g., [25, 39], which can be used to solve quasilinear elliptic PDEs with the boundary data partly given.

2.2 An inverse scattering problem

As mentioned in Section 1, Problem 1.1 arises from nonlinear inverse problems. We present here an important example in the context of inverse scattering problems in the frequency domain. Let c:d[1,)c:\mathbb{R}^{d}\to[1,\infty) be the spatially distributed dielectric constant of the medium and [k¯,k¯][\underline{k},\overline{k}] be an interval of wavenumbers with k¯>0\underline{k}>0. Since the dielectric constant of the air is 1, we set c(𝐱)=1c({\bf x})=1 for all 𝐱dΩ.{\bf x}\in\mathbb{R}^{d}\setminus\Omega. For each k[k¯,k¯],k\in[\underline{k},\overline{k}], let w(𝐱,k)w({\bf x},k), 𝐱d{\bf x}\in\mathbb{R}^{d}, be the wave field generated by a point source at 𝐱0dΩ{\bf x}_{0}\in\mathbb{R}^{d}\setminus\Omega with wavenumber kk. The function ww is governed by the Helmholtz equation and the Sommerfeld radiation condition

{Δw(𝐱,k)+k2c(𝐱)w(𝐱,k)=δ(𝐱𝐱0)𝐱d,(|𝐱|ik)w(𝐱,k)=o(|𝐱|(1d)/2)|𝐱|.\left\{\begin{array}[]{ll}\Delta w({\bf x},k)+k^{2}c({\bf x})w({\bf x},k)=-\delta({\bf x}-{\bf x}_{0})&{\bf x}\in\mathbb{R}^{d},\\ (\frac{\partial}{\partial{|{\bf x}|}}-{\rm i}k)w({\bf x},k)=o(|{\bf x}|^{(1-d)/2})&|{\bf x}|\to\infty.\end{array}\right. (2.5)

The inverse scattering problem is formulated as follows.

Problem 2.1 (Inverse Scattering Problem).

Compute the function c(𝐱)c({\bf x}), 𝐱Ω{\bf x}\in\Omega, from the measurements of

f1(𝐱,k)=w(𝐱,k)andf2(𝐱,k)=νw(𝐱,k)f_{1}({\bf x},k)=w({\bf x},k)\quad\mbox{and}\quad f_{2}({\bf x},k)=\partial_{\nu}w({\bf x},k) (2.6)

for all 𝐱Ω{\bf x}\in\partial\Omega, k[k¯,k¯].k\in[\underline{k},\overline{k}].

The knowledge of the function cc partly provides the internal structure of the domain Ω\Omega. In other words, solving the inverse scattering problem allows us to examine a domain from external measurements, which has applications in security, sonar imaging, geographical exploration, medical imaging, near-field optical microscopy, nano-optics, see, e.g., [7] and references therein for more details. There have been many important methods to solve inverse scattering problems in the literature. Each method has its own advantages and disadvantages. A common drawback of the widely-used method based on optimization to solve inverse scattering problems is the need of a good initial guess of the true solution cc. We recall from [31] a method to solve the above inverse scattering problem in which such a need is relaxed. Denote by

z(𝐱,k)=1k2log(w(𝐱,k)w0(𝐱,k)) for all 𝐱Ω,k[k¯,k¯],z({\bf x},k)=\frac{1}{k^{2}}\log\Big{(}\frac{w({\bf x},k)}{w_{0}({\bf x},k)}\Big{)}\quad\text{ for all }{\bf x}\in\Omega,k\in[\underline{k},\overline{k}],

where w0(𝐱,k)=eik|𝐱𝐱0|4π|𝐱𝐱0|.w_{0}({\bf x},k)=\frac{e^{{\rm i}k|{\bf x}-{\bf x}_{0}|}}{4\pi|{\bf x}-{\bf x}_{0}|}. Then, zz satisfies

Δz(𝐱,k)+k2|z(𝐱,k)|2+2z(𝐱,k)w0(𝐱,k)w0(𝐱,k)=c(𝐱)+1\Delta z({\bf x},k)+k^{2}|\nabla z({\bf x},k)|^{2}+\frac{2\nabla z({\bf x},k)\cdot\nabla w_{0}({\bf x},k)}{w_{0}({\bf x},k)}=-c({\bf x})+1

for all 𝐱Ω,k[k¯,k¯].{\bf x}\in\Omega,k\in[\underline{k},\overline{k}]. Let {Ψn}n1\{\Psi_{n}\}_{n\geq 1} be the orthonormal basis of L2(k¯,k¯)L^{2}(\underline{k},\overline{k}) introduced in [20] and define

zn(𝐱)=k¯k¯z(𝐱,k)Ψn(k)𝑑k for n1,𝐱Ω.z_{n}({\bf x})=\int_{\underline{k}}^{\overline{k}}z({\bf x},k)\Psi_{n}(k)\,dk\quad\text{ for }n\geq 1,{\bf x}\in\Omega. (2.7)

We approximate

v(𝐱,k)=i=1zi(𝐱)Ψi(k)i=1Nzi(𝐱)Ψi(k),v({\bf x},k)=\sum_{i=1}^{\infty}z_{i}({\bf x})\Psi_{i}(k)\approx\sum_{i=1}^{N}z_{i}({\bf x})\Psi_{i}(k),

for a suitable cut-off number NN\in\mathbb{N}. Then, the vector ZN=(z1,z2,,zN)Z_{N}=(z_{1},z_{2},\dots,z_{N}) “approximately” satisfies the system

i=1NsliΔzi(𝐱)+i,j=1Nalijzi(𝐱)zj(𝐱)+i=1NBli(𝐱)zi(𝐱)=0\sum_{i=1}^{N}s_{li}\Delta z_{i}(\mathbf{x})+\sum_{i,j=1}^{N}a_{lij}\nabla z_{i}(\mathbf{x})\cdot\nabla z_{j}(\mathbf{x})+\sum_{i=1}^{N}B_{li}(\mathbf{x})\cdot\nabla z_{i}(\mathbf{x})=0 (2.8)

where

{sli=k¯k¯Ψi(k)Ψl(k)𝑑k,alij=2k¯k¯(k2Ψi(k)Ψj(k)+kΨi(k)Ψj(k))Ψl(k)𝑑k,Bli(𝐱)=2k¯k¯(Ψi(k)w0(𝐱,k)w0(𝐱,k)+Ψi(k)kw0(𝐱,k)w0(𝐱,k))Ψl(k)𝑑k\left\{\begin{array}[]{l}s_{li}=\displaystyle\int_{\underline{k}}^{\overline{k}}\Psi_{i}^{\prime}(k)\Psi_{l}(k)\,dk,\\ a_{lij}=\displaystyle 2\int_{\underline{k}}^{\overline{k}}\Big{(}k^{2}\Psi_{i}(k)\Psi_{j}^{\prime}(k)+k\Psi_{i}(k)\Psi_{j}(k)\Big{)}\Psi_{l}(k)\,dk,\\ B_{li}(\mathbf{x})=\displaystyle 2\int_{\underline{k}}^{\overline{k}}\Big{(}\Psi_{i}^{\prime}(k)\frac{\nabla w_{0}(\mathbf{x},k)}{w_{0}(\mathbf{x},k)}+\Psi_{i}(k)\partial_{k}\frac{\nabla w_{0}(\mathbf{x},k)}{w_{0}(\mathbf{x},k)}\Big{)}\Psi_{l}(k)\,dk\end{array}\right.

for all i,j,l{1,,N}i,j,l\in\{1,\dots,N\} and 𝐱Ω\mathbf{x}\in\Omega, see [31, Section 6] for details. The Dirichlet and Neumann boundary conditions for ZNZ_{N} can be computed by the knowledges of f1f_{1}, f2f_{2}, and (2.7). Solving the system of quasilinear elliptic equations (2.8) with the provided Dirichlet and Neumann data is basically a goal of Problem 1.1. Doing so is the key step to compute cc. See [31] for convexification method to compute ZNZ_{N} and the procedure to obtain cc from the knowledge of ZN.Z_{N}.

2.3 Electrical impedance tomography

We present here another potential application of our study in this paper to the 3D electrical impedance tomography (EIT) problem, so-called the 3D Calderón problem, with only a part of the Dirichlet to Neumann map is given. Let Ω=(R,R)3\Omega=(-R,R)^{3} for some R>0R>0. Let the line of source be defined as

Lsc={𝐱α=(α,0,R):|α|R}Ω.L_{\rm sc}=\{{\bf x}_{\alpha}=(\alpha,0,-R):|\alpha|\leq R\}\subset\partial\Omega.

For each α[R,R]\alpha\in[-R,R], let 𝐱α=(α,0,R)Lsc{\bf x}_{\alpha}=(\alpha,0,-R)\in L_{\rm sc} and w=w(𝐱,𝐱α)w=w({\bf x},{\bf x}_{\alpha}), 𝐱Ω{\bf x}\in\Omega, be the solution to

{div(a(𝐱)w(𝐱,𝐱α))=0𝐱Ω,w(𝐱,𝐱α)=f1(𝐱,𝐱α)𝐱Ω.\left\{\begin{array}[]{rcll}{\rm div}\big{(}a({\bf x})\nabla w({\bf x},{\bf x}_{\alpha})\big{)}&=&0&{\bf x}\in\Omega,\\ w({\bf x},{\bf x}_{\alpha})&=&f_{1}({\bf x},{\bf x}_{\alpha})&{\bf x}\in\partial\Omega.\end{array}\right. (2.9)

Here, f1(𝐱,𝐱α)>0f_{1}({\bf x},{\bf x}_{\alpha})>0 is a smooth approximation of the Dirac delta δ0(𝐱𝐱α)\delta_{0}({\bf x}-{\bf x}_{\alpha}). In EIT, f1(𝐱,𝐱α)f_{1}({\bf x},{\bf x}_{\alpha}) represents the boundary electric voltage. The EIT problem is formulated as follows.

Problem 2.2 (Electrical impedance tomography).

Determine the electric conductivity a(𝐱)a({\bf x}), 𝐱Ω,{\bf x}\in\Omega, from the boundary measurement of the electric current f2(𝐱,𝐱α)=νw(𝐱,𝐱α)f_{2}({\bf x},{\bf x}_{\alpha})=\partial_{\nu}w({\bf x},{\bf x}_{\alpha}) for all 𝐱Ω,{\bf x}\in\partial\Omega, 𝐱αLsc.{\bf x}_{\alpha}\in L_{\rm sc}.

Remark 2.2.

Problem 2.2 only requests the data generated by the source moving on LscL_{\rm sc}. The dimension of this set of data is 3, including 1 dimension of the orbit LscL_{\rm sc} of the source and 2 dimensions of the measurement surface Ω\partial\Omega. This feature makes Problem 2.2 not over-determined because our goal is to reconstruct a 3D function aa. This is unlike most of the works studying the EIT problem that request the whole Dirichlet to Neumann map Γ:H1/2(Ω)H1/2(Ω)\Gamma:H^{1/2}(\partial\Omega)\to H^{-1/2}(\partial\Omega). Since H1/2(Ω)H^{1/2}(\partial\Omega), the domain of Γ\Gamma, has uncountably infinity dimensions, the data requested by the Dirichlet to Neumann map is highly over-determined.

The EIT problem arises from bio-medical imaging; especially, in detecting early cancerous tumors in living tissues without operation. Most publications for the EIT problem studied the question how to reconstruct the electric conductivity a(𝐱)a({\bf x}), 𝐱Ω{\bf x}\in\Omega for some domain Ω\Omega, from the Dirichlet to Neumann or the Neumann to Dirichlet map. We refer the reader to [5, 27, 35, 46] for some well-known uniqueness results of the EIT problem. We list here a few publications for some effective approaches to numerically solve the EIT problem: the D-bar method [26, 33, 34], the methods based on optimization [41, 48] and the convexification method [24]. Although effective, these methods have drawbacks. Firstly, the D-bar method is designed only for 2D. Secondly, the methods based on optimization request good initial guess of the true solution. And finally, the convexification method is time consuming. Naturally, a new computational method should be studied in this direction. Our potential method to solve the 3D EIT problem is to reduce this inverse problem to a problem of the form (1.1). Introduce the change of variable

z(𝐱,𝐱α)=log(a(𝐱)w(𝐱,𝐱α))for all 𝐱Ω,𝐱αLsc.z({\bf x},{\bf x}_{\alpha})=\log\big{(}\sqrt{a({\bf x})}w({\bf x},{\bf x}_{\alpha})\big{)}\quad\mbox{for all }{\bf x}\in\Omega,{\bf x}_{\alpha}\in L_{\rm sc}. (2.10)

Let {Ψn}n=1\{\Psi_{n}\}_{n=1}^{\infty} be the orthonormal basis of L2(R,R)L^{2}(-R,R) first introduced in [20]. Like in Section 2.2, we approximate the function z(𝐱,𝐱α)z({\bf x},{\bf x}_{\alpha}) as

z(𝐱,𝐱α)=n=1zn(𝐱)Ψn(α)n=1Nzn(𝐱)Ψn(α)z({\bf x},{\bf x}_{\alpha})=\sum_{n=1}^{\infty}z_{n}({\bf x})\Psi_{n}(\alpha)\simeq\sum_{n=1}^{N}z_{n}({\bf x})\Psi_{n}(\alpha)\quad (2.11)

for all 𝐱Ω,𝐱αLsc{\bf x}\in\Omega,{\bf x}_{\alpha}\in L_{\rm sc}, for some cut-off number NN where

zn(𝐱)=RRz(𝐱,𝐱α)Ψn(α)𝑑αz_{n}({\bf x})=\int_{-R}^{R}z({\bf x},{\bf x}_{\alpha})\Psi_{n}(\alpha)d\alpha (2.12)

for all n1.n\geq 1. We choose NN such that the approximation in (2.11) “numerically” holds for all 𝐱Ω{\bf x}\in\partial\Omega where the data are known. Then, it is not hard to verify, see [24], that the vector ZN=(z1,,zN)Z_{N}=(z_{1},\ldots,z_{N}) satisfies the system

n=1NsmnΔzn(𝐱)+n,l=1Namnlzn(𝐱)zl(𝐱)=0for all 𝐱Ω,\sum_{n=1}^{N}s_{mn}\Delta z_{n}({\bf x})+\sum_{n,l=1}^{N}a_{mnl}\nabla z_{n}({\bf x})\cdot\nabla z_{l}({\bf x})=0\quad\mbox{for all }{\bf x}\in\Omega, (2.13)

for each m{1,,N}m\in\{1,\dots,N\}, where

smn\displaystyle s_{mn} =RRΨn(α)Ψm(α)𝑑α,\displaystyle=\int_{-R}^{R}\Psi_{n}^{\prime}(\alpha)\Psi_{m}(\alpha)d\alpha,
amnl\displaystyle a_{mnl} =2RRΨn(α)Ψl(α)Ψm(α)𝑑α.\displaystyle=2\int_{-R}^{R}\Psi_{n}(\alpha)\Psi_{l}^{\prime}(\alpha)\Psi_{m}(\alpha)d\alpha.

The Dirichlet and Neumann boundary conditions for ZNZ_{N} can be computed by the knowledges of the Dirichlet condition f1f_{1}, the given Neumann data f2f_{2}, (2.10) and (2.12). Solving the system of quasilinear elliptic equations (2.13) with the provided Dirichlet and Neumann data is basically a goal of Problem 1.1. We refer the reader to [24] the step of computing aa from the knowledge of the vector ZNZ_{N}.

The discussion above for the inverse scattering problem and the EIT problem motivates us to study Problem 1.1. Since these inverse problems are out of the scope of this paper, we only mention them here to explain the significance of Problem 1.1. In future works, we will use our solver for Problem 1.1 to solve these two important inverse problems.

3 The iteration and linearization approach for Problem 1.1

Our approach to solve (1.1) is based on linearization and iteration. Assume that the solution uu^{*} to (1.1) is in the space Hp(Ω)H^{p}(\Omega) for some p>d/2+2p>\lceil d/2\rceil+2 where d/2\lceil d/2\rceil is the smallest integer that is greater than d/2.d/2. Define the set of admissible solutions

W={φHp(Ω):φ|Ω=f,νφ|Ω=g}.W=\big{\{}\varphi\in H^{p}(\Omega)\;:\,\varphi|_{\partial\Omega}=f,\partial_{\nu}\varphi|_{\partial\Omega}=g\big{\}}. (3.1)

Then, the assumption in Problem 1.1 implies that WW\neq\emptyset, and uWu^{*}\in W. We now construct a sequence {un}n0\{u_{n}\}_{n\geq 0} that converges to the solution uu^{*} to (1.1). Take a function u0W.u_{0}\in W. Assume by induction that we have the knowledge of unu_{n} for some n0n\geq 0. We find un+1u_{n+1} as follows. Assume that u=un+hu^{*}=u_{n}+h for some hH0h\in H_{0} where

H0={φHp(Ω):φ|Ω=0,νφ|Ω=0}.H_{0}=\{\varphi\in H^{p}(\Omega)\,:\,\varphi|_{\partial\Omega}=0,\partial_{\nu}\varphi|_{\partial\Omega}=0\}.

Plugging u=un+hu^{*}=u_{n}+h into (1.1), we have

div(A(𝐱)(un+h)(𝐱))+F(𝐱,un(𝐱)+h(𝐱),un(𝐱)+h(𝐱))=0-{\rm div}(A({\bf x})\nabla(u_{n}+h)({\bf x}))+F({\bf x},u_{n}({\bf x})+h({\bf x}),\nabla u_{n}({\bf x})+\nabla h({\bf x}))=0 (3.2)

for all 𝐱Ω.{\bf x}\in\Omega. Heuristically, we assume at this moment that hh is small, that is, hC1(Ω)1\|h\|_{C^{1}(\Omega)}\ll 1.

Remark 3.1.

The temporary assumption hC1(Ω)1\|h\|_{C^{1}(\Omega)}\ll 1 is imposed for the suggestion in establishing a numerical scheme to find un+1u_{n+1} while this condition is completely relaxed in the proof of the convergence theorem.

By Taylor’s expansion, we approximate (3.2) as

div(Aun)div(Ah)+F(𝐱,un(𝐱),un(𝐱))+D(un)h=0-{\rm div}(A\nabla u_{n})-{\rm div}(A\nabla h)+F({\bf x},u_{n}({\bf x}),\nabla u_{n}({\bf x}))+D\mathcal{F}(u_{n})h=0 (3.3)

for all 𝐱Ω{\bf x}\in\Omega where

D(v)h=Fs(𝐱,v(𝐱),v(𝐱))h+𝐩F(𝐱,v(𝐱),v(𝐱))h(𝐱)D\mathcal{F}(v)h=F_{s}({\bf x},v({\bf x}),\nabla v({\bf x}))h+\nabla_{{\bf p}}F({\bf x},v({\bf x}),\nabla v({\bf x}))\cdot\nabla h({\bf x})

for all vHp(Ω)v\in H^{p}(\Omega). Here, FsF_{s} and 𝐩F\nabla_{{\bf p}}F are the partial derivative of FF with respect to its second variable and its gradient vector with respect to the third variable, respectively.

The next step is to compute a function hH0h\in H_{0} satisfying (3.3). Since there is no guarantee for the existence of such a function hh, we only compute a “best fit” function hh by the Carleman-based quasi-reversibility method described below. For each vHp(Ω)v\in H^{p}(\Omega), λ>1\lambda>1, β>0\beta>0 and η>0\eta>0, define the functional Jvλ,β,η:H0J_{v}^{\lambda,\beta,\eta}:H_{0}\to\mathbb{R} as

Jvλ,β,η(φ)=Ωe2λμβ(𝐱)|div(Aφ)div(Av)+F(𝐱,v(𝐱),v(𝐱))+D(v)φ|2d𝐱+ηv+φHp(Ω)2.J_{v}^{\lambda,\beta,\eta}(\varphi)=\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\Big{|}-{\rm div}(A\nabla\varphi)-{\rm div}(A\nabla v)+F({\bf x},v({\bf x}),\nabla v({\bf x}))\\ +D\mathcal{F}(v)\varphi\Big{|}^{2}\,d{\bf x}+\eta\|v+\varphi\|^{2}_{H^{p}(\Omega)}.

Here, the function μβ\mu_{\beta} is defined in (2.1) and ηv+φHp(Ω)2\eta\|v+\varphi\|^{2}_{H^{p}(\Omega)} is a regularization term.

Proposition 3.1.

For all λ>1,β>0\lambda>1,\beta>0 and η>0\eta>0, for each vHp(Ω),v\in H^{p}(\Omega), the functional Jvλ,β,η:H0H0J_{v}^{\lambda,\beta,\eta}:H_{0}\to H_{0} has a unique minimizer.

Proof.

It is obvious that Jvλ,β,ηJ_{v}^{\lambda,\beta,\eta} is coercive and weakly lower semicontinuous. Thus, it has a minimizer in H0H_{0}. The uniqueness of the minimizer can be deduced from the strict convexity of Jvλ,β,ηJ_{v}^{\lambda,\beta,\eta} in H0H_{0}. ∎

For each n0n\geq 0, thanks for Proposition 3.1, we can minimize Junλ,β,η(φ)J_{u_{n}}^{\lambda,\beta,\eta}(\varphi) on H0H_{0}. The unique minimizer is the desired function h.h. We then set

un+1=un+h.u_{n+1}=u_{n}+h. (3.4)

The construction of the sequence {un}n0\{u_{n}\}_{n\geq 0} above is summarized in Algorithm 1. We will prove that the sequence {un}n0\{u_{n}\}_{n\geq 0} converges to uu^{*} in Section 4 as nn\to\infty and η0\eta\to 0. The presence of the Carleman weight e2λμβ(𝐱)e^{2\lambda\mu_{\beta}({\bf x})} is a key point for us to prove this convergence result.

Algorithm 1 The procedure to compute the numerical solution to (1.1)
1:   Choose a threshold number 0<κ01.0<\kappa_{0}\ll 1. Choose an arbitrary initial solution u0W.u_{0}\in W.
2:  Set n=0n=0.
3:   Solve the linear equation (3.3) for a function hH0h\in H_{0} by minimizing Junλ,β,η(φ)J_{u_{n}}^{\lambda,\beta,\eta}(\varphi) on H0H_{0}. Set un+1=un+hu_{n+1}=u_{n}+h.
4:  if un+1unL2(Ω)>κ0\|u_{n+1}-u_{n}\|_{L^{2}(\Omega)}>\kappa_{0} then
5:     Replace nn by n+1.n+1.
6:     Go back to Step 3.
7:  else
8:     Set the computed solution ucomp=un+1.u_{\rm comp}=u_{n+1}.
9:  end if

4 The convergence analysis

In this section, we prove that the sequence {un}n0\{u_{n}\}_{n\geq 0} generated by Algorithm 1 converges to the solution uu^{*} to (1.1). The following result is the main theorem in this paper.

Theorem 4.1.

Assume that FC2(Ω,,d)\|F\|_{C^{2}(\Omega,\mathbb{R},\mathbb{R}^{d})} is a finite number. Assume further that (1.1) has a unique solution uWu^{*}\in W. Let {un}n0\{u_{n}\}_{n\geq 0} be the sequence generated by Algorithm 1, where u0Wu_{0}\in W is chosen arbitrarily. Then, there exist λ0>0\lambda_{0}>0 and θ(0,1)\theta\in(0,1) such that, for all λ>λ0\lambda>\lambda_{0},

un+1uλ,μ,β2θn+1u0uλ,μ,β2+ηθ1θn+11θuHp(Ω)2.\|u_{n+1}-u^{*}\|_{\lambda,\mu,\beta}^{2}\leq\theta^{n+1}\|u_{0}-u^{*}\|_{\lambda,\mu,\beta}^{2}+\eta\theta\frac{1-\theta^{n+1}}{1-\theta}\|u^{*}\|_{H^{p}(\Omega)}^{2}. (4.1)

Here,

vλ,μ,β=[Ωe2λμβ(𝐱)(|v|2+|v|2)𝑑𝐱]12for all vH1(Ω).\|v\|_{\lambda,\mu,\beta}=\Big{[}\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\big{(}|v|^{2}+|\nabla v|^{2}\big{)}\,d{\bf x}\Big{]}^{\frac{1}{2}}\quad\mbox{for all }v\in H^{1}(\Omega).
Proof.

Fix n0n\geq 0. Let h=un+1unh=u_{n+1}-u_{n}. Due to Step 3 in Algorithm 1, hh is the minimizer of Junλ,β,ηJ_{u_{n}}^{\lambda,\beta,\eta}. By the variational principle, we have

Ωe2λμβ(𝐱)(div(Ah)div(Aun)+F(𝐱,un,un)+D(un)h)(div(Aφ)+D(un)φ)d𝐱+ηun+h,φHp(Ω)=0\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\big{(}-{\rm div}(A\nabla h)-{\rm div}(A\nabla u_{n})+F({\bf x},u_{n},\nabla u_{n})+D\mathcal{F}(u_{n})h\big{)}\\ \big{(}-{\rm div}(A\nabla\varphi)+D\mathcal{F}(u_{n})\varphi\big{)}\,d{\bf x}+\eta\langle u_{n}+h,\varphi\rangle_{H^{p}(\Omega)}=0 (4.2)

for all φH0.\varphi\in H_{0}. Since un+1=un+hu_{n+1}=u_{n}+h, we can rewrite (4.2) as

Ωe2λμβ(𝐱)(div(Aun+1)+F(𝐱,un,un)+D(un)(un+1un))(div(Aφ)+D(un)φ)d𝐱+ηun+1,φHp(Ω)=0\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\big{(}-{\rm div}(A\nabla u_{n+1})+F({\bf x},u_{n},\nabla u_{n})+D\mathcal{F}(u_{n})(u_{n+1}-u_{n})\big{)}\\ \big{(}-{\rm div}(A\nabla\varphi)+D\mathcal{F}(u_{n})\varphi\big{)}\,d{\bf x}+\eta\langle u_{n+1},\varphi\rangle_{H^{p}(\Omega)}=0 (4.3)

for all φH0.\varphi\in H_{0}. As uu^{*} is the solution to (1.1), we have

Ωe2λμβ(𝐱)(div(Au)+F(𝐱,u,u))(div(Aφ)+D(un)φ)d𝐱=0\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\big{(}-{\rm div}(A\nabla u^{*})+F({\bf x},u^{*},\nabla u^{*})\big{)}\\ \big{(}-{\rm div}(A\nabla\varphi)+D\mathcal{F}(u_{n})\varphi\big{)}\,d{\bf x}=0 (4.4)

for all φH0.\varphi\in H_{0}. It follows from (4.3) and (4.4) that for all φH0\varphi\in H_{0}

Ωe2λμβ(𝐱)(div(A(un+1u))+F(𝐱,un,un)F(𝐱,u,u)+D(un)(un+1un))(div(Aφ)+D(un)φ)d𝐱+ηun+1,φHp(Ω)=0.\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\big{(}-{\rm div}(A\nabla(u_{n+1}-u^{*}))+F({\bf x},u_{n},\nabla u_{n})-F({\bf x},u^{*},\nabla u^{*})\\ +D\mathcal{F}(u_{n})(u_{n+1}-u_{n})\big{)}\big{(}-{\rm div}(A\nabla\varphi)+D\mathcal{F}(u_{n})\varphi\big{)}\,d{\bf x}\\ +\eta\langle u_{n+1},\varphi\rangle_{H^{p}(\Omega)}=0. (4.5)

Take φ=un+1uH0\varphi=u_{n+1}-u^{*}\in H_{0}. It follows from (4.5) that

Ωe2λμβ(𝐱)[|div(Aφ)|2D(un)φdiv(Aφ)(F(𝐱,un,un)F(𝐱,u,u))div(Aφ)+(F(𝐱,un,un)F(𝐱,u,u))D(un)φdiv(Aφ)D(un)(un+1un)+D(un)(un+1un)D(un)φ]d𝐱=ηun+1,φHp(Ω).\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}\Big{[}|{\rm div}(A\nabla\varphi)|^{2}-D\mathcal{F}(u_{n})\varphi\,{\rm div}(A\nabla\varphi)-\big{(}F({\bf x},u_{n},\nabla u_{n})\\ -F({\bf x},u^{*},\nabla u^{*})\big{)}{\rm div}(A\nabla\varphi)+\big{(}F({\bf x},u_{n},\nabla u_{n})-F({\bf x},u^{*},\nabla u^{*})\big{)}D\mathcal{F}(u_{n})\varphi\\ -{\rm div}(A\nabla\varphi)D\mathcal{F}(u_{n})(u_{n+1}-u_{n})+D\mathcal{F}(u_{n})(u_{n+1}-u_{n})D\mathcal{F}(u_{n})\varphi\Big{]}\,d{\bf x}\\ =-\eta\langle u_{n+1},\varphi\rangle_{H^{p}(\Omega)}. (4.6)

As FC2(Ω,,d)=C<+\|F\|_{C^{2}(\Omega,\mathbb{R},\mathbb{R}^{d})}=C<+\infty, we can estimate

|D(un)φ|C(|φ|+|φ|),\displaystyle|D\mathcal{F}(u_{n})\varphi|\leq C(|\varphi|+|\nabla\varphi|),
|F(𝐱,un,un)F(𝐱,u,u)|C(|unu|+|(unu)|),\displaystyle|F({\bf x},u_{n},\nabla u_{n})-F({\bf x},u^{*},\nabla u^{*})|\leq C(|u_{n}-u^{*}|+|\nabla(u_{n}-u^{*})|),
|D(un)(un+1un)|=|D(un)(un+1u+uun)|\displaystyle|D\mathcal{F}(u_{n})(u_{n+1}-u_{n})|=|D\mathcal{F}(u_{n})(u_{n+1}-u^{*}+u^{*}-u_{n})|
C(|un+1u|+|(un+1u)|)+C(|unu|+|(unu)|),\displaystyle\hskip 28.45274pt\leq C(|u_{n+1}-u^{*}|+|\nabla(u_{n+1}-u^{*})|)+C(|u_{n}-u^{*}|+|\nabla(u_{n}-u^{*})|),
and
ηun+1,φHp(Ω)=ηφ+u,φHp(Ω)\displaystyle-\eta\langle u_{n+1},\varphi\rangle_{H^{p}(\Omega)}=-\eta\langle\varphi+u^{*},\varphi\rangle_{H^{p}(\Omega)}
=ηφHp(Ω)2ηu,φHp(Ω)η2uHp(Ω)2.\displaystyle\hskip 82.51282pt=-\eta\|\varphi\|_{H^{p}(\Omega)}^{2}-\eta\langle u^{*},\varphi\rangle_{H^{p}(\Omega)}\leq\frac{\eta}{2}\|u^{*}\|_{H^{p}(\Omega)}^{2}.

These estimates, together with the inequality |ab|4a2+b2/8|ab|\leq 4a^{2}+b^{2}/8, imply

Ωe2λμβ(𝐱)|div(Aφ)|2d𝐱C[Ωe2λμβ(𝐱)(|φ|2+|φ|2)d𝐱+Ωe2λμβ(𝐱)(|unu|2+|(unu)|2)d𝐱]+η2uHp(Ω)2.\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|{\rm div}(A\nabla\varphi)|^{2}\,d{\bf x}\leq C\Big{[}\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}(|\varphi|^{2}+|\nabla\varphi|^{2})\,d{\bf x}\\ +\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}(|u_{n}-u^{*}|^{2}+|\nabla(u_{n}-u^{*})|^{2})\,d{\bf x}\Big{]}+\frac{\eta}{2}\|u^{*}\|_{H^{p}(\Omega)}^{2}. (4.7)

Applying the Carleman estimate (2.3) for the function φ\varphi, we have

Ωe2λμβ(𝐱)|div(Aφ)|2𝑑𝐱CλΩe2λμβ(𝐱)|φ(𝐱)|2𝑑𝐱+Cλ3Ωe2λμβ(𝐱)|φ(𝐱)|2𝑑𝐱.\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|{\rm div}(A\nabla\varphi)|^{2}\,d{\bf x}\\ \geq C\lambda\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|\nabla\varphi({\bf x})|^{2}\,d{\bf x}+C\lambda^{3}\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|\varphi({\bf x})|^{2}\,d{\bf x}. (4.8)

Combining (4.7) and (4.8), we have

λΩe2λμβ(𝐱)|φ(𝐱)|2𝑑𝐱+λ3Ωe2λμβ(𝐱)|φ(𝐱)|2𝑑𝐱C[Ωe2λμβ(𝐱)(|φ|2+|φ|2)𝑑𝐱+Ωe2λμβ(𝐱)(|unu|2+|(unu)|2)𝑑𝐱]+η2uHp(Ω)2.\lambda\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|\nabla\varphi({\bf x})|^{2}\,d{\bf x}+\lambda^{3}\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}|\varphi({\bf x})|^{2}\,d{\bf x}\\ \leq C\Big{[}\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}(|\varphi|^{2}+|\nabla\varphi|^{2})\,d{\bf x}+\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}(|u_{n}-u^{*}|^{2}+|\nabla(u_{n}-u^{*})|^{2})\,d{\bf x}\Big{]}\\ +\frac{\eta}{2}\|u^{*}\|_{H^{p}(\Omega)}^{2}. (4.9)

Letting λ\lambda be sufficiently large, we can simplify (4.9) as

λΩe2λμβ(𝐱)(|φ|2+|φ|2)𝑑𝐱CΩe2λμβ(𝐱)(|unu|2+|(unu)|2)𝑑𝐱+η2uHp(Ω)2.\lambda\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}(|\varphi|^{2}+|\nabla\varphi|^{2})\,d{\bf x}\leq C\int_{\Omega}e^{2\lambda\mu_{\beta}({\bf x})}(|u_{n}-u^{*}|^{2}+|\nabla(u_{n}-u^{*})|^{2})\,d{\bf x}\\ +\frac{\eta}{2}\|u^{*}\|_{H^{p}(\Omega)}^{2}. (4.10)

Recall that φ=un+1u\varphi=u_{n+1}-u^{*}. We get from (4.10) that

un+1uλ,μ,β2Cλunuλ,μ,β2+η2λuHp(Ω)2.\|u_{n+1}-u^{*}\|_{\lambda,\mu,\beta}^{2}\leq\frac{C}{\lambda}\|u_{n}-u^{*}\|_{\lambda,\mu,\beta}^{2}+\frac{\eta}{2\lambda}\|u^{*}\|_{H^{p}(\Omega)}^{2}. (4.11)

Applying (4.11) for n1n-1 and denoting θ=C/λ(0,1)\theta=C/\lambda\in(0,1), we have

un+1uλ,μ,β2\displaystyle\|u_{n+1}-u^{*}\|_{\lambda,\mu,\beta}^{2} θ[θun1uλ,μ,β2+ηθuHp(Ω)2]+ηθuHp(Ω)2\displaystyle\leq\theta\Big{[}\theta\|u_{n-1}-u^{*}\|_{\lambda,\mu,\beta}^{2}+\eta\theta\|u^{*}\|_{H^{p}(\Omega)}^{2}\Big{]}+\eta\theta\|u^{*}\|_{H^{p}(\Omega)}^{2}
=θ2un1uλ,μ,β2+ηθ(1+θ)uHp(Ω)2.\displaystyle=\theta^{2}\|u_{n-1}-u^{*}\|_{\lambda,\mu,\beta}^{2}+\eta\theta(1+\theta)\|u^{*}\|_{H^{p}(\Omega)}^{2}.

By induction, we have

un+1uλ,μ,β2θn+1u0uλ,μ,β2+ηθi=0nθnuHp(Ω)2,\|u_{n+1}-u^{*}\|_{\lambda,\mu,\beta}^{2}\leq\theta^{n+1}\|u_{0}-u^{*}\|_{\lambda,\mu,\beta}^{2}+\eta\theta\sum_{i=0}^{n}\theta^{n}\|u^{*}\|_{H^{p}(\Omega)}^{2}, (4.12)

which implies (4.1). The proof is complete. ∎

Remark 4.1 (Removing the boundedness condition of FF in C2C^{2} in Theorem 4.1).

In the case when FC2(Ω××d+1)=\|F\|_{C^{2}(\Omega\times\mathbb{R}\times\mathbb{R}^{d+1})}=\infty, we need to assume that we know in advance that the true solution uu^{*} to (1.1) belongs to the ball BMB_{M} of C1(Ω)C^{1}(\Omega) for some M>0M>0. This assumption does not weaken the result since MM can be arbitrarily large. Define the cut-off function χC(Ω××d)\chi\in C^{\infty}(\Omega\times\mathbb{R}\times\mathbb{R}^{d}) as

χ(𝐱,s,𝐩)={1|s|+|𝐩|<M,0|s|+|𝐩|>2M\chi({\bf x},s,{\bf p})=\left\{\begin{array}[]{ll}1&|s|+|{\bf p}|<M,\\ 0&|s|+|{\bf p}|>2M\end{array}\right.

and set F~=χF.\widetilde{F}=\chi F. Since |u|+|u|<M|u^{*}|+|\nabla u^{*}|<M, it is obvious that uu^{*} solves the problem

{div(A(𝐱)u(𝐱))+F~(𝐱,u(𝐱),u(𝐱))=0𝐱Ω,u(𝐱)=f(𝐱)𝐱Ω,νu(𝐱)=g(𝐱)𝐱Ω.\left\{\begin{array}[]{ll}-{\rm div}(A({\bf x})\nabla u({\bf x}))+\widetilde{F}({\bf x},u({\bf x}),\nabla u({\bf x}))=0&{\bf x}\in\Omega,\\ u({\bf x})=f({\bf x})&{\bf x}\in\partial\Omega,\\ \partial_{\nu}u({\bf x})=g({\bf x})&{\bf x}\in\partial\Omega.\end{array}\right. (4.13)

Now, we can apply Algorithm 1 for (4.13) to compute uu^{*}.

Remark 4.2.

Theorem 4.1 and estimate (4.1) rigorously guarantee that each sequence generated by Algorithm 1 converges to uu^{*} at the exponential rate. This fact is numerically confirmed by our numerical results in Section 5.

Remark 4.3.

As seen in the proof of Theorem 4.1, the efficiency of Algorithm 1 is guaranteed by Carleman estimate (2.3). Therefore, we call the proposed method described in Algorithm 1 the second generation of Carleman-based numerical methods. This second generation includes the method in [30, 38] in which the iteration scheme is developed based on the contraction principle and Carleman estimates. The first generation of Carleman-based numerical method was developed in [21], which is called the convexification. See [1, 25, 30, 31] for some following up results. Like the convexification method, Algorithm 1 can be used to compute solutions to nonlinear PDEs without requesting an initial good guess. The advantage of our new method is the fast convergence rate, see Remark 4.2.

5 Numerical study

In this section, we present some numerical results obtained by Algorithm 1. For simplicity, we set d=2d=2 and Ω=(1,1)2\Omega=(-1,1)^{2}. On Ω,\Omega, we arrange an N×NN\times N grid points

𝒢={(xi=1+(i1)δ,yj=1+(j1)δ):1i,jN}\mathcal{G}=\{(x_{i}=-1+(i-1)\delta,y_{j}=-1+(j-1)\delta):1\leq i,j\leq N\}

where δ=2/(N1).\delta=2/(N-1). In our numerical scripts, N=80N=80.

In Step 1 of Algorithm 1, we choose a function u0Wu_{0}\in W. It is natural to find a function u0u_{0} satisfying the equation obtained by removing from (1.1) the nonlinearity F(𝐱,u,u)F({\bf x},u,\nabla u). We apply the Carleman-based quasi-reversibility method to do so. That means, u0u_{0} is the minimizer of

J0λ,β,η(φ)=Ωeλμβ(𝐱)|div(Aφ)|2𝑑𝐱+ηφH2(Ω)2J_{0}^{\lambda,\beta,\eta}(\varphi)=\int_{\Omega}e^{\lambda\mu_{\beta}({\bf x})}|{\rm div}(A\nabla\varphi)|^{2}d{\bf x}+\eta\|\varphi\|^{2}_{H^{2}(\Omega)} (5.1)

subject to the boundary conditions in (5.7). To simplify the efforts in implementation, the norm in the regularization term is the H2(Ω)H^{2}(\Omega)-norm rather than the Hp(Ω)H^{p}(\Omega)-norm. This change does not affect the performance of Algorithm 1. Algorithm 1 still provides satisfactory solutions to (1.1).

Remark 5.1.

We employ the Carleman-based quasi-reversibility method to find u0u_{0} for the consistency to Step 3 of Algorithm 1. Since u0Wu_{0}\in W can be chosen arbitrarily, one can use the quasi-reversibility method without the presence of the Carleman weight function eλμβ(𝐱)e^{\lambda\mu_{\beta}({\bf x})}. In our computation for all numerical examples below, λ=4\lambda=4, β=10\beta=10 and 𝐱0=(4,0).{\bf x}_{0}=(-4,0). The regularized parameter is η=104\eta=10^{-4}. The threshold number κ0=106.\kappa_{0}=10^{-6}.

We minimize J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} on H0H_{0} by the least square MATLAB command “lsqlin”. The implementation for the quasi-reversibility method to minimize more general functionals than J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} was described in [30, §5.3] and in [37, §5]. We do not repeat this process here. In Step 3 of Algorithm 1, given unW,u_{n}\in W, we minimize the functional Junλ,β,ηJ^{\lambda,\beta,\eta}_{u_{n}} on H0H_{0}. Again, we refer the reader to [30, §5.3] and in [37, §5] for details in implementation. The scripts for other steps of Algorithm 1 can be written easily.

5.1 Quasilinear elliptic equations

The convexification method, first introduced in [21], was used to numerically solve quasilinear elliptic equations in [1, 25, 31]. Our approach here is of course different based on iteration and linearization. In particular, Step 3 in Algorithm 1 is very efficient as we only need to solve the linear equation (3.3) as opposed to solving directly a nonlinear quasilinear elliptic equation. In this subsection, we present two (2) numerical tests. In both tests, we choose the matrix AA to be

A=[2112].A=\left[\begin{array}[]{ll}2&1\\ 1&2\end{array}\right].

That means, div(Au)=2uxx+2uxy+2uyy.{\rm div}(A\nabla u)=2u_{xx}+2u_{xy}+2u_{yy}.

In test 1, we solve (1.1) when

F(𝐱,s,𝐩)=s+|𝐩|(x2+2y2+4x2+16y24)F({\bf x},s,{\bf p})=s+|{\bf p}|-\big{(}-x^{2}+2y^{2}+\sqrt{4x^{2}+16y^{2}}-4\big{)} (5.2)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\Omega, ss\in\mathbb{R} and 𝐩2{\bf p}\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=x2+2y2,νu(𝐱)=2x,4yνu({\bf x})=-x^{2}+2y^{2},\quad\partial_{\nu}u({\bf x})=\langle-2x,4y\rangle\cdot\nu (5.3)

for all 𝐱=(x,y)Ω.{\bf x}=(x,y)\in\partial\Omega. The true solution of (1.1) is the function utrue(x,y)=x2+2y2u_{\rm true}(x,y)=-x^{2}+2y^{2}.

Refer to caption
(a) The function uu^{*}
Refer to caption
(b) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω)\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}
Refer to caption
(c) unun1L2(Ω).\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration n.n.
Figure 1: Test 1. The numerical solution to (1.1) where FF is given in (5.2) and the boundary data are given in (5.3).

It is evident from Figure 1 that Algorithm 1 provides out of expectation solution for test 1. The relative error utrueucompL(Ω)/utrueL(Ω)=1.23×105\|u_{\rm true}-u_{\rm comp}\|_{L^{\infty}(\Omega)}/\|u_{\rm true}\|_{L^{\infty}(\Omega)}=1.23\times 10^{-5}. One can see from Figure 1c that Algorithm 1 converges at the third iteration.

In test 2, we solve (1.1) when

F(𝐱,s,𝐩)=|𝐩|[[π2cos(π2(x+y))+ex]2+π24cos2(π2(x+y))+3π22sin(π2(x+y))2ex]F({\bf x},s,{\bf p})=|{\bf p}|-\Big{[}\sqrt{\Big{[}\frac{\pi}{2}\cos\big{(}\frac{\pi}{2}(x+y)\big{)}+e^{x}\Big{]}^{2}+\frac{\pi^{2}}{4}\cos^{2}\big{(}\frac{\pi}{2}(x+y)\big{)}}\\ +\frac{3\pi^{2}}{2}\sin\big{(}\frac{\pi}{2}(x+y)\big{)}-2e^{x}\Big{]} (5.4)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\Omega, ss\in\mathbb{R} and 𝐩2{\bf p}\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=sin(π2(x+y))+ex,νu(𝐱)=π2cos(π2(x+y))+ex,cos(π2(x+y))νu({\bf x})=\sin\big{(}\frac{\pi}{2}(x+y)\big{)}+e^{x},\quad\partial_{\nu}u({\bf x})=\big{\langle}\frac{\pi}{2}\cos\big{(}\frac{\pi}{2}(x+y)\big{)}+e^{x},\cos\big{(}\frac{\pi}{2}(x+y)\big{)}\big{\rangle}\cdot\nu (5.5)

for all 𝐱=(x,y)Ω.{\bf x}=(x,y)\in\partial\Omega. The true solution of (1.1) is the function utrue(x,y)=sin(π2(x+y))+exu_{\rm true}(x,y)=\sin\big{(}\frac{\pi}{2}(x+y)\big{)}+e^{x}.

Refer to caption
(a) The function uu^{*}
Refer to caption
(b) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω)\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}
Refer to caption
(c) unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn
Figure 2: Test 2. The numerical solution to (1.1) where FF is given in (5.4) and the boundary data are given in (5.5).

It is evident from Figure 2 that Algorithm 1 provides out of expectation solution for test 2. The relative error utrueucompL(Ω)/utrueL(Ω)=4.19×105\|u_{\rm true}-u_{\rm comp}\|_{L^{\infty}(\Omega)}/\|u_{\rm true}\|_{L^{\infty}(\Omega)}=4.19\times 10^{-5}. One can see from Figure 2c that Algorithm 1 converges at the fifth iteration.

5.2 Application to first-order Hamilton-Jacobi equations

Our aim in this subsection is to solve numerically

F(𝐱,u,u)=0for all 𝐱Ω.F({\bf x},u,\nabla u)=0\quad\mbox{for all }{\bf x}\in\Omega. (5.6)

with the boundary conditions

u|Ω=fandνu|Ω=g.u|_{\partial\Omega}=f\quad\mbox{and}\quad\partial_{\nu}u|_{\partial\Omega}=g. (5.7)

Basically, we use the vanishing viscosity process to approximate solutions to (5.6). For ϵ>0\epsilon>0, consider

ϵΔu+F(𝐱,u,u)=0for all 𝐱Ω-\epsilon\Delta u+F({\bf x},u,\nabla u)=0\quad\mbox{for all }{\bf x}\in\Omega (5.8)

with boundary conditions (5.7). Again, (5.8)–(5.7) is an over-determined boundary value problem. For ϵ>0\epsilon>0 sufficiently small, assume that (5.8)–(5.7) has a solution uϵWu^{\epsilon}\in W. Then, uϵu^{\epsilon} approximates uu, solution to (5.6)–(5.7), quite well under some appropriate assumptions on FF. In our numerical tests, we choose ϵ=ϵ0=103\epsilon=\epsilon_{0}=10^{-3}.

In this part, we provide six (6) numerical tests, in which we compute the viscosity solution to some Hamilton-Jacobi equations of the form (5.6) given Cauchy boundary data. That means, by applying Algorithm 1, we numerically find a function ucompu_{\rm comp} satisfying (5.8)-(5.7) when FF, ff and gg are given. The verification that uu^{*} is the correct viscosity solution can be done in a similar manner as that in [47, 25].

Test 1. In this test, we solve (5.8)-(5.7) when

F(𝐱,s,𝐩)=s+|𝐩|+|x|1.F({\bf x},s,{\bf p})=s+|{\bf p}|+|x|-1. (5.9)

for all 𝐱=(x,y)Ω,s,𝐩2{\bf x}=(x,y)\in\Omega,s\in\mathbb{R},{\bf p}\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=|x|,νu(𝐱)=sign(x),0νu({\bf x})=-|x|,\quad\partial_{\nu}u({\bf x})=\langle-\mbox{sign}(x),0\rangle\cdot\nu (5.10)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\partial\Omega. In this case, the true solution is u=|x|.u^{*}=-|x|. The numerical result is displayed in Figure 3.

Refer to caption
(a) The true solution uu^{*} to the Hamilton-Jacobi equation where the Hamiltonian is given in (5.9).
Refer to caption
(b) The initial solution u0u_{0} computed by minimizing J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} defined in (5.1)
Refer to caption
(c) The computed solution ucompu_{\rm comp}.
Refer to caption
(d) unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn.
Refer to caption
(e) The true and computed solutions on the line {(x=0,y)Ω}\{(x=0,y)\in\Omega\}
Refer to caption
(f) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω).\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}.
Figure 3: Test 1. The true viscosity solution to (5.8)-(5.7) and the computed one. The Hamiltonian and the boundary conditions are given in (5.9)-(5.10).

It is evident from Figure 3 that we successfully compute the solution ucompu_{\rm comp}. The procedure described in Algorithm 1 converges after four iterations. The relative error uucompL(Ω)uL(Ω)=5.33%.\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}}=5.33\%.

Test 2. We find the viscosity solution to the eikonal equation. In this test, we solve (5.8)-(5.7) when the Hamiltonian is

F(𝐱,s,𝐩)=|𝐩|2(1+(1+sign(x+y))2)F({\bf x},s,{\bf p})=|{\bf p}|^{2}-(1+(1+\mbox{sign}(x+y))^{2}) (5.11)

for all 𝐱=(x,y)Ω,s,𝐩2{\bf x}=(x,y)\in\Omega,s\in\mathbb{R},{\bf p}\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=|x+y|y,νu(𝐱)=(sign(x+y),sign(x+y)1)νu({\bf x})=-|x+y|-y,\quad\partial_{\nu}u({\bf x})=(-\mbox{sign}(x+y),-\mbox{sign}(x+y)-1)\cdot\nu (5.12)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\partial\Omega. The true solution is u(𝐱)=|x+y|yu^{*}({\bf x})=-|x+y|-y. The graphs of uu^{*} and ucompu_{\rm comp} are displayed in Figure 4.

Refer to caption
(a) The true solution uu^{*} to the Hamilton-Jacobi equation where the Hamiltonian is given in (5.11).
Refer to caption
(b) The initial solution u0u_{0} computed by minimizing J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} defined in (5.1)
Refer to caption
(c) The computed solution ucompu_{\rm comp}.
Refer to caption
(d) unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn.
Refer to caption
(e) The true and computed solutions on the line {(x=0.5,y)Ω}\{(x=0.5,y)\in\Omega\}
Refer to caption
(f) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω).\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}.
Figure 4: Test 2. The true viscosity solution to (5.8)-(5.7) and the computed one. The Hamiltonian and the boundary conditions are given in (5.11)-(5.12).

The relative error uucompL(Ω)uL(Ω)=3.6%.\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}}=3.6\%.

Test 3. We test the case when the Hamiltonian is not convex with respect its third variable. The Hamiltonian in this test is given by

F(𝐱,s,𝐩)=20s+|p1||p2|[20(|x+0.5|+ecos(2π(x2+y2)))+|sign(x+0.5)+4πxsin(2π(x2+y2))ecos(2π(x2+y2))||4πysin(2π(x2+y2))ecos(2π(x2+y2))|]F({\bf x},s,{\bf p})=20s+|p_{1}|-|p_{2}|-\Big{[}20(-|x+0.5|+e^{\cos(2\pi(x^{2}+y^{2}))})\\ +|\mbox{sign}(x+0.5)+4\pi x\sin(2\pi(x^{2}+y^{2}))e^{\cos(2\pi(x^{2}+y^{2}))}|\\ -|4\pi y\sin(2\pi(x^{2}+y^{2}))e^{\cos(2\pi(x^{2}+y^{2}))}|\Big{]} (5.13)

for all 𝐱=(x,y)Ω,s,𝐩=(p1,p2)2{\bf x}=(x,y)\in\Omega,s\in\mathbb{R},{\bf p}=(p_{1},p_{2})\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=|x+0.5|+ecos(2π(x2+y2))for all 𝐱=(x,y)Ωu({\bf x})=-|x+0.5|+e^{\cos(2\pi(x^{2}+y^{2}))}\quad\mbox{for all }{\bf x}=(x,y)\in\partial\Omega (5.14)

and

νu(𝐱)=(sign(x+0.5)4πxsin(2π(x2+y2))ecos(2π(x2+y2)),4πysin(2π(x2+y2))ecos(2π(x2+y2)))ν\partial_{\nu}u({\bf x})=\big{(}-\mbox{sign}(x+0.5)-4\pi x\sin(2\pi(x^{2}+y^{2}))e^{\cos(2\pi(x^{2}+y^{2}))},\\ -4\pi y\sin(2\pi(x^{2}+y^{2}))e^{\cos(2\pi(x^{2}+y^{2}))}\big{)}\cdot\nu (5.15)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\partial\Omega. The true solution is u(𝐱)=|x+0.5|+ecos(2π(x2+y2))u^{*}({\bf x})=-|x+0.5|+e^{\cos(2\pi(x^{2}+y^{2}))}. The graphs of uu^{*} and ucompu_{\rm comp} are displayed in Figure 5.

Refer to caption
(a) The true solution uu^{*} to the Hamilton-Jacobi equation where the Hamiltonian is given in (5.13).
Refer to caption
(b) The initial solution u0u_{0} computed by minimizing J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} defined in (5.1)
Refer to caption
(c) The computed solution ucompu_{\rm comp}.
Refer to caption
(d) unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn.
Refer to caption
(e) The true and computed solutions on the line {(x=0,y)Ω}\{(x=0,y)\in\Omega\}
Refer to caption
(f) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω).\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}.
Figure 5: Test 3. The true viscosity solution to (5.8)-(5.7) and the computed one. The Hamiltonian and the boundary conditions are given in (5.13)-(5.15).

The relative error uucompL(Ω)uL(Ω)=10.65%.\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}}=10.65\%. Although the relative error in this test is large, the numerical result is acceptable. In fact, we observe from Figure 5f that uucompL(Ω)uL(Ω)\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}} is small almost everywhere while it is large at only two small places near the left edge of the domain.

Test 4. We test the case when the Hamiltonian is not increasing in the second variable and is not convex in the third variable. The Hamiltonian in this test is given by

F(𝐱,s,𝐩)=40s+||𝐩|10|+40(|x+y0.5|+sin(x22+y2))|([sign(x+y0.5)+xcos(x22+y2)]2+[sign(x+y0.5)+2ycos(x22+y2)]2)1/210|F({\bf x},s,{\bf p})=-40s+||{\bf p}|-10|+40\Big{(}|x+y-0.5|+\sin\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)}\Big{)}\\ -\Big{|}\Big{(}\Big{[}\mbox{sign}(x+y-0.5)+x\cos\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)}\Big{]}^{2}+\\ \Big{[}\mbox{sign}(x+y-0.5)+2y\cos\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)}\Big{]}^{2}\Big{)}^{1/2}-10\Big{|} (5.16)

for all 𝐱=(x,y)Ω,s,𝐩=(p1,p2)2{\bf x}=(x,y)\in\Omega,s\in\mathbb{R},{\bf p}=(p_{1},p_{2})\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=|x+y0.5|+sin(x22+y2)for all 𝐱=(x,y)Ωu({\bf x})=|x+y-0.5|+\sin\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)}\quad\mbox{for all }{\bf x}=(x,y)\in\partial\Omega (5.17)

and

νu(𝐱)=(sign(x+y0.5)+xcos(x22+y2),sign(x+y0.5)+2ycos(x22+y2))ν\partial_{\nu}u({\bf x})=\Big{(}\mbox{sign}(x+y-0.5)+x\cos\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)},\\ \mbox{sign}(x+y-0.5)+2y\cos\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)}\Big{)}\cdot\nu (5.18)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\partial\Omega. The true solution is u(𝐱)=|x+y0.5|+sin(x22+y2)u^{*}({\bf x})=|x+y-0.5|+\sin\Big{(}\frac{x^{2}}{2}+y^{2}\Big{)}. The graphs of uu^{*} and ucompu_{\rm comp} are displayed in Figure 6.

Refer to caption
(a) The true solution uu^{*} to the Hamilton-Jacobi equation where the Hamiltonian is given in (5.16).
Refer to caption
(b) The initial solution u0u_{0} computed by minimizing J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} defined in (5.1)
Refer to caption
(c) The computed solution ucompu_{\rm comp}.
Refer to caption
(d) The relative error unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn.
Refer to caption
(e) The true and computed solutions on the line {(x=0,y)Ω}\{(x=0,y)\in\Omega\}
Refer to caption
(f) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω).\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}.
Figure 6: Test 4. The true viscosity solution to (5.8)-(5.7) and the computed one. The Hamiltonian and the boundary conditions are given in (5.16)-(5.18).

The relative error uucompL(Ω)uL(Ω)=0.95%.\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}}=0.95\%.

Test 5. We solve a G-equation. The Hamiltonian in this test is given by

F(𝐱,s,𝐩)=5s+|𝐩|xp1+[5(|x0.5|+|y|)xsign(x0.5)2]F({\bf x},s,{\bf p})=5s+|{\bf p}|-xp_{1}+\Big{[}5(|x-0.5|+|y|)-x\mbox{sign}(x-0.5)-\sqrt{2}\Big{]} (5.19)

for all 𝐱=(x,y)Ω,s,𝐩=(p1,p2)2{\bf x}=(x,y)\in\Omega,s\in\mathbb{R},{\bf p}=(p_{1},p_{2})\in\mathbb{R}^{2}. The boundary conditions are given by

u(𝐱)=|x0.5||y|for all 𝐱=(x,y)Ωu({\bf x})=-|x-0.5|-|y|\quad\mbox{for all }{\bf x}=(x,y)\in\partial\Omega (5.20)

and

νu(𝐱)=(sign(x0.5),sign(y))ν\partial_{\nu}u({\bf x})=-\big{(}\mbox{sign}(x-0.5),\mbox{sign}(y)\big{)}\cdot\nu (5.21)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\partial\Omega. The true solution is u(𝐱)=|x0.5||y|u^{*}({\bf x})=-|x-0.5|-|y|. The graphs of uu^{*} and ucompu_{\rm comp} are displayed in Figure 5.

Refer to caption
(a) The true solution uu^{*} to the Hamilton-Jacobi equation where the Hamiltonian is given in (5.19).
Refer to caption
(b) The initial solution u0u_{0} computed by minimizing J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} defined in (5.1)
Refer to caption
(c) The computed solution ucompu_{\rm comp}.
Refer to caption
(d) The relative error unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn.
Refer to caption
(e) The true and computed solutions on the line {(x=0,y)Ω}\{(x=0,y)\in\Omega\}
Refer to caption
(f) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω).\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}.
Figure 7: Test 5. The true viscosity solution to (5.8)-(5.7) and the computed one. The Hamiltonian and the boundary conditions are given in (5.19)-(5.21).

The relative error uucompL(Ω)uL(Ω)=0.98%.\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}}=0.98\%.

Test 6. The Hamiltonian in this test is given by

F(𝐱,s,𝐩)=20s+min{|𝐩|,||𝐩|10|+6}[20(|x|+sin(π(x2+y2)))+min{h(x,y),|h(x,y)10|+6}]F({\bf x},s,{\bf p})=20s+\min\{|{\bf p}|,||{\bf p}|-10|+6\}-\Big{[}20(-|x|+\sin(\pi(x^{2}+y^{2})))\\ +\min\Big{\{}h(x,y),|h(x,y)-10|+6\Big{\}}\Big{]} (5.22)

for all 𝐱=(x,y)Ω,s,𝐩=(p1,p2)2{\bf x}=(x,y)\in\Omega,s\in\mathbb{R},{\bf p}=(p_{1},p_{2})\in\mathbb{R}^{2} where

h(x,y)=[sign(x)+2πcos(π(x2+y2))]2+[2πcos(π(x2+y2)]2.h(x,y)=\sqrt{[-\mbox{sign}(x)+2\pi\cos(\pi(x^{2}+y^{2}))]^{2}+[2\pi\cos(\pi(x^{2}+y^{2})]^{2}}.

The boundary conditions are given by

u(𝐱)=|x|+sin(π(x2+y2))for all 𝐱=(x,y)Ωu({\bf x})=-|x|+\sin(\pi(x^{2}+y^{2}))\quad\mbox{for all }{\bf x}=(x,y)\in\partial\Omega (5.23)

and

νu(𝐱)=(sign(x)+2πxcos(π(x2+y2)),2πycos(π(x2+y2)))ν\partial_{\nu}u({\bf x})=\big{(}-\mbox{sign}(x)+2\pi x\cos(\pi(x^{2}+y^{2})),2\pi y\cos(\pi(x^{2}+y^{2}))\big{)}\cdot\nu (5.24)

for all 𝐱=(x,y)Ω{\bf x}=(x,y)\in\partial\Omega. The true solution is u(𝐱)=|x|+sin(π(x2+y2))u^{*}({\bf x})=-|x|+\sin(\pi(x^{2}+y^{2})). The graphs of uu^{*} and ucompu_{\rm comp} are displayed in Figure 8.

Refer to caption
(a) The true solution uu^{*} to the Hamilton-Jacobi equation where the Hamiltonian is given in (5.22).
Refer to caption
(b) The initial solution u0u_{0} computed by minimizing J0λ,β,ηJ_{0}^{\lambda,\beta,\eta} defined in (5.1)
Refer to caption
(c) The computed solution ucompu_{\rm comp}.
Refer to caption
(d) The relative error unun1L2(Ω)\|u_{n}-u_{n-1}\|_{L^{2}(\Omega)}. The horizontal axis is the number of iteration nn.
Refer to caption
(e) The true and computed solutions on the line {(x=0,y)Ω}\{(x=0,y)\in\Omega\}
Refer to caption
(f) The relative error |u(𝐱)ucomp(𝐱)|utrueL(Ω).\frac{|u^{*}({\bf x})-u_{\rm comp}({\bf x})|}{\|u_{\rm true}\|_{L^{\infty}(\Omega)}}.
Figure 8: Test 6. The true viscosity solution to (5.8)-(5.7) and the computed one. The Hamiltonian and the boundary conditions are given in (5.22)-(5.24).

The relative error uucompL(Ω)uL(Ω)=4.8%.\frac{\|u^{*}-u_{\rm comp}\|_{L^{\infty}(\Omega)}}{\|u^{*}\|_{L^{\infty}(\Omega)}}=4.8\%.

Remark 5.2.

The LL^{\infty} relative errors in all tests above for first-order Hamilton-Jacobi equations are compatible with max{O(η),O(ϵ0)}3%.\max\{O(\sqrt{\eta}),O(\sqrt{\epsilon_{0}})\}\simeq 3\%.

6 Concluding remarks

We have developed a new globally convergent numerical method to solve over-determined boundary value problems of quasilinear elliptic equations. The key point of the method is to repeatedly solve the linearization of the given PDE by the Carleman weighted quasi-reversibility method (Algorithm 1). As the result, we obtain a sequence of functions converging to the solution thanks to the Carleman estimate (Lemma 2.1). The strength of our new method includes (1) the global convergence property and (2) the fast convergence rate, which is described in Theorem 4.1. Some numerical results for quasilinear elliptic equations and first-order Hamilton-Jacobi equations are presented to show the effectiveness of our method.

Acknowledgement

The works of TTL and LHN were supported by US Army Research Laboratory and US Army Research Office grant W911NF-19-1-0044 and , in part, by funds provided by the Faculty Research Grant program at UNC Charlotte, Fund No. 111272. The work of HT is supported in part by NSF CAREER grant DMS-1843320 and a Simons fellowship.

References

  • [1] A. B. Bakushinskii, M. V. Klibanov, and N. A. Koshev. Carleman weight functions for a globally convergent numerical method for ill-posed Cauchy problems for some quasilinear PDEs. Nonlinear Anal. Real World Appl., 34:201–224, 2017.
  • [2] G. Barles and P. E. Souganidis. Convergence of approximation schemes for fully nonlinear second order equations. Asymptotic Anal., 4(3):271–283, 1991.
  • [3] L. Beilina and M. V. Klibanov. Approximate Global Convergence and Adaptivity for Coefficient Inverse Problems. Springer, New York, 2012.
  • [4] A. L. Bukhgeim and M. V. Klibanov. Uniqueness in the large of a class of multidimensional inverse problems. Soviet Math. Doklady, 17:244–247, 1981.
  • [5] A. P. Calderón. On an inverse boundary value problem. In Seminar on Numerical Analysis and its Applications to Continuum Physics, Río de Janeiro. Soc. Brasileira de Matemática, 1980.
  • [6] T. Carleman. Sur les systèmes linéaires aux derivées partielles du premier ordre a deux variables. C. R. Acad. Sci. Paris, 197:471–474, 1933.
  • [7] David Colton and Rainer Kress. Inverse acoustic and electromagnetic scattering theory. Applied Mathematical Sciences. Springer, New York, 3rd edition, 2013.
  • [8] M. G. Crandall, L. C. Evans, and P.-L. Lions. Some properties of viscosity solutions of Hamilton-Jacobi equations. Trans. Amer. Math. Soc., 282(2):487–502, 1984.
  • [9] M. G. Crandall and P.-L. Lions. Viscosity solutions of Hamilton-Jacobi equations. Trans. Amer. Math. Soc., 277(1):1–42, 1983.
  • [10] M. G. Crandall and P.-L. Lions. Two approximations of solutions of Hamilton-Jacobi equations. Math. Comp., 43(167):1–19, 1984.
  • [11] M. Falcone and R. Ferretti. Semi-Lagrangian schemes for Hamilton-Jacobi equations, discrete representation formulae and Godunov methods. J. Comput. Phys., 175(2):559–575, 2002.
  • [12] M. Falcone and R. Ferretti. Semi-Lagrangian approximation schemes for linear and Hamilton-Jacobi equations. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 2014.
  • [13] V. A. Khoa, G. W. Bidney, M. V. Klibanov, L. H. Nguyen, L. Nguyen, A. Sullivan, and V. N. Astratov. Convexification and experimental data for a 3D inverse scattering problem with the moving point source. Inverse Problems, 36:085007, 2020.
  • [14] V. A. Khoa, G. W. Bidney, M. V. Klibanov, L. H. Nguyen, L. Nguyen, A. Sullivan, and V. N. Astratov. An inverse problem of a simultaneous reconstruction of the dielectric constant and conductivity from experimental backscattering data. Inverse Problems in Science and Engineering, 29(5):712–735, 2021.
  • [15] V. A. Khoa, M. V. Klibanov, and L. H. Nguyen. Convexification for a 3D inverse scattering problem with the moving point source. SIAM J. Imaging Sci., 13(2):871–904, 2020.
  • [16] M. V. Klibanov. Global convexity in a three-dimensional inverse acoustic problem. SIAM J. Math. Anal., 28:1371–1388, 1997.
  • [17] M. V. Klibanov. Global convexity in diffusion tomography. Nonlinear World, 4:247–265, 1997.
  • [18] M. V. Klibanov. Carleman estimates for the regularization of ill-posed Cauchy problems. Applied Numerical Mathematics, 94:46–74, 2015.
  • [19] M. V. Klibanov. Carleman weight functions for solving ill-posed Cauchy problems for quasilinear PDEs. Inverse Problems, 31:125007, 2015.
  • [20] M. V. Klibanov. Convexification of restricted Dirichlet to Neumann map. J. Inverse and Ill-Posed Problems, 25(5):669–685, 2017.
  • [21] M. V. Klibanov and O. V. Ioussoupova. Uniform strict convexity of a cost functional for three-dimensional inverse scattering problem. SIAM J. Math. Anal., 26:147–179, 1995.
  • [22] M. V. Klibanov, T. T. Le, L. H. Nguyen, A. Sullivan, and L. Nguyen. Convexification-based globally convergent numerical method for a 1D coefficient inverse problem with experimental data. to appear on Inverse Problems and Imaging, DOI: https://www.aimsciences.org/article/doi/10.3934/ipi.2021068, 2021.
  • [23] M. V. Klibanov and J. Li. Inverse Problems and Carleman Estimates: Global Uniqueness, Global Convergence and Experimental Data. De Gruyter, 2021.
  • [24] M. V. Klibanov, J. Li, and W. Zhang. Convexification of electrical impedance tomography with restricted Dirichlet-to-Neumann map data. Inverse Problems, 35:035005, 2019.
  • [25] M. V. Klibanov, L. H. Nguyen, and H. V. Tran. Numerical viscosity solutions to Hamilton-Jacobi equations via a Carleman estimate and the convexification method. Journal of Computational Physics, 451:110828, 2022.
  • [26] K. Knudsen, M. Lassas, J. L. Mueller, and S. Siltanen. Regularized D-bar method for the inverse conductivity problem. Inverse Problems, 3:599–624, 2009.
  • [27] R. V. Kohn and M. Vogelius. Determining conductivity by boundary measurements II. Interior results. Comm. Pure & Applied Math., 38:643–667, 1985.
  • [28] R. Lattès and J. L. Lions. The Method of Quasireversibility: Applications to Partial Differential Equations. Elsevier, New York, 1969.
  • [29] M. M. Lavrent’ev, V. G. Romanov, and S. P. Shishat\cdotskiĭ. Ill-Posed Problems of Mathematical Physics and Analysis. Translations of Mathematical Monographs. AMS, Providence: RI, 1986.
  • [30] T. T. Le and L. H. Nguyen. A convergent numerical method to recover the initial condition of nonlinear parabolic equations from lateral Cauchy data. Journal of Inverse and Ill-posed Problems, DOI: https://doi.org/10.1515/jiip-2020-0028, 30(2):265–286, 2022.
  • [31] T. T. Le and L. H. Nguyen. The gradient descent method for the convexification to solve boundary value problems of quasi-linear PDEs and a coefficient inverse problem. to appear in Journal of Scientific Computing, preprint Arxiv:2103.04159, 2022.
  • [32] T. T. Le, L. H. Nguyen, T-P. Nguyen, and W. Powell. The quasi-reversibility method to numerically solve an inverse source problem for hyperbolic equations. Journal of Scientific Computing, 87:90, 2021.
  • [33] J. L. Mueller and S. Siltanen. The D-bar method for electrical impedance tomography-demystified. Inverse Problems, 36:093001, 2020.
  • [34] E. K. Murphy and J. L. Mueller. Effect of domain shape modeling and measurement errors on the 2-D D-bar method for EIT. IEEE Trans. Med. Imaging, 28:1576–1584, 2009.
  • [35] A. I. Nachman. Global uniqueness for a two-dimensional inverse boundary value problem. Ann. of Math., 143:71–96, 1996.
  • [36] H. M. Nguyen and L. H. Nguyen. Cloaking using complementary media for the Helmholtz equation and a three spheres inequality for second order elliptic equations. Transaction of the American Mathematical Society, 2:93–112, 2015.
  • [37] L. H. Nguyen. A new algorithm to determine the creation or depletion term of parabolic equations from boundary measurements. Computers and Mathematics with Applications, 80:2135–2149, 2020.
  • [38] L. H. Nguyen and M. V. Klibanov. Carleman estimates and the contraction principle for an inverse source problem for nonlinear hyperbolic equations. Inverse Problems, 38:035009, 2022.
  • [39] L. H. Nguyen, Q. Li, and M. V. Klibanov. A convergent numerical method for a multi-frequency inverse source problem in inhomogenous media. Inverse Problems and Imaging, 13:1067–1094, 2019.
  • [40] S. Osher and R. Fedkiw. Level set methods and dynamic implicit surfaces, volume 153 of Applied Mathematical Sciences. Springer-Verlag, New York, 2003.
  • [41] G. PGonzález, V. Kolehmainen, and A. Seppänen. Isotropic and anisotropic total variation regularization in electrical impedance tomography. Computers & Mathematics with Applications, 74:564–576, 2017.
  • [42] M. H. Protter. Unique continuation for elliptic equations. Trans. Amer. Math. Soc., 95(1):81–91, 1960.
  • [43] J. A. Scales, M. L. Smith, and T. L. Fischer. Global optimization methods for multimodal inverse problems. J. Computational Physics, 103:258–268, 1992.
  • [44] J. A. Sethian. Level set methods and fast marching methods, volume 3 of Cambridge Monographs on Applied and Computational Mathematics. Cambridge University Press, Cambridge, second edition, 1999. Evolving interfaces in computational geometry, fluid mechanics, computer vision, and materials science.
  • [45] P. E. Souganidis. Approximation schemes for viscosity solutions of Hamilton-Jacobi equations. J. Differential Equations, 59(1):1–43, 1985.
  • [46] J Sylvester and G. Uhlmann. A global uniqueness theorem for an inverse boundary value problem. The Annals of Mathematics, 125(1):153–169, 1987.
  • [47] H.V. Tran. Hamilton–Jacobi Equations: Theory and Applications. Graduate Studies in Mathematics. American Mathematical Society, 2021.
  • [48] Z. Zhou, G. Sato dos Santos, T. Dowrick, J. Avery, Z. Sun, H. Xu, and D. Holder. Comparison of total variation algorithms for electrical impedance tomography. Physiological measurement, 36(6):1193, 2015.