A Carleman-based numerical method for quasilinear elliptic equations with over-determined boundary data and applications
Abstract
We propose a new iterative scheme to compute the numerical solution to an over-determined boundary value problem for a general quasilinear elliptic PDE. The main idea is to repeatedly solve its linearization by using the quasi-reversibility method with a suitable Carleman weight function. The presence of the Carleman weight function allows us to employ a Carleman estimate to prove the convergence of the sequence generated by the iterative scheme above to the desired solution. The convergence of the iteration is fast at an exponential rate without the need of an initial good guess. We apply this method to compute solutions to some general quasilinear elliptic equations and a large class of first-order Hamilton-Jacobi equations. Numerical results are presented.
Key words: numerical methods; Carleman estimate; linearization; boundary value problems; quasilinear elliptic equations; Hamilton-Jacobi equations; viscosity solutions; vanishing viscosity process.
AMS subject classification: 35D40, 35F30, 35J62, 35N25, 65N12, 78A46.
1 Introduction
The main aim of this paper is to develop a numerical method based on Carleman estimates to solve quasilinear elliptic PDEs with over-determined boundary data. We consider this new method as the second generation of Carleman-based numerical methods while the first generation is called the convexification, which will be mentioned in detail later. Let be an open and bounded domain in , , with smooth boundary . Let and be two smooth functions defined on . Let be a function in the class Let be , symmetric, and positive definite, that is,
for some fixed . The following problem is of our interests.
Problem 1.1.
Assume that the over-determined boundary value problem
(1.1) |
has a solution in . Compute the function .
Problem 1.1 is motivated by a class of nonlinear inverse problems in PDEs, in which and are the data that can be measured. One important goal of inverse problems is to reconstruct the internal structure of a domain from boundary measurements, which allow us to impose both Dirichlet and Neumann data of the unknown in (1.1). Recently, a unified framework to solve such inverse problems was developed by the research group of the first and second authors, which has two main steps. In the first step, by introducing a change of variables, one derives a PDE of the form (1.1) from the given inverse problem, in which and can be computed directly by the given boundary data. In the second step, one numerically solves (1.1) to find . The knowledge of directly yields that of the solution to the corresponding inverse problem under consideration. See [15, 30] and the references therein for some works in this framework. Moreover, this unified framework was successfully tested with experimental data in [13, 14, 22]. Another motivation to study Problem 1.1 is to seek solutions to Hamilton-Jacobi equations under the circumstance that the Neumann data of the unknown can be computed by its Dirichlet data and the given form of the Hamiltonian, see e.g., [25, Assumption 1.1 and Remark 1.1]. Since inverse problems are out of the scope of this paper, we only focus on the applications in solving quasilinear elliptic PDEs and first-order Hamilton-Jacobi equations.
A natural approach to solve (1.1) is based on optimization. That means one sets the computed solution to (1.1) as a minimizer of a mismatch functional, e.g.,
subject to the Cauchy boundary conditions and The methods based on optimization are widely used in the scientific community, especially in computational mathematics, physics and engineering. Although effective and popular, the optimization-based approaches have some drawbacks:
-
1.
In general, it is not clear that the obtained minimizer approximates the true solution to (1.1).
-
2.
The mismatch functional is not convex, and it might have multiple minima and ravines (see an example in [43] for illustration). To deliver reliable numerical solutions, one must know some good initial guesses of the true solutions.
-
3.
The computation is expensive and time consuming.
Drawbacks # 1 and #2 can be treated by the convexification method, which is designed to globalize the optimization methods. The main idea of the convexification method is to employ some suitable Carleman weight functions to convexify the mismatch functionals. The convexity of weighted mismatch functionals is rigorously proved by Carleman estimates. Several versions of the convexification method have been developed in [1, 15, 16, 17, 19, 22] since it was first introduced in [21]. Moreover, we recently discovered that the convexification method can be used to solve a large class of first-order Hamilton-Jacobi equations [25].
In this paper, we introduce a new method to solve (1.1) based on linearization and Carleman estimates. Like the convexification method, our method delivers a reliable solution to (1.1) without requiring a good initial guess. This fact is rigorously proved. Unlike the convexification method which is time consuming, our new method quickly provides the desired solutions. Its converge rate is as for some
We find the numerical solution to (1.1) by repeatedly solving the linearization of (1.1) by a new “Carleman weighted” quasi-reversibility method. The classical quasi-reversibility method was first proposed in [28], and it has been studied intensively since then (see [18] for a survey). By a Carleman weighted quasi-reversibility method, we mean that we let a suitable Carleman weight function involve in the cost functional suggested by the classical quasi-reversibility method. The presence of the Carleman weight function is the key for us to prove our convergence theorem. Our process to solve Problem 1.1 is as follows. We first choose any initial solution that might be far away from the true one. Denote this initial solution by the function . Linearizing (1.1) about , we obtain a linear PDE. We then solve this linear PDE by the Carleman weighted quasi-reversibility method to obtain an updated solution . Using the Carleman weighted quasi-reversibility method rather than the classical one in this step is the key to our success. By iteration, we repeat this step to construct a sequence . The convergence of this sequence to the true solution to (1.1) is proved by using Carleman estimates. We then apply this method to numerically solve some quasilinear elliptic equations. It is important to note that our approach works well for systems of quasilinear elliptic PDEs too.
Remark 1.1.
In general, (1.1) is over-determined and it might have no solution, especially when the boundary data contains some noise. Our iteration and linearization method using the Carleman weighted quasi-reversibility method in each step still delivers a function that “most fits” (1.1).
On the other hand, (1.1) with only Dirichlet boundary condition, that is, the equation
(1.2) |
might have many solutions since we do not impose structural conditions on . For example, in case for some and , (1.2) becomes an eigenvalue problem with possibly many solutions as eigenfunctions. In such cases, requiring the additional Neumann boundary condition is then natural, and (1.1) is not over-determined.
Next, we use our method to solve some first-order Hamilton-Jacobi equations. More precisely, to find viscosity solutions to the first-order equation
we use the vanishing viscosity procedure and consider, for ,
with given Cauchy boundary data. Our new method is robust in the sense that it works for general nonlinearity that might not be convex in We refer the readers to [9, 8, 47] and the references therein for the theory of viscosity solutions. A weakness of our new approach in computing the viscosity solutions to Hamilton-Jacobi equations is that we need to require both Dirichlet and Neumann data of the unknown . See [25, Remark 1.1] for some circumstances that this requirement is fulfilled. There have been many important methods to solve Hamilton-Jacobi equations in the literature. For finite difference monotone and consistent schemes of first-order equations and applications, see [2, 10, 40, 44, 45] for details and recent developments. If is convex in and satisfies appropriate conditions, it is possible to construct some semi-Lagrangian approximations by the discretization of the Dynamical Programming Principle associated to the problem, see [11, 12] and the references therein.
The paper is organized as follows. In Section 2, we recall a Carleman estimate and two examples about inverse problems in which Problem 1.1 appears. In Section 3, we introduce the new iterative method based on linearization and Carleman estimates. In Section 4, we prove the convergence of our method. Some numerical results for quasilinear elliptic equations and first-order Hamilton-Jacobi equations are presented in Section 5. Concluding remarks are given in Section 6.
2 Preliminaries
In this section, we recall a Carleman estimate, which plays a key role for the proof of the convergence theorem in this paper. We then present an inverse scattering problem in which Problem 1.1 appears.
2.1 A Carleman estimate
In this section, we present a simple form of Carleman estimates. Carleman estimates were first employed to prove the unique continuation principle, see e.g., [6, 42], and they quickly became a powerful tool in many areas of PDEs afterwards. Let be a point in such that for all For each , define
(2.1) |
We have the following lemma.
Lemma 2.1 (Carleman estimate).
There exist positive constants depending only on , , , and such that for all function satisfying
(2.2) |
the following estimate holds true
(2.3) |
for all and . Here, depends only on the listed parameters.
Proof.
Lemma 2.1 is a direct consequence of [36, Lemma 5]. Let and such that . Here, for Extend to the whole such that for all . Using a change of variable and [36, Lemma 5], there exists a number depending on , and such that for all and , we have
(2.4) |
for some constant depending only on and . Since on and since , allowing to depend on and , we deduce the Carleman estimate (2.3) from (2.4). ∎
An alternative way to obtain (2.3) is to apply the Carleman estimate in [29, Chapter 4, §1, Lemma 3] for general parabolic operators. The arguments to obtain (2.3) using [29, Chapter 4, §1, Lemma 3] are similar to that in [32, Section 3] with the Laplacian replaced by the operator .
Remark 2.1.
We specially draw the reader’s attention to different forms of Carleman estimates for all three main kinds of differential operators (elliptic, parabolic and hyperbolic) and their applications in inverse problems and computational mathematics [3, 4, 23]. It is worth mentioning that some Carleman estimates hold true for all functions satisfying and where is a part of , see e.g., [25, 39], which can be used to solve quasilinear elliptic PDEs with the boundary data partly given.
2.2 An inverse scattering problem
As mentioned in Section 1, Problem 1.1 arises from nonlinear inverse problems. We present here an important example in the context of inverse scattering problems in the frequency domain. Let be the spatially distributed dielectric constant of the medium and be an interval of wavenumbers with . Since the dielectric constant of the air is 1, we set for all For each let , , be the wave field generated by a point source at with wavenumber . The function is governed by the Helmholtz equation and the Sommerfeld radiation condition
(2.5) |
The inverse scattering problem is formulated as follows.
Problem 2.1 (Inverse Scattering Problem).
Compute the function , , from the measurements of
(2.6) |
for all ,
The knowledge of the function partly provides the internal structure of the domain . In other words, solving the inverse scattering problem allows us to examine a domain from external measurements, which has applications in security, sonar imaging, geographical exploration, medical imaging, near-field optical microscopy, nano-optics, see, e.g., [7] and references therein for more details. There have been many important methods to solve inverse scattering problems in the literature. Each method has its own advantages and disadvantages. A common drawback of the widely-used method based on optimization to solve inverse scattering problems is the need of a good initial guess of the true solution . We recall from [31] a method to solve the above inverse scattering problem in which such a need is relaxed. Denote by
where Then, satisfies
for all Let be the orthonormal basis of introduced in [20] and define
(2.7) |
We approximate
for a suitable cut-off number . Then, the vector “approximately” satisfies the system
(2.8) |
where
for all and , see [31, Section 6] for details. The Dirichlet and Neumann boundary conditions for can be computed by the knowledges of , , and (2.7). Solving the system of quasilinear elliptic equations (2.8) with the provided Dirichlet and Neumann data is basically a goal of Problem 1.1. Doing so is the key step to compute . See [31] for convexification method to compute and the procedure to obtain from the knowledge of
2.3 Electrical impedance tomography
We present here another potential application of our study in this paper to the 3D electrical impedance tomography (EIT) problem, so-called the 3D Calderón problem, with only a part of the Dirichlet to Neumann map is given. Let for some . Let the line of source be defined as
For each , let and , , be the solution to
(2.9) |
Here, is a smooth approximation of the Dirac delta . In EIT, represents the boundary electric voltage. The EIT problem is formulated as follows.
Problem 2.2 (Electrical impedance tomography).
Determine the electric conductivity , from the boundary measurement of the electric current for all
Remark 2.2.
Problem 2.2 only requests the data generated by the source moving on . The dimension of this set of data is 3, including 1 dimension of the orbit of the source and 2 dimensions of the measurement surface . This feature makes Problem 2.2 not over-determined because our goal is to reconstruct a 3D function . This is unlike most of the works studying the EIT problem that request the whole Dirichlet to Neumann map . Since , the domain of , has uncountably infinity dimensions, the data requested by the Dirichlet to Neumann map is highly over-determined.
The EIT problem arises from bio-medical imaging; especially, in detecting early cancerous tumors in living tissues without operation. Most publications for the EIT problem studied the question how to reconstruct the electric conductivity , for some domain , from the Dirichlet to Neumann or the Neumann to Dirichlet map. We refer the reader to [5, 27, 35, 46] for some well-known uniqueness results of the EIT problem. We list here a few publications for some effective approaches to numerically solve the EIT problem: the D-bar method [26, 33, 34], the methods based on optimization [41, 48] and the convexification method [24]. Although effective, these methods have drawbacks. Firstly, the D-bar method is designed only for 2D. Secondly, the methods based on optimization request good initial guess of the true solution. And finally, the convexification method is time consuming. Naturally, a new computational method should be studied in this direction. Our potential method to solve the 3D EIT problem is to reduce this inverse problem to a problem of the form (1.1). Introduce the change of variable
(2.10) |
Let be the orthonormal basis of first introduced in [20]. Like in Section 2.2, we approximate the function as
(2.11) |
for all , for some cut-off number where
(2.12) |
for all We choose such that the approximation in (2.11) “numerically” holds for all where the data are known. Then, it is not hard to verify, see [24], that the vector satisfies the system
(2.13) |
for each , where
The Dirichlet and Neumann boundary conditions for can be computed by the knowledges of the Dirichlet condition , the given Neumann data , (2.10) and (2.12). Solving the system of quasilinear elliptic equations (2.13) with the provided Dirichlet and Neumann data is basically a goal of Problem 1.1. We refer the reader to [24] the step of computing from the knowledge of the vector .
The discussion above for the inverse scattering problem and the EIT problem motivates us to study Problem 1.1. Since these inverse problems are out of the scope of this paper, we only mention them here to explain the significance of Problem 1.1. In future works, we will use our solver for Problem 1.1 to solve these two important inverse problems.
3 The iteration and linearization approach for Problem 1.1
Our approach to solve (1.1) is based on linearization and iteration. Assume that the solution to (1.1) is in the space for some where is the smallest integer that is greater than Define the set of admissible solutions
(3.1) |
Then, the assumption in Problem 1.1 implies that , and . We now construct a sequence that converges to the solution to (1.1). Take a function Assume by induction that we have the knowledge of for some . We find as follows. Assume that for some where
Plugging into (1.1), we have
(3.2) |
for all Heuristically, we assume at this moment that is small, that is, .
Remark 3.1.
The temporary assumption is imposed for the suggestion in establishing a numerical scheme to find while this condition is completely relaxed in the proof of the convergence theorem.
By Taylor’s expansion, we approximate (3.2) as
(3.3) |
for all where
for all . Here, and are the partial derivative of with respect to its second variable and its gradient vector with respect to the third variable, respectively.
The next step is to compute a function satisfying (3.3). Since there is no guarantee for the existence of such a function , we only compute a “best fit” function by the Carleman-based quasi-reversibility method described below. For each , , and , define the functional as
Here, the function is defined in (2.1) and is a regularization term.
Proposition 3.1.
For all and , for each the functional has a unique minimizer.
Proof.
It is obvious that is coercive and weakly lower semicontinuous. Thus, it has a minimizer in . The uniqueness of the minimizer can be deduced from the strict convexity of in . ∎
For each , thanks for Proposition 3.1, we can minimize on . The unique minimizer is the desired function We then set
(3.4) |
The construction of the sequence above is summarized in Algorithm 1. We will prove that the sequence converges to in Section 4 as and . The presence of the Carleman weight is a key point for us to prove this convergence result.
4 The convergence analysis
In this section, we prove that the sequence generated by Algorithm 1 converges to the solution to (1.1). The following result is the main theorem in this paper.
Theorem 4.1.
Proof.
Fix . Let . Due to Step 3 in Algorithm 1, is the minimizer of . By the variational principle, we have
(4.2) |
for all Since , we can rewrite (4.2) as
(4.3) |
for all As is the solution to (1.1), we have
(4.4) |
for all It follows from (4.3) and (4.4) that for all
(4.5) |
Take . It follows from (4.5) that
(4.6) |
As , we can estimate
and | ||
These estimates, together with the inequality , imply
(4.7) |
Applying the Carleman estimate (2.3) for the function , we have
(4.8) |
Combining (4.7) and (4.8), we have
(4.9) |
Letting be sufficiently large, we can simplify (4.9) as
(4.10) |
Recall that . We get from (4.10) that
(4.11) |
Applying (4.11) for and denoting , we have
By induction, we have
(4.12) |
which implies (4.1). The proof is complete. ∎
Remark 4.1 (Removing the boundedness condition of in in Theorem 4.1).
In the case when , we need to assume that we know in advance that the true solution to (1.1) belongs to the ball of for some . This assumption does not weaken the result since can be arbitrarily large. Define the cut-off function as
and set Since , it is obvious that solves the problem
(4.13) |
Remark 4.2.
Remark 4.3.
As seen in the proof of Theorem 4.1, the efficiency of Algorithm 1 is guaranteed by Carleman estimate (2.3). Therefore, we call the proposed method described in Algorithm 1 the second generation of Carleman-based numerical methods. This second generation includes the method in [30, 38] in which the iteration scheme is developed based on the contraction principle and Carleman estimates. The first generation of Carleman-based numerical method was developed in [21], which is called the convexification. See [1, 25, 30, 31] for some following up results. Like the convexification method, Algorithm 1 can be used to compute solutions to nonlinear PDEs without requesting an initial good guess. The advantage of our new method is the fast convergence rate, see Remark 4.2.
5 Numerical study
In this section, we present some numerical results obtained by Algorithm 1. For simplicity, we set and . On we arrange an grid points
where In our numerical scripts, .
In Step 1 of Algorithm 1, we choose a function . It is natural to find a function satisfying the equation obtained by removing from (1.1) the nonlinearity . We apply the Carleman-based quasi-reversibility method to do so. That means, is the minimizer of
(5.1) |
subject to the boundary conditions in (5.7). To simplify the efforts in implementation, the norm in the regularization term is the -norm rather than the -norm. This change does not affect the performance of Algorithm 1. Algorithm 1 still provides satisfactory solutions to (1.1).
Remark 5.1.
We employ the Carleman-based quasi-reversibility method to find for the consistency to Step 3 of Algorithm 1. Since can be chosen arbitrarily, one can use the quasi-reversibility method without the presence of the Carleman weight function . In our computation for all numerical examples below, , and The regularized parameter is . The threshold number
We minimize on by the least square MATLAB command “lsqlin”. The implementation for the quasi-reversibility method to minimize more general functionals than was described in [30, §5.3] and in [37, §5]. We do not repeat this process here. In Step 3 of Algorithm 1, given we minimize the functional on . Again, we refer the reader to [30, §5.3] and in [37, §5] for details in implementation. The scripts for other steps of Algorithm 1 can be written easily.
5.1 Quasilinear elliptic equations
The convexification method, first introduced in [21], was used to numerically solve quasilinear elliptic equations in [1, 25, 31]. Our approach here is of course different based on iteration and linearization. In particular, Step 3 in Algorithm 1 is very efficient as we only need to solve the linear equation (3.3) as opposed to solving directly a nonlinear quasilinear elliptic equation. In this subsection, we present two (2) numerical tests. In both tests, we choose the matrix to be
That means,
In test 1, we solve (1.1) when
(5.2) |
for all , and . The boundary conditions are given by
(5.3) |
for all The true solution of (1.1) is the function .



It is evident from Figure 1 that Algorithm 1 provides out of expectation solution for test 1. The relative error . One can see from Figure 1c that Algorithm 1 converges at the third iteration.
In test 2, we solve (1.1) when
(5.4) |
for all , and . The boundary conditions are given by
(5.5) |
for all The true solution of (1.1) is the function .



5.2 Application to first-order Hamilton-Jacobi equations
Our aim in this subsection is to solve numerically
(5.6) |
with the boundary conditions
(5.7) |
Basically, we use the vanishing viscosity process to approximate solutions to (5.6). For , consider
(5.8) |
with boundary conditions (5.7). Again, (5.8)–(5.7) is an over-determined boundary value problem. For sufficiently small, assume that (5.8)–(5.7) has a solution . Then, approximates , solution to (5.6)–(5.7), quite well under some appropriate assumptions on . In our numerical tests, we choose .
In this part, we provide six (6) numerical tests, in which we compute the viscosity solution to some Hamilton-Jacobi equations of the form (5.6) given Cauchy boundary data. That means, by applying Algorithm 1, we numerically find a function satisfying (5.8)-(5.7) when , and are given. The verification that is the correct viscosity solution can be done in a similar manner as that in [47, 25].
Test 1. In this test, we solve (5.8)-(5.7) when
(5.9) |
for all . The boundary conditions are given by
(5.10) |
for all . In this case, the true solution is The numerical result is displayed in Figure 3.






It is evident from Figure 3 that we successfully compute the solution . The procedure described in Algorithm 1 converges after four iterations. The relative error
Test 2. We find the viscosity solution to the eikonal equation. In this test, we solve (5.8)-(5.7) when the Hamiltonian is
(5.11) |
for all . The boundary conditions are given by
(5.12) |
for all . The true solution is . The graphs of and are displayed in Figure 4.






The relative error
Test 3. We test the case when the Hamiltonian is not convex with respect its third variable. The Hamiltonian in this test is given by
(5.13) |
for all . The boundary conditions are given by
(5.14) |
and
(5.15) |
for all . The true solution is . The graphs of and are displayed in Figure 5.






The relative error Although the relative error in this test is large, the numerical result is acceptable. In fact, we observe from Figure 5f that is small almost everywhere while it is large at only two small places near the left edge of the domain.
Test 4. We test the case when the Hamiltonian is not increasing in the second variable and is not convex in the third variable. The Hamiltonian in this test is given by
(5.16) |
for all . The boundary conditions are given by
(5.17) |
and
(5.18) |
for all . The true solution is . The graphs of and are displayed in Figure 6.






The relative error
Test 5. We solve a G-equation. The Hamiltonian in this test is given by
(5.19) |
for all . The boundary conditions are given by
(5.20) |
and
(5.21) |
for all . The true solution is . The graphs of and are displayed in Figure 5.






The relative error
Test 6. The Hamiltonian in this test is given by
(5.22) |
for all where
The boundary conditions are given by
(5.23) |
and
(5.24) |
for all . The true solution is . The graphs of and are displayed in Figure 8.






The relative error
Remark 5.2.
The relative errors in all tests above for first-order Hamilton-Jacobi equations are compatible with
6 Concluding remarks
We have developed a new globally convergent numerical method to solve over-determined boundary value problems of quasilinear elliptic equations. The key point of the method is to repeatedly solve the linearization of the given PDE by the Carleman weighted quasi-reversibility method (Algorithm 1). As the result, we obtain a sequence of functions converging to the solution thanks to the Carleman estimate (Lemma 2.1). The strength of our new method includes (1) the global convergence property and (2) the fast convergence rate, which is described in Theorem 4.1. Some numerical results for quasilinear elliptic equations and first-order Hamilton-Jacobi equations are presented to show the effectiveness of our method.
Acknowledgement
The works of TTL and LHN were supported by US Army Research Laboratory and US Army Research Office grant W911NF-19-1-0044 and , in part, by funds provided by the Faculty Research Grant program at UNC Charlotte, Fund No. 111272. The work of HT is supported in part by NSF CAREER grant DMS-1843320 and a Simons fellowship.
References
- [1] A. B. Bakushinskii, M. V. Klibanov, and N. A. Koshev. Carleman weight functions for a globally convergent numerical method for ill-posed Cauchy problems for some quasilinear PDEs. Nonlinear Anal. Real World Appl., 34:201–224, 2017.
- [2] G. Barles and P. E. Souganidis. Convergence of approximation schemes for fully nonlinear second order equations. Asymptotic Anal., 4(3):271–283, 1991.
- [3] L. Beilina and M. V. Klibanov. Approximate Global Convergence and Adaptivity for Coefficient Inverse Problems. Springer, New York, 2012.
- [4] A. L. Bukhgeim and M. V. Klibanov. Uniqueness in the large of a class of multidimensional inverse problems. Soviet Math. Doklady, 17:244–247, 1981.
- [5] A. P. Calderón. On an inverse boundary value problem. In Seminar on Numerical Analysis and its Applications to Continuum Physics, Río de Janeiro. Soc. Brasileira de Matemática, 1980.
- [6] T. Carleman. Sur les systèmes linéaires aux derivées partielles du premier ordre a deux variables. C. R. Acad. Sci. Paris, 197:471–474, 1933.
- [7] David Colton and Rainer Kress. Inverse acoustic and electromagnetic scattering theory. Applied Mathematical Sciences. Springer, New York, 3rd edition, 2013.
- [8] M. G. Crandall, L. C. Evans, and P.-L. Lions. Some properties of viscosity solutions of Hamilton-Jacobi equations. Trans. Amer. Math. Soc., 282(2):487–502, 1984.
- [9] M. G. Crandall and P.-L. Lions. Viscosity solutions of Hamilton-Jacobi equations. Trans. Amer. Math. Soc., 277(1):1–42, 1983.
- [10] M. G. Crandall and P.-L. Lions. Two approximations of solutions of Hamilton-Jacobi equations. Math. Comp., 43(167):1–19, 1984.
- [11] M. Falcone and R. Ferretti. Semi-Lagrangian schemes for Hamilton-Jacobi equations, discrete representation formulae and Godunov methods. J. Comput. Phys., 175(2):559–575, 2002.
- [12] M. Falcone and R. Ferretti. Semi-Lagrangian approximation schemes for linear and Hamilton-Jacobi equations. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 2014.
- [13] V. A. Khoa, G. W. Bidney, M. V. Klibanov, L. H. Nguyen, L. Nguyen, A. Sullivan, and V. N. Astratov. Convexification and experimental data for a 3D inverse scattering problem with the moving point source. Inverse Problems, 36:085007, 2020.
- [14] V. A. Khoa, G. W. Bidney, M. V. Klibanov, L. H. Nguyen, L. Nguyen, A. Sullivan, and V. N. Astratov. An inverse problem of a simultaneous reconstruction of the dielectric constant and conductivity from experimental backscattering data. Inverse Problems in Science and Engineering, 29(5):712–735, 2021.
- [15] V. A. Khoa, M. V. Klibanov, and L. H. Nguyen. Convexification for a 3D inverse scattering problem with the moving point source. SIAM J. Imaging Sci., 13(2):871–904, 2020.
- [16] M. V. Klibanov. Global convexity in a three-dimensional inverse acoustic problem. SIAM J. Math. Anal., 28:1371–1388, 1997.
- [17] M. V. Klibanov. Global convexity in diffusion tomography. Nonlinear World, 4:247–265, 1997.
- [18] M. V. Klibanov. Carleman estimates for the regularization of ill-posed Cauchy problems. Applied Numerical Mathematics, 94:46–74, 2015.
- [19] M. V. Klibanov. Carleman weight functions for solving ill-posed Cauchy problems for quasilinear PDEs. Inverse Problems, 31:125007, 2015.
- [20] M. V. Klibanov. Convexification of restricted Dirichlet to Neumann map. J. Inverse and Ill-Posed Problems, 25(5):669–685, 2017.
- [21] M. V. Klibanov and O. V. Ioussoupova. Uniform strict convexity of a cost functional for three-dimensional inverse scattering problem. SIAM J. Math. Anal., 26:147–179, 1995.
- [22] M. V. Klibanov, T. T. Le, L. H. Nguyen, A. Sullivan, and L. Nguyen. Convexification-based globally convergent numerical method for a 1D coefficient inverse problem with experimental data. to appear on Inverse Problems and Imaging, DOI: https://www.aimsciences.org/article/doi/10.3934/ipi.2021068, 2021.
- [23] M. V. Klibanov and J. Li. Inverse Problems and Carleman Estimates: Global Uniqueness, Global Convergence and Experimental Data. De Gruyter, 2021.
- [24] M. V. Klibanov, J. Li, and W. Zhang. Convexification of electrical impedance tomography with restricted Dirichlet-to-Neumann map data. Inverse Problems, 35:035005, 2019.
- [25] M. V. Klibanov, L. H. Nguyen, and H. V. Tran. Numerical viscosity solutions to Hamilton-Jacobi equations via a Carleman estimate and the convexification method. Journal of Computational Physics, 451:110828, 2022.
- [26] K. Knudsen, M. Lassas, J. L. Mueller, and S. Siltanen. Regularized D-bar method for the inverse conductivity problem. Inverse Problems, 3:599–624, 2009.
- [27] R. V. Kohn and M. Vogelius. Determining conductivity by boundary measurements II. Interior results. Comm. Pure & Applied Math., 38:643–667, 1985.
- [28] R. Lattès and J. L. Lions. The Method of Quasireversibility: Applications to Partial Differential Equations. Elsevier, New York, 1969.
- [29] M. M. Lavrent’ev, V. G. Romanov, and S. P. Shishatskiĭ. Ill-Posed Problems of Mathematical Physics and Analysis. Translations of Mathematical Monographs. AMS, Providence: RI, 1986.
- [30] T. T. Le and L. H. Nguyen. A convergent numerical method to recover the initial condition of nonlinear parabolic equations from lateral Cauchy data. Journal of Inverse and Ill-posed Problems, DOI: https://doi.org/10.1515/jiip-2020-0028, 30(2):265–286, 2022.
- [31] T. T. Le and L. H. Nguyen. The gradient descent method for the convexification to solve boundary value problems of quasi-linear PDEs and a coefficient inverse problem. to appear in Journal of Scientific Computing, preprint Arxiv:2103.04159, 2022.
- [32] T. T. Le, L. H. Nguyen, T-P. Nguyen, and W. Powell. The quasi-reversibility method to numerically solve an inverse source problem for hyperbolic equations. Journal of Scientific Computing, 87:90, 2021.
- [33] J. L. Mueller and S. Siltanen. The D-bar method for electrical impedance tomography-demystified. Inverse Problems, 36:093001, 2020.
- [34] E. K. Murphy and J. L. Mueller. Effect of domain shape modeling and measurement errors on the 2-D D-bar method for EIT. IEEE Trans. Med. Imaging, 28:1576–1584, 2009.
- [35] A. I. Nachman. Global uniqueness for a two-dimensional inverse boundary value problem. Ann. of Math., 143:71–96, 1996.
- [36] H. M. Nguyen and L. H. Nguyen. Cloaking using complementary media for the Helmholtz equation and a three spheres inequality for second order elliptic equations. Transaction of the American Mathematical Society, 2:93–112, 2015.
- [37] L. H. Nguyen. A new algorithm to determine the creation or depletion term of parabolic equations from boundary measurements. Computers and Mathematics with Applications, 80:2135–2149, 2020.
- [38] L. H. Nguyen and M. V. Klibanov. Carleman estimates and the contraction principle for an inverse source problem for nonlinear hyperbolic equations. Inverse Problems, 38:035009, 2022.
- [39] L. H. Nguyen, Q. Li, and M. V. Klibanov. A convergent numerical method for a multi-frequency inverse source problem in inhomogenous media. Inverse Problems and Imaging, 13:1067–1094, 2019.
- [40] S. Osher and R. Fedkiw. Level set methods and dynamic implicit surfaces, volume 153 of Applied Mathematical Sciences. Springer-Verlag, New York, 2003.
- [41] G. PGonzález, V. Kolehmainen, and A. Seppänen. Isotropic and anisotropic total variation regularization in electrical impedance tomography. Computers & Mathematics with Applications, 74:564–576, 2017.
- [42] M. H. Protter. Unique continuation for elliptic equations. Trans. Amer. Math. Soc., 95(1):81–91, 1960.
- [43] J. A. Scales, M. L. Smith, and T. L. Fischer. Global optimization methods for multimodal inverse problems. J. Computational Physics, 103:258–268, 1992.
- [44] J. A. Sethian. Level set methods and fast marching methods, volume 3 of Cambridge Monographs on Applied and Computational Mathematics. Cambridge University Press, Cambridge, second edition, 1999. Evolving interfaces in computational geometry, fluid mechanics, computer vision, and materials science.
- [45] P. E. Souganidis. Approximation schemes for viscosity solutions of Hamilton-Jacobi equations. J. Differential Equations, 59(1):1–43, 1985.
- [46] J Sylvester and G. Uhlmann. A global uniqueness theorem for an inverse boundary value problem. The Annals of Mathematics, 125(1):153–169, 1987.
- [47] H.V. Tran. Hamilton–Jacobi Equations: Theory and Applications. Graduate Studies in Mathematics. American Mathematical Society, 2021.
- [48] Z. Zhou, G. Sato dos Santos, T. Dowrick, J. Avery, Z. Sun, H. Xu, and D. Holder. Comparison of total variation algorithms for electrical impedance tomography. Physiological measurement, 36(6):1193, 2015.