Local well-posedness for quasilinear problems: a primer
Abstract.
Proving local well-posedness for quasilinear problems in pde’s presents a number of difficulties, some of which are universal and others of which are more problem specific. On one hand, a common standard for what well-posedness should mean has existed for a long time, going back to Hadamard. On the other hand, in terms of getting there, there are by now both many variations, but also many misconceptions.
The aim of these expository notes is to collect a number of both classical and more recent ideas in this direction, and to assemble them into a cohesive road-map that can be then adapted to the reader’s problem of choice.
Key words and phrases:
quasilinear evolutions, local well-posedness, frequency envelopes1991 Mathematics Subject Classification:
Primary: 35L45, 35L50, 35L60.1. Introduction
Local well-posedness is the first question to ask for any evolution problem in partial differential equations. These notes, prepared by the authors for a summer school at MSRI [13] in 2020, aim to discuss ideas and strategies for local well-posedness in quasilinear and fully nonlinear evolution equations, primarily of hyperbolic type. We hope to persuade the reader that the structure presented here should be adopted as the standard for proving these results. Of course, there are many possible variations, and we try to point out some of them in our many remarks. While a few of the ideas here can be found in several of the classical books, see e.g. [30],[10], [3],[24], some of the others have appeared only in articles devoted to specific problems, and have never been collected together, to the best of our knowledge.
1.1. Nonlinear evolutions
For our exposition we will adopt a two track structure, where we will broadly discuss ideas for a general problem, and in parallel implement these ideas on a simple, classical concrete example.
Our general problem will be a nonlinear partial differential equation of the form
(1.1) |
i.e. a first order system in time, where we think of as a scalar or a vector valued function belonging to a scale of either real or complex Sobolev spaces. This scale will be chosen to be for the purpose of this discussion, though in practice it often has to be adapted to the class of problems to be considered. The nonlinearity represents a nonlinear function of and its derivatives,
where we will refer to as the order of the evolution. Here typical examples include (hyperbolic equations), (Schrödinger type evolutions) and (KdV type evolutions). But many other situations arise in models which are nonlocal, e.g. in water waves one encounters for gravity waves respectively for capillary waves.
Some problems are most naturally formulated as second order evolutions in time, for instance nonlinear wave equations. While some such problems admit also good first order in time formulations (e.g. the compressible Euler flow), it is sometimes better to treat them as second order. Regardless, our road-map still applies, with obvious adjustments.
Our model problem will be a classical first order symmetric hyperbolic system in , of the form
(1.2) |
where takes values in and the matrices are symmetric, and smooth as functions of . Here the order of the nonlinearity is , and the scale of Sobolev spaces to be used is indeed the Sobolev scale.
1.2. What is well-posedness ?
To set the expectations for our problems, we recall the classical Hadamard standard for well-posedness, formulated relative to our chosen scale of spaces.
Definition 1.1.
The problem (1.1) is locally well-posed in a Sobolev space if the following properties are satisfied:
- (i) Existence:
-
For each there exists some time and a solution .
- (ii) Uniqueness:
-
The above solution is unique in .
- (iii) Continuous dependence:
-
The data to solution map is continuous from to .
As a historical remark, we note that Hadamard primarily discussed the question of well-posedness in the context of linear pde’s, specifically for the Laplace and wave equation, beginning with an incipient form in [8], and a more developed form in [9]. It is in the latter reference where the continuous dependence is discussed, seemingly inspired by Cauchy’s theorem for ode’s.
The above definition should not be taken as universal, but rather as a good starting point, which may need to be adjusted depending on the problem. Consider for instance the uniqueness statement, which, as given in (ii), is in the strongest form, which is often referred to as unconditional uniqueness. Often this may need to be relaxed somewhat, particularly when low regularity solutions are concerned. Some common variations concerning uniqueness are as follows:
-
a)
The solutions in (i) are shown to belong to a smaller space, , and then the uniqueness in (ii) holds in the same class.
-
b)
Unconditional uniqueness holds apriori only in a more regular class with , but the data to solution map extends continuously as a map from to .
Since we are discussing nonlinear equations here, the lifespan of the solutions need not be infinite, i.e. there is always the possibility that solutions may blow up in finite time. In particular, in the context of well-posed problems it is natural to consider the notion of maximal lifespan, which is the largest for which the solution exists in ; here the limit of as approaches cannot exist, or else the solution may be continued further.
In this context, the last property in the definition should be interpreted to mean in particular that, for a solution , small perturbations of the initial data yield solutions which are also defined in . This in turn implies that the maximal lifespan is lower semicontinuous as a function of .
In view of the above discussion, it is always interesting to provide more precise assertions about the lifespan of solutions, or, equivalently, continuation (or blow-up) criteria for the solutions. Some interesting examples are as follows:
-
a)
The lifespan is bounded from below uniformly for data in a bounded set,
This implies a blow-up criteria as follows:
-
b)
The blow-up may be characterized in terms of weaker bounds,
relative to a Banach topology , or perhaps a time integrated version thereof
To conclude our discussion of the above definition, we note that many well-posedness statements also provide additional properties for the flow:
- Higher regularity:
-
if the initial data has more regularity with , then this regularity carries over to the solution, , with bounds and lifespan bounds depending only on the size of the data.
- Weak Lipschitz bounds:
-
on bounded sets in , the flow is Lipschitz in a weaker topology (e.g. up to in our model problem).
Both of these properties are often an integral part of a complete theory, and frequently also serve as intermediate steps in establishing the main well-posedness result.
In all of the above discussion, a common denominator remains the fact that the solution to data map is locally continuous, but not uniformly continuous. It is very natural indeed to redefine (expand) the notion of quasilinear evolution equations to include all flows which share this property.
In many problems of this type, one is interested not only in local well-posedness in some Sobolev space , but also in lowering the exponent as much as possible. We will refer to such solutions as rough solutions. Then, a natural question is what kind of regularity thresholds should one expect or aim for in such problems? One clue in this direction comes from the scaling symmetry, whenever available. As an example, our model problem exhibits the scaling symmetry
The scale invariant initial data Sobolev space corresponding to this symmetry is the homogeneous space , where . This space is called the critical Sobolev space, and should heuristically be thought of as an absolute lower bound for any reasonable well-posedness result. Whereas in some semilinear dispersive evolutions one can actually reach this threshold, in nonlinear flows it seems to be out of reach in general.
1.3. A set of results for the model problem
In order to state the results, we begin with a discussion of control parameters. We will use two such control parameters. The first one is
This is a scale invariant quantity, which appears in the implicit constants in all of our bounds. Our second control parameter is
which instead will be shown to control the energy growth in all the energy estimates. Precisely, the norm , plays the role of the norm mentioned in the discussion above.
The primary well-posedness result for the model problem is as follows:
Theorem 1.
The equation (1.2) is locally well-posed in in the Hadamard sense for .
The reader will notice that this result is one derivative above scaling. It is also optimal in some cases, including the scalar case (where the problem can be solved locally using the method of characteristics), but not optimal in many other cases where the system is dispersive.
For the uniqueness result we have in effect a stronger statement, which only requires Lipschitz bounds for . This however does not improve the scaling comparison relative to the critical spaces:
Theorem 2.
Uniqueness holds in the Lipschitz class, and we have the difference bound
(1.3) |
This is exactly the kind of weak Lipschitz bound discussed earlier. With a bit of additional effort, for the solutions in Theorem 1 this may be extended to a larger range of Sobolev spaces,
(1.4) |
The small price to pay here is that now the implicit constant in the estimate depends not only on and but also on the norms of and in .
A key role in the proof of the well-posedness result is played by the energy estimates, which are also of independent interest:
Theorem 3.
The following bounds hold for for solutions to (1.2) for all :
(1.5) |
Finally, as a corollary of the last result, we obtain a continuation criteria for solutions:
Theorem 4.
Solutions can be continued in for as long as remains finite.
Theorem 1 has been first proved by Kato [16], borrowing ideas from nonlinear semigroup theory, see e.g. Barbu’s book [4]. The existence and uniqueness part, as well as the energy estimates, can also be found in standard references, e.g. in the books of Taylor [30], Hörmander [10] and Sogge [24] (in the last two the wave equation is considered, but the idea is similar). However, interestingly enough, the continuous dependence part is missing in all these references. We did find presentations of continuous dependence arguments inspired from Kato’s work in Chemin’s book [3], and also on Tao’s blog [26].
Our objective for remainder of the paper will be to provide complete proofs for the four theorems above, which the reader may take as a guide for his problem of choice. While these results are not new in the model case we consider, to the best of our knowledge this is the first time when the proofs of these results are presented in this manner. Along the way, we will also provide extensive comments and pointers to alternative methods developed along the years.
In particular, we would emphasize the frequency envelope approach for the regularization and continuous dependence parts, as well as the time discretization approach for the existence proof. The frequency envelope approach has been repeatedly used by the authors, jointly with different collaborators, in a number of papers, see e.g. [23], [29],[18], [12], [15], with some of the ideas crystalizing along the way. The version of the existence proof based on a time discretization is in some sense very classical, going back to ideas which have originally appeared in the context of semigroup theory; however, its implementation is inspired from the authors’ recent work [15], though the situation considered here is considerably simpler.
1.4. An outline of these notes
Our strategy will be, in each section, to provide some ideas and a broader discussion in the context of the general equation (1.1), and then show how this works in detail in the context of our chosen example (1.2).
In the next section we introduce the paradifferential form of our equations, both the main equation and its linearization. This is an idea that goes back to work of Bony [6], and helps clarify the roles played by different frequency interaction modes in the equation. Another very useful reference here is Metivier’s more recent book [21].
Section 3 is devoted to the energy estimates, in multiple contexts. These are presented both for the full equation, for its linearization, for its associated linear paradifferential flow, and for differences of solutions. The latter, in turn, yields the uniqueness part of the well-posedness theorem. A common misconception here has been that for well-posedness it suffices to prove energy estimates for the full equation. Instead, in our presentation we regard the bound for the linearized problem as fundamental, though, at the implementation level, it is the paradifferential flow bound which can be found at the core.
Section 4 provides two approaches for the existence part of the well-posedness theorem. The first one, more classical, is based on an iteration scheme, which works well on our model problem but may run into implementation issues in more complex problems. The second approach, which we regard as more robust, relies on a time discretization, and is somewhat related to nonlinear semigroup theory, which also inspired Kato’s work. Two other possible strategies, which have played a role historically, are briefly outlined.
Section 5 introduces Tao’s notion of frequency envelopes (see for example [27]), which is very well suited to track the flow of energy as time progresses. This is then used to show how rough solutions can be obtained as uniform limits of smooth solutions. This is a key step in many well-posedness arguments, and helps decouple the regularity for the initial existence result from the rough data results.
Finally the last section of the paper is devoted to the continuous dependence result, where we provide the modern, frequency envelopes based approach. At the same time, for a clean, elegant reinterpretation of Kato’s original strategy we refer the reader to Tao’s blog [26].
1.5. Acknowledgements
The first author was supported by a Luce Professorship, by the Sloan Foundation, and by an NSF CAREER grant DMS-1845037. The second author was supported by the NSF grant DMS-1800294 as well as by a Simons Investigator grant from the Simons Foundation. Both authors are extremely grateful to MSRI for their full support in holding the graduate summer school “Introduction to water waves” in a virtual format due to the less than ideal circumstances.
2. A menagerie of related equations
While ultimately one would want all the results stated in terms of the full nonlinear equation, any successful approach to quasilinear problems needs to also consider a succession of closely related linear equations, as well as associated reformulations of the nonlinear flow. Here we aim to motivate and describe these related flows, stripping away technicalities.
2.1. The linearized equation
This plays a key role in comparing different solutions; we will write it in the form
(2.1) |
where stands for the differential of , which in our setting is a partial differential operator of order . One may also reinterpret the equation for the difference of two solutions as a perturbed linearized equation with a quadratic source term; some caution is required here, because often some structure is lost in doing this, and the question is whether that is not too much.
In the particular case of (1.2), the linearized equation takes the form
(2.2) |
2.2. The linear paradifferential equation
One distinguishing feature of quasilinear evolutions is that the nonlinearity cannot be interpreted as perturbative. Nevertheless, one may seek to separate parts of the nonlinearity which can be seen as perturbative, at least at high regularity, in order to better isolate and understand the nonperturbative part.
To narrow things down, consider a nonlinear term which is quadratic, say of the form , and consider the three modes of interaction between these terms, according to the Littlewood-Paley trichotomy, or paraproduct decomposition,
where the three terms represent the -, - respectively the - frequency interactions. The high-high interactions in the last term are always perturbative at high regularity, so are placed into the perturbative box. But one cannot do the same with the low-high or high-low interactions, which are kept on the nonperturbative side. This is closely related to the linearization, and indeed, at the end of the day, we are left with a paradifferential style nonperturbative part of our evolution, which we can formally write as
(2.3) |
Here, one can naively use Bony’s notion of paraproduct [6] to define the linear operator as
where is a placeholder for the argument of the nonlinearity . However there are also other related choices one can make, see for instance the discussion at the end of this subsection. For a discussion on the use of paradifferential calculus in nonlinear PDE’s (though not the above notation) we refer the reader to Metivier’s book [21].
One can think of the above evolution as a linear evolution of high frequency waves on a low frequency background. Then one can interpret solving the nonperturbative part of our evolution as an infinite dimensional triangular system, where each dyadic frequency of the solution is obtained at some step by solving a linear system with coefficients depending only on the lower components, and in turn it affects the coefficients of the equations for the higher frequency components. Of course, this should only be understood in a philosophical sense, because a variable coefficient flow in general does not preserve frequency localizations. This can sometimes be achieved with careful choices of the paraproduct quantizations, but it never seems worthwhile to implement, as the perturbative terms will mix frequencies anyway and add tails.
Turning to our model problem, in a direct interpretation the associated paradifferential equation will have the form
(2.4) |
However, upon closer examination one may see several choices that could be made. Considering for instance the first paraproduct, which of the following expressions would make the better choice at frequency ?
The last one may seem the most complicated, but it is also the most accurate. In many cases, including our model problem, it makes no difference in practice. However, one should be aware that often a simpler choice, which is made for convenience in one problem, might not work anymore in a more complex setting.
Remark 2.1.
Here the frequency gap, which was set to be equal to in the above formulas, is chosen rather arbitrarily; its role is simply to enforce the frequency separation between the coefficients and the leading term. On occasion, particularly in large data problems, it is also useful to work instead with a large frequency gap as a proxy for smallness, see e.g. [25].
2.3. The paradifferential formulation of the main equations
Consider first our general equation (1.1), which we can write in the form
(2.5) |
Here one would hope that the paradifferential source term can be seen as perturbative, in the sense that
Similarly we can write the linearized equation (2.1) in the same format,
(2.6) |
with the appropriate nonlinearity . This is still based on the paradifferential equation (2.5), but can no longer be interpreted as the direct paralinearization of the linearized equation. This is because the expression also contains some low-high interactions, precisely those where is the low frequency factor.
3. Energy estimates
Energy estimates are a critical part of any well-posedness result, even if they do not tell the entire story. In this section we begin with a heuristic discussion of several ideas in the general case, and then continue with some more concrete analysis in the model case.
3.1. The general case
Consider first the energy estimates for the general problem (1.1), where it is simpler to think of this in the paradifferential formulation (2.3). An energy estimate for this problem is an estimate that allows us to control the time evolution of the Sobolev norms of the solution. In the simplest formulation, the idea would be to prove that
with a constant that at the very least depends on the norm of .
There are two points that one should take into account when considering such estimates. The first is that it is often useful to strenghten such bounds by relaxing the dependence of the constant on . Heuristically, the idea is that this constant measures the effect of nonlinear interactions, which are strongest when our functions are pointwise large, not only large in an sense. Thus, it is often possible to replace the constant with an analogue of the uniform control norm in the model case, perhaps with some additional implicit dependence on another scale invariant uniform control parameter . See however the discussion in Remark 3.2.
A second point is that, although it is tempting to try to work directly with the norm, it is often the case that the straight norm is not well adapted to the structure of the problem; see e.g. what happens in water waves [2], [12]. Then it is useful to construct energy functionals adapted to the problem at hand. For these energies we should aim for the following properties
-
i)
Energy equivalence:
(3.1) -
ii)
Energy propagation
(3.2) where the control parameter satisfies
(3.3)
Now consider our main equation written in the form (2.3). For the perturbative part of the nonlinearity we hope to have some boundedness,
(3.4) |
This in turn allows us to reduce nonlinear energy bounds of the form (3.2) to similar bounds for the linear paradifferential equation (2.5). One may legitimately worry here that some structure is lost when we decouple the paradifferential coefficients from the evolution variable; however, the point is that these two objects are indeed separate, as they represent different frequencies of the solution.
Remark 3.1.
In our discussion here we took the simplified view that bounds for begin at . But this is not always the case in practice, and often one needs to identify the lower range for where this works; see e.g. the nonlinear wave equation [23], the wave map equation [29], or the water wave problem considered in [1].
Now consider the paradifferential evolution (2.5), and begin with the case by setting . Then we need to produce a linearized type energy so that the solutions satisfy
(3.5) |
Then the associated nonlinear energy at would be
If , then the bound (3.5) would simply require that the paradifferential operator is essentially antisymmetric in . If that is not true, then the backup plan is to find an equivalent Hilbert norm on so that the antisymmetry holds. Some care is however needed; if this norm depends on , then this dependence needs to be mild.
The next step is to consider a larger . By interpolation it suffices to work with integer , in which case one might simply differentiate (2.3),
Here we would be done if the last commutator is bounded from into . In principle that would be the case almost automatically at least when the order of is at most one. One can heuristically associate this with the finite speed of propagation in the high frequency limit.
Remark 3.2.
The case , which corresponds to an infinite speed of propagation, is often more delicate; see e.g. [18, 17, 19] for quasilinear Schrödinger flows, or [14] for capillary waves. There one needs to further develop the function space structure based on either dispersive properties of solutions, or on normal forms analysis.
3.2. Coifman-Meyer and Moser type estimates
Before considering our model problem, we briefly review some standard bilinear and nonlinear estimates that play a role later on. In the context of bilinear estimates, a standard tool is to consider the Littlewood-Paley paraproduct type decomposition of the product of two functions, which leads to Coifman-Meyer type estimates, see [7], [22]:
Proposition 3.3.
Using the standard paraproduct notations, one has the following estimates
(3.6) | ||||
as well as the commutator bound
(3.7) |
Here is the Littlewood-Paley projection onto frequencies .
These results are standard in the harmonic/microlocal analysis community. For nonlinear expressions we use Moser type estimates instead:
Proposition 3.4.
The following Moser estimate holds for a smooth function , with , and :
Of course many more extensions of both the bilinear and the nonlinear estimates above are available.
3.3. The model case
We now turn our attention to our model problem, where, if we adopt the expression (2.4) for the paradifferential flow, the source term is given by
(3.8) |
We can rewrite this in the form
(3.9) |
For this expression we can show that it always plays a perturbative role:
Proposition 3.5.
The above nonlinearity satisfies the following bounds:
i) Sobolev bounds
(3.10) |
ii) Difference bounds
(3.11) |
as well as
(3.12) |
The next to last bound shows in particular that is Lipschitz in for . The simplification in the case is also useful in order to bound differences of solutions in the topology.
Proof.
i) We use the expression (3.9) for . The first term can be estimated using a version of the Coifman-Meyer estimates and Moser estimates by
For the second term we use again paraproduct bounds and Moser estimates to get
The third term is similar to the second.
ii) First, we note the representation
which we use to separate factors. Here is a smooth function of and . Then taking differences in the first term of , we need two estimates
respectively
noting that for the last term can be avoided.
Similarly we have two estimates corresponding to the second term in , namely
where
respectively
both with obvious simplifications if . Finally, the bounds for the third term in are similar to the ones for the second.
∎
Remark 3.6.
Next we consider the paradifferential equation:
Proposition 3.7.
Assume that and (i.e. ). Then the paradifferential equation (2.4) is well-posed in all spaces, , and
(3.13) |
Proof.
We first consider the energy estimate, where we work with the corresponding inhomogeneous equation,
(3.14) |
The bound is easiest; we have
In the second term we simply estimate the para-coefficient in . In the first term we commute and integrate by parts, to arrive at
where due to the symmetry of the matrices we have the bound
(3.15) |
which shows that the corresponding paraproduct operators are self-adjoint at leading order. Here we use the ∗ notation to denote the adjoint of an operator. Hence we obtain
which further by Gronwall’s inequality yields
(3.16) |
This by itself does not prove well-posedness in , it only proves uniqueness. However, a similar bound will hold for the backward adjoint system in the same spaces; this is because the adjoint system coincides with the direct system modulo bounded terms. Together, these two pieces of information yield well-posedness for the paradifferential system in . This is a standard linear duality argument, where the solutions are constructed by a direct application of the Hahn-Banach Theorem. In a nutsell, one has the following equivalencies, see for instance [11]:
Exactly the same argument applies in , with the small change that now the the adjoint system should be considered in . There the bound (3.15) is replaced by
(3.17) |
∎
3.4. The linearized equation
Next, we turn our attention to the linearized equation, which we also write in a paradifferential form
(3.18) |
where
We note here that the equation (3.18) is not exactly a true paralinearization of the linearized equation, as does contain low-high interactions. This difference is observed in the estimates satisfied by the two terms.
On one hand, the term satisfies good bounds in all Sobolev spaces,
(3.19) |
so it can be seen as a true perturbative term. This a simple, Coifman-Meyer type estimate which is left for the reader.
On the other hand, assuming we know that , the term can at best be estimated in , and there of course we could not use the control norms, instead we would have to use the full norm of . However, we can use the control norms for bounds to directly estimate
(3.20) |
Combining the last two estimates with Proposition 3.7 we perturbatively obtain
Proposition 3.8.
Assume that and that . Then the linearized equation (2.2) is well-posed in , with bounds
(3.21) |
We observe the obvious fact that one does not need paradifferential calculus in order to prove this proposition; a simple integration by parts suffices. However, it is instructive to dissect the terms in the equation and understand their respective roles. Also, it is interesting to observe that in appropriate settings, the linearized equation can be thought of as a perturbation of the associated paradifferential equation.
Remark 3.9.
Well-posedness and bounds for the linearized equation can be also obtained in all spaces for . However, this can no longer be done in terms of our control parameters; for instance if then we need to use the full norm of the solutions. While interesting, this observation will not be needed for the rest of the paper.
3.5. Difference bounds and uniqueness
The easiest way to compare two solutions and for (1.1) is to subtract their respective equations, to obtain an equation for . In the general case, using the form (2.5) of the equation, we obtain
Here we identify this equation as the paradifferential equation associated to , but with two source terms, which we would like to interpret as perturbative in a low regularity Sobolev space, say . That would yield a bound of the form
(3.22) |
where , , with , and , for .
Let us see how this works out in our model problem. We will show that
Proposition 3.10.
Proof.
We have already seen in Proposition 3.7 that the paradifferential evolution is well-posed in , and in Proposition 3.5 that we have a good Lipschitz bound for . It remains to bound the remaining difference
For this we write
For the first term we have a Coifman-Meyer type bound
The second term is even easier,
and the third term is similar. Finally, in the fourth term we can use Coifman-Meyer to rebalance again the derivatives and obtain
concluding as before. ∎
Remark 3.11.
The observant reader may have noticed that for our model problem the difference bound can be directly proved using a simple integration by parts, without any need for paradifferential calculus, and may wonder why we are doing it this way. There are three reasons for this: (i) to show that it works, (ii) to show how both the bound for the full equation and the bound for the difference equation can be seen as two sides of the same coin, and (iii) to provide a guide for the reader for situations where a simpler approach does not work.
Remark 3.12.
In the same vein as in Remark 3.9, bounds for the difference equation can be also obtained in all spaces for .
Remark 3.13.
In our particular example it was easy to cast the difference equation in a form which is very much like the linearized equation. However, this is not always the case. For this reason, we point out that there is another way one can think of difference bounds, namely by viewing the two initial data and as being connected via a one parameter family of data where . Then we can interpret the difference as
where are the solutions with data . Here the integrand represents a solution to the linearized equation around . Hence difference bounds for can be obtained by integrating bounds for the linearized equation. The only downside to such an argument is that such bounds will require the control parameters for the entire family of solutions, rather than just the end-points.
4. Existence of solutions
Here we consider the question of existence of solutions for the evolution (1.1) with initial data in , where will be taken sufficiently large. The idea here is to construct a good sequence of approximate solutions , which will eventually be shown to converge in a weaker topology. The tricky bit is to choose the correct iteration scheme.
Naively, one might think of trying to base such a scheme on the linearized flow, setting
where the expression on the right represents the error at step . Here one can eliminate the time derivative of and rewrite this as
This would be akin to a Nash-Moser scheme, which, even when it works, loses derivatives. That may be reasonable in a small divisor situation, but not so much if our goal is to obtain a Hadamard style well-posedness result. Nevertheless, Nash-Moser schemes have been used on occasion to produce solutions for quasilinear evolutions, though often they prove to be unnecessary.
Remark 4.1.
We observe that for the existence of solutions one does not need to work from the start at low regularity. As we will see, rough solutions can be constructed later on as limits of smooth solutions. This is strictly speaking not necessary in our model problem, but for more nonlinear, geometric problems it does seem to make a difference. This is because in such situations it is often easier to compare exact solutions via the linearized equation which is a geometric object, instead of working with approximate solutions where the geometric character might be lost.
We will present two strategies to prove existence, and at the end we point out several other methods which have been successfully used in existence proofs.
4.1. Take 1: an iterative/fixed point construction
In order not to lose derivatives in the approximation scheme, the idea here is to carefully choose how to distribute and in the iteration. A key observation is that, whereas solving the linearized equation would cause a loss of derivatives, solving the paradifferential equation does not in general. Then, a good starting point would be the formulation (2.3) of the equations, which would suggest the following iteration scheme:
(4.1) |
We will apply this scheme on a time interval , with sufficiently small depending on the initial data size
For the above sequence the aim would be to inductively prove two uniform bounds in :
(4.2) |
and
(4.3) |
where is a fixed large constant. In the last bound, the time interval size is used in order to gain smallness for the constant, which is needed in order to obtain convergence. Together, these two bounds imply convergence in to some function , as well as regularity for the limit. This in general suffices in order to show that the limit solves the equation.
To obtain uniform bounds for this evolution one would need two pieces of information:
-
(1)
Well-posedness of the paradifferential equation (2.3) in and more generally in all spaces. Heuristically, the two should be equivalent, as the operator does not change the dyadic frequency localization. In practice though it might not be as easy, as leakage to other frequencies may occur, and in particular even the associated Hamilton flow might not preserve the dyadic localization on a unit time scale.
-
(2)
Lipschitz property of in Sobolev spaces. More generally, a bound of the form
(4.4) which should be thought of as a Moser type inequality.
In addition to uniform bounds in a strong norm , one would also like to have convergence in a weaker topology, say for the purpose of this presentation. The difference equation reads
(4.5) |
Here energy estimates in would follow from (1) and (2) above, provided that the last difference has a good bound
This is in general relatively straightforward if is large enough.
Remark 4.2.
The argument above yields solutions which are apriori only in as opposed to , as desired. Getting continuity in for is relatively straightforward by interpolation, but proving continuity in requires considerable extra work111e.g. by showing continuity in time of solutions to the linear paradifferential equation. if one wants a direct argument. The easy way out is to rely on the arguments in the next section, where we show that all solutions can be seen as uniform limits of smooth solutions.
Remark 4.3.
The above iterative argument can be rephrased as a fixed point argument as follows. For we define as the solution to
Then the desired solution has to be a fixed point for . Solutions to this fixed point problem may often be obtained using the contraction principle in the right topology. Precisely, the strategy is to choose the domain of to be the ball in , but endow this ball with a weaker topology, e.g. . Then both the mapping properties of and the small Lipschitz constant can be achieved by choosing the time sufficiently small. Here for the domain we have to choose rather than continuity in order to guarantee completeness.
We now implement this scheme for our model problem. Denoting , we will prove inductively that for fixed large enough and small enough , we have the bound
Taking this as induction hypothesis we have the following bounds for the control parameters and associated to :
Then we can estimate in by combining Proposition 3.7 and Proposition 3.5 to obtain
and by Gronwall’s inequality we arrive at
with a universal implicit constant. This completes the induction if we first choose large enough (to dominate the implicit constant), and then small enough (depending on and ).
On the other hand, in order to prove the convergence in we use the equation (4.5) for the difference , and claim that the following estimate holds:
(4.6) |
Assuming this is true, by Gronwall’s inequality we obtain
which gives us the small Lipschitz constant if is sufficiently small, depending only on .
It remains to prove (4.6). For the paradifferential equation we can use Proposition 3.7 and for the difference we can use Proposition 3.5, so it remains to examine the last term in (4.5), and show that
In the case of the model problem the difference on the left reads
For the first term we have the obvious bound
The second term is split into three parts,
where the first two parts are easy to estimate. A similar bound follows for the third term after we move the derivative onto the high frequency factor, using an estimate of the form
which is a corollary of the second bound in (3.6).
4.2. Take 2: a time discretization method
Here the idea is to discretize time at a small scale , and to construct approximate discrete solutions with the following properties:
-
i)
Uniform bounds
(4.7) -
ii)
Approximate solution
(4.8)
Once this is done, if is large enough222For instance in our model case case suffices. then it is a relatively straightforward matter to show that a uniform limit exists333Here one may extend to all times by linear interpolation. on a subsequence as , by applying the Arzela-Ascoli theorem. This works in a time interval with . By passing to the limit in the above bounds in a weak topology, it follows the limit solves the equation and has regularity
The nice feature of this method is that one really only needs to carry out one single step. Precisely, given with size , and , one needs to find (which corresponds to above) with the following properties:
-
i)’
Uniform bounds
(4.9) -
ii)’
Approximate solution
(4.10)
Reiterating this, the bound (4.7) follows by applying a discrete form of Gronwall’s inequality.
Remark 4.4.
The bound in ii)’ can be harmlessly replaced by with a small constant .
Remark 4.5.
Sometimes the square norm of is not the correct quantity to propagate in time, and one needs to replace it with appropriate equivalent energies in property (ii)’.
Remark 4.6.
The remaining question is how to construct the single iterate satisfying properties (i)’,(ii)’ above. The obvious choice would be Euler’s method, which is to set
but this does not work because it loses derivatives.
Inspired by the nonlinear semigroup theory [4], one may choose instead to solve
This idea has potential at least when this is an elliptic equation. Alternatively one may opt for a paradifferential version
which has the advantage that one only needs to solve a linear elliptic equation. However, ellipticity is not guaranteed.
Instead, here we will adopt a two steps approach, which has the advantage that no partial differential equation needs to be solved. Precisely, our steps are as follows:
STEP 1: Regularization. Here we take the initial data and we regularize it on an dependent scale. Precisely, if is the order of the nonlinearity , then it is natural to choose the spatial truncation frequency scale to be , which corresponds to an order parabolic regularization; this regularization scale is needed in order to be able to bound the error in the Euler step. Then our regularization would have the following properties:
-
(a)
Regularization
(4.11) -
(b)
Energy bound
(4.12) -
(c)
Approximate solution
(4.13)
STEP 2: Euler iteration. Here we simply set
(4.14) |
so that the approximate solution bound (4.10) becomes relatively straightforward, and the energy bound (4.9) becomes akin to proving the energy estimate; see the example below.
We now implement the above strategy on our chosen model problem. Here our chosen energy is simply the Sobolev norm,
Our equation has order , so the proper regularization scale is . Hence, we use a Littlewood-Paley projector to simply define
and the three properties (a), (b) and (c) above are trivially satisfied.
Next we turn our attention to the Euler iteration (4.14) for which we need to establish the properties (i)’ and (ii)’. We begin with (i)’, where it suffices to compare the energies of and . For we have
If , then in the second term on the right we have at most derivatives, so this term has size in the norm
and we can neglect it.
It remains to consider . Then we can separate the terms with no more than derivatives and estimate them as above, using appropriate interpolation inequalities,
Hence we have
and, neglecting terms, we compute norms,
The last norm has size in view of property (a) above. In the integral, on the other hand, we use the symmetry of to integrate by parts,
which can again be estimated by . Thus we obtain
as desired, as can be estimated by the Sobolev norm of by Sobolev embeddings.
It remains to consider (ii)’, where, by (c) above, it suffices to show that
This is a soft argument, where we simply write
where the norm on the right is bounded by interpolating (c) above with the uniform bound provided by (b). This requires .
4.3. Other strategies
Most of the other strategies to prove existence of solutions are based on constructing approximate flows, and solutions are obtained as limits of solutions to the approximate flows. There are two such methods which are more widely used.
a) Parabolic regularization. Here one uses a parabolic regularization of the original flow (1.1), defining the approximate solutions by
where the correct choice for the parabolic term seems to be to double the order of the original equation. These problems can often be solved for a short, dependent time, as semilinear problems, with a direct, fixed point argument. However, in doing this, the main challenge is to prove uniform in bounds for these approximate flows. This sometimes requires more careful choices of the regularization term, to make it fit better with the geometry of the problem.
b) Galerkin approximation. Here the idea is to work with a low frequency projector in the equation, e.g. of the type
with , see e.g. the example in [30]. The local solvability for this evolution becomes trivial as this evolution is an ordinary differential equation in a Hilbert space, but the challenge is again to prove uniform in bounds for these approximate flows. The double use of the projector above is a choice that usually facilitates achieving this objective. Depending on the problem, this may require careful choices for the frequency projectors, adapted to the problem.
5. Rough solutions as limits of smooth solutions
Here we explore the idea of constructing rough solutions as limits of smooth solutions. There are at least two good reasons to do this, which we discuss in order:
-
(1)
In quasilinear problems one does not expect any sort of uniformly continuous dependence of solutions on the initial data, so the continuity of the flow map becomes a purely qualitative assertion. However, one can still ask for a quantitative way of comparing solutions, and such a quantitative venue is found by using the regular approximations as a convenient proxy. This is discussed in the last section.
-
(2)
It is also often the case that more regular solutions are sometimes easier to produce, and in such situations, obtaining the rough solutions as limits of smooth solutions might be the only option. This is particularly the case in problems where the state space is not a linear space, such as Schrödinger maps [20], Yang-Mills or other problems with a nontrivial gauge structure. See also [15] for an implementation of this idea in a free boundary problem. This is because in such problems it is always easier to obtain estimates for the linearized equations, or at least to compare exact solutions, rather than to cook up a constructive scheme which is consistent with the geometry.
To make this analysis quantitative, it is very useful to track the flow of energy between different frequencies. Whereas energy cascades (energy migration to higher frequencies) have long been associated with blow-up phenomena, well-posedness should correspond to a lack thereof. To quantify this, we will use Tao’s notion of frequency envelopes.
5.1. Frequency envelopes
Frequency envelopes, introduced by Tao (see for example [27]), are a very useful device in order to track the evolution of the energy of solutions between dyadic energy shells. As there is always nearby leakage between the dyadic shells in nonlinear flows, on needs to do this in a more stable way, rather than look directly at the exact amount of energy in every shell.
This is realized via the following definition:
Definition 5.1.
We say that is a frequency envelope for a function in if we have the following two properties:
a) Energy bound:
(5.1) |
b) Slowly varying
(5.2) |
Here represent the standard Littlewood-Paley projectors, and is a positive constant, which is taken small enough in order to account for energy leakage between nearby frequencies.
One can also try to limit from above the size of a frequency envelope, for instance by requiring that
We call such envelopes sharp. Such frequency envelopes always exist, for instance one can take
For a better understanding see Figure 1 below, where the actual dyadic norms, indicated by red bullets on a logarithmic scale, are lifted (based on the above formula) to a slowly varying frequency envelope, indicated by the green circles.
We will use frequency envelopes in order to track the evolution of energy in time as follows: we start with a sharp frequency envelope for the initial data, and then seek to show that we can propagate this frequency envelope to the solutions to our quasilinear flow, at least for a short time.
Remark 5.2.
One alternative here is to unbalance the choice of in (5.2), asking for a small if , but replacing with a large constant for . This heuristically corresponds to a better control of leakage to higher frequencies, and it is useful in order to deal with higher regularity properties also within the frequency envelope set-up.
5.2. Regularized data
Consider an initial data with size , and let be a sharp frequency envelope for in . For we consider a family of regularizations at frequencies where is a dyadic frequency parameter. This parameter can be taken either discrete or continuous, depending on whether we have access to difference bounds or only to the linearized equation. Suppose we work with differences. Then the family can be taken to have similar properties to Littlewood-Paley truncations:
-
i)
Uniform bounds:
(5.3) -
ii)
High frequency bounds:
(5.4) -
iii)
Difference bounds:
(5.5) -
iv)
Limit as :
(5.6)
Correspondingly, we obtain a family of smooth solutions .
Here in the simplest setting where the phase space is linear one may simply choose , which would have all the above properties. However, in geometric settings where the phase space is nonlinear, a more complex regularization method may be needed, for instance using a corresponding geometric heat flow, see [28] or a variable scale regularization as in [15].
5.3. Uniform bounds
Corresponding to the above family of regularized data, we obtain a family of smooth solutions . For this we can use the energy estimates as in Theorem 3 to propagate Sobolev regularity for solutions as well as difference bounds as in Proposition 3.10. This yields a time interval where all these solutions exist, and whose size depends only on , where we have the following properties:
-
i)
High frequency bounds:
(5.7) -
ii)
Difference bounds:
(5.8)
From (5.7) one may obtain a similar bound for the difference . Interpolating this with (5.8), we also have
(5.9) |
One may use these bounds to establish uniform frequency envelope bounds for ,
(5.10) |
on the same time interval which depends only on the initial data size. This is a direct consequence of (5.7) for , while if we can use the telescopic expansion
and use (5.7) for the first term and (5.8) for the differences.
5.4. The limiting solution
Consider now the convergence of as . From the difference bounds (5.8) we obtain convergence in to a limit , with
On the other hand, expanding the difference as a telescopic sum we get
where, in view of the above bounds (5.7) and (5.8), each summand is essentially concentrated at frequency , with size and exponentially decreasing tails. This leads to
(5.11) |
so we also have convergence in .
This type of argument plays multiple roles:
-
(1)
It produces rough solutions as smooth solutions, justifying the earlier assertion that it often suffices to carry out the initial construction of solutions only in a smooth setting.
-
(2)
It establishes the continuity of solutions as valued flows, which is sometimes missing from the constructive proof of existence.
-
(3)
It provides the quantitative bound (5.11) for the difference between the rough and the smooth solutions, which plays a key role in the continuous dependence proof in the next section.
6. Continuous dependence
Here we use frequency envelopes in order to prove continuous dependence of the solution as a function of the initial data , and also discuss some historical alternatives.
6.1. The continuous dependence proof
Consider a sequence of initial data
and the corresponding solutions , which exist with a uniform lifespan , where depends only on the initial data size . We will prove that in . Once we have this property, it automatically extends to any larger time interval where the solution is defined and satisfies . This should be understood in the sense that for all large enough , the solutions are defined in , with similar regularity, and the convergence holds as .
The difference bounds in Proposition 3.10 guarantee that in . Since are uniformly bounded in , this also implies convergence in for every , but not for .
It remains to consider the convergence in the strong topology, i.e. in . Rather than trying to compare the solutions and directly, we will use as a proxy the approximate solutions , respectively . For these, we will take advantage of the fact that their initial data converge in all Sobolev norms,
Hence, according to the preceding discussion, we have convergence of the regular solutions in all Sobolev norms,
To compare the solutions and themselves, we use the triangle inequality,
(6.1) |
The first term goes to zero as for fixed , while the second goes to zero as , but does not depend on . It is the third term which is the problem, and for which we need to gain some smallness uniformly in .
However, in the previous section we have learned to estimate such differences using frequency envelopes. Precisely, let , respectively be frequency envelopes for the initial data , respectively in . Then, as we saw in the previous section, we can estimate the last two terms above in terms of frequency envelopes and obtain
(6.2) |
The important observation is that the convergence in allows us to choose the frequency envelopes , respectively so that
This implies that
Hence, passing to the limit in the relation (6.1), we obtain
(6.3) |
and finally letting we obtain
as desired.
6.2. Comparison with Kato and Bona-Smith
The more classical approach for continuous dependence goes back to Kato [16] as well as a variation due to Bona-Smith [5]. We will briefly describe this approach using our notations and set-up; we caution the reader that the original arguments in these papers are not self-contained and are instead mixed with the other parts of well-posedness proofs, so it is not exactly easy to correlate the papers with the description below. In effect our discussion below is more closely based on the interpretations of Kato’s work provided by Chemin [3] and, even closer, by Tao [26].
This also relies on the use of some sort of approximate solutions . However, in this approach one aims to directly estimate the difference in in terms of the corresponding initial data. One might at first hope to directly track the difference , but this cannot work without knowledge that the low frequencies of the difference (i.e. below ) are better controlled. So the better object to track turns out to be a norm of the form
(6.4) |
where we recall that is the order of our nolinearity. Here the second part can be estimated directly for any two solutions, see Remark 3.12, so one can think of this as decoupled as a two step process. To better understand why this works, it is useful to write the equation for the difference in a paradifferential form
(6.5) |
which should essentially be thought of as a perturbation of the linear paradifferential flow, which can be estimated in all Sobolev spaces. The difference is tame because admits Lipschitz bounds in all Sobolev spaces, so the issue is the last term.
There there is seemingly a loss of derivatives, but these derivatives are applied to , which has higher regularity bounds, so they yield losses of at most a factor. But this factor can be absorbed by the lower frequency paradifferential coefficients given by , in view of the factor in (6.4). Here it is important that we wrote the equation using rather on the left, which allows us to use as the argument in the last term on the right.
In Kato’s argument the same principle is used to get bounds not only for the difference but also for for an arbitrary solution . In Bona-Smith’s, version, on the other hand, one estimates only , but the proof is more roundabout in that is not only assumed to have regularized data, but also to solve a regularized equation, thus combining the existence and the continuous dependence arguments.
In our opinion, working with frequency envelopes has definite advantages:
-
•
It provides more accurate information on the solutions.
-
•
It does not require any direct difference bounds in the strong topology.
-
•
By working with a continuous, rather than a discrete family of regularizations one can fully replace difference estimates by bounds for the linearized equation, which is to be preferred in many cases, in particular in geometric contexts where the state space is an infinite dimensional manifold.
References
- [1] A. Ai, M. Ifrim, and D. Tataru. Two dimensional gravity waves at low regularity I: Energy estimates. arXiv e-prints, page arXiv:1910.05323, Oct. 2019.
- [2] T. Alazard, N. Burq, and C. Zuily. On the Cauchy problem for gravity water waves. Invent. Math., 198(1):71–163, 2014.
- [3] H. Bahouri, J.-Y. Chemin, and R. Danchin. Fourier Analysis and Nonlinear Partial Differential Equations, volume 343 of Grundlehren der mathematischen Wissenschaften. Springer-Verlag Berlin Heidelberg, 2011.
- [4] V. Barbu. Nonlinear semigroups and differential equations in Banach spaces. Editura Academiei Republicii Socialiste România, Bucharest; Noordhoff International Publishing, Leiden, 1976. Translated from the Romanian.
- [5] J. L. Bona and R. Smith. The initial-value problem for the Korteweg-de Vries equation. Philos. Trans. Roy. Soc. London Ser. A, 278(1287):555–601, 1975.
- [6] J.-M. Bony. Calcul symbolique et propagation des singularités pour les équations aux dérivées partielles non linéaires. Ann. Sci. École Norm. Sup. (4), 14(2):209–246, 1981.
- [7] R. R. Coifman and Y. Meyer. Au delà des opérateurs pseudo-différentiels, volume 57 of Astérisque. Société Mathématique de France, Paris, 1978. With an English summary.
- [8] J. Hadamard. Sur les problèmes aux dérivés partielles et leur signification physique. Princeton University Bulletin, 13:49–52, 1902.
- [9] J. Hadamard. Lectures on Cauchy’s Problem in Linear Partial Differential Equations. Mrs. Hepsa Ely Silliman memorial lectures. Yale University Press, 1923.
- [10] L. Hörmander. Lectures on nonlinear hyperbolic differential equations, volume 26 of Mathématiques & Applications (Berlin) [Mathematics & Applications]. Springer-Verlag, Berlin, 1997.
- [11] L. Hörmander. The analysis of linear partial differential operators. III. Classics in Mathematics. Springer, Berlin, 2007. Pseudo-differential operators, Reprint of the 1994 edition.
- [12] J. K. Hunter, M. Ifrim, and D. Tataru. Two dimensional water waves in holomorphic coordinates. Comm. Math. Phys., 346(2):483–552, 2016.
- [13] M. Ifrim and D. Tataru. Lectures, Summer Graduate School: Introduction to water waves; MSRI, Summer 2020. https://www.msri.org/summer_schools/910/schedules. Accessed:2020-08-10.
- [14] M. Ifrim and D. Tataru. The lifespan of small data solutions in two dimensional capillary water waves. Arch. Ration. Mech. Anal., 225(3):1279–1346, 2017.
- [15] M. Ifrim and D. Tataru. The compressible Euler equations in a physical vacuum: a comprehensive Eulerian approach. arXiv e-prints, page arXiv:2007.05668, July 2020.
- [16] T. Kato. The Cauchy problem for quasi-linear symmetric hyperbolic systems. Arch. Rational Mech. Anal., 58(3):181–205, 1975.
- [17] J. L. Marzuola, J. Metcalfe, and D. Tataru. Quasilinear Schrödinger equations I: Small data and quadratic interactions. Adv. Math., 231(2):1151–1172, 2012.
- [18] J. L. Marzuola, J. Metcalfe, and D. Tataru. Quasilinear Schrödinger equations, II: Small data and cubic nonlinearities. Kyoto J. Math., 54(3):529–546, 2014.
- [19] J. L. Marzuola, J. Metcalfe, and D. Tataru. Quasilinear Schrödinger equations III: Large Data and Short Time. arXiv e-prints, page arXiv:2001.01014, Jan. 2020.
- [20] H. McGahagan. An approximation scheme for Schrödinger maps. Comm. Partial Differential Equations, 32(1-3):375–400, 2007.
- [21] G. Métivier. Para-differential calculus and applications to the Cauchy problem for nonlinear systems, volume 5 of Centro di Ricerca Matematica Ennio De Giorgi (CRM) Series. Edizioni della Normale, Pisa, 2008.
- [22] C. Muscalu and W. Schlag. Classical and multilinear harmonic analysis. Vol. II, volume 138 of Cambridge Studies in Advanced Mathematics. Cambridge University Press, Cambridge, 2013.
- [23] H. F. Smith and D. Tataru. Sharp local well-posedness results for the nonlinear wave equation. Ann. of Math. (2), 162(1):291–366, 2005.
- [24] C. D. Sogge. Lectures on non-linear wave equations. International Press, Boston, MA, second edition, 2008.
- [25] J. Sterbenz and D. Tataru. Energy dispersed large data wave maps in dimensions. Comm. Math. Phys., 298(1):139–230, 2010.
- [26] T. Tao. Blog, Local well-posedness for the Euler equations. https://terrytao.wordpress.com/2018/10/09/254a-notes-3-local-well-posedness-for-the-euler-equations/#apb-diff. Accessed:2020-08-10.
- [27] T. Tao. Global regularity of wave maps. II. Small energy in two dimensions. Comm. Math. Phys., 224(2):443–544, 2001.
- [28] T. Tao. Geometric renormalization of large energy wave maps. In Journées “Équations aux Dérivées Partielles”, pages Exp. No. XI, 32. École Polytech., Palaiseau, 2004.
- [29] D. Tataru. Rough solutions for the wave maps equation. Amer. J. Math., 127(2):293–377, 2005.
- [30] M. E. Taylor. Partial differential equations III. Nonlinear equations, volume 117 of Applied Mathematical Sciences. Springer, New York, second edition, 2011.