Differential Elimination and Algebraic Invariants of Polynomial Dynamical Systems
Abstract
Invariant sets are a key ingredient for verifying safety and other properties of cyber-physical systems that mix discrete and continuous dynamics. We adapt the elimination-theoretic Rosenfeld-Gröbner algorithm to systematically obtain algebraic invariants of polynomial dynamical systems without using Gröbner bases or quantifier elimination. We identify totally real varieties as an important class for efficient invariance checking.
keywords:
algebraic invariants , algebraic varieties , elimination theory , differential algebra , polynomial dynamical systems , Rosenfeld-Gröbner algorithm , characteristic set , Gröbner basis , totally real varieties , computational complexity1 Introduction
Computers are increasingly hitting the road, taking to the skies, and interacting with people and the physical environment in unprecedented ways. Engineering concerns like realistic models and reliable sensors are critical, but just as important are the complex mathematical problems that lie at the heart of making cyber-physical systems (CPS) safe [1, 2]. One of these central problems is computing invariant sets [3] of continuous dynamical systems (Section 2.3), where an invariant set is a collection of states from which any trajectory starting in the set will never leave. Given a system of ordinary differential equations (ODEs) and a set of unsafe states, we must identify initial states that will never lead to an unsafe state under the specified dynamics. If an invariant does not intersect the unsafe set, then every state in the invariant set is a safe starting point. Beyond safety [4], invariant reasoning is an important part of other significant properties like stability [5] and liveness [6]. Due to its significance, the problem of invariant generation for ODEs has received substantial interest [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22] and even has a dedicated tool [23], but most methods have nontrivial heuristic search parts.
In this article we employ strategies from elimination theory [24] to give new algorithms for systematically generating invariants and checking candidates for invariance. Elimination theory draws on algebra, geometry, and logic to give algorithmic procedures for understanding the polynomial systems that arise in many applications. In a narrow sense, elimination is the process of exposing explicit relationships and removing variables. The prototypical example of an elimination method is Gaussian elimination, which converts a system of linear equations into a simpler system with the same solutions from which those solutions can be read off easily. However, linear algebraic equations are merely the tip of the iceberg. Crucially, many elimination tools are extremely general. They apply to nonlinear problems of diverse forms, allowing an unusual degree of reuse across theories and applications. In this paper, our problems require methods for symbolically analyzing systems of ODEs that have real algebraic [25] constraints; i.e., are defined by polynomial equations over the real numbers (Section 2.3). In recent decades, researchers in differential algebra [26] have extended classical algebraic elimination methods to polynomial differential equations (Section 2.4). Broadly speaking, differential elimination [27] provides tools for finding all relations implied by a polynomial differential system, regardless of the form. The elimination procedures we develop in Sections 3 and 4 are intrinsically mathematical, but we show that the output corresponds directly to invariants of the input system, giving a novel, complementary view on systematically computing and checking ODE invariants.
As always, there are trade-offs involved. Elimination methods are exact, general and remarkably versatile, but used naively they can have high theoretical and practical costs [28]. We emphasize alternatives within elimination theory that have better computational complexity than standard choices.For example, characteristic sets [29] carry less information than the more ubiquitous Gröbner bases [30] but they have singly exponential complexity [31] in the number of variables instead of Gröbner bases’ doubly exponential growth [32]. We work extensively with regular systems (Section 2.6), a weak form of characteristic sets with lower complexity. Similarly, real number solutions are the main goal in applications, but obtaining them is often significantly more expensive than working over the complex numbers [33, 34]. We give a promising workaround that can equivalently replace real algebraic computation with more efficient complex111Ironically, computation over complex-valued structures is often less complex than over real-valued structures. alternatives in many situations (Sections 2.2 and 4.4).
Computation is our main motivation for restricting to polynomial differential and algebraic equations. Such equations are general [4], already powerful for applications [35] and undecidability typically ensues when we leave the polynomial setting [36].
Structure of the Paper
Section 2 reviews the algebraic geometry and differential algebra used to rigorously justify our results in Sections 3 and 4. If we do not have a reference at hand or the proof is short, we provide a proof without claiming originality.
Most of the material in Section 2 is standard and is self-contained assuming familiarity with basic linear algebra and multivariate calculus. That said, the ideas are nontrivial and, depending on background, readers may need to study most or all of Section 2 in order to fully understand the new results and proofs in Sections 3 and 4. However, skimming Section 2 for notation and definitions should provide a good initial picture even without carefully reading Section 2.
Sections 3 and 4 are the heart of the paper. In Section 3 the main contributions are Theorems 86 and 88. Theorem 86 identifies a novel sufficient condition for a system of polynomial equations and inequations to implicitly determine an invariant. Theorem 88 gives multiple criteria, frequently met in practice, that allow to define the associated invariant more explicitly. A detailed example (Section 3.3) illustrates the necessary computations while postponing a full explanation until Section 4.2.
Section 4.1 introduces , our new variant of the Rosenfeld-Gröbner algorithm (RGA) of Boulier et al. [37]. takes in a system of ordinary differential equations and polynomial equations and systematically extracts a structured, maximal invariant satisfying the equations. In particular, Theorem 98 shows that produces output that meets the requirements of Theorem 86 for an invariant. Section 4.2 revisits the example of Section 3.3 in the context of RGA. In Section 4.3 we analyze the computational complexity of and find an explicit upper bound (Theorem 109) on the degrees of polynomials in the output. (Prior works on RGA have either ignored complexity or only found bounds on the order of intermediate differential equations.)Our focus in Section 4.4 is a simple and efficient—compared to standard methods from the literature—test for invariance (Theorem 112) based on totally real varieties. These geometric objects provide a bridge that lets us draw conclusions about real number solutions using tools that are naturally suited to the complex numbers with their computational advantages.
2 Mathematical Background
2.1 Ideals and Varieties
We work extensively with multivariate polynomials over various fields [38] like the rational numbers , the real numbers , and the complex numbers . We often write or for an unspecified field. (A field is essentially a set equipped with some “addition” and “multiplication” operations that satisfy the usual properties of arithmetic in , or . In particular, we have the ability to divide by nonzero elements.) In Section 2.4 we consider differential fields, which are fields having additional structure inspired by calculus. We sometimes restrict to polynomials over the real or complex numbers even when results apply more generally. The fact that has an ordering (given by the usual relation) is important, but we try to work algebraically as much as possible (in which case is particularly convenient.) Occasionally we use the Euclidean metric properties of and [39].
Unless indicated otherwise, we use for the number of variables and write for the -tuple ; then the polynomial ring is the set of all polynomials in variables over field . We denote by the collection of all -tuples of elements from .
2.1.1 Ideals
Definition 1.
Let be a field. If , , and for all and , then we say is an ideal of and write . If and , we call a proper ideal and write .
If , then the collection of all finite sums , where and is any polynomial in , is the ideal generated by in . We write or just to denote this ideal of .
Our interest in ideals comes from the fact that they correspond to systems of polynomial equations. If , we associate with the equation . The definition of an ideal mirrors the behavior of polynomial equation solutions: given , if and are simultaneously 0 when evaluated at , then and are also 0 for any .
Note that is the minimal ideal of containing . Elements of an ideal are linear combinations of generators of the ideal, although the “coefficients” multiplied by the generators are arbitrary polynomials over , not necessarily elements of .
Even though nonzero ideals in are infinite, they can always be represented in a finite way:
Theorem 2 (Hilbert’s basis theorem [30, Thm. 4, p. 77]).
Let be a field. Every ideal in the multivariate polynomial ring is finitely generated (has a finite generating set).
We generally rely on context to indicate the polynomial ring of which or is an ideal (namely, the ring mentioned whenever the ideal’s name is introduced), but we use the following notation when changing fields:
Notation 3.
Let be fields and let . We write for the ideal generated by in .
For instance, if , then is the ideal generated by (or equivalently by ) in .
Lemma 4.
Let be fields. If and for , then .
Proof.
This is immediate from the definition of ideal membership and the linear algebra fact that linear equations with coefficients from a field are solvable in if and only if they are solvable in the extension . ∎
Two types of ideals play a particularly fundamental role in relating geometry (specifically, solution sets of polynomial equations) to algebra.
Definition 5.
A prime ideal is a proper ideal such that for all for which , either or .
A radical ideal has the property that for all such that for some natural number , already . The radical of an arbitrary ideal is the intersection of all radical ideals containing .
The radical is a radical ideal and consists precisely of those polynomials such that for some . Ideal is radical if and only if . Prime ideals are radical, but not necessarily vice versa (for instance, is radical but not prime).
2.1.2 Algebraic Varieties and Constructible Sets
Sets definable by polynomial equations, as well as the corresponding ideals, are central players in our work:
Definition 6.
Let be a field, , and . The zero set of , denoted , is the set of all such that for all . If for some , we call a variety, algebraic set, or Zariski closed set over . We refer to subsets of that are themselves zero sets of polynomials in as subvarieties, algebraic subsets, or Zariski closed subsets over .
Even if is not algebraic, we define the vanishing ideal of , denoted , to be the collection of all such that for all . (Note that is a radical ideal.) If is understood from context, we may write or for the zero set and vanishing ideal, respectively. If contains only one polynomial , we write instead of .
Example 7.
Trivial examples of algebraic sets are the empty set and all of . A more interesting example is , where is the determinant of the matrix corresponding to . The points of represent singular matrices.
Lemma 8 ([30, Thm. 15, p. 196]).
Let be a field and let . Then .
Remark 10.
Lemma 9 suggests an inverse relationship between and . A deeper companion result that fills out this intuition is Hilbert’s Nullstellensatz (Theorem 20). We will review several similar results involving various fields ( and differential fields (Section 2.4)), so for convenience we list all of them in Table 1 on p. 1. (The table is referenced prior to the statement of each such theorem.) These theorems are instances of Galois connections [40, 41].
Any field : • (Lemma 9) Let . Then . • (Lemma 15) An algebraic variety over is irreducible if and only if its vanishing ideal is a prime ideal. (or any algebraically closed field): • (Theorem 20, Hilbert’s Nullstellensatz) Polynomial vanishes at every point in the complex zero set of if and only if . • (Corollary 25, complex algebra-geometry dictionary) Let . Then if and only if . • (Theorem 29, Hilbert’s Nichtnullstellensatz) Let , and let be finite. A polynomial vanishes at every complex solution of if and only if . (or any real closed field [25]): • (Theorem 24, real Nullstellensatz) Polynomial vanishes at every point in the real zero set of if and only if . • (Corollary 25, real algebra-geometry dictionary) Let . Then if and only if . Any differential field K (characteristic 0): • (Theorem 61, differential Nullstellensatz) Given , a differential polynomial vanishes at every point of if and only if . • (Corollary 62, differential algebra-geometry dictionary) Let . Then if and only if . • (Theorem 63, differential Nichtnullstellensatz) Let , and let be finite. A differential polynomial vanishes at every solution of in every differential extension field of if and only if . |
Complements of algebraic sets (i.e., points not satisfying certain polynomial equations) are also important for our results in Sections 3 and 4.
Definition 11.
Let be a field. -constructible sets are Boolean combinations (i.e., complements, finite intersections, and finite unions) of Zariski closed sets over . We write constructible sets if is understood.
A typical example of a constructible set that might not be an algebraic set is the set difference of two algebraic sets . However, finite unions and arbitrary intersections of algebraic sets are still algebraic [42, p. 24]. Said differently, Zariski open sets (complements of Zariski closed sets) form a topology on .
2.1.3 Zariski Closure
Even if a constructible set is not algebraic, it can be augmented to one that is equivalent from the perspective of polynomial equations.
Definition 12.
Let be a field. Given a set , the intersection of all Zariski closed sets over that contain is the Zariski closure of over . We denote the Zariski closure by , or simply if is understood.
Being an intersection of Zariski closed sets, the Zariski closure is itself a Zariski closed set. As suggested above, the Zariski closure is the best algebraic overapproximation (using coefficients from ) to the original set. In particular, two sets having the same Zariski closure are indistinguishable using polynomial equations. (For a more precise statement, see Lemma 17.)
Over or , Zariski open sets form a topology that is coarser than the usual Euclidean topology. That is, Zariski open sets are Euclidean open, but not necessarily vice versa (for instance, the open interval is Euclidean open but not Zariski open because its complement is not the set of roots of a univariate polynomial). However, the following is true:
Lemma 13.
Real (respectively, complex) Zariski-closed sets are closed in the real (respectively, complex) Euclidean topology, and the real (respectively, complex) Zariski closure of a real (respectively, complex) constructible set is the real (respectively, complex) Euclidean closure of the set.
Proof.
See [43, Cor. 4.20]. The cited proof shows that the complex Euclidean closure of the image of a complex variety under a polynomial map equals the complex Zariski closure of the image, but the constructible set version (real or complex) reduces to that result. (Overview of the idea: Use the identity map. Since the complex Euclidean and Zariski closures are equal by [43, Cor. 4.20], it now suffices to show i) the real Zariski closure is the restriction of the complex Zariski closure to , and ii) likewise for the real versus the complex Euclidean topologies. Claim i) is Lemma 19, which is not dependent on the current lemma. Claim ii) is immediate using the Euclidean metric on and .)∎
To distinguish between the Zariski and the Euclidean closures of a set over or , we write or for the Euclidean closures.
Definition 14.
Let be a field and where is Zariski closed over . We say that is -Zariski dense (or just Zariski dense) in if .
Here the difference between the Zariski and Euclidean topologies is evident: for instance, is -Zariski dense in (because the only univariate polynomial with infinitely many roots is ) but is not dense in with respect to the Euclidean topology since points of are not limit points of . This does not violate Lemma 13 because is not a -constructible set.
The next lemma concerns irreducible algebraic sets, i.e., algebraic sets over a field that are not the union of two smaller algebraic subsets over .
Lemma 15 ([42, p. 35]; Table 1).
An algebraic variety over is irreducible if and only if its vanishing ideal is a prime ideal.
A typical example is an algebraic set defined by a single polynomial that is irreducible in . Some authors require an algebraic set to be irreducible to qualify for the name “variety”, but we do not impose that restriction.
Lemma 16.
Let be a field and let be Zariski closed sets. If is irreducible over and is nonempty, then .
Proof.
If were not all of , then would be a decomposition of that contradicts irreducibility. ∎
Lemma 17.
Let be a field.
-
1.
If and , then vanishes at every point of if and only if vanishes at every point of the Zariski closure of ; i.e., .
-
2.
Given , we have if and only .
Proof.
1. In both cases is a Zariski-closed set containing .
2. If , then by 1. Conversely, suppose . Then similarly by 1. Since , are Zariski closed, we have by Lemma 9. ∎
Definition 18.
Let be fields and let . We call the -points of or the restriction of to . We denote this set by .
Lemma 19.
Let . Then the real Zariski closure equals , the restriction of the complex Zariski closure to the reals.
Proof.
: We must show that contains and is real Zariski closed. Containment of is clear. For real Zariski closedness, note that is the zero set of the real and imaginary parts of the defining polynomials of . (That is, suppose for some . For all , distribute over the real and imaginary parts of each monomial’s coefficient to write where . We only care about real solutions here, so if and only if .)
: We must show that is contained in every real Zariski closed set that contains . Observe that is a complex Zariski closed set containing and thus . Then and so , with the last equality following from Lemma 9. ∎
2.1.4 The Nullstellensatz: Connecting geometry and algebra
The following well-known result is fundamental for going back and forth between varieties and ideals over :
Theorem 20 (Hilbert’s Nullstellensatz [30, Thm. 2, p. 179]; Table 1).
Polynomial vanishes at every point in the complex zero set of if and only if . Equivalently, for .
Real algebraic geometry requires a more refined (but computationally less tractable) notion than the radical, the real radical:
Definition 21.
Let . The real radical of is the set of all polynomials such that for some natural number and polynomials , the sum belongs to .
Remark 22.
An expression of the form is a sum-of-squares. Sums-of-squares are important for real algebraic geometry because is zero at if and only if for all .
The following straightforward properties follow immediately from the definitions and Remark 22. We may use them without comment but record them here for completeness.
Proposition 23.
-
1.
If , then is a radical ideal that contains (possibly strictly) .
-
2.
If (respectively, ), then (respectively, and ).
-
3.
If (respectively, ) and (respectively, ), then
-
(a)
(respectively, ) and
-
(b)
(respectively, ).
-
(a)
A much deeper result known as the real Nullstellensatz tells us that is the largest collection of polynomials over that has the same real zero set as :
Theorem 24 (Real Nullstellensatz [25, Cor. 4.1.8]; Table 1).
Polynomial vanishes at every point in the real zero set of if and only if . Equivalently, for .
With the Nullstellensatz and real Nullstellensatz, we may give an “algebra-geometry dictionary” connecting ideals and algebraic sets :
Corollary 25 (Algebra-geometry dictionary [30, Thm. 7, p. 183]; Table 1).
Let (respectively, ). Then (respectively, ) if and only if (respectively, ).
The cited result only deals with the complex case, but the argument is identical given the real Nullstellensatz in place of the Nullstellensatz.
2.1.5 Inequations and Saturation
Just as the definition of an ideal corresponds to the behavior of equations, we need an operation on ideals that reflects the presence of inequations. Note that and if two polynomials are not 0 at some point, then is not zero at the point either. These simple observations motivate the saturation of an ideal by a multiplicative set:
Definition 26.
Let be a field. A subset of is a multiplicative set if , and for all . If , the saturation of by multiplicative set is the set of all such that for some we have . We write to denote the saturation of by .
We think of the elements of as equations and the elements of as inequations; i.e., for all and we include in the system of simultaneous equations and inequations. Lemma 28 (2) makes this interpretation precise. The superscript in the notation refers to the fact that is closed under multiplication and typically the elements of are arbitrary products of some finite subset of elements.
Note that the saturation is an ideal containing and equals the entire polynomial ring if and only if contains some element of .
Usually the saturation adds to or “saturates” with new elements, but sometimes the ideal is unchanged:
Lemma 27.
If is a prime ideal containing no element of , then .
Proof.
: By definition we have for some if , but for prime either or must then belong to . If , then .
: Immediate. ∎
We make the convention that if , then by we mean the saturation of by the multiplicative set generated by (i.e., the set containing and every finite product of elements from ). When we write we automatically assume that even if we do not explicitly state this. Note that if , then .
Our main interest is the case where is finite. In this case it is equivalent to saturate by the single element defined by the product of the . Let . We write to denote the saturation of by the multiplicative set that is generated by the single element (note that is not 0 because ).
The solution set (points that make all elements of , but no elements of , vanish) is closely related to . The following lemma will help prove our central results in Section 3 by making the connection between saturation ideals, equations, and inequations (and hence constructible sets):
Lemma 28.
Let and let be finite. Then
-
1.
and
-
2.
.
Proof.
-
1.
: If , then for some and . Multiplying by appropriate powers of the to equalize the exponents, we see that (recall that ideals are closed under multiplication by anything). Hence .
: This direction holds because is in the multiplicative set generated by .
-
2.
Part 1 implies the first equality. For the last equality:
: Let . By definition of Zariski closure, it suffices to show that for every that vanishes everywhere on .
We claim that . This is equivalent to . (By definition, means that for some natural numbers . Multiplying by appropriate powers of and as in the proof of part 1, we have , whence .) By the Nullstellensatz (Theorem 20), we must prove that vanishes at every point of . But this is true since vanishes everywhere on except possibly at some points of (where vanishes by definition). Hence and so .
: Since is -Zariski closed, it already contains if contains the smaller set . Hence we only need to show that for every and . This holds because for some and so ; we have by assumption, so it must be that .
∎
Lemma 28 (1) holds for any field. In Lemma 28 (2) we work over because the proof uses the Nullstellensatz. The algebraic version of Lemma 28 (2) is:
Theorem 29 (Hilbert’s Nichtnullstellensatz; Table 1).
Let , and let be finite. A polynomial vanishes at every complex solution of if and only if .
Proof.
The case in the proof of Lemma 28 (2) (in particular, the argument there showing ) establishes that vanishing at every complex solution of () implies membership in .
For the other direction, if , then for some natural number . Substitute this for the that appears in the proof of the case in Lemma 28 (2). Follow the argument given there to show that , and hence , vanishes at every complex solution of . ∎
The next result will help prove correctness (Theorem 93) of an important elimination algorithm in Section 4.
Theorem 30 (Splitting, algebraic case [44, Cor. 5]).
Let , and let be finite. If , then
Proof.
The claim follows from Theorem 29 and the fact that every point either makes vanish or not. ∎
In Section 3 we will need the following fact about zero sets of saturations and extending ideals of to :
Lemma 31.
For and multiplicative set , .
Proof.
: Clear because .
: Assume and . Note that the real and imaginary parts of both belong to . Hence both vanish at and as needed. ∎
2.2 Totally Real Varieties: Keeping complex things real
As suggested earlier, real radicals are usually considered computationally intractable [34, 33]. A recurring theme of our work in Sections 3 and 4 is that we can sometimes circumvent the complications of real algebraic geometry by working over . For instance, we would like to use, without loss of precision, complex radical ideals instead of unwieldy real radical ideals. The following is an important condition that, if satisfied, makes this possible.
Definition 32.
Let be a complex variety that is defined over . (That is, for some .) We say is totally real if the real points of are -Zariski dense in ; i.e., .
The key intuition about a totally real variety is that it has “enough real points” for the real variety to closely resemble . More precisely, the real points are not contained in a proper complex subvariety of and in that sense are algebraically indistinguishable from the strictly complex points. Proposition 38 below characterizes this phenomenon in terms of dimension. If is totally real, we can often transfer simpler proofs or more efficient algorithms for to the real variety [45].
The following fact illustrates that being totally real allows us to replace real radicals with complex radical ideals:
Proposition 33 ([46]).
is real radical (i.e.,) if and only if is radical and is totally real.
We need an alternative characterization that is more amenable to computation. To state it, we require the notions of irreducible component, dimension of a variety, and nonsingular point.
Definition 34.
Let be a complex variety (not necessarily defined over ). An irreducible component of is a maximal irreducible complex subvariety of (i.e., is not strictly contained in any irreducible complex subvariety of ).
Example 35.
The complex variety is not irreducible, but it has two irreducible components defined by and (these are lines in ).
Definition 36.
Let be a nonempty complex variety, and let . We define the (complex) dimension of at (written ) to be if i) there are distinct irreducible complex subvarieties of such that and ii) this is the longest such sequence of subvarieties in . If there is no such , we define to be 0. The dimension of is the maximal dimension of at any of its points.
By the correspondence between prime ideals and irreducible varieties, this definition is equivalent to the maximal length of a chain of prime ideals in the coordinate ring of the variety (the quotient of by the ideal of all polynomials in that vanish at every complex point of the variety; this ideal is radical) [42, p. 25]. This version is often referred to as the Krull dimension.
Example 37.
If is finite, where , then . If for some non-constant polynomial , then at all points of . The dimension of at any point is
If a variety is reducible, then different components can have different dimensions. For instance, is a union of a line (, dimension 1) and a plane (, dimension 2). At the intersection point , has dimension 2.
The same definitions are used, mutatis mutandis, for the real dimension of a real variety [25, Def. 2.8.1]; we just replace with and “complex” with “real” in Definition 36. For the prime ideal/Krull dimension version, we use the real coordinate ring (the quotient of by all polynomials over that vanish at the real points of the variety; this ideal is real radical and not just radical). The real dimension could be smaller than the complex if the variety is not totally real (see Example 42), but not otherwise:
Proposition 38 ([47, Thm. 12.6.1 (3)][48, Thm. 2.4]).
Let be an irreducible complex variety defined over . Then is totally real if and only if the real dimension of equals the complex dimension.
Whether or not this happens is determined by nonsingular points.
Definition 39.
Let where and is a radical ideal; recall . The Jacobian matrix corresponding to is the -matrix whose -th entry is the formal partial derivative of polynomial with respect to variable . We say is a nonsingular point of the variety if the matrix (that is, with substituted for ) has rank equal to . (In other words, there is a maximal choice of rows of that form a linearly independent set of -vectors in .) Otherwise we say is a singular point of .
If and both and are radical, then nonsingularity of is independent of the choice of or in the sense that and have the same rank. This follows from the Nullstellensatz (which implies in our case that and [49, Thm. 5.1, p. 32].
Example 40.
Suppose for some square-free (has no repeated factors) ; then is radical. Moreover, matrix has the form and either has rank 0 (if all vanish at ; then is singular) or rank (in which case is a nonsingular point of ).
While the definition is technical, the intuitive picture is simpler: at a nonsingular point the variety looks “smooth”, while at a singular point we find cusps/sharp corners, self-intersections, etc. For instance, points in the intersection of two irreducible components are always singular [42, Thm. 2.9]. Importantly, smooth points give us a characterization of totally real varieties that we can calculate with:
Proposition 41 ([47, Thm. 12.6.1 (4)][50, p. 736]).
Let be a complex variety defined over . Then is totally real if and only if every irreducible component of contains a nonsingular real point of .
In light of Proposition 38, the equivalence given by Proposition 41 is not surprising because the real dimension of a variety (defined over ) at a smooth real point is the same as the complex dimension [51, p. 185].
Decomposition into irreducible components [52] and finding smooth points [48] are both well-studied (albeit computationally intensive) problems in algorithmic algebraic geometry, so we can decide if any given complex variety is totally real.
Example 42.
As a simple example, the complex variety defined by is not totally real because the lone real point is singular, being the intersection of two complex lines (the irreducible components of ). Alternatively, the partial derivatives and vanish simultaneously at ; note that is square-free so is radical. Nevertheless, the equations define the same real points and the corresponding singleton set is a totally real complex variety (see Proposition 43). In terms of dimension, all this happens because has real dimension 0 but has complex dimension 1. We give further examples in Section 3.
In general, totally real varieties contain non-real points as well; the definition only requires that the complex solutions to the defining equations form the smallest algebraic set containing all the real solutions.
The next result shows more generally than Example 42 that being totally real depends on the complex variety and not only on the set of real points.
Proposition 43.
For any , there exists a finite set such that is totally real and .
Proof.
By Hilbert’s basis theorem there is a finite generating set of , the real radical of the ideal generated by . The definition of the real radical implies that .
To show that is totally real, we must establish . By Lemma 17 (2) it suffices to show that any that vanishes everywhere on also vanishes on (i.e, ; the reverse containment is immediate). Splitting into real and imaginary parts (see the proof of Lemma 19), it follows from the real Nullstellensatz (Theorem 24) that both vanish everywhere on and so must belong to (since is real radical). Thus (and consequently ) vanish at each point of .∎
Proposition 43 shows that any real variety has a representation (i.e., some choice of defining polynomial equations) such that the corresponding complex variety is totally real. Theoretically, this suggests that we can always assume our real varieties are totally real as complex varieties. Computationally, though, this is a nontrivial assumption because it may require computing generators of the real radical to transform the representation if the original equations have “too many complex solutions”/define a complex variety that is not totally real.
However, in our experience most systems of polynomial equations over that arise from applications already define a totally real variety. Indeed, the only counterexamples we are aware of have the form where and for all (for instance, sums of squares like in Example 42). The property of being totally real is common and even appears to be the typical situation. Evidence of this is the following theorem, sometimes called the “sign-changing criterion” in the literature:
Theorem 44 ([47, Thm. 12.7.1]).
Let be irreducible. Then is totally real if and only if there exist such that and .
In Section 4.4 we discuss the intriguing relationship between this result and efficiently verifying algebraic invariants (Definition 48) of polynomial dynamical systems.
We extend the notion of totally real varieties to constructible sets. (We are not aware of this definition appearing in the literature, but it is only a minor generalization of the concept for varieties.) This will enable us to use results about saturation ideals over (in particular, Lemma 28 (2)) when handling inequations over .
Definition 45.
Let . We say that the constructible set is a totally real constructible set if ; i.e., the real constructible set is Zariski dense in .
The main way to obtain a totally real constructible set is to take the set difference of a totally real variety with an arbitrary variety:
Lemma 46.
Let . If is totally real, then is a totally real constructible set.
Proof.
By Lemma 17 (2) it suffices to show that if vanishes on , then vanishes on . Let with the goal of showing . Note that for every the product vanishes on all of since vanishes on and vanishes on . It follows by Lemma 17 (1) that vanishes on because is totally real by assumption and so is Zariski dense in . Since , we have for some . Thus and as needed. ∎
2.3 Lie Derivatives and Algebraic Invariants of Polynomial Vector Fields
A system of ODEs
defines a vector field [53] that describes the motion of a hypothetical object moving according to the given differential equations. (In particular, is the velocity vector at .) Abusing terminology, we also refer to the corresponding system as a vector field. In this paper we assume that the are multivariate polynomials over , so satisfies the Picard-Lindelöf existence and uniqueness theorem for ODEs at all points of [54, Thm. 10.VI]. We write for the set .
The rate of change of a function along vector is given by the Lie derivative of with respect to at :
We repeat the procedure on to get higher-order Lie derivatives , (take to be and to be ).
To emphasize the particular ODEs, we sometimes write instead of . If the vector field is understood, we simply write for the Lie derivative . As with , we only consider .
The following is straightforward:
Lemma 47.
Lie derivatives with respect to a vector field obey the sum and product rules: and .
Lie derivatives have a close relationship with invariant sets of a vector field .
Definition 48.
Let be a vector field on defined by a system of ODEs . A subset is invariant with respect to if for every and solution to the initial value problem , we have for all such that is defined. ( could be a proper open subset of if does not exist for all time.)
Intuitively, if is invariant, then an object starting at any point of will remain in as the object follows the dynamics described by . We restrict our focus to invariant sets that are real algebraic varieties. The following well-known result (Theorem 50) characterizes such invariant sets. Its statement uses Lie derivatives of all orders as well as the concept of invariant ideals. An ideal is an invariant ideal with respect to if for all , the Lie derivative belongs to (i.e., is closed under Lie differentiation, written ).
Notation 49.
Given a subset and polynomial vector field , we write (or simply if the vector field is understood) to denote the ideal generated by the collection of higher Lie derivatives for all and . (Note that is distinct from , which is the set of first Lie derivatives of the elements of .)
Observe that is an invariant ideal of by construction because the given generating set is closed under Lie differentiation with respect to .
Theorem 50 (Characterization of algebraic invariants [33, Lemma 5][55, Lemma 2.1]).
Let be a polynomial vector field and let . The following are equivalent:
-
1.
is an algebraic invariant set of .
-
2.
For all such that , we have for all , , and .
-
3.
There exists such that and is an invariant ideal with respect to (i.e., ).
Proof.
: Part (ii) of [55, Lemma 2.1] proves that if is invariant with respect to , then for . Hence vanishes at every point of . Reapply part (ii) of [55, Lemma 2.1] to , which is still the same invariant . This shows that second-order Lie derivatives of the vanish on . Continue the process for any order . (See also [56, Thm. 1] for a variant of .)
: Assuming (2), if then we have ; as noted above, is an invariant ideal.
: This is precisely [33, Lemma 5]. ∎
Corollary 51.
Let be a polynomial vector field and let . If for all , then is an algebraic invariant set of .
Proof.
The useful Lemma 52 and Proposition 53 below are probably folklore but we know of no specific references.
Lemma 52.
If is an invariant ideal of , then is also an invariant ideal.
Proof.
We use all three equivalent statements in Theorem 50. Since is an invariant ideal of , we know is an invariant set by in Theorem 50. We have by Proposition 23 (3b). By Hilbert’s basis theorem (Theorem 2), there exist that generate . Since is invariant and , implies that for all , , and . By the real Nullstellensatz (Theorem 24), each order of Lie derivative belongs to (since the generate ) (Proposition 23 (2)). Hence as needed. ∎
Finally, we show that is the “largest” invariant contained in :
Proposition 53.
Let and let be a polynomial vector field. The real variety is invariant with respect to , is contained in , and contains every invariant set of such that .
Proof.
The ideal is an invariant ideal, so is an invariant set by Theorem 50. We have by Corollary 25 because . For the last claim, by Theorem 50 there is an invariant ideal such that . By Corollary 25, since . Note that is invariant by Lemma 52 and that since . Because is contained in every invariant ideal containing , we have and thus by Corollary 25 . ∎
2.4 Differential Polynomial Rings and Differential Ideals
Differential algebra extends commutative algebra and algebraic geometry to settings that involve differentiation. This makes the theory a natural choice for exploring algebraic invariants of polynomial dynamical systems. In particular, differential elimination [27] is the algorithmic engine that powers our main results, Theorem 86 in Section 3 and Theorem 98 in Section 4.
Definition 54.
An (ordinary) differential field is a field equipped with a single derivation operator ′ that satisfies and for all .
Examples are standard fields with the trivial zero derivation, the fraction field of with derivation given by the quotient rule, and the field of meromorphic functions (quotients of complex analytic functions) with the usual complex derivative [59]. Many of the definitions and results in Sections 2.4,2.5,and 2.6 apply to more general differential fields, but in this paper we restrict to ordinary differential fields of characteristic 0 (repeated addition of never yields 0; equivalently, the field contains the rational numbers as a subfield). Henceforth we simply refer to differential fields, with these restrictions being understood even if not stated explicitly. We generally leave the differential field unspecified and assume that it contains all the solutions that interest us (e.g., real analytic functions solving systems of ODEs with rational number coefficients).
Definition 55.
Let be a differential field and let be an indeterminate. The differential polynomial ring with differential indeterminate and coefficients is the polynomial ring having infinitely many algebraic indeterminates (named in a suggestive way). The derivation ′ on extends to all of by defining , , and for . We call the -th derivative of . The largest such appearing for any variable in a differential polynomial is the order of the polynomial. If we call a proper derivative of . For uniformity we call a derivative (just not a proper one.) Differential polynomial rings having several differential indeterminates are defined analogously and are denoted by . If an element has no proper derivatives of (i.e., ), we say is a nondifferential polynomial.
Intuitively, differential polynomials are standard polynomials except for the presence of variables representing “derivatives” of the (finitely many) differential indeterminates. For instance, is an order 2 differential polynomial of total degree 4 (from the term ) in two differential variables .
The derivatives are formal and do not necessarily represent limits of difference quotients like in calculus; we only require that they obey the sum and product rules. However, just as we can substitute elements of a field for the variables of a nondifferential polynomial, we can substitute differentiable functions into a differential polynomial and treat ′ as the usual analytic derivative. For example, is a differential polynomial and is an element of the differential field of complex meromorphic functions (which contains the elements of considered as constant functions). Substituting for and interpreting ′ as the usual complex derivative, we find that , the zero function (which is the zero element in this differential field).
Note that is also 0 : and . More generally, if for differential field and , then . For instance, implies . Commutativity of differentiation and substitution is analogous to commutativity of polynomial addition and substitution; e.g., .
2.4.1 Differential Ideals
Just as ideals give us an algebraic way to approach polynomial equations, differential ideals correspond to systems of polynomial differential equations.
Definition 56.
Let be a differential field. A differential ideal is an ideal that is also closed under differentiation: implies . A radical differential ideal is a differential ideal that is also a radical ideal in .
Let . We write or just to denote the differential ideal generated by in : the collection of all finite sums where , is the -th derivative of , and is any differential polynomial.
We write or to denote the ideal generated by in viewed as a polynomial ring. (That is, consists of sums where and is any differential polynomial.) If there is risk of confusion, we specify whether is the ideal generated by in or in , but context usually makes it clear (for instance, this could only be an issue if ).
The ideal is defined similarly to , except that in forming we do not allow differentiation of the elements of . Hence but generally . Note that is the minimal differential ideal containing .
We remark that invariant ideals with respect to a polynomial vector field are essentially differential ideals in the sense of Definition 56 (the only difference being that is a polynomial ring and not a differential polynomial ring). By Lemma 47 (sum and product rule for the Lie derivative), the operator is a derivation on the polynomial ring . Since for an invariant ideal , such an ideal is a differential ideal with respect to the derivation . See Lemma 66 for the converse relating a given differential ideal to an invariant ideal.
Radical differential ideals are theoretically and practically more tractable than general differential ideals. We need the following straightforward properties:
Lemma 57.
Let be a differential field and let be a differential ideal.
-
1.
The radical of differential ideal in is also a differential ideal.
-
2.
as ideals in .
Proof.
-
1.
[59, Lemma 1.15].
-
2.
Immediate from the definition of the radical of an ideal. Note that is the radical of taken in while the is the radical of taken in .
∎
Remark 58.
By Lemma 57 (1), the radical ideal is a differential ideal and hence the smallest radical differential ideal containing . This radical differential ideal is often denoted by in the differential algebra literature. Because we work so often with finite sets containing differential polynomials, we do not follow this usage and by we mean the set of differential polynomials with elements , not the radical differential ideal .
2.4.2 Differential Nullstellensatz and Differential Saturation Ideals
On the “geometric” side, we are interested in zero sets of collections of differential polynomials. In the differential setting, there is no specific field playing the role of that contains solutions to all polynomial differential equations with coefficients in the field. (Differentially closed fields [60], named in analogy to algebraically closed fields like , do exist but we do not explicitly need them in this paper. Moreover, there are no known natural examples.) Instead we use the following terminology and notation.
Definition 59.
Let be a differential field with . Let be any differential field extending and let satisfy for all . Then we say that is a point (or element) of the differential zero set . (We write if is understood.) If every point of is also a point of , we write (i.e., for all differential field extensions of , all solutions of in are solutions of .) If also , we write and say that and have the same differential zero sets (or the same differential solutions). (See Remark 60 about our use of the term “set” in this context.)
We also refer to differential zero sets as differential varieties, differential algebraic sets, or Kolchin closed sets (over , when we specify a differential field containing the coefficients of the defining differential polynomials).
Analogous definitions hold for differential constructible sets (i.e., Boolean combinations of Kolchin closed sets). See, for instance, Example 71.
Remark 60.
We could have defined algebraic varieties in an analogous way, having points in any extension field containing the coefficients of the defining polynomials. In that case Hilbert’s Nullstellensatz (Theorem 20) would read as follows: Let be a field and let . Then a polynomial belongs to if and only if for all fields extending and all such that for all , we have . However, is algebraically closed and thus has the property that a system of polynomial equations over has a solution in some extension field if and only if it has a solution in [61]. Hence there is no need to mention extensions in the algebraic case. We merely do so in the differential case because there is no natural analogue of available.
Strictly speaking, the collection of all points of is not a set, but rather a proper class [62] containing the solutions of from all differential fields extending . However, in this paper we never need to treat it as a single completed object and so do not risk set-theoretic difficulties. We simply use as shorthand for universal quantification over differential field extensions. The terms “differential zero set” and “differential algebraic set” are harmless abuses of terminology that we use to mirror the corresponding algebraic concepts. We introduce the notion so that we can conveniently state the differential Nullstellensatz without a technical detour into differentially closed fields. Like its algebraic counterpart, the differential Nullstellensatz connects algebra and geometry, giving us a correspondence between radical differential ideals and differential varieties. In this paper, the differential Nullstellensatz is mainly used in the form of Theorems 63 and 64. In turn, these will help prove correctness (Theorem 96) of our main algorithm in Section 4.
Theorem 61 (Differential Nullstellensatz [44, Thm. 2]; Table 1).
Let be a differential field of characteristic 0. Given , a differential polynomial vanishes at every point of if and only if . (Here is the radical in of the differential ideal generated by .)
Phrased differently, consists precisely of those differential polynomials over that vanish at all differential solutions of in any differential extension field of .
The differential Nullstellensatz immediately implies a differential analogue of the algebra-geometry dictionary from Corollary 25:
Corollary 62 (Differential algebra-geometry dictionary; Table 1).
Let , where is a differential field of characteristic 0. Then if and only if .
Given an ideal and multiplicative set , the saturation ideal is defined as before by . We have . If is a differential ideal, then is also a differential ideal (called a differential saturation ideal) [59, Lemma 1.3].
Much like algebraic saturation ideals, differential saturation ideals represent systems of polynomial differential equations and inequations. (Example 71 and Lemma 94 show that inequations, and not just equations, are important for differential elimination.) The differential Nichtnullstellensatz makes the following important connection between solutions of differential polynomial (in)equations and radicals of differential saturation ideals. (See Theorem 29 for the algebraic version.)
Theorem 63 (Differential Nichtnullstellensatz [44, Cor. 3]; Table 1).
Let be a differential field of characteristic 0, let , and let be finite. A differential polynomial vanishes at every solution of in every differential extension field of if and only if .
The last result in this subsection is a differential analogue of Theorem 30:
Theorem 64 (Splitting, differential case [44, Cor. 5]).
Let be a differential field of characteristic 0, let , and let be finite. If , then
.
2.4.3 Explicit Form and Lie Derivatives in Differential Ideals
Since we intend to compute algebraic invariants of polynomial vector fields, we naturally focus on differential polynomials of the following form:
Definition 65.
Let be a differential field. We say a differential polynomial of order 1 is in explicit form (or is explicit) if for some and . (In particular, no proper derivative appears in .) We also say the corresponding differential equation is in explicit form. (Note that explicit form is desirable because it essentially replaces a derivative with a polynomial.)
In the following lemma and thereafter, if we write as shorthand indicating that each element of belongs to . This property connects differential ideals to invariant ideals and allows us to use differential algebra to find sets invariant with respect to :
Lemma 66.
If and is a differential ideal, then for any we have . (In other words, is an invariant ideal with respect to .)
Proof.
We have because is a differential ideal. Since , all monomials of have the form for some and . (This is one of the summands produced by the product rule applied to a monomial in .) Subtracting from produces an element of (since by assumption) that replaces the monomial with . Doing this for each in monomial and summing the output replaces in with the Lie derivative of . It follows from Lemma 47 that substituting for in this way yields . (Recall that by definition.) ∎
2.5 Rankings and Reduction
As discussed in the introduction, the central aim of this paper is to use differential elimination to algorithmically generate algebraic invariants of polynomial dynamical systems. Elimination identifies the core content of a system of polynomial (or differential polynomial) equations by “reducing” some polynomials with respect to others. In addition to Gaussian elimination, another classic example is long division of one univariate polynomial by another. If the divisor does not evenly divide the dividend, we are left with a nonzero remainder. One way to extend this to multivariate differential polynomials is to use a differential ranking (or ranking, if the context is clear) [63]. Differential rankings identify a “leading derivative” in any differential polynomial (e.g., to determine if or is eliminated first) and ensure termination of algorithms by eliminating “large terms” first. Rankings are analogous to monomial orderings in the theory of Gröbner bases [30, Sect. 2.2, Def. 1]. However, rankings only apply to individual variables and their derivatives, whereas monomial orderings concern monomials. Many of the following results do not depend on the choice of ranking (the main exceptions are algorithm on p. 4.1 and the proof of Theorem 96) and we do not specify a ranking except when necessary.
Definition 67.
A differential ranking is a well-founded linear ordering of derivatives (i.e., every nonempty subset of derivatives has a least element with respect to ) that goes up with differentiation and respects ′: and implies .
There are different rankings for different purposes. Two prominent classes are elimination rankings (sort variables lexicographically irrespective of their order; e.g., if , then for any ) and orderly rankings (the derivative of highest order ranks highest, regardless of the base variable; ties in order are decided according to the ordering on the variables). For instance, if for an orderly ranking, then if and if .
We identify several important components of a differential polynomial with respect to a ranking. These constituents are useful for specifying the control flow of algorithms involving systems of differential polynomials. (For instance, in Section 4 we split cases by either setting a given separant equal to 0 or not.)
Definition 68.
Fix a differential ranking. The highest-ranking derivative that appears in a non-constant differential polynomial is the leader of (denoted ). Note that is a derivative of a single differential indeterminate and does not involve powers or multiple variables. The initial of (denoted ) is the coefficient (viewing as a univariate polynomial in with coefficients in that do not involve ) of the highest power of . The separant of (denoted ) is the initial of any derivative where (equivalently, the formal partial derivative of with respect to ). Unlike the initial , the separant could contain , but with a lower degree than has in . We do not define leaders, initials, or separants for constants (i.e., elements of the field ).
Example 69.
Fix any differential ranking such that and let . Then
The leader of is (and not , , etc.), the initial is the coefficient of , and the separant is .
Remark 70.
Rankings induce a well-partial-ordering on differential polynomials ( if the leader of is greater than the leader of or if they are the same and the degree of the leader is greater in than in ) and, in turn, on sets of polynomials. See [58, Sect. I.10] and [29, Ch. 5] for exact definitions; also see our remarks about the RGA algorithm at the beginning of Section 4.
Rankings for nondifferential polynomial rings are simply linear orderings of the variables. The leader, initial, and separant of a polynomial are defined as in the differential case.
2.5.1 Differential Pseudodivision
Rankings give us a systematic way of generalizing long division via differential pseudodivision (or just pseudodivision; we also say differential pseudoreduction or Ritt reduction). Here we view multivariate differential polynomials as univariate differential polynomials whose coefficients are differential polynomials in one fewer variable. Differential pseudodivision is like univariate long division except the coefficients are differential polynomials and we usually cannot divide without introducing fractions. We give a concrete example before describing the process more formally.
Example 71.
Choose any differential ranking with and let and . Observe that is the leader of and is the highest derivative of in . Pseudodividing/pseudoreducing by consists of first differentiating to match the and then “dividing” by (in quotation marks because we must premultiply by something in order to divide by without fractions). This eliminates . If the resulting pseudoremainder contains , we “divide” the pseudoremainder by to obtain another pseudoremainder. The final reduced pseudoremainder does not contain any proper derivative of and has degree less than 2 in .
-
•
Differentiate to find . (Note that has degree 1 in .)
-
•
Then premultiply by because is the coefficient of in and .
-
•
Now subtract from to obtain pseudoremainder
Notice that vanish precisely when either all vanish or vanish and does not. Using the notation of Definition 59 we have . In other words, using splitting and differential pseudodivision to eliminate differential polynomials splits differential varieties into a union of differential constructible sets. The same idea appears in Equations 3,4 on p. 4.
-
•
With one more round of premultiplication and subtraction, we can eliminate in . Premultiply by (this is the initial of and doesn’t already contain any factors of ) and subtract to obtain the final answer . This pseudoremainder has no variables that can be eliminated using .
In general, let be differential polynomials ( non-constant so that leaders, etc., are well defined; in particular, ). Pseudodividing by involves the following steps (see [58, Sect. I.9] and [64, Sect. 2.2] for other versions):
:
-
1.
If a proper derivative of the leader of appears in :
-
•
Let , , be the highest proper derivative of appearing in . Compute .
-
•
Let . (Note that the initial of is the separant of since .) Then return .
(Note that the call must proceed to step 2 because is the leader of . Also note that the degree of in is 1.)
-
•
-
2.
If (but no proper derivative thereof) appears in and the highest power of that appears in is at least the degree of in :
-
•
Let be the coefficient of in . Multiply by an appropriate factor of the initial of to ensure that is divisible by . This is called premultiplication. It suffices to premultiply by , where is the GCD of and .
-
•
Now subtract the necessary multiple of (namely, ) to eliminate from and obtain a pseudoremainder whose highest power of is less than :
Note that and are both differential polynomials and not fractions because divides both and by definition. Also note that the total degree of is at most the sum of the total degrees of and . (This is because degrees of products are additive and the degrees of and are at most and , respectively. Hence , but for simplicity we use the larger bound .)
-
•
Return .
If (but no proper derivative thereof) appears in and , return .
-
•
-
3.
If neither nor any proper derivative of appears in , return .
Remark 72.
The algorithm terminates because step 1 reduces the order of the highest proper derivative of that appears in and step 2 reduces the degree of in to a value below . (In particular, is eliminated by the call because the degree of in is .)
The same concepts apply in the nondifferential case (i.e., rankings on variables in ). Algebraic pseudodivision is the same process as , except that step 1 never applies because there are no derivatives.
If , we say is the (differential) pseudoremainder resulting from (differential) pseudodivision of by . The core component of pseudodivision is a single round of premultiplication and subtraction (the first and second items in step 2 of ). We refer to this as a “single step” of pseudodivision and call the result a pseudoremainder even though it is an intermediate element that may be used for further steps of pseudodivision until the final pseudoremainder is reached. See Remark 91 immediately preceding the description of algorithm .
The next proposition makes explicit the differential-algebraic relationship between , and when . (The analogue for division of integer by nonzero integer is the equation , where is the quotient and is the remainder.)
Proposition 73.
Let . Then for some a product of factors of , a product of factors of , and we have . In particular, we have if is 0.
In the nondifferential case (or if contains no proper derivatives of the leader of ) we have , with now, and if is 0.
Proof.
This follows from the form of operations in . If contains a highest proper derivative of the leader of , then the first intermediate pseudoremainder (from the call using step 2 is
where is the initial of , is the highest power of in , is the coefficient of in , and is the GCD of and . Note that has the claimed property. (Consequently, Proposition 73 also applies to a single step of pseudodivision as defined in Remark 72.) By induction on and we may assume that for some a product of factors of , a product of factors of , and . Multiplying by and subtracting , the two preceding equations imply that
which has the correct form.
Analogous arguments hold for the case that contains no proper derivative of and the nondifferential case. The assertions for follow from the definition of a saturation ideal. (Multiply both sides by appropriate factors of so that and become powers of and not just arbitrary products of their factors.) ∎
Just as the process in Example 71 eliminated and cut the degree of from 2 to 1, more generally we have the following definition:
Definition 74.
Given a differential polynomial ring , a differential ranking, and , we say is partially reduced with respect to if no proper derivative of the leader of appears in . (In the nondifferential case there are no derivatives and so partial reducedness holds vacuously.) We say is reduced with respect to if is partially reduced with respect to and any instances of in have strictly lower degree than the maximum degree of in . (In particular, it is impossible to pseudodivide by any further.) We say that a subset is partially reduced if all members of are pairwise partially reduced. We likewise call autoreduced if the elements of are pairwise reduced. Similarly, we say is (partially) reduced with respect to if is (partially) reduced with respect to each element of .
It follows from Remark 72 that the pseudoremainder resulting from pseudodividing by is reduced with respect to . After a “single step” of pseudoreduction as defined in Remark 72, the pseudoremainder is not necessarily reduced yet with respect to , but it does have degree in strictly less than that of .
2.6 Regular Systems
Systems of polynomial equations and inequations can “hide” their information in the sense that deciding the system’s properties may require substantial computation. We focus on structured collections known as regular systems and regular sets that are easier to analyze. The basic intuition is that “enough” pseudodivision has been done beforehand so that there is no redundancy left to obscure relations between the system’s elements.
Definition 75 ([44, Def. 1]).
Let be a field, let be finite with , and fix a ranking. The equation/inequation pair (or simply ) denoting and for each and , respectively, is a regular algebraic system over if
-
1.
the elements of have distinct leaders and
-
2.
contains the separant of each element of .
Definition 76 ([44, Def. 7]).
Let be a differential field, let be finite with , and fix a differential ranking. The equation/inequation pair is a regular differential system over if
-
1.
is partially reduced and the elements of have distinct leaders and
-
2.
the elements of are partially reduced with respect to , and contains the separant of each element of .
Our statement is simpler than the general definition [44] because we restrict ourselves to a single derivation. The “coherence property” in [44] holds vacuously in the ordinary differential case. Definitions 75 (1), 76 (1) force to be finite (since there are only indeterminates), so there is no loss of generality in assuming is finite to begin with.
We often omit “over ” when the field is clear from context. Also, nonzero constant multiples in are irrelevant because we interpret the elements of as inequations. Thus, for example, if is the separant of and , we take that as satisfying the requirement that contain . Similarly, we do not explicitly show nonzero scalars when listing the elements of even if, like in , the separant is 1.
An equation/inequation pair from (respectively, ) is an algebraic system (respectively, differential system) if it is not necessarily regular.
While an arbitrary differential system need not be regular, there always exists a decomposition into one or more regular differential systems (Theorem 90). Our main algorithm from Section 4 yields such a decomposition if the differential equations have explicit form (Theorem 96).
Definition 77.
Let be a differential field and let . The restrictions of and , respectively, to are and . If is a differential system, we call the restriction of to .
Every regular differential system naturally contains a regular algebraic system.
Lemma 78.
Let ) be a regular differential system. Then the restriction of to is a regular algebraic system (with respect to the ranking inherited from that of ).
Proof.
The elements of have distinct leaders, as do all elements of . By definition ; this implies that the separants of elements of belong to because contains the separants of elements of and the separants of elements of belong to . ∎
For some purposes regular systems are not strong enough. In particular, they do not suffice for testing membership in or . For that we need the related notion of a regular set. We first give the definition, but the connection to saturation ideal membership is contained in Theorem 81. This in turn plays an important role in the proof of Theorem 97 in Section 3.
Definition 79 ([65, Def. 2]).
Let be a field, let , and fix a ranking. For let and let be the multiplicative set generated by the initials of . We say is a regular set (or regular chain) if
-
1.
the elements of have distinct leaders and
-
2.
for , if and does not belong to , then .
In the usual terminology of the area, is a triangular set (having distinct leaders is suggestive of the triangular shape of an invertible matrix in row echelon form) and for the initial of is regular (i.e., a non-zerodivisor) in the quotient ring .
2.6.1 Key Results for Regular Systems
To round out the necessary mathematical background, we cite three deeper properties of regular systems and sets that we need in Sections 3 and 4. The first is a technical lemma that, in light of the Nullstellensatz, allows us to go back and forth between zero sets and ideals when dealing with regular algebraic systems.
Lemma 80 (Lazard’s lemma [44, Thm. 1.1 and Thm. 4]).
Let be a regular algebraic system. Then is a radical ideal; i.e., . Likewise, if is a regular differential system, then is a radical differential ideal; i.e., .
The second gives a decision procedure for membership in saturation ideals determined by regular sets.
Theorem 81 ([65, Prop. 11][66, Thm. 6.1]).
Let be a regular set over field and let be the multiplicative set generated by the initials of the elements of . Then for any , we have if and only if the pseudoremainder of with respect to is 0 (i.e., the pseudoremainder after reducing as much as possible with respect to all elements of ).
This is analogous to how Gröbner bases decide membership in arbitrary polynomial ideals [30, Sect. 2.6, Cor. 2]. See Theorem 90 for a differential version of Theorem 81.
The last result provides a link between differential and algebraic ideals. This is critical because computation becomes less complicated when we do not have to keep differentiating. Moreover, theory and tools for symbolic computation are more developed in the algebraic case than the differential.
Lemma 82 (Rosenfeld’s lemma [44, Thm. 3]).
Let ) be a regular differential system. Then for all differential polynomials partially reduced with respect to , we have if and only if .
Remark 83.
Since nondifferential polynomials are (trivially) partially reduced with respect to any set, Rosenfeld’s lemma implies that .
The key lesson of Rosenfeld’s lemma is that, given a regular differential system, a partially reduced polynomial belongs to the differential saturation for essentially algebraic reasons; we do not have to differentiate to prove it.
3 Regular Differential Systems and Algebraic Invariants of Polynomial Vector Fields
3.1 Explicit Regular Differential Systems
We have now covered the background needed for our new method that uses differential elimination to generate algebraic invariants of polynomial dynamical systems. The principal results of the current section are Theorems 86 and 88. Example 3.3 demonstrates these theorems using the well-known Lorenz system.
To abbreviate theorem statements, we make the following definition that specifies our systems of interest. As stated earlier, we restrict to differential polynomials in explicit form (Definition 65) because we want to apply the theory to finding algebraic invariants of polynomial vector fields. Moreover, this application naturally concerns nondifferential inequations (which we can use, for instance, to indicate “unsafe” locations/states).
Definition 84.
Let be a differential field and let ) be a differential system over . We say is an explicit differential system over with nondifferential inequations (or just an explicit system with nondifferential inequations if and the differential system are understood) if i) all elements of that have a proper derivative are in explicit form and ii) all elements of are nondifferential polynomials (in which case ).
We prove an important technical lemma that, along with Rosenfeld’s lemma (Lemma 82), implies the central result of this section (Theorem 86).
Lemma 85.
Let ) be an explicit regular differential system over with nondifferential inequations. Then .
Proof.
The reverse containment is automatic. For the forward containment, we must show that if is a nondifferential polynomial, then . Each explicit differential polynomial in has the form for some variable and nondifferential polynomial . Let be the subset of indices for which such an belongs to . The remaining elements of belong to . Since and , there exist nondifferential polynomials and as well as differential polynomials such that .
We claim that vanishes at all complex solutions of the restriction . (In other words, . Recall that is the product of the (finitely many) elements of .) Let be such that and for all and (in particular, the and from the previous paragraph). By Peano’s existence theorem for (complex-valued) ODEs [54, Thm. X, p. 110], there exists a solution to the initial value problem (IVP) , , . (If , for the IVP simply let . For such , the choice does not affect the following argument.)
Substituting into the various differential and nondifferential polynomials, we obtain
Evaluating at , we find
(We have written instead of because are differential polynomials and the function must be substituted into and differentiated before evaluating at . We similarly write instead of , which is simply 0.) Since and (by assumption on , and ) and (because solves the IVP), this proves that and establishes the claim that .
Lemma 17 (1) now implies that . Lemma 28 (2) converts this to , which equals by the Nullstellensatz (Theorem 20). It follows from the definitions of radical and saturation ideals that for some , so by Lemma 4 we have . Lazard’s lemma (Lemma 80) gives since is a regular algebraic system by Lemma 78. This completes the proof. ∎
The proof of Lemma 85 illustrates the theme of proving things about real polynomial systems by first going up to the complex numbers (see Definition 32 and the comments preceding it). This strategy also appears in Theorem 88, Section 3.3, and Section 4.4.
We are ready to give the connection between explicit regular differential systems and algebraic invariants. For convenience going forward, we sometimes refer to this result as the “regular invariant theorem”.
Theorem 86 (Regular invariant theorem).
Let ) be an explicit regular differential system over with nondifferential inequations.
Let be a polynomial vector field such that . Then is an algebraic invariant set of .
Proof.
By Lemma 50, it suffices to prove that is an invariant ideal. Let with the goal of showing (recall that is the Lie derivative of with respect to ). The saturation is a differential ideal, , and by assumption , so by Lemma 66 we obtain . As is nondifferential and thus partially reduced with respect to , Rosenfeld’s lemma (Lemma 82) yields . Because , Lemma 85 implies as desired. ∎
While it is not necessary that for the regular invariant theorem to hold (see the example in Section 3.3), the hypothesis that implies that is a differential-algebraic consequence of (by the differential Nichtnullstellensatz, Theorem 63.) How we obtain explicit regular differential systems in the first place is a central topic in Section 4.
3.2 Alternate Representations from Additional Hypotheses
As we will see in Section 3.3 and Section 4, Theorem 86 opens the door to novel ways of finding and analyzing algebraic invariants of polynomial vector fields. However, it poses the challenge of finding generators of the saturation ideal if we want to explicitly write equations for the invariant. While this is possible using Gröbner bases [30, p. 205], it adds complexity to the process (Remark 114). We would much prefer to read off the invariant directly from and . To explore this possibility, we start by noting that is sandwiched between two more convenient sets. (The following lemma does not depend on regular systems, so we use generic names in place of and .)
Lemma 87.
If , with finite, then .
Proof.
We have by Lemma 13 since is a real constructible set. (Recall that denotes the Euclidean closure of a set . Though in this case the closures coincide, we invoke the Euclidean topology because of its visually intuitive nature compared to the Zariski topology.)
The containment holds because is a real Zariski closed set containing (this follows quickly from the definition of a saturation ideal; see the case in the proof of Lemma 28 (2)). The last containment holds because . ∎
It is possible that both and , in addition to , must be invariant under the hypotheses of the regular invariant theorem. We cannot yet prove or disprove this conjecture. However, we can prove invariance of the various sets using additional hypotheses that are commonly satisfied (see also Theorem 112, the discussion following it, and Remark 114):
Theorem 88 (Alternative criteria for regular invariants).
Let ) be an explicit regular differential system over with nondifferential inequations. Let be a polynomial vector field such that . Then each of the following conditions is sufficient for the indicated set to be an algebraic invariant set of .
-
1.
If is a totally real constructible set, then is invariant.
-
2.
If is irreducible over and is nonempty (i.e., has a real solution), then is invariant.
- 3.
Proof.
-
1.
The following chain of equalities establishes the claim:
- 2.
-
3.
As indicated, the hypotheses are chosen to mimic the proof of the regular invariant theorem using instead of . In particular, let with the goal of showing . (By Lemma 47 it suffices to consider generators of .) Now and it follows from the assumption about monomials in derivatives of elements of that . (Replace each in with and distribute. The assumption about monomials implies that .) Since is nondifferential and by assumption , we conclude that .
∎
Remark 89.
3.3 Example: Lorenz equations
The ODEs comprise the famous Lorenz equations [67]. Depending on the parameters, this nonlinear system can display widely varying behavior, including chaotic dynamics [68]. The literature also contains studies of algebraic invariants for the Lorenz system [69, 70] . The benchmark collection from [71] considers the parameters , yielding the particular equations that we abbreviate as .
Fix an orderly ranking with . We analyze . Thus . We discuss how we obtained these sets in Section 4.2; for now we take them as given and confirm the hypotheses of our preceding theorems.
We claim satisfies the regular invariant theorem (Theorem 86) and each part of Theorem 88. In particular, the strongest conclusion holds: is invariant with respect to .
(Theorem 86 applies) Clearly, is an explicit regular differential system over with nondifferential inequations. Differential polynomials and belong to , but the presence of implies that cannot also be in lest not be partially reduced. However, we can check with standard computer algebra systems (CAS) by confirming , where are new algebraic indeterminates. That is, we consider ideal membership in the nondifferential polynomial ring . This establishes and Theorem 86 implies that is an algebraic invariant of .
(Theorem 88 applies) To check the additional hypotheses in Theorem 88, we first show that is irreducible over . (The weaker condition of irreducibility over suffices for parts 2 and 3 of Theorem 88, but we prefer to prove the stronger result that is helpful for part 1.) In this case it is simple to work directly: reducibility would imply that for some . But distributing and comparing to the coefficients of gives an inconsistent system: (Inconsistency is conveniently shown by using a CAS to conclude . Thus the polynomial and the variety are irreducible over .
-
1.
(Part 1) Since only has one irreducible component, we just need to find one real smooth point to prove that (and hence , by Lemma 46) is totally real. An obvious choice is , which is smooth because is prime (and hence radical) and evaluated at is . Alternatively, is totally real by Theorem 44 since is irreducible, , and . It follows that satisfies Theorem 88 (1).
-
2.
(Part 2) Irreducibility over implies irreducibility over and , so satisfies part 2.
-
3.
(Part 3) Ideal is prime and is empty since does not divide . Remark 89 thus implies that .
Lastly, since , we must show that , and belong to . This was done above in our proof of .
4 The Rosenfeld-Gröbner Algorithm for Algebraic Invariants
A modified version of the Rosenfeld-Gröbner algorithm (RGA) of Boulier et al. [44, 37] is our main tool for differential elimination. We explain the basic ideas behind RGA and then formulate , our handcrafted version that produces explicit regular differential systems with nondifferential inequations (and hence invariants by the regular invariant theorem). The subscript o stands for “ordinary” because our setting involves ODEs of a special form.
4.1 for Explicit Systems
Unlike Gröbner basis techniques that output generators of an ideal, RGA uses a differential generalization of characteristic sets. Given a ranking, a characteristic set of ideal is by definition a minimal-rank autoreduced subset of (Remark 70). Such a is not necessarily unique and might not be a generating set of , but is nonempty, finite, and pseudoreduces every element of to zero [29, p. 175]. The same definition and properties hold for differential characteristic sets if we use differential rankings, differential ideals, and differential pseudoreduction. Given a differential ranking and a system of polynomial differential equations and inequations, RGA outputs a finite collection of special differential characteristic sets called regular differential chains [72] or simply chains; see parts 2, 3 of Theorem 90. (Recall Definition 79 and Theorem 81 for the nondifferential version.) Multiple chains in the output come from case splits over the vanishing of initials and separants. For this reason, RGA typically outputs a “generic” chain (obtained by placing initials and separants with the inequations as much as possible while maintaining consistency) and several more specific ones; see [73] for precise definitions and Section 4.2 for an example that continues the one in Section 3.3.
Regular differential chains contain important “geometric” information. Take an input differential system of equations and inequations and perform RGA. Then a differential polynomial is zero at all points that satisfy if and only if is differentially pseudoreduced to zero by each chain that RGA returns given input [44, Cor. 3 and Thm. 9]. By the differential Nichtnullstellensatz (Theorem 63), this gives an algorithm for testing radical differential saturation ideal membership. This is especially noteworthy because, unlike the algebraic case, general differential ideal membership is undecidable (even for finitely generated differential ideals) [74].
We cite the important properties guaranteed by RGA. Compare Theorem 90 below to Theorems 96 and 98, our main results in this section:
Theorem 90 ([44, Thm. 9]).
Let be a differential field of characteristic 0, let be finite with , and fix a differential ranking.
Then RGA applied to returns differential systems such that
-
1.
,
-
2.
each is a regular differential system (together with item 3, this makes each a regular differential chain), and
-
3.
for all and , we have if and only if the differential pseudoremainder of with respect to is 0.
Strictly speaking, for differential elimination results to be computationally meaningful we must be able to algorithmically add, multiply, divide, differentiate, and check equality with 0. We always tacitly assume this about the finitely many differential field elements that appear during a computation. (Our inputs are finite sets of differential polynomials and so there are only finitely many coefficients; all intermediate elements result from these via arithmetic operations or differentiation.) The assumption is mild because in practice coefficients typically belong to .
The original authors of RGA gave two versions of the algorithm [37, pp. 162-3] [44, p. 111]. RGA has been implemented in the Maple computer algebra system, where it forms the heart of the package [75]. While this tool is convenient (for example, we use it in the example from Section 4.2), the proprietary nature of Maple impedes a full analysis of the implementation and its performance. As an alternative, Boulier makes freely available the C libraries on which the Maple implementation is based [76]. RGA has proven its versatility by admitting refinements and extensions over the past two decades [77, 78, 44, 79, 80, 81].
Published applications of RGA include parameter estimation for continuous dynamical systems [82, 83] and preprocessing of systems for later numerical solution [84]. The literature contains various case studies from control theory [85], medicine [86], and mathematical biology [64].
We give a modified version of RGA that we call and that is well-suited for analyzing systems in explicit form. In some ways our algorithm is simpler than those in the literature. In particular, we produce regular differential systems but do not guarantee that they are regular differential chains. This helps us control the form of the output and more easily obtain explicit bounds. Our application to algebraic invariants (culminating in Theorem 98) does not require radical differential ideal membership testing, so regular differential systems suffice for our use case.
First we describe a purely algebraic algorithm, , that we use as a subroutine in . The general structure of mirrors that of but does not have to deal with derivatives. After presenting both algorithms, proving them correct, and interpreting the output, we analyze their complexity.
Intuitively, is a recursive divide-and-conquer algorithm that decomposes the radical of a saturation ideal into an intersection of radical ideals determined by regular algebraic systems. (The same high-level description also applies to , but in that case using differential polynomial rings and ideals.) Case splits are based on assigning initials and separants either to the set of equations or inequations.
Remark 91.
:
-
•
Choose the ranking . (The particular choice is not essential; we specify a ranking for concreteness.)
-
•
Input: a finite set of polynomials that determines the equations . Let be a finite set of polynomials in corresponding to inequations. (That is, the input is the algebraic system ). We abuse terminology slightly by calling itself a set of equations and itself a set of inequations.
-
•
Output: Pairs such that are finite sets of polynomials and
where is a regular algebraic system (i.e., is triangular and contains at least the separants of ).
-
1.
If is already triangular and separant belongs to for every , then return .
Otherwise, choose the highest-ranking leader that either appears in multiple elements of or appears in only one but . We call the target variable for the pair . Choose some that has minimal degree in among all elements of having leader ; if there are multiple such polynomials, pick one that has least total degree among those with minimal degree in (again there might be several).
We define two auxiliary sets that we need in steps 2 and 3.
-
•
Let , where is the tail of (i.e., what is left of if we substitute zero for the initial of ).
The notation represents the degree of variable in polynomial . The set is finite; the parentheses around separate the union from the set difference and do not indicate an ideal. Note that the solutions of are the same as those of . This replacement is not technically pseudodivision, but it behaves similarly and in Theorem 93 (correctness of ) and Lemma 94 we analyze this case together with the pseudodivision steps.
Note that does not appear in and . Also note that the target variable of could still be or might have strictly lower rank, but cannot have higher rank than .
-
•
Let . (Note that is not 0.)
-
•
-
2.
If appears in multiple elements of , choose some that has maximal degree in among all elements of having leader . If there are multiple such polynomials, pick any one that has greatest total degree among those with maximal degree in (again there might be several). Now pseudodivide by and let be the resulting pseudoremainder. Update the equations by omitting and including ; let .
Note that and , but we do not guarantee because we use a single step of pseudodivision (Remark 91) and is not necessarily reduced with respect to .
Return the union
.(Note: It is convenient to describe these recursive calls in terms of splitting the computation into different branches. Here in step 2 we have split twice. On one branch we included with the equations, left the inequations unchanged, and called on the updated system . On the other branch we included with the inequations, pseudodivided by , and then split again over including with the equations or the inequations. This led to and , respectively.)
-
3.
If is the lone element of with leader , then we consider two cases:
-
(a)
If , return the union .
-
(b)
If , then because if does not appear in , then and . Pseudodivide by and let be the resulting pseudoremainder; note that .
Return the union .
(Note: As in step 2, we say that we have split over including with the equations or inequations, and then likewise over .)
-
(a)
Remark 92.
Though not necessary for correctness of , in practice it is essential to trim inconsistent branches (i.e., pairs that have no solution in the reals) as the algorithm progresses. Discarding such branches does not lose any solutions. Analogously, state-of-the-art versions of Buchberger’s algorithm for Gröbner bases avoid redundant -polynomial calculations by finding an appropriate subset of the possibilities [30, Sect. 2.10]. A simple optimization of would be to omit branches that add a nonzero constant to the equations or 0 to the inequations. Likewise, we can spot inconsistency by inspection if a nonzero constant multiple of a polynomial in the equations shows up in the inequations or vice versa. In general, though, it may be necessary to compute a regular chain (see Theorem 81) or a Gröbner basis of to determine if is solvable in (in would require even more, such as real quantifier elimination [25, Prop. 5.2.2]). These calculations can be expensive, so good judgment is required to pick a strategy for efficiently detecting most of the inconsistent branches. In any case, the worst-case complexity (Theorem 101) is not affected by the lack of consistency checking in since we must always take into account the possibility of splitting.
We now prove termination and correctness of . The termination argument is somewhat delicate because many different cases can emerge during a run of the algorithm. To ensure termination we desire a well ordering that strictly decreases in all cases we could encounter. This demands multiple criteria for ranking sets of equations and inequations and requires a cleverly crafted ordering of the list at the beginning of Theorem 93’s proof.
Theorem 93 (Termination and correctness of ).
Given finite sets of polynomials , algorithm terminates and the output satisfies
where , and is a regular algebraic system (i.e., is triangular and contains at least the separants of ).
Proof.
We first prove termination. This follows from a partial well ordering on pairs where are finite sets of polynomials. Given such pairs , we say that has higher rank than and write if one of the following conditions holds. (Recall that the target variable of pair is the variable of highest rank such that either contains multiple elements having leader or has only one element with leader and the separant of does not belong to .)
-
1.
The target variable of has higher rank than the target variable of .
-
2.
The two pairs have the same target variable , but appears with strictly larger degree in .
-
3.
The target variable and maximal degree in the target are the same for both pairs, but has strictly more elements with maximal degree in the target.
-
4.
All the previous quantities are the same for both pairs, but has strictly more elements whose leader is the target.
-
5.
All the previous quantities are the same for both pairs, but an element of with minimal nonzero degree in the target variable has strictly higher degree in than the corresponding element of .
More formally, we are using ordered lexicographically with the standard orders on and . The subscript of the target variable belongs to and the four copies of represent the highest degree in the target variable, the number of elements with the highest degree in the target, the number of elements having the target as their leader, and the minimal degree in the target, respectively.
recursively transforms the input pair into more refined pairs via finitely many splitting and pseudodivision operations. Thus to show termination it suffices to show that each recursive call in the body of acts on a pair of strictly lower rank than that of .
We first note that in all cases the target variable can only remain the same or decrease. This is because we never remove anything from the inequations and the only variables that can appear in the elements added to the equations are the original target variable or variables of lower rank. If the target variable remains the same, then the maximal degree in the target must remain the same or decrease because anything we add to the equations has degree in the target that is strictly less than the original maximal degree in that variable. Hence we must show that one of the remaining quantities decreases if the quantities that lexicographically precede it stay the same.
-
•
Step 2 of deals with the case that has multiple elements with leader . First we produce the pair , where , is the target variable of , and is an element of having leader and minimal degree in . To see that , we observe that
-
–
If is maximal as well as minimal, then the number of elements of maximal degree in decreases because we removed , the initial does not contain and .
-
–
If is not maximal, then the number of elements of maximal degree in stays the same. If does not appear in , then the number of elements with as the leader decreases (we replaced with and does not appear in ). If does appear in , then the number of elements with as the leader stays the same, but the minimal degree in decreases (because .
Step 2 also produces the pairs and , where and for having maximal degree in and the pseudoremainder upon pseudodividing by . The number of elements having maximal degree in decreases because and have strictly lower degree in than and we remove .
-
–
-
•
Step 3 of deals with the case that is the lone element of with leader .
-
–
If , then step 3 produces the pair . (We already have shown that , so here we just look at ). In this situation the target variable must decrease because is the only element of with leader and now belongs to the inequations (so by definition is no longer a target variable).
-
–
If , then step 3 produces the pairs and , where appears in and is the pseudoremainder upon pseudodividing by . In the first instance, because either the target variable decreases or it remains the same and the maximal degree in decreases (since , and was the lone element of with leader ). In the second, because the target variable must decrease ( was the only element of with leader and now belongs to the inequations).
-
–
This proves termination. We now establish the properties of the output systems .
Each is triangular and contains the separants of all elements of (i.e., is a regular algebraic system) because this is the base case that returns an explicit answer instead of a recursive call. (Note that because we only ever add or to the inequations, and these are never the zero polynomial.)
Lastly, we show that the ideals decompose as claimed. Observe that because each is radical by Lazard’s lemma (Lemma 80). Each operation during a run of converts a system into another system or two systems . By induction on the number of splitting and pseudodivision operations during a run of , it suffices to confirm at each step that or .
-
1.
(Splitting over an initial or separant) Steps 2 and 3 both split the computation into branches, one where the initial or separant of a chosen polynomial is included with the equations and one where it is included with the inequations . Thus by Theorem 30, we have, respectively,
or
-
2.
(Pseudodivision) Steps 2 and 3b of replace some with its pseudoremainder upon pseudodividing by some , where and the initial . (In step 2, is and is while in step 3b, is and is . In the latter case, the initial of is a nonzero constant multiple of (which does belong to the inequations ) because separants are partial derivatives with respect to the leader.)
We must show
(1) By Proposition 73 (which makes explicit the relationship between a pseudoremainder and the original polynomials) there exist and a product of factors of such that . Thus for any point we see that causes both and to vanish but not if and only if causes and to vanish but not . (If , then because is a product of factors of .) Symbolically,
(2)
This proves the claimed properties of . ∎
We can now describe the full algorithm. eliminates ODEs in explicit form using differential pseudodivision; moreover, no new ODEs are introduced. Between pseudodivision steps, we apply to the nondifferential elements of the current system. The computation recursively splits into multiple branches that yield different parts of the desired differential radical decomposition.
Importantly, the eliminated ODEs still belong to all differential saturation ideals that appear during the computation. This is ensured by a technical assumption on that is given in the input description below. (The correctness proof for , Theorem 96, shows how the assumption accomplishes this.) Without the assumption, the operations of could produce differential saturation ideals that do not satisfy the hypotheses of the regular invariant theorem (Theorem 86). In turn, that would break the connection between and algebraic invariants that we establish in Theorem 98.
:
-
•
Choose the orderly differential ranking such that . (We will use the fact that the ranking is orderly to prove correctness, but the particular sequence of variables is not essential.)
-
•
Input: Let be a polynomial vector field. Let be a finite set where is a subset of and the are nondifferential polynomials. Also, we have a finite set of nondifferential polynomials . As in , the elements of correspond to equations and the elements of to inequations. Lastly, the “technical assumption” mentioned above: we assume that for each member of we either have (equivalently, ) or
for some , and . (This is a specific case of . The key point is that for all , either explicitly belongs to the equations or is implied by and the inequations .)
-
•
Output: Pairs such that are finite sets and
where each is a regular differential system. That is, is partially reduced (no element contains a proper derivative of the leader of another element) and triangular (no two elements have the same leader), and is partially reduced with respect to and contains at least the separants of . Moreover, the elements of containing proper derivatives form a subset of , and for each and we either have or
for some , and . (The proof of Theorem 96 shows that the technical assumption on at the start is preserved by each operation of , and hence holds of the output.)
-
1.
If is already partially reduced and triangular, and separant belongs to for every , then return .
-
2.
If is not triangular or separant for some nondifferential , compute . (Recall the are the nondifferential elements of . Also, an element of the form has separant 1, so as usual we assume such a separant belongs to the inequations .) This yields a finite collection of regular algebraic systems involving only nondifferential polynomials. Return the union over all returned by . (That is, for each such , re-include with the equations the elements of that contain proper derivatives and call on the resulting system. This produces a finite collection of differential systems; take the union of all these collections.)
-
3.
If is triangular and for all but is not partially reduced, choose the highest-ranking leader such that and there exists having leader . (Such and exist because is not partially reduced. Because is triangular and is the highest-ranking variable with the desired property, and are unique.) For concision we write as .
Recalling that by assumption in the current case, differentiate and pseudodivide by . Let be the resulting pseudoremainder; note that is not present in because the degree of is 1 in and . (In this case, the pseudoremainder is reduced with respect to after a single step of pseudodivision.) However, derivatives of other variables in may be present. Let be such a variable; there is some such that is a member of . Replace every instance of in with . (We justify this move in Theorem 96, the correctness proof for .) Doing so for each proper derivative present in produces a nondifferential polynomial . (Abusing terminology, we also refer to as a pseudoremainder.) Update the equations by omitting and including ; let . Note that the elements of having proper derivatives form the set because is nondifferential.
Now compute . Return the union over all returned by .
The next result is an important ingredient of the correctness proof for (Theorem 96). The lemma performs the critical job of uniting the algebraic part (calls to ) and differential part (differential saturation ideals) of .
Lemma 94.
Let be a polynomial vector field. Let be a finite set where is a subset of the differential polynomials and the are nondifferential polynomials. Let be a finite set of nondifferential polynomials. Let be the output of ; by correctness of (Theorem 93) we know that
Then we have
Proof.
The key observation is that the two operations of (namely, splitting and nondifferential pseudodivision) preserve the differential solutions of the augmented system that includes the ODEs with the equations. (See Remark 95 for a metamathematical issue concerning this approach.) Moreover, these operations only involve nondifferential polynomials. For instance, if we have
after a splitting step because the solutions of in are the union of the solutions of and in (see Theorem 30). The analogous differential decomposition holds if we include the ODEs and consider differential solutions in any differential extension field. By definition of and by Theorem 64 we have
(3) | ||||
If and is the pseudoremainder from pseudodividing by some other element of whose initial belongs to , then much like Equation 2 in the proof of Theorem 93 we have
(4) | ||||
Theorem 63 then gives
Thus re-inserting the ODEs after a single splitting or pseudodivision step preserves the differential decomposition. The full decomposition follows by induction on the number of splitting and pseudodivision operations during a run of . (In other words, since the ODEs are not involved in the operations of , it is equivalent to perform the entire run of on the nondifferential elements and then re-introduce the .) ∎
The system is a regular algebraic system and so by Lazard’s lemma (Lemma 80). However, might not be partially reduced so we need the radicals in the differential decomposition given by Lemma 94.
Remark 95.
Though the proof of Lemma 94 is straightforward, a subtle issue lurks nearby. We initially hoped to use more efficient radical ideal decomposition methods (e.g., that of Szántó [87, 88], which we discuss in Section 4.3, p. 109) in place of . This provides an algebraic decomposition of the desired form
(5) | ||||
but requires methods beyond simple splitting and (nondifferential) pseudodivision. We were unable to prove Lemma 94 using only the algebraic decomposition in Equation 5 without the fact that splitting and pseudodivision preserve differential solutions. This illustrates a major challenge of differential algebra: even when differential polynomials have a simple form, drawing conclusions about differential ideals from restrictions to algebraic ideals is nontrivial. It also explains why we stopped with regular differential systems in instead of continuing with the calculations needed to get regular differential chains.
Other published versions of RGA [37, pp. 162-3] [79, p. 587] [44, p. 111] use analogues of (i.e., they are based on recursive splitting and pseudoreduction), so they must have nonelementary [89, pp. 419-23] worst-case complexity like we find for (Theorems 107,109). The Wu-Ritt process, a close relative of , has nonelementary complexity [90, p. 121] similar to what we show in Theorem 106.
Theorem 96 (Termination and correctness of ).
Let be a polynomial vector field. Let be a finite set where is a subset of the differential polynomials and the are nondifferential polynomials. Let be a finite set of nondifferential polynomials. Lastly, assume that for each member of we either have (equivalently, ) or
for some , and .
Then the algorithm terminates on input and the output satisfies
where are finite sets such that each is a regular differential system. Moreover, the elements of containing proper derivatives form a subset of , and for each and we either have or
for some , and . (That is, the “technical assumption” on that we made of the input also holds of the output.)
Proof.
We first argue that terminates. On any branch of the computation, once the input has the property that is triangular and for all , this property continues to hold for each new call to along that branch. This is so because the property activates step 3 and step 3 calls (which terminates by Theorem 93) right before recursively calling . Hence every subsequent call to either terminates immediately or proceeds to step 3 and eliminates one of the ODEs . Note that at most of the original equations contain proper derivatives and no eliminated proper derivatives are ever re-introduced to the set of equations ( only involves nondifferential polynomials and the pseudoremainder introduced by step 3 of is nondifferential). It follows that, at most, a branch of calls and times (the first time in step 2 to make the equations triangular and put separants in the inequations; thereafter only step 3 applies) before terminating.
We justify the decomposition by showing that after each intermediate step we have a decomposition with the claimed form and properties, except possibly partial reducedness and being radical. However, these will both hold of the final output; see the final paragraph of the proof. During a run of , new differential systems are produced either by calling on the nondifferential equations (followed by re-inserting the ODEs) or performing differential pseudodivision. By Lemma 94, a call to , followed by re-inserting the ODEs, decomposes into a finite intersection of radicals of differential ideals of the right form (except possibly for partial reducedness)
where is the output of . In particular, the equations with proper derivatives are exactly . We must confirm that each preserves the technical assumption on the elements of . If belongs to , then remains in because the proper derivatives are unchanged. Otherwise, we have
(6) |
for some , and . These same are now in . By the algebraic decomposition
guaranteed by correctness of (Theorem 93), we know that each belongs to each . It follows that for all and hence .
Multiplying both sides of equation 6 by an appropriate element of , we thus obtain
(7) |
for some (recall that because on a given branch no inequations are ever removed), and . This shows that preserves the technical assumption on the elements of .
We now check the differential pseudodivision from step 3 of . Step 3 of produces a pseudoremainder by pseudodividing some by , where the separant and has leader . Recall that is the initial of and belongs to . Because the initial of is 1 and the degree of is 1 in both and , we have . (This is an instance of Proposition 73.) Step 3 of goes on to substitute nondifferential polynomials (using the relations ) for the remaining proper derivatives (which necessarily have lesser rank than ) in . This produces such that for some and . (As in the proof of Lemma 66, we replace in with . Then distribute to get an equation of the claimed form for .) Hence . Let ; note that contains all nondifferential polynomials of and of the others only omits . Step 3 of then proceeds using the system . If is no longer triangular or does not contain , the next call to will restore those properties.
We show that preserves the technical assumption on the elements of and that . Though , by definition . However, by assumption each in the equation
(8) |
either belongs to (and hence to since ) or can be written as
(9) |
for some , and . Since , we have and so . Multiplying both sides of 8 by appropriate elements of , using 9, and using the fact that belong to , we see that the technical assumption holds of with respect to . We now consider all such that . If , then still belongs to . If , by assumption
(10) |
for some , and . Here one of the might be , but we showed above that has the right form with respect to . Hence we can multiply both sides of 10 by appropriate elements of , substitute for , and conclude that has the form required by the technical assumption.
To prove that , it suffices to show that and . Both memberships follow from equations 8 and 9 (recall , and ). Thus differential pseudodivision replaces a system with an equivalent one whose corresponding differential saturation ideal is the same (in particular, the decomposition does not change).
Lastly, we explain why each output is a regular differential system. (Each is then radical by Lazard’s lemma (Lemma 80), which is why .) Both and are finite because only finitely many elements can enter at any step. Initials and separants are never identically zero, so 0 was not added to the inequations during the computation (i.e., ). Termination only occurs when is partially reduced, so that property is assured. The nondifferential polynomials in form a triangular set by the action of . The remaining differential polynomials form a subset of . The orderly ranking ensures that an element of has leader , so the entire set (i.e., the union of the nondifferential elements of and a subset of the original ODEs) remains triangular. The separant of any element of is and the separants of nondifferential elements of belong to by . Since has no proper derivatives, it is partially reduced with respect to . Thus is a regular differential system. This completes the proof.
∎
We need the following theorem to interpret the output of . The result is intuitively natural but slightly fussy to prove. It connects differential ideals to invariant ideals of polynomial vector fields: in particular, in the presence of an explicit system , the nondifferential polynomials in a differential ideal form the smallest invariant ideal containing the nondifferential polynomials in the original set that generated the differential ideal.
Theorem 97.
Let with . Then , where is the polynomial vector field .
Proof.
(): Immediate from Lemma 66.
() Let . Then for some , we have a representation
where . Proceeding as in the proof of Lemma 66 (i.e., replacing with ), for we may replace with plus a sum of multiples of the various . Regrouping and renaming the coefficient polynomials , etc., as necessary, we may thus assume has the form
Because and the various and belong to , we may assume that no derivative of order greater than appears on the right-hand side (i.e., ). Finally, we may assume that and that each has order no greater than . To see this, note that any in ( in the case of and for ) may be replaced with where has order less than . The general form of the representation of is preserved but the resulting replacements for contain fewer derivatives of order (and no higher ones were introduced), so the claim follows by induction.
We now have a form that is conducive to proving . We induct on the maximum order of the (in particular, for the inductive hypothesis we do not need the order of the Lie derivatives of to match ; we just require the assumptions that and that has order no greater than ). Subtracting from both sides, we obtain
(11) |
Denote the left-hand side of equation 4.1 by and choose an orderly ranking on (so all derivatives of order are greater than any derivative of lower order). Observe that is a regular set when we consider the entries as elements of : each entry in has a distinct leader and every initial is 1 (whence Definition 79 (2) trivially holds). Moreover, is reduced with respect to (by our assumptions, none of the leaders appears in ). Equation 4.1 witnesses that (since the multplicative set generated by the initials of is simply ). However, by Theorem 81, if and only if pseudoreduces to zero. Since is already reduced with respect to , this implies that both and the right-hand side are identically zero. Hence
which has lower maximum order of the but the same overall form and preserves our assumptions on . By induction we may eliminate all the , leaving
This proves that .
∎
Our work on culminates with Theorem 98, which interprets the output as a canonical algebraic invariant of a polynomial vector field . In particular, parts 1 and 3 connect differential ideals and the maximal invariant contained in . Part 3 also indicates the structure of this invariant in terms of smaller invariant sets. These are the major contributions of Theorem 98. Part 2 is implicit in [91] because there the authors obtain , from which membership in the radical can be tested. (See our remarks on [91] in Section 5.1. We discuss radical ideal membership in Section 4.4 and the two paragraphs preceding it.) However, our mechanism is different (regular differential systems instead of Gröbner bases) and shows that to test membership in it is unnecessary to first find generators of .
Theorem 98.
Let with and let be the polynomial vector field . If are the regular differential systems returned by given as input, then
-
1.
-
2.
Membership in is decidable.
-
3.
is the largest invariant with respect to that is contained in . Moreover, is a finite union of invariants of the form .
Proof.
Note that our Theorems 96,98 have more specialized hypotheses than the RGA theorem of Boulier et al. (Theorem 90) and do not guarantee that all elements of are differentially pseudoreduced to 0 by . However, Theorem 90 does not identify the dynamical meaning of the nondifferential polynomials in the output. Moreover, our Theorem 98 still allows for testing membership in . The greater specificity of Theorem 98 may also lead to better computational complexity; see Remark 114.
4.2 Algebraic Invariants from RGA Using Parameters
We now explain how we obtained the Lorenz system invariant in Section 3.3. Recall that the ODEs in question are . Our goal is to identify at least one nontrivial algebraic invariant of this system using the algorithms and theorems from Section 4.1. Consider the parameterized polynomial constraint , where are unknown constants. (We call a template.) To settle on this choice, we first restricted to polynomials of degree 2 and looked at possibilities containing but also having additional monomials. By trial-and-error we found that other parameters seem to be 0 for interesting invariants, so here we only present . The input system is , where we interpret the polynomials as equations; the inequation set is empty. The terms express that are constant and do not depend on time like . Computing RGA in Maple (using the command) on the system with an orderly ranking yields five regular differential chains:
-
1.
,
-
2.
,
-
3.
,
-
4.
,
-
5.
.
Even though we used a black-box commercial version of RGA for the calculation, Section 3.3 shows that the hypotheses of Theorems 86 and 88 hold of the output. (Specifically, chain 2; the others give invariants by inspection. See the next paragraph.) By default, Maple suppresses the inequations associated to each chain; however, the user can print them with the command. When implemented, should give equivalent output (but explicitly showing the inequations).
In the output we interpret as parameters and the invariants as algebraic sets in defined by polynomials in . Chain 1 defines the trivial invariant set because causes to vanish at every point in . Chain 2 requires and , which is the relationship of coefficients in . In other words, chain 2 defines the two-dimensional surface , the invariant we confirmed in Section 3.3. (Any nonzero scalar multiple of , or equivalently any nonzero solution of , would define the same surface.) This appears to be the “most generic chain” or “general solution” produced by RGA (p. 4.1; roughly, no additional equations hold beyond those implied by the input). Chain 3 requires to constantly be zero; the final term of vanishes because . Thus chain 3 defines the -axis, a one-dimensional curve. Chain 4 is even simpler, defining the (zero-dimensional) points and . Lastly, chain 5 defines the origin, .
Theorem 98 implies that the union of these sets contains every algebraic invariant of the form in for the vector field . (Moreover, invariants that are proper subsets of are contained in the union of the invariants given by the final four chains.) This is essentially a completeness result describing the sense in which finds all invariants having a given parameterized form. Moreover, the output invariant is structured, being comprised of smaller invariants of various dimensions. The example suggests that a generic chain corresponds to an invariant that has maximal dimension among the proper invariants.
4.3 Bounds on
We now find upper bounds (Theorems 107,109) on the complexity of . The maximal degree of polynomials in the output is a common measure of complexity for algorithms that operate on polynomials; in this paper we use the term degree complexity. Like many authors we consider degree complexity as a function of the maximal degree of the input polynomials and the number of variables . Algorithms involving polynomial ideals frequently have degree complexity that is singly or doubly exponential in and polynomial in (e.g., [92]. Thus it is reasonable to ignore matters like time complexity of addition and multiplication, the number of polynomials in the input, etc., that affect the output in only polynomial or singly exponential fashion.
We start by analyzing . This involves a recursively defined function that bounds the degree complexity. It is not surprising that the Fibonacci numbers appear in the definition of because a single step of pseudodivision produces a pseudoremainder whose degree is at most the sum of the degrees of the dividend and divisor (see step 2 of on p. • ‣ 2).
Notation 99.
We write for the -th Fibonacci number, defined recursively by for .
Proposition 100 (Binet’s formula [93, p. 92]).
where and .
It is somewhat tricky to pin down the degree complexity of for the same reason that termination (Theorem 93) was delicate to establish: depending on the inputs, must accommodate many different patterns of degrees throughout the computation. This includes the difficulty that multiple intermediate polynomials can have the same degree, so the degree of a given output is not necessarily tied to how many pseudodivision steps the calculation required. To handle this, we induct on the number of times that the minimal and maximal degrees in the target variable change.
Theorem 101 (Degree complexity of ).
Given finite sets of polynomials in variables such that each polynomial has total degree at most , algorithm returns polynomials that have total degree at most , where is defined recursively by , for .
Proof.
Throughout a run of , the inequations are updated with initials and separants of updated equations. Hence it suffices to bound the total degrees of the equations. Equations result from including initials or separants, which doesn’t increase the degree, or from a single pseudodivision step. Hence to find an upper bound we may assume that each new equation results from pseudodivision and adds the degrees of the two polynomials involved.
For we claim that is a bound on the total degree of any equation after eliminating target variables, which implies the theorem statement. (That is, is a bound after ensures that each of those target variables is the leader in at most one equation and the separant the corresponding polynomial belongs to the inequations.) Initially, total degrees are at most because we haven’t done anything yet (no variables eliminated). We prove that is a bound on the total degree after eliminating one target variable. This is sufficient because the process is identical going from to . In particular, if bounds total degrees after we have eliminated variables, the same argument (just replacing with ) shows that is a bound after eliminating variables. The latter expression is by definition.
We now show that is a bound on the total degree after eliminating one target variable. Let be the first target variable. We introduce the following terminology to assist with the proof. We refer to the polynomial chosen in step 1 of as the minimal pseudodivisor for the next pseudodivision step. (Recall that by definition has minimal degree in among the equations with leader and also has minimal total degree among equations with that minimal degree in .) We refer to the polynomial chosen in step 2 of as the maximal pseudodividend ( has maximal total degree among all equations of maximal degree in ). Suppose we have performed pseudodivision steps so far. The key values are the current minimal nonzero degree in of any equation (call it ) and the current maximal degree in of any equation (call it ). Then the current minimal pseudodivisor has degree in and the maximal pseudodividend has degree in . Let be the current maximal total degree of any equation (not necessarily having as its leader). For instance, (the initial bound before any pseudodivisions).
As long as elements of degree in are used to pseudodivide elements of degree in , we get pseudoremainders of total degree at most . (In other words, as long as and do not change.) This is because any polynomials not present initially are pseudoremainders that have degree in strictly less than . Hence they have not been used as pseudodividends yet. Also, any polynomial with degree in that becomes the minimal pseudodivisor in place of the original must have total degree no greater than that of the original minimal pseudodivisor (which was at most ). Otherwise it would not be minimal. Hence as soon as or for the first time, we still have . Note that . Note also that if exactly one of or occurs, then at least one of the subsequent minimal pseudodivisor and maximal pseudodividend still has total degree at most . This is because the subsequent minimal pseudodivisor has total degree at most if . If , then one of the original equations of maximal degree in is still present and the subsequent maximal pseudodividend has total degree at most . If both and occur, then both the new minimal pseudodivisor and maximal pseudodividend have degree at most . From now on we omit the numerical subscripts from and because the important quantity is how many times can increase, not how many pseudodivision steps are performed in all. The amount that can change in any individual step is bounded (e.g., earlier in this paragraph we saw that is at most after the first change in or ), so bounding the number of changes allows us to bound the final value of .
In particular, the crucial observation is that cannot increase more times than the combined number of decreases in and , plus 1. Each of can decrease at most times and keep a nonzero value. So after decreasing a combined number of times, at most . Then each subsequent pseudodivision eliminates , so the pseudoremainders are not used as long as is still the target variable. Hence at this stage there can be no more changes to and (they will both still be 1 when only one polynomial with leader remains) and can only increase one more time (due to pseudodivision involving elements having degree 1 in ), making for increases at most.
As alluded to immediately before the theorem statement, we induct on the combined number of decreases in and to show that never exceeds . In particular, we assert that after total decreases in and (with ), we satisfy two conditions: 1. we have and 2. at least one of the current minimal pseudodivisor and maximal pseudodividend has total degree at most . The paragraph before last establishes the base case . The reasoning is analogous for the inductive case: suppose after decreases we have and at least one of the current minimal pseudodivisor and maximal pseudodividend has total degree at most . As in the base case, we perform pseudodivision until we decrease either one or both of and for a total of or decreases. Since by assumption either the pseudodivisor or pseudodividend has total degree at most and neither has total degree greater than , the pseudoremainder has total degree at most . This proves the assertion about . If stays the same, then the subsequent minimal pseudodivisor has total degree at most . If stays the same, then the subsequent maximal pseudodivisor has total degree at most . If both and decrease, the condition on minimal pseudodivisors and maximal pseudodividends continues to hold: we now have decreases and both the new minimal pseudodivisor and new maximal pseudodividend have total degree at most . This proves the inductive case.
Hence when (necessarily forcing ), we have a current minimal pseudodivisor of total degree at most and current maximal pseudodividend of total degree at most . The final pseudodivision steps at most add and to yield the claimed value . This completes the proof. ∎
We now find an explicit function that bounds the recursively defined function . We start with a couple of straightforward inequalities.
Lemma 102.
For all , the -th Fibonacci number satisfies .
Proof.
(Binet’s formula, Proposition 100) | ||||
∎
Lemma 103.
Let with . Then .
Proof.
Note that if and only if . This holds by calculus: for and the derivatives satisfy when . ∎
Notation 104.
For natural numbers and we define
For and we define
where the right-hand side is an exponent tower of height (i.e., the tower consists of nested exponents with copies of 2 followed by a final exponent ).
Lemma 105.
All natural numbers satisfy .
Proof.
First let . Then
Now let . We find
The net effect is to add 1 to the exponent of the third 2 from the bottom of the tower. Repeat the cycle of “adding 1 to an exponent is less than multiplying the exponent by 2, which adds 1 to the next exponent up” as long as the exponent is 2. This propagates addition of 1 up the tower to the final exponent, whence
∎
Theorem 106 (Explicit bounds on degree complexity of ).
Let be the recursive function from Theorem 101 that bounds the degree complexity of the output of . The following inequality holds for natural numbers :
(See Notation 104 for the definition of .)
Proof.
We induct on , starting with . By definition, . We observe the following:
(Lemma 102) | ||||
This proves the base case. Now suppose the inequality holds for ; i.e., . The inductive case is similar but we now have an additional power of 2, allowing us to use Lemma 103. By definition, . We obtain the following:
This completes the proof. ∎
Theorem 107 (Degree complexity of ).
Let satisfy the hypotheses of Theorem 96, the termination and correctness result for . Suppose further that each element of has degree at most . Then every nondifferential polynomial returned by has degree at most , defined recursively by , for , and . (The values give bounds on intermediate polynomials’ degrees after has recursively called itself times out of the maximum possible . The outputs containing a proper derivative form a subset of , so their degree is bounded by .)
Proof.
As noted in the correctness proof (Theorem 96), alternates between calls to and to itself. The first possible call is to , producing nondifferential polynomials of degree at most . Then in step 3 an ODE may be pseudodivided by the derivative of a nondifferential polynomial ; this at most adds the degrees of the dividend and divisor, yielding . (Recall that , where is the pseudoremainder. The total degree of is at most . In fact, is a strict upper bound because the degree of the separant actually goes down, but we ignore this. The proper derivatives in all come from , which has total degree at most . After replacing the proper derivatives using the relations and obtaining , the total degree in the transformed version of is also at most .)
Another call to yields and restarts the process (beginning with pseudodivision in step 3 since the equations are now triangular and have their separants in the set of inequations). After iterations have eliminated ODEs, the degree is at most , whence pseudodivision and another call to yield .∎
Notation 108.
For natural numbers and we define
where the right-hand side is an exponent tower of height (i.e., the tower consists of nested exponents with copies of 2 followed by a final exponent ).
We also define
where the right-hand side is an exponent tower of height .
Theorem 109 (Explicit bounds on degree complexity of ).
Let be the recursive function from Theorem 107 that bounds the degree complexity of the output of . The following inequality holds for natural numbers :
(See Notation 108 for the definition of .)
Proof.
We use induction on for to bound (recall that . For we have by Theorem 106; this equals , which is a tower with copies of 2 followed by exponent .
Now let and suppose . We show that .
(Theorem 106) | ||||
∎
After expanding the top exponent of , the bound on becomes an exponent tower with copies of 2 followed by a final exponent .
While the nonelementary bounds in Theorem 109 are enormous from a practical perspective, there are important compensating factors. As noted in Remark 95, other published versions of RGA use the same mechanism of recursive splitting and pseudodivision that blows up the bounds on and hence on . In spite of RGA’s apparently high worst-case complexity, it has nonetheless been used profitably in applications (p. 4.1). Likewise, our experience (e.g., Sections 3.3, 4.2) shows that RGA has potential for studying algebraic invariants due to its systematic nature and universal results.
More theoretically, the problem of getting a differential radical decomposition of an explicit system with nondifferential inequations is close in spirit to the non-differential radical ideal membership problem. The effective Nullstellensatz [94] shows that radical ideal membership in polynomial rings over fields is strictly easier than general ideal membership in those rings. This opens the door for specialized approaches to radical ideal membership that outperform general Gröbner basis methods in terms of asymptotic complexity. (Möller and Mora [95] adapted a previous example [32] to prove doubly exponential lower bounds for Gröbner bases. In particular, they demonstrated ideals with generators of degree in variables such that all Gröbner bases using certain monomial orderings must have an element of degree doubly exponential in and polynomial in .) Szántó [87, 88] developed a characteristic set method of singly exponential complexity (in , polynomial in ) [96] for radical ideal membership. (See also [97, 98, 66, 99].) Szántó’s method reduces radical ideal membership testing to checking whether polynomials pseudoreduce to zero modulo certain characteristic sets. (In fact, the theory is closely related to that of RGA.)
It may be possible to generalize this to the differential case and prove Lemma 94 while replacing with a method of singly exponential complexity. The results of [100] and [91] imply that the invariant yielded by can be obtained by other techniques with doubly exponential complexity (Section 5.2). This further supports the idea that some modification of might dispense with the nonelementary bounds. Lastly, Theorem 112 and the resulting analysis in Section 4.4 suggest that the closely related problem of checking invariance has a very rare worst case and that singly exponential complexity is actually the norm.
4.4 Invariance Checking, Totally Real Varieties, and Complexity
In this section we connect totally real varieties to invariance checking. Theorem 112 below and the subsequent commentary are also relevant to complexity and bounds.
Proposition 110.
Let , let be a polynomial vector field, and let be an invariant set with respect to . Then is an invariant ideal. If in addition is totally real, then is an invariant ideal.
Proof.
Proposition 111.
Let and let be a polynomial vector field. If for some , then is not an invariant set with respect to . If is totally real and for some , then is not invariant.
Proof.
The next theorem generalizes to multiple polynomials Lie’s criterion for invariance of smooth algebraic manifolds [91, Fig. 1][101, Thm. 2.8]: let and assume that the real variety has no singular points. Lie’s criterion then implies that is invariant with respect to polynomial vector field if the first Lie derivative vanishes everywhere on .
Theorem 112.
Let and let be a polynomial vector field. If is real radical (i.e., ), then is an invariant set with respect to if and only if for all . Equivalently, if is totally real and is radical (i.e., ), then is an invariant set with respect to if and only if for all .
Theorem 112 raises some important questions. How common is it for to be totally real and to be radical, and can we check those properties efficiently? Regarding the prevalence of totally real varieties, we make the following observations about Theorem 44, which states that the complex variety corresponding to an irreducible real algebraic variety defined by a single polynomial is totally real if and only if assumes both positive and negative values as varies over .
Fix the degree . Blekherman ([102, Thm. 1.1]) has given an asymptotic, probabilistic result showing that for large , a randomly chosen polynomial almost certainly attains both positive and negative values. (More precisely, [102] puts a probability measure on the space of polynomials and shows that as , the ratio of the measure of the set of nonnegative polynomials and the measure of the set of all polynomials goes to 0. See the first inequality in Theorem 1.1 of [102], which gives an upper bound on this ratio for some constant .) Also, if the space of reducible polynomials of at most a given bounded degree has strictly lower dimension than the space of all polynomials with that bound on the degree (see Example 113 below). Hence for large a randomly chosen polynomial will almost certainly define a totally real complex variety by Theorem 44 because will almost certainly be irreducible and take on both positive and negative values. We suspect that a similar result holds for varieties that are defined by multiple, possibly reducible, polynomials (i.e., the generic situation should be that a complex variety defined over is totally real).
We are not aware of specific results on the prevalence of radical ideals among all ideals, but we expect that a “typical” ideal (with respect to some reasonable probability distribution) is radical. Consequently, we hypothesize that for almost all (in a precise sense that would require additional work to specify) complex varieties defined over the reals, checking invariance with respect to a polynomial vector field reduces by Theorem 112 to checking membership of the first Lie derivatives in the ideal generated by the defining polynomials of the variety.
This leads to an interesting situation with regard to the computational complexity of checking for invariance. Given , if is totally real and is radical, then we can check invariance of with complexity singly exponential in the number of variables. (This requires a singly exponential test for radical ideal membership like Szántó’s [87, 88].) As explained in the preceding paragraph, we have reason to believe that this is possible for almost all choices of .
However, we are left with the “measure zero” cases where is not totally real or is not radical. Determining whether a given candidate has these properties appears, in the worst case, to be doubly exponential in the number of variables. For instance, decomposing a variety into irreducible components (which seems unavoidable for determining if the variety is totally real) has doubly exponential lower bounds in [103, Thm. 1]. (In particular, Chistov shows that for sufficiently large there are polynomials in variables of degree less than defining a reducible variety with a component whose corresponding prime ideal has no set of generators with all elements having degree less than a certain doubly exponential bound.) Likewise, checking if a variety is radical is closely related to primary ideal decomposition [30, Sect. 4.8], which has doubly exponential complexity (following again from [103, Thm. 1]).
In summary, then, we suspect that with singly exponential complexity we can almost always correctly guess whether a given is invariant, but proving that we have the correct answer may require doubly exponential complexity. (This is similar, but not identical, to the hypothesis–also possibly true–that proving invariance of an algebraic set has singly exponential average-case complexity but doubly exponential worst-case complexity.) However, we cannot discard the possibility that the worst case is also singly exponential; settling the matter appears difficult.
Example 113.
As part of the preceding analysis we claimed that the set of reducible polynomials of at most a given bounded degree has smaller dimension than the set of all such polynomials, and hence there is a precise sense in which almost all polynomials are irreducible. If , then over every polynomial of degree is reducible, so we only care about the case . Rather than prove a formal theorem with full details, we illustrate with a simple case that mirrors the general result (the example is based on an argument from [104]).
Let be an algebraically closed field and consider , where is the set of all nonzero polynomials over in two variables having degree at most 2. An arbitrary element of has the form , with not all being zero. We can thus identify with the 6-tuple . Irreducibility is not affected by nonzero scalar multiples, so it is natural to view as a point in , the five-dimensional projective space over [30, Sect. 8.2]. By definition, is the quotient of by the equivalence relation such that if and only if there exists such that . (Going forward we omit the in .) For instance, both and correspond to .
Points of that correspond to reducible polynomials in are contained in the image of the map defined by
This point represents the factorization , where and . The image of has dimension at most . These values are justified by Examples 1.30, 1.33 (p. 67) and Theorem 1.25 (p. 75) of [42]. Hence every polynomial corresponding to a point in the “large” set (the complement of a lower-dimensional set) is irreducible.
Remark 114.
We note that the prevalence of totally real and/or irreducible varieties is also relevant to the complexity of generating equations for algebraic invariants. In particular, Theorem 88 implies that these and other conditions make it easier to interpret the output of . For instance, suppose returns a collection of regular differential systems and each satisfies the hypotheses of Theorem 88 (2). Then we may ignore the (doubly exponential) complication of finding generators of the saturation ideals because the invariant is simply and the inequations are superfluous. (Compare to Theorem 98 (3).)
5 Related work
5.1 Invariant Generation and Checking
Several recent algebraic algorithms for generating invariants are close in spirit to our use of . In particular, all are template methods that find invariants having the form of a parameterized polynomial input (see the example in Section 4.2). However, ’s basic elimination method (namely, differential pseudodivision) is unique in the class. This leads to a major advantage of our approach: avoids both Gröbner bases and real quantifier elimination. In particular, we obtain the largest possible invariant that is consistent with the input system without having to test ideal membership or the vanishing of Lie derivatives.
Liu, Zhan, and Zhao [105] proved that invariance of semialgebraic sets (i.e., defined by polynomial equations and inequalities over the real numbers) with respect to a polynomial vector field is decidable and gave an algorithm–referred to here as the LZZ algorithm–for the problem of checking invariance. By enforcing an invariance criterion on a template, the algorithm gives a way of generating invariants of a given form. The chief downside of LZZ is that it uses two expensive tools, Gröbner bases and real quantifier elimination. (The generation version only uses real quantifier elimination, but the template might need to be large to find a nontrivial invariant.)
Ghorbal and Platzer [56] contributed a procedure that here we refer to as DRI (“differential radical invariant”). Given a homogeneous template polynomial with parameter coefficients, assume an order and write a symbolic matrix representing membership of the -th Lie derivative of in the ideal generated by and the preceding Lie derivatives. DRI seeks an invariant of maximal dimension with this form by minimizing the rank of the symbolic matrix.
One virtue of DRI is completeness: any algebraic invariant that exists can be produced by choosing and the degree of to be sufficiently large [56, Cor. 1]. Another is its reliance on linear algebra (similar to [106, 107], which DRI partly generalizes), an area with many efficient computational tools. DRI does not use Gröbner bases, but nonlinear real arithmetic/quantifier elimination is generally required to find parameter values that minimize the rank. DRI has considerable power, as demonstrated by a number of nontrivial case studies from the aerospace domain [56, Sect. 6]. Nonetheless, it is significantly slower than the following two algorithms [33, Sect. 6].
Kong et al. [108] use a similar idea for invariant generation, though they restrict themselves to the case of invariants that are Darboux polynomials. This means that the Lie derivative for some polynomial (i.e., divides its own Lie derivative with respect to the vector field in question). This prevents completeness, but also contributes to better empirical performance when it applies [108, Table 1]. However, the implementation still uses Gröbner bases and nonlinear real arithmetic [108, Remark 3].
Recent work of Boreale [33] has strong ties to both DRI [56] and [108]. The POST algorithm [33, p. 9] builds on the same theory as [56] but focuses on invariants containing a given initial set. Moreover, POST uses Gröbner bases and an iterative process involving descending chains of vector spaces and ascending chains of ideals. In some sense, Boreale’s work is intermediate between that of [56] and [108]: POST offers completeness guarantees missing from [108] but in general is slower than [108] and faster than DRI. Boreale also uses POST and a generic point of the initial set to study the semialgebraic case.
In [91] Ghorbal, Sogokon, and Platzer give an invariant checking algorithm that is related to LZZ but is restricted to algebraic sets and improves efficiency in that case. Like LZZ, the method in [91] uses Gröbner bases and nonlinear real arithmetic. The algorithm computes successive Lie derivatives of the input and checks ideal membership (where Gröbner bases come in) until the first-order Lie derivatives of all preceding elements belong to the ideal generated by those elements. This will eventually happen by Hilbert’s basis theorem (Theorem 2), so and the subsequent Lie derivatives computed up to that stage generate (see our comments immediately preceding Theorem 98). Real arithmetic checks that each Lie derivative computed vanishes on all of . This holds if and only if is invariant with respect to (Corollary 51).
The recent ESE invariant checking algorithm of [109] refines and extends both LZZ and [91]. ESE again uses Gröbner bases and real arithmetic, but it structures quantifier elimination subproblems more efficiently than does LZZ and, unlike [91], applies to semialgebraic sets. Based on the topological notion of exit sets [110] (which concern vector field trajectories exiting the boundary of a set in ), the ESE algorithm can verify or disprove invariance of a candidate semialgebraic set. ESE is often much faster than an enhanced version of LZZ that is also described in [109].
We lastly mention a new approach of Wang et al. [111] for generating invariant barrier certificates (a concept closely tied to invariants; see [111, Thm. 4]). The method employs Gröbner bases to find a sufficient number of Lie derivatives and uses them to satisfy conditions [111, Def. 4] that guarantee a template is an invariant barrier certificate. However, [111] then translates the criterion into a numerical optimization problem. Experiments show promising performance, but the method’s reliance on local extrema occasionally results in erroneous output [111, Table 1]. (Wang et al. provide the option of searching for global extrema using a branch-and-bound technique, but this can increase the complexity substantially.) This contrasts with the trade-offs made by fully symbolic methods like for generating or checking invariants.
5.2 Complexity
As indicated in Section 5.1, the order of Lie derivatives needed to check invariants is an important factor in the complexity of the algorithms from [105, 91, 33, 111, 109]. The best upper bound known is doubly exponential in the number of variables [100, Thm. 4], but both [109] and [100] suggest this is overly conservative. This intuition is consistent with our discussion in Section 4.4, where Theorem 112 collapses invariance checking to a single round of radical ideal membership testing (assuming a radical ideal that defines a totally real variety; we argued that this should almost always hold). Whether or not the worst-case complexity is greater than singly exponential, it is necessarily considerable since generating algebraic invariant sets is NP-hard [56, p. 289].)
In another direction, Section 4.3 demonstrates that the complexity of RGA is nontrivial to analyze. While degree bounds for RGA have not previously appeared, upper bounds on the order of intermediate and output differential polynomials have been worked out [79, 112]. The order bounds on RGA from [79] amount to for ODEs of the form in variables. However, this result is excessive for because we differentiate at most times. Moreover, in we only handle differential polynomials of order at most 1 since we use the explicit form to substitute nondifferential polynomials for proper derivatives.
Degree bounds for differential characteristic sets appear in [113, Thm. 5.48], but these are again excessive for our situation.
6 Conclusion
In this paper we have adapted the Rosenfeld-Gröbner algorithm to generate algebraic invariants of polynomial differential systems. Finding invariants is a crucial and computationally difficult task that our method tackles with a new and systematic approach using tools from algebraic geometry, differential algebra, and the theory of regular systems. Our algorithm provides an alternative to the traditional Gröbner bases and real quantifier elimination while giving a novel representation of the largest algebraic invariant contained in an input algebraic set. We have also highlighted the prevalence and importance of totally real varieties for reducing the computational complexity of both generating and checking invariants of differential equations.
The unique aspects of provide multiple directions for future work. One important goal is modifying the subroutine to preserve differential solutions while obtaining a regular algebraic decomposition with singly exponential complexity. Beyond theoretical complexity improvements, our framework would greatly benefit from effective heuristics (e.g., for detecting totally real varieties) and optimizations like pruning superfluous branches in . Some rankings are much easier to compute than others [114], so translating from one differential ranking to another could be an important strategy. Partial runs of are another interesting possibility. In our experience, the final iterations of the algorithm tend to be drastically more expensive than the early ones while yielding less additional information (e.g., giving relatively trivial, low-dimensional chains when we already know the generic chain). The example in 4.2 demonstrates this; we are most interested in the generic Chain 2 and would like the option of only computing it.
The treatment of parameters is also a critical matter. Our example used auxiliary variables with constant derivative, but this is suboptimal. Fakouri et al. [77] describe a modified version of RGA designed to more efficiently identify all possible regular chains corresponding to different values of parameters. Adapting this algorithm to our problems could be important for a more scalable .
While purely algebraic invariants are important for continuous dynamical systems and enjoy strong decidability properties [4], many polynomial vector fields have no such invariants that are nontrivial (not isolated points, coordinate axes, etc.). The Van der Pol oscillator is an example that nonetheless has interesting semialgebraic invariants [115, 116]. In the literature there are several encodings of inequalities as equations using auxiliary variables (e.g., [117, 118]). Our initial investigation suggests that this idea, together with RGA, can yield insights into semialgebraic invariants of nonlinear systems with non-obvious dynamics.
In light of these prospects, we view our results in this paper as laying the groundwork for an enhanced method that produces easily verified invariants, combines optimal theoretical complexity with strong practical performance, and is broadly applicable. Progress on these fronts could lead to inclusion in practical tools for CPS verification such as the theorem prover KeYmaera X [119] with its automatic invariant-generating utility, Pegasus [71].
Acknowledgements
We are grateful to François Boulier, Khalil Ghorbal, and Yong Kiam Tan for helpful exchanges. This work was supported by the AFOSR under grants FA9550-18-1-0120 and FA9550-16-1-0288.
References
- [1] A. Platzer, Logical Analysis of Hybrid Systems: Proving Theorems for Complex Dynamics, Springer, Heidelberg, 2010. doi:10.1007/978-3-642-14509-4.
-
[2]
A. Platzer, Logical
Foundations of Cyber-Physical Systems, Springer, 2018.
doi:10.1007/978-3-319-63588-0.
URL https://doi.org/10.1007/978-3-319-63588-0 - [3] A. Goriely, Integrability and nonintegrability of dynamical systems, Vol. 19, World Scientific, 2001.
- [4] A. Platzer, Y. K. Tan, Differential equation invariance axiomatization, J. ACM 67 (1) (2020) 6:1–6:66. doi:10.1145/3380825.
- [5] Y. K. Tan, A. Platzer, Deductive stability proofs for ordinary differential equations, in: TACAS (2), Vol. 12652 of Lecture Notes in Computer Science, Springer, 2021, pp. 181–199.
- [6] Y. K. Tan, A. Platzer, An axiomatic approach to existence and liveness for differential equations, Formal Aspects of Computing (2021) 1–58.
- [7] E. Rodríguez-Carbonell, A. Tiwari, Generating polynomial invariants for hybrid systems, in: M. Morari, L. Thiele (Eds.), HSCC, Vol. 3414 of LNCS, Springer, 2005, pp. 590–605.
- [8] S. Sankaranarayanan, H. B. Sipma, Z. Manna, Constructing invariants for hybrid systems, Form. Methods Syst. Des. 32 (1) (2008) 25–55. doi:10.1007/s10703-007-0046-1.
- [9] A. Tiwari, Generating box invariants, in: M. Egerstedt, B. Mishra (Eds.), HSCC, Vol. 4981 of LNCS, Springer, 2008, pp. 658–661.
- [10] A. Platzer, E. M. Clarke, Computing differential invariants of hybrid systems as fixedpoints, Form. Methods Syst. Des. 35 (1) (2009) 98–120. doi:10.1007/s10703-009-0079-8.
- [11] S. Sankaranarayanan, Automatic invariant generation for hybrid systems using ideal fixed points, in: K. H. Johansson, W. Yi (Eds.), HSCC, ACM, 2010, pp. 221–230. doi:10.1145/1755952.1755984.
- [12] S. Sankaranarayanan, A. Tiwari, Relational abstractions for continuous and hybrid systems, in: G. Gopalakrishnan, S. Qadeer (Eds.), CAV, Vol. 6806 of LNCS, Springer, Berlin, 2011, pp. 686–702. doi:10.1007/978-3-642-22110-1_56.
- [13] A. Taly, S. Gulwani, A. Tiwari, Synthesizing switching logic using constraint solving, STTT 13 (6) (2011) 519–535. doi:10.1007/s10009-010-0172-8.
- [14] A. Platzer, A differential operator approach to equational differential invariants, in: L. Beringer, A. Felty (Eds.), ITP, Vol. 7406 of LNCS, Springer, 2012, pp. 28–48.
- [15] B. Wu, X. Zou, Computing invariants for hybrid systems, in: Electronic System-Integration Technology Conference (ESTC), 2012 4th, 2012, pp. 203–206. doi:10.1109/ESTC.2012.6485572.
- [16] K. Ghorbal, A. Platzer, Characterizing algebraic invariants by differential radical invariants, in: E. Ábrahám, K. Havelund (Eds.), TACAS, Vol. 8413 of LNCS, Springer, 2014, pp. 279–294. doi:10.1007/978-3-642-54862-8_19.
-
[17]
A. Sogokon, K. Ghorbal, P. B. Jackson, A. Platzer,
A method for invariant
generation for polynomial continuous systems, in: B. Jobstmann, K. R. M.
Leino (Eds.), Verification, Model Checking, and Abstract Interpretation -
17th International Conference, VMCAI 2016, St. Petersburg, FL, USA, January
17-19, 2016. Proceedings, Vol. 9583 of Lecture Notes in Computer Science,
Springer, 2016, pp. 268–288.
doi:10.1007/978-3-662-49122-5\_13.
URL https://doi.org/10.1007/978-3-662-49122-5_13 - [18] P. Roux, Y. Voronin, S. Sankaranarayanan, Validating numerical semidefinite programming solvers for polynomial invariants, in: X. Rival (Ed.), Static Analysis - 23rd International Symposium, SAS 2016, Edinburgh, UK, September 8-10, 2016, Proceedings, Vol. 9837 of LNCS, Springer, 2016, pp. 424–446. doi:10.1007/978-3-662-53413-7_21.
- [19] S. Sankaranarayanan, Change-of-bases abstractions for non-linear hybrid systems, Nonlinear Analysis: Hybrid Systems 19 (2016) 107 – 133.
- [20] H. Kong, S. Bogomolov, C. Schilling, Y. Jiang, T. A. Henzinger, Safety verification of nonlinear hybrid systems based on invariant clusters, in: G. Frehse, S. Mitra (Eds.), Proceedings of the 20th International Conference on Hybrid Systems: Computation and Control, HSCC 2017, Pittsburgh, PA, USA, April 18-20, 2017, ACM, 2017, pp. 163–172.
- [21] M. Boreale, Complete algorithms for algebraic strongest postconditions and weakest preconditions in polynomial ode’s, in: A. M. Tjoa, L. Bellatreche, S. Biffl, J. van Leeuwen, J. Wiedermann (Eds.), SOFSEM 2018: Theory and Practice of Computer Science - 44th International Conference on Current Trends in Theory and Practice of Computer Science, Krems, Austria, January 29 - February 2, 2018, Proceedings, Vol. 10706 of LNCS, Springer, 2018, pp. 442–455.
-
[22]
K. Ghorbal, A. Sogokon,
Characterizing positively
invariant sets: Inductive and topological methods, J. Symb. Comput. 113
(2022) 1–28.
doi:10.1016/j.jsc.2022.01.004.
URL https://doi.org/10.1016/j.jsc.2022.01.004 - [23] A. Sogokon, S. Mitsch, Y. K. Tan, K. Cordwell, A. Platzer, Pegasus: Sound continuous invariant generation, Form. Methods Syst. Des. 58 (1) (2022) 5–41, special issue for selected papers from FM’19. doi:10.1007/s10703-020-00355-z.
-
[24]
D. Wang, Elimination Methods,
Texts & Monographs in Symbolic Computation, Springer, 2001.
doi:10.1007/978-3-7091-6202-6.
URL https://doi.org/10.1007/978-3-7091-6202-6 - [25] J. Bochnak, M. Coste, M.-F. Roy, Real algebraic geometry, Vol. 36 of Ergebnisse der Mathematikund ihrer Grenzgebiete, Springer, 1998.
- [26] E. R. Kolchin, Differential algebra and algebraic groups, Pure and Applied Mathematics, Vol. 54, Academic Press, New York-London, 1973.
-
[27]
W. Li, C. Yuan, Elimination
theory in differential and difference algebra, J. Syst. Sci. Complex. 32 (1)
(2019) 287–316.
doi:10.1007/s11424-019-8367-x.
URL https://doi.org/10.1007/s11424-019-8367-x - [28] L. M. Pardo, How lower and upper complexity bounds meet in elimination theory, in: AAECC, Vol. 948 of Lecture Notes in Computer Science, Springer, 1995, pp. 33–69.
- [29] B. Mishra, Algorithmic algebra, Texts and Monographs in Computer Science, Springer-Verlag, New York, 1993.
- [30] D. Cox, J. Little, D. O’Shea, Ideals, varieties, and algorithms, 4th Edition, Undergraduate Texts in Mathematics, Springer, 2015.
- [31] G. Gallo, B. Mishra, Efficient algorithms and bounds for Wu-Ritt characteristic sets, in: Effective methods in algebraic geometry, Springer, 1991, pp. 119–142.
- [32] E. W. Mayr, A. R. Meyer, The complexity of the word problems for commutative semigroups and polynomial ideals, Advances in Mathematics 46 (3) (1982) 305–329.
-
[33]
M. Boreale, Complete
algorithms for algebraic strongest postconditions and weakest preconditions
in polynomial ODEs, Sci. Comput. Program. 193 (2020) 102441.
doi:10.1016/j.scico.2020.102441.
URL https://doi.org/10.1016/j.scico.2020.102441 - [34] E. Becker, R. Neuhaus, Computation of real radicals of polynomial ideals, in: Computational algebraic geometry, Springer, 1993, pp. 1–20.
- [35] L. Menini, C. Possieri, A. Tornambè, Algebraic Geometry for Robotics and Control Theory, World Scientific, 2021.
- [36] D. Richardson, Some undecidable problems involving elementary functions of a real variable, The Journal of Symbolic Logic 33 (4) (1969) 514–520.
- [37] F. Boulier, D. Lazard, F. Ollivier, M. Petitot, Representation for the radical of a finitely generated differential ideal, in: ISSAC, ACM, 1995, pp. 158–166.
- [38] S. Lang, Algebra, 3rd Edition, Vol. 211 of Graduate Texts in Mathematics, Springer-Verlag, New York, 2002.
- [39] W. Rudin, Real and Complex Analysis, 3rd Ed., McGraw-Hill, Inc., USA, 1987.
- [40] S. Mac Lane, Categories for the working mathematician, 2nd Edition, Vol. 5 of Graduate Texts in Mathematics, Springer Science & Business Media, 2013.
- [41] O. Ore, Galois connexions, Transactions of the American Mathematical Society 55 (3) (1944) 493–513.
- [42] I. R. Shafarevich, Basic algebraic geometry 1, 3rd Edition, Springer-Verlag, Berlin Heidelberg, 2013, translated from the 2007 Russian edition by Miles Reid. doi:https://doi.org/10.1007/978-3-642-37956-7.
- [43] M. Michałek, B. Sturmfels, Invitation to nonlinear algebra, Vol. 211 of Graduate Studies in Mathematics, American Mathematical Soc., 2021.
-
[44]
F. Boulier, F. Ollivier, D. Lazard, M. Petitot,
Computing representations
for radicals of finitely generated differential ideals, Appl. Algebra Engrg.
Comm. Comput. 20 (1) (2009) 73–121.
doi:10.1007/s00200-009-0091-7.
URL http://dx.doi.org/10.1007/s00200-009-0091-7 - [45] F. Sottile, Real algebraic geometry for geometric constraints, in: M. Sitharam, A. S. John, J. Sidman (Eds.), Handbook of geometric constraint systems principles, CRC Press, 2019, pp. 273–286.
- [46] T. Sander, Aspects of algebraic geometry over non algebraically closed fields, Technical report, International Comp. Sci. Institute (1996).
- [47] M. Marshall, Positive polynomials and sums of squares, no. 146, American Mathematical Soc., 2008.
- [48] K. Harris, J. D. Hauenstein, Á. Szántó, Smooth points on semi-algebraic sets, ACM Commun. Comput. Algebra 54 (3) (2020) 105–108.
- [49] R. Hartshorne, Algebraic Geometry, Graduate Texts in Mathemematics 52. Springer-Verlag, 1977.
- [50] G. Blekherman, R. Sinn, G. G. Smith, M. Velasco, Sums of squares: A real projective story, Notices of the American Mathematical Society 68 (5) (2021).
- [51] P. Conti, C. Traverso, A case of automatic theorem proving in euclidean geometry: the Maclane 8 theorem, in: AAECC, Vol. 948 of Lecture Notes in Computer Science, Springer, 1995, pp. 183–193.
- [52] D. Wang, Irreducible decomposition of algebraic varieties via characteristic sets and Gröbner bases, Comput. Aided Geom. Des. 9 (6) (1992) 471–484.
- [53] J. H. Hubbard, B. B. Hubbard, Vector calculus, linear algebra, and differential forms: a unified approach, Matrix Editions, 2015.
- [54] W. Walter, Ordinary Differential Equations, 1st Edition, Graduate Texts in Mathematics, 182, Springer New York, 1998.
- [55] C. Christopher, J. Llibre, C. Pantazi, S. Walcher, Inverse problems for multiple invariant curves, Proceedings of the Royal Society of Edinburgh Section A: Mathematics 137 (6) (2007) 1197–1226.
- [56] K. Ghorbal, A. Platzer, Characterizing algebraic invariants by differential radical invariants, in: TACAS, Vol. 8413 of Lecture Notes in Computer Science, Springer, 2014, pp. 279–294.
- [57] J. F. Ritt, Differential Algebra, Dover Publications, New York, 1950.
- [58] E. R. Kolchin, Differential Algebra and Algebraic Groups, Academic Press, New York, 1976.
- [59] D. Marker, Model theory of differential fields, Model Theory of Fields, eds. D. Marker, M. Messmer, & A. Pillay 5 (2005) 41–109.
- [60] A. Robinson, On the concept of a differentially closed field, Office of Scientific Research, US Air Force, 1959.
- [61] D. Marker, Introduction to the model theory of fields, Model Theory of Fields, eds. D. Marker, M. Messmer, & A. Pillay 5 (2005) 1–37.
- [62] H. B. Enderton, Elements of set theory, Academic Press [Harcourt Brace Jovanovich Publishers], New York, 1977.
- [63] W. Y. Sit, The Ritt–Kolchin theory for differential polynomials, in: W. Y. Sit (Ed.), Differential algebra and related topics, World Scientific, 2002, pp. 1–70.
- [64] F. Boulier, Differential elimination and biological modelling, in: Workshop D2. 2 of the Special Semester on Gröbner Bases and Related Methods, Vol. 2, de Gruyter, 2006, pp. 111–139.
- [65] F. Boulier, F. Lemaire, M. M. Maza, A. Poteaux, A short contribution to the theory of regular chains, Math. Comput. Sci. 15 (2) (2021) 177–188.
- [66] P. Aubry, D. Lazard, M. M. Maza, On the theories of triangular sets, J. Symb. Comput. 28 (1-2) (1999) 105–124.
- [67] E. N. Lorenz, Deterministic nonperiodic flow, Journal of atmospheric sciences 20 (2) (1963) 130–141.
- [68] C. Sparrow, The Lorenz equations: bifurcations, chaos, and strange attractors, Vol. 41, Springer Science & Business Media, 2012.
- [69] J. Llibre, X. Zhang, Invariant algebraic surfaces of the Lorenz system, Journal of Mathematical Physics 43 (3) (2002) 1622–1645.
- [70] S. P. Swinnerton-Dyer, The invariant algebraic surfaces of the Lorenz system, in: Mathematical Proceedings of the Cambridge Philosophical Society, Vol. 132, Cambridge University Press, 2002, pp. 385–393.
- [71] A. Sogokon, S. Mitsch, Y. K. Tan, K. Cordwell, A. Platzer, Pegasus: sound continuous invariant generation, Formal Methods Syst. Des. 58 (1-2) (2021) 5–41.
- [72] F. Boulier, F. Lemaire, A normal form algorithm for regular differential chains, Math. Comput. Sci. 4 (2-3) (2010) 185–201.
- [73] E. Hubert, Essential components of an algebraic differential equation, J. Symb. Comput. 28 (4-5) (1999) 657–680.
- [74] U. Umirbaev, Algorithmic problems for differential polynomial algebras, Journal of Algebra 455 (2016) 77–92.
- [75] M. (Waterloo Maple Inc.), Overview of the DifferentialAlgebra package (online documentation), https://www.maplesoft.com/support/help/maple/view.aspx?path=DifferentialAlgebra, Accessed: 2021-09-01.
- [76] F. Boulier, BLAD software package, https://www.lifl.fr/%7Eboulier/pmwiki/pmwiki.php?n=Main.BLAD, Accessed: 2021-09-01.
- [77] S. Fakouri, S. Rahmany, A. Basiri, A new algorithm for computing regular representations for radicals of parametric differential ideals, Cogent Mathematics & Statistics 5 (1) (2018) 1507131.
- [78] A. Hashemi, Z. Touraji, An improvement of Rosenfeld-Gröbner algorithm, in: ICMS, Vol. 8592 of Lecture Notes in Computer Science, Springer, 2014, pp. 466–471.
- [79] O. Golubitsky, M. V. Kondratieva, M. M. Maza, A. Ovchinnikov, A bound for the Rosenfeld-Gröbner algorithm, J. Symb. Comput. 43 (8) (2008) 582–610.
- [80] D. Bouziane, A. Kandri-Rody, H. Maarouf, Unmixed-dimensional decomposition of a finitely generated perfect differential ideal, J. Symb. Comput. 31 (6) (2001) 631–649.
- [81] E. Hubert, Factorization-free decomposition algorithms in differential algebra, J. Symb. Comput. 29 (4-5) (2000) 641–662.
- [82] R. Ushirobira, D. V. Efimov, P. Blirnan, Estimating the infection rate of a SIR epidemic model via differential elimination, in: ECC, IEEE, 2019, pp. 1170–1175.
- [83] N. Verdière, S. Zhu, L. Denis-Vidal, A distribution input-output polynomial approach for estimating parameters in nonlinear models: application to a chikungunya model, J. Comput. Appl. Math. 331 (2018) 104–118.
- [84] F. Boulier, F. Lemaire, Differential algebra and system modeling in cellular biology, Algebraic Biology (2008) 22–39.
- [85] H. A. Harrington, K. L. Ho, N. Meshkat, A parameter-free model comparison test using differential algebra, Complex. 2019 (2019) 6041981:1–6041981:15.
- [86] J.-Y. Hong, K. Hara, J.-W. Kim, E. F. Sato, E. B. Shim, K.-H. Cho, Minimal systems analysis of mitochondria-dependent apoptosis induced by cisplatin, The Korean Journal of Physiology & Pharmacology 20 (4) (2016) 367–378.
- [87] Á. Szántó, Complexity of the Wu-Ritt decomposition, in: PASCO, ACM, 1997, pp. 139–149.
- [88] A. Szántó, Computation with polynomial systems, PhD dissertation, Cornell University, 1999.
- [89] A. V. Aho, J. E. Hopcroft, J. D. Ullman, The Design and Analysis of Computer Algorithms, Addison-Wesley, 1974.
- [90] G. Gallo, B. Mishra, Wu-Ritt characteristic sets and their complexity, in: Discrete and Computational Geometry, Vol. 6 of DIMACS Series in Discrete Mathematics and Theoretical Computer Science, DIMACS/AMS, 1990, pp. 111–136.
- [91] K. Ghorbal, A. Sogokon, A. Platzer, Invariance of conjunctions of polynomial equalities for algebraic differential equations, in: SAS, Vol. 8723 of Lecture Notes in Computer Science, Springer, 2014, pp. 151–167.
- [92] A. Dickenstein, N. Fitchas, M. Giusti, C. Sessa, The membership problem for unmixed polynomial ideals is solvable in single exponential time, Discret. Appl. Math. 33 (1-3) (1991) 73–94.
- [93] W. D. Wallis, J. C. George, Introduction to combinatorics, Chapman and Hall/CRC, 2010.
- [94] W. D. Brownawell, Bounds for the degrees in the Nullstellensatz, Annals of Mathematics 126 (3) (1987) 577–591.
- [95] H. M. Möller, F. Mora, Upper and lower bounds for the degree of Groebner bases, in: EUROSAM, Vol. 174 of Lecture Notes in Computer Science, Springer, 1984, pp. 172–183.
- [96] E. Amzallag, M. Sun, G. Pogudin, T. N. Vo, Complexity of triangular representations of algebraic sets, Journal of Algebra 523 (2019) 342–364.
- [97] P. Bürgisser, P. Scheiblechner, On the complexity of counting components of algebraic varieties, J. Symb. Comput. 44 (9) (2009) 1114–1136.
- [98] F. Boulier, F. Lemaire, M. M. Maza, Well known theorems on triangular systems and the D5 principle, in: Transgressive Computing 2006, J.-G. Dumas, Université Joseph Fourier, Grenoble, France, 2006, pp. 79–91.
- [99] M. Kalkbrener, Prime decompositions of radicals in polynomial rings, J. Symb. Comput. 18 (4) (1994) 365–372.
- [100] D. Novikov, S. Yakovenko, Trajectories of polynomial vector fields and ascending chains of polynomial ideals, in: Annales de l’institut Fourier, Vol. 49, 1999, pp. 563–609.
- [101] P. J. Olver, Applications of Lie groups to differential equations, 2nd Edition, Vol. 107 of Graduate Texts in Mathematics, Springer, 2000.
- [102] G. Blekherman, There are significantly more nonegative polynomials than sums of squares, Israel Journal of Mathematics 153 (1) (2006) 355–380.
- [103] A. Chistov, Double-exponential lower bound for the degree of any system of generators of a polynomial prime ideal, St. Petersburg Mathematical Journal 20 (6) (2009) 983–1001.
- [104] Q. Yuan, Are most polynomials reducible or irreducible?, Mathematics Stack Exchange, https://math.stackexchange.com/q/459881 (version: 2013-08-05, accessed 2022-05-30).
- [105] J. Liu, N. Zhan, H. Zhao, Computing semi-algebraic invariants for polynomial dynamical systems, in: EMSOFT, ACM, 2011, pp. 97–106.
- [106] R. Rebiha, A. V. Moura, N. Matringe, Generating invariants for non-linear hybrid systems, Theor. Comput. Sci. 594 (2015) 180–200.
- [107] N. Matringe, A. V. Moura, R. Rebiha, Generating invariants for non-linear hybrid systems by linear algebraic methods, in: SAS, Vol. 6337 of Lecture Notes in Computer Science, Springer, 2010, pp. 373–389.
- [108] H. Kong, S. Bogomolov, C. Schilling, Y. Jiang, T. A. Henzinger, Safety verification of nonlinear hybrid systems based on invariant clusters, in: HSCC, ACM, 2017, pp. 163–172.
- [109] K. Ghorbal, A. Sogokon, Characterizing positively invariant sets: Inductive and topological methods, J. Symb. Comput. 113 (2022) 1–28.
- [110] C. C. Conley, Isolated invariant sets and the Morse index, no. 38, American Mathematical Soc., 1978.
- [111] Q. Wang, M. Chen, B. Xue, N. Zhan, J. Katoen, Synthesizing invariant barrier certificates via difference-of-convex programming, in: CAV (1), Vol. 12759 of Lecture Notes in Computer Science, Springer, 2021, pp. 443–466.
- [112] R. Gustavson, A. Ovchinnikov, G. Pogudin, New order bounds in differential elimination algorithms, J. Symb. Comput. 85 (2018) 128–147.
- [113] W. Simmons, H. Towsner, Proof mining and effective bounds in differential polynomial rings, Advances in Mathematics 343 (2019) 567–623.
- [114] F. Boulier, Efficient computation of regular differential systems by change of rankings using kähler differentials, in: MEGA 2000, 2000.
- [115] M. Jones, M. M. Peet, Using SOS for optimal semialgebraic representation of sets: Finding minimal representations of limit cycles, chaotic attractors and unions, in: ACC, IEEE, 2019, pp. 2084–2091.
- [116] K. Odani, The limit cycle of the van der Pol equation is not algebraic, Journal of Differential Equations 115 (1) (1995) 146–152.
- [117] A. Platzer, J. Quesel, P. Rümmer, Real world verification, in: CADE, Vol. 5663 of Lecture Notes in Computer Science, Springer, 2009, pp. 485–501.
- [118] B. D. O. Anderson, R. W. Scott, Output feedback stabilization—solution by algebraic geometry methods, Proceedings of the IEEE 65 (6) (1977) 849–861.
- [119] N. Fulton, S. Mitsch, J. Quesel, M. Völp, A. Platzer, Keymaera X: an axiomatic tactical theorem prover for hybrid systems, in: CADE, Vol. 9195 of Lecture Notes in Computer Science, Springer, 2015, pp. 527–538.