This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Computer-assisted proofs of Hopf bubbles and degenerate Hopf bifurcations

Kevin Church [email protected] Université de Montréal, Montreal, Canada. Elena Queirolo Corresponding author, [email protected] Technische Universität München, Munich, Germany.
Abstract

We present a computer-assisted approach to prove the existence of Hopf bubbles and degenerate Hopf bifurcations in ordinary and delay differential equations. We apply the method to rigorously investigate these nonlocal bifurcation structures in the FitzHugh-Nagumo equation, the extended Lorenz-84 model and a time-delay SI model.

1 Introduction

The Hopf bifurcation is a classical mechanism leading to the birth of a periodic orbit in a dynamical system. In the simplest setting of an ordinary differential equation (ODE), a Hopf bifurcation generically occurs when the linearization at a fixed point has a single pair of complex-conjugate imaginary eigenvalues. In this case, a perturbation by way of a scalar parameter would be expected, but not guaranteed, to result in a Hopf bifurcation. To prove the existence of the Hopf bifurcation, some non-degeneracy conditions must be checked. One such non-degeneracy condition pertains to the eigenvalues at the linearization: on variation of the distinguished scalar parameter, they must cross the imaginary axis transversally. We refer the reader to the papers [6, 9, 13, 19, 25] for some classical (and more recent) background concerning the Hopf bifurcation in the context of infinite-dimensional dynamical systems. A standard reference for the ODE case is the book of Marsden & McCracken [21].

1.1 Hopf bubbles

The Hopf bifurcation is generic in one-parameter vector fields of ODEs. Roughly speaking, this means it is prevalent in a topological sense among all vector fields that possess an equilibrium point with a pair of complex-conjugate imaginary eigenvalues. In the world of two-parameter vector fields, the richness of possible bifurcations is much greater. In this paper, we are primarily interested in a type of degenerate Hopf bifurcation that results from a particular failure of the eigenvalue transversality condition. The failure we are interested in is where the curve of eigenvalues has a quadratic tangency with the imaginary axis, so the eigenvalues do not cross transversally and, importantly, do not cross the imaginary axis at all.

Before surveying some literature about this bifurcation pattern, let us construct a minimal working example. Consider the ODE system

x˙\displaystyle\dot{x} =βxyx(x2+y2+α2)\displaystyle=\beta x-y-x(x^{2}+y^{2}+\alpha^{2}) (1)
y˙\displaystyle\dot{y} =x+βyy(x2+y2+α2).\displaystyle=x+\beta y-y(x^{2}+y^{2}+\alpha^{2}). (2)

The reader familiar with the normal form of the Hopf bifurcation should find this familiar, but might be be unnerved by the α2\alpha^{2} term, which is not present in the typical normal form. The linearization at the equilibrium (0,0)(0,0), produces the matrix

(βα211βα2),\left(\begin{array}[]{cc}\beta-\alpha^{2}&-1\\ 1&\beta-\alpha^{2}\end{array}\right),

which has the pair of complex-conjugate eigenvalues λ=βα2±i\lambda=\beta-\alpha^{2}\pm i. Treating α\alpha as being a fixed constant, we have a supercritical Hopf bifurcation as β\beta passes through α2\alpha^{2}. Conversely, if β>0\beta>0 is fixed and we interpret α\alpha as the parameter, there are two supercritical Hopf bifurcations: as α\alpha passes through ±β\pm\sqrt{\beta}. However, something problematic happens if β=0\beta=0: the eigenvalue branches αα2±i\alpha\mapsto-\alpha^{2}\pm i are tangent to the imaginary axis at α=0\alpha=0 and do not cross at all, so the non-degeneracy condition of the Hopf bifurcation fails.

More information can be gleaned by transforming to polar coordinates. Setting x=rcosθx=r\cos\theta and y=rsinθy=r\sin\theta results in the equation for the radial component

r˙=r(βr2α2),\dot{r}=r(\beta-r^{2}-\alpha^{2}),

while θ˙=1\dot{\theta}=1. There is a nontrivial periodic orbit (in fact, limit cycle) if βα2>0\beta-\alpha^{2}>0, with amplitude βα2\sqrt{\beta-\alpha^{2}}. Alternatively, there is a surface of periodic orbits described by the graph of β=r2+α2\beta=r^{2}+\alpha^{2}. For fixed β>0\beta>0, the amplitude rr of the periodic orbit, as a function of α\alpha, is αβα2\alpha\mapsto\sqrt{\beta-\alpha^{2}}, whose graph is half of an ellipse with semi-major axis 2β2\sqrt{\beta} and semi-minor axis β\sqrt{\beta}.

This non-local bifurcation structure, characterized by the connection of two Hopf bifurcations by a one-parameter branch of periodic orbits, has been given several names in different scientific fields. In intracellular calcium, it is frequently called a Hopf bubble [7, 16, 22, 27]. In infectious-disease modelling, where the bifurcation typically occurs at an endemic equilibrium, the accepted term is endemic bubble [8, 17, 18, 23, 26]. A more general definition of a (parametrically) non-local structure called bubbling is given in [15]. In the present work, we will refer to the structure as a Hopf bubble, since that name is descriptive of the geometric picture, the bifurcation involved, and is sufficiently general to apply in different scenarios in a model-independent way.

Hopf bubbles can, in many instances, be understood as being generated by a codimension-two bifurcation; see Section 1.3. However, this is not to say that they are rare. Aside from the applications in calcium dynamics and infectious-disease modelling in the previous paragraph, they have been observed numerically in models of neurons [1], condensed-phase combustion [20], predator-prey models [3], enzyme-catalyzed reactions [12], and a plant-water ecosystem model [31]. A recent computer-assisted proof also established the existence of a Hopf bubble in the Lorenz-84 model [30]. We will later refer to the degenerate Hopf bifurcation that gives rise to Hopf bubbles as a bubble bifurcation. The literature seems to not have an accepted name for this bifurcation, so we have elected to give it one here.

1.2 Bautin bifurcation

Bubble bifurcations are one type of dgenerate Hopf bifurcation. Another is the Bautin bifurcation, which occurs at a Hopf bifurcation whose first Lyapunov coefficient vanishes [14]. Like the bubble bifurcation, the Bautin bifurcation is a codimension-two bifurcation of periodic orbits. To illustrate this bifurcation, consider the planar normal form

x˙\displaystyle\dot{x} =βxyx(x2+y2)(αx2y2)\displaystyle=\beta x-y-x(x^{2}+y^{2})(\alpha-x^{2}-y^{2}) (3)
y˙\displaystyle\dot{y} =x+βyy(x2+y2)(αx2y2).\displaystyle=x+\beta y-y(x^{2}+y^{2})(\alpha-x^{2}-y^{2}). (4)

Transforming to polar coordinates, the angular component decouples producing θ˙=1\dot{\theta}=1, while the radial component gives

r˙=r(β+αr2r4).\dot{r}=r(\beta+\alpha r^{2}-r^{4}).

Nontrivial periodic orbits are therefore determined by the zero set of rβ+αr2r4r\mapsto\beta+\alpha r^{2}-r^{4}, which defines a two-dimensional smooth manifold. See later Figure 15 for a triangulation of (part of) this manifold.

1.3 Degenerate Hopf bifurcations and multi-parameter continuation

The bubble bifurcation has been fully characterized by LeBlanc [17], using the center manifold reduction and normal form theory for functional differential equations [9] and a prior classification of degenerate Hopf bifurcations for ODEs by Golubitsky and Langford [11]. The study of the Bautin bifurcation goes back to 1949 with the work of Bautin [2], and a more modern derivation based on normal form theory appears in the textbook of Kuznetsov [14].

Normal form theory and centre manifold reduction are incredibly powerful, providing both the direction of the bifurcation and criticality of bifurcating periodic orbits. The drawback is that they are computationally intricate and inherently local. In the present work, we advocate for an analysis of degenerate Hopf bifurcations in ODE and delay differential equations (DDE) by way of two-parameter continuation, desingularization and computer-assited proofs. We will discuss the latter topics in the next section.

The intuition behind our continuation idea can be pictorially seen in Figure 1, and understood analytically by way of our minimal working example, (1)–(2), and its surface of periodic orbits described by the equation β=α2+r2\beta=\alpha^{2}+r^{2}. There is an implicit relationship between two scalar parameters (α\alpha and β\beta) and the set of periodic orbits of the dynamical system. At a Hopf bifurcation, one of these periodic orbits should retract onto a steady state. In other words, their amplitudes (relative to steady state) should become zero. Using a desingularization technique to topologically separate periodic orbits from steady states that they bifurcate from, we can use two-parameter continuation to numerically compute and continue periodic orbits as they pass through a degenerate Hopf bifurcation. With computer-assisted proof techniques, we can prove that the numerically-computed objects are close to true periodic orbits, with explicit error bounds. Further a posteriori analysis can then be used to prove the existence of a degenerate Hopf bifurcation based on the output of the computer-assisted proof of the continuation.

Refer to caption
Figure 1: Near a Hopf bifurcation, the bifurcating periodic orbit is close to a pure cosine (red solution) with small amplitude. Far from the bifurcation, the amplitude tends to grow and more Fourier modes are represented (blue solution). In a Hopf bubble, the amplitude grows from zero and then decreases to zero as a parameter (α\alpha in this figure) is varied montonically. Variation of a control parameter (β\beta) can result in a smaller range of amplitudes and admissible interval over which the periodic orbit persists. If variation of the control parameter ultimately results in a collapse of the amplitude and admissibility ranges to a point, the result is a degenerate Hopf bifurcation.

1.4 Rigorous numerics

Computer-assisted proofs of Hopf bifurcations have been completed in [30] using a desingularization (sometimes called blow-up) approach, in conjunction with an a posteriori Newton-Kantorvich-like theorem. We lean heavily on these ideas in the present work. The former desingularization idea, which we adapt to delay equations in Section 3.1, is used to resolve the fact that, at the level of a Fourier series, a periodic orbit that limits (as a parameter varies) to a fixed point at a Hopf bifurcation is, itself, indistinguishable from that fixed point. Our methods of computer-assisted proof are based on contraction mappings, and it is critical that the objects we prove are isolated as fixed points. The desingularization idea exploits the fact that an amplitude-based change of variables can be used to develop an equivalent problem where the representative of a periodic orbit consists of a Fourier series, a fixed point and a real number. The latter real number represents a signed amplitude. This reformulation results in fixed points being spatially islolated from periodic orbits, thereby allowing contraction-based computer-assisted proofs to succeed.

The other big point of inspiration in this work is validated multi-parameter continuation. The technique was developed in [10] for continuation in general Banach spaces, and applied to some steady state problems for PDEs. We will overview this method in Section 2. It is based on a combination of a two-dimensional analogue of pseudo-arclength continuation and a posteriori, uniform validation of zeroes of nonlinear maps by a Newton-Kantorovich-like theorem applied to a Newton-like operator.

1.5 Contributions and applications

The main contribution of the present work is a general-purpose code for computing and proving two-parameter families of periodic orbits in polynomial delay differential equations. Equations of advanced or mixed-type can similarly be handled; there is no restriction whether delays are positive or negative. Ordinary differential equations can also be handled as a special case. Orbits can be proven in the vicinity of (and at) Hopf bifurcations, whether these are non-degenerate or degenerate. The first major release of the library BiValVe (Bifurcation Validation Venture, [4]) is being made in conjunction with the present work, and builds on an earlier version of the code associated to the work [29, 30]. Some non-polynomial delay differential equations can be handled using the polynomial embedding technique. The existence of Hopf bifurcation curves and degenerate Hopf bifurcations can then be completed by post-processing of the output of the computer-assisted proof.

To explore the applicability of our validated numerical methods, we explore Hopf bifurcations and degenerate Hopf bifurcations in

  • the extended Lorenz-84 model (ODE)

  • a time-delay SI model (DDE)

  • the FitzHugh-Nagumo equation (ODE)

  • an ODE with complicated branches of periodic orbits (ODE)

The first two examples replicate and extend some of the analysis appearing in [17, 30] using our computational scheme. The third examples provides, to our knowledge, the first analytically verified results on degenerate Hopf bifurcations and Hopf bubbles in that equation. The final example is a carefully designed ODE that exhibits the degenerate Hopf bifurcation in addition to folding and pinching in some projections of the manifold.

1.6 Overview of the paper

Section 2 serves as an overview of two-parameter continuation, both in the finite-dimensional and infinite-dimensional case. We introduce the continuation scheme for periodic orbits near Hopf bifurcations in Section 3 in the general case of delay differential equations. Some technical bounds for the computer-assisted proofs in Section 4. A specification to ordinary differential equations is presented in Section 5. In Section 6, we connect the computer-assisted proofs of the manifold of periodic orbits to analytical proofs of Hopf bifurcation curves, Hopf bubbles, and degenerate Hopf bifurcations. Our examples are presented in Section 7, and we complete a discussion and comment on future research directions in Section 8.

1.7 Notation

Given nn\in\mathbb{N}, denote n\mathbb{C}^{\mathbb{Z}}_{n} the vector space of \mathbb{Z}-indexed sequences of elements of n.\mathbb{C}^{n}. Denote Symm(n)\mbox{Symm}(\mathbb{C}^{\mathbb{Z}}_{n}) the proper subspace of n\mathbb{C}^{\mathbb{Z}}_{n} consisting of symmetric sequences; zSymm(n)z\in\mbox{Symm}(\mathbb{C}^{\mathbb{Z}}_{n}) if and only if znz\in\mathbb{C}^{\mathbb{Z}}_{n} and zk=zk¯z_{k}=\overline{z_{-k}} for all kk\in\mathbb{Z}. For any subspace UnU\subset\mathbb{C}^{\mathbb{Z}}_{n} closed under (componentwise) complex conjugation, denote Symm(U)=USymm(n)\operatorname{Symm}(U)=U\cap\operatorname{Symm}(\mathbb{C}^{\mathbb{Z}}_{n}). We will sometimes drop the subscript nn on n\mathbb{C}^{\mathbb{Z}}_{n} when the context is clear.

Given ν>1\nu>1, we denote ν1(n)\ell_{\nu}^{1}(\mathbb{C}^{n}) the subspace of n\mathbb{C}_{n}^{\mathbb{Z}} whose elements zz satisfy zν=defk|zk|ν|k|<||z||_{\nu}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\sum_{k\in\mathbb{Z}}|z_{k}|\nu^{|k|}<\infty. |||\cdot| will always be taken to be the norm on n\mathbb{C}^{n} induced by the standard inner product. Denote Kν(n)K_{\nu}(\mathbb{C}^{n}) the subspace of n\mathbb{C}_{n}^{\mathbb{Z}} whose elements zz satisfy zν,K=defk(ν|k|/|k|)|zk|<.||z||_{\nu,K}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\sum_{k\in\mathbb{Z}}(\nu^{|k|}/|k|)|z_{k}|<\infty. Introduce a bilinear form on ,\langle\cdot,\cdot\rangle on ν1()\ell_{\nu}^{1}(\mathbb{C}) as follows:

v,w=kvkwk¯.\langle v,w\rangle=\sum_{k\in\mathbb{Z}}v_{k}\overline{w_{k}}.

For v,w1v,w\in\mathbb{C}^{\mathbb{Z}}_{1}, define their convolution vwv*w by

(vw)k=k1+k2=kvk1wk2\displaystyle(v*w)_{k}=\sum_{k_{1}+k_{2}=k}v_{k_{1}}w_{k_{2}} (5)

whenever this series converges. Convolution is commutative and associative for sequences in ν1()\ell_{\nu}^{1}(\mathbb{C}). In this case, we define multiple convolutions (e.g. triple convolutions abca*b*c) inductively, by associativity.

A function g:ν1(n)ν1()g:\ell_{\nu}^{1}(\mathbb{C}^{n})\rightarrow\ell_{\nu}^{1}(\mathbb{C}) is a convlution polynomial of degree qq if

g(z)\displaystyle g(z) =c0+k=1qpkcp(zp1zpk),\displaystyle=c_{0}+\sum_{k=1}^{q}\sum_{p\in\mathcal{M}_{k}}c_{p}(z_{p_{1}}*\cdots*z_{p_{k}}),

for cpc_{p}\in\mathbb{C}, where k\mathcal{M}_{k} denotes the set of kk-element multisets of {1,,n}\{1,\dots,n\}, and each multiset pp is identified with the unique tuple (p1,,pk){1,,n}k(p_{1},\dots,p_{k})\{1,\dots,n\}^{k} such that pipi+1p_{i}\leq p_{i+1} for all i=1,,k1i=1,\dots,k-1. Analogously, g:ν1(n)ν1(m)g:\ell_{\nu}^{1}(\mathbb{C}^{n})\rightarrow\ell_{\nu}^{1}(\mathbb{C}^{m}) is a convolution polynomial of degree qq if zg(z),jν1()z\mapsto g(z)_{\cdot,j}\in\ell_{\nu}^{1}(\mathbb{C}) is a convolution polynomial of degree qq, for j=1,mj=1\dots,m.

If XX is a vector space, 0X0_{X} will denote the zero vector in that space. If XX is a metric space and UXU\subset X, we denote UU^{\circ} its interior, U\partial U its boundary, and U¯\overline{U} its closure.

An interval vector in k\mathbb{R}^{k} for some k1k\geq 1 is a subset of the form v=[a1,b1]××[ak,bk]v=[a_{1},b_{1}]\times\cdots\times[a_{k},b_{k}] for real scalars aj,bja_{j},b_{j}\in\mathbb{R}, j=1,,kj=1,\dots,k. If k\mathbb{R}^{k} is equipped with a norm ||||||\cdot|| we define v=supwvw||v||=\sup_{w\in v}||w||. Similarly, an interval vector in k\mathbb{C}^{k} is a product v=A1××Akv=A_{1}\times\cdots\times A_{k}, where each AjA_{j} is a closed disc in \mathbb{C}. We define the norm of a complex interval vector as v=supwvw||v||=\sup_{w\in v}||w|| whenever ||||||\cdot|| is a norm on k\mathbb{C}^{k}. In each case, be it real or comlex, the unit interval vector is the unique interval vector VV with the property that V=1||V||=1, and for any other interval vector vv with v=1||v||=1, the inclusion vVv\subseteq V is satisfied.

2 Validated two-parameter continuation

In this section, we review validated two-parameter continuation. Our presentation will loosely follow [10]. Some noteworthy changes compared to the references are that we work in a complexified (as opposed to strictly real) vector space, which causes some minor difficulties at the level of implementation.

We first review the continuation algorithm as it applies to finite-dimensional vector spaces in Section 2.1. We make comments concerning implementation in Section 2.2. We then describe how it is extended to general Banach spaces in Section 2.3. Validated continuation (i.e. computer-assisted proof) is discussed in Section 2.4.

2.1 Continuation in a finite-dimensional space

Let 𝒳\mathcal{X} and 𝒴\mathcal{Y} be finite-dimensional vector spaces over the field \mathbb{R}, with dim𝒳=dim𝒴+2\dim\mathcal{X}=\dim\mathcal{Y}+2, and consider a map 𝒢:𝒳𝒴\mathcal{G}:\mathcal{X}\rightarrow\mathcal{Y}. We are interested in the zero set of 𝒢\mathcal{G}. Given the codimension of 𝒢\mathcal{G}, we expect zeroes to be in a two-dimensional manifold in 𝒳\mathcal{X}. In what follows, equality with zero should be understood in the sense of numerical precision. For example, if we write 𝒢(x)=0\mathcal{G}(x)=0, what we mean is that 𝒢(x)||\mathcal{G}(x)|| is machine precision small.

Let x^0𝒳\hat{x}_{0}\in\mathcal{X} satisfy G(x^0)=0G(\hat{x}_{0})=0, and suppose D𝒢(x^0)D\mathcal{G}(\hat{x}_{0}) has two-dimensional kernel. This property is generic. Let {Φ^1,Φ^2}\{\hat{\Phi}_{1},\hat{\Phi}_{2}\} span the kernel. The tangent plane 𝒯x^0\mathcal{T}_{\hat{x}_{0}}\mathcal{M} at x^0\hat{x}_{0} of the two-dimensional solution manifold \mathcal{M} is therefore spanned by {Φ^1,Φ^2}\{\hat{\Phi}_{1},\hat{\Phi}_{2}\}. If ϵ1,ϵ2\epsilon_{1},\epsilon_{2} are small, we have

𝒢(x^0+ϵ1Φ^1+ϵ2Φ^2)=𝒢(x^0)+ϵ1D𝒢(x^0)Φ^1+ϵ2D𝒢(x^0)Φ^2+o(|ϵ|)=o(|ϵ|)\mathcal{G}(\hat{x}_{0}+\epsilon_{1}\hat{\Phi}_{1}+\epsilon_{2}\hat{\Phi}_{2})=\mathcal{G}(\hat{x}_{0})+\epsilon_{1}D\mathcal{G}(\hat{x}_{0})\hat{\Phi}_{1}+\epsilon_{2}D\mathcal{G}(\hat{x}_{0})\hat{\Phi}_{2}+o(|\epsilon|)=o(|\epsilon|)

by Taylor’s theorem, provided 𝒢\mathcal{G} is differentiable at x^0\hat{x}_{0}, and |ϵ||\epsilon| is any Euclidean norm of ϵ=(ϵ1,ϵ2)\epsilon=(\epsilon_{1},\epsilon_{2}). If 𝒢\mathcal{G} is twice continuously differentiable, the error is improved to O(|ϵ|2)O(|\epsilon|^{2}). Therefore, new candidate zeroes of 𝒢\mathcal{G} can be computed using x^0\hat{x}_{0} and a basis for the tangent space. This idea is at the heart of the continuation.

The continuation from x^0\hat{x}_{0} is done by way of iterative triangulation of the manifold \mathcal{M}. First, we compute an orthonormal basis of 𝒯x^0\mathcal{T}_{\hat{x}_{0}}\mathcal{M} by applying the Gram-Schmidt process to {Φ^1,Φ^2}\{\hat{\Phi}_{1},\hat{\Phi}_{2}\}; see Section 2.2 for some technical details. Using this orthonormal basis to define a local coordinate system, six vertices of a regular hexagon are computed around x^0\hat{x}_{0} at a specified distance σ\sigma from x^0\hat{x}_{0}; see Figure 2. Let these vertices be denoted x^1\hat{x}_{1} through x^6\hat{x}_{6}, arranged in counterclockwise order (relative to the local coordinate system) around x^0\hat{x}_{0}. Note that this means

x^i=x^0+ϵj,1Φ^1+ϵj,2Φ^2\hat{x}_{i}=\hat{x}_{0}+\epsilon_{j,1}\hat{\Phi}_{1}+\epsilon_{j,2}\hat{\Phi}_{2}

for some small ϵj,1,ϵj,2\epsilon_{j,1},\epsilon_{j,2}, which means that G(x^j)0G(\hat{x}_{j})\approx 0 for j=1,,6j=1,\dots,6. Each of these candidate zeroes x^j\hat{x}_{j} are then refined by applying Newton’s method to the map

x𝒢j(x)=(𝒢(x)Φ^1,xx^jΦ^2,xx^j).\displaystyle x\mapsto\mathcal{G}_{j}(x)=\left(\begin{array}[]{c}\mathcal{G}(x)\\ \langle\hat{\Phi}_{1},x-\hat{x}_{j}\rangle\\ \langle\hat{\Phi}_{2},x-\hat{x}_{j}\rangle\end{array}\right). (9)

The two added inner product equations ensure isolation of the solution (hence quadratic convergence of the Newton iterates) and that the Newton correction is perpendicular to the tangent plane. This map can similarly be used to refine the original zero x^0\hat{x}_{0}.

This initial hexagonal “patch” is itself formed by six triangles; see Figure 2. The continuation algorithm proceeds by selecting one of the boundary vertices (i.e. one of the vertices x^1\hat{x}_{1} through x^6\hat{x}_{6}) and attempting to “grow” the manifold further. We describe this “growth” phase below. However, first, some terminology. The vertices will now be referred to as nodes. An edge is a line segment connecting two nodes, and they will be denoted by pairs of nodes: {x^i,x^j}\{\hat{x}_{i},\hat{x}_{j}\} for node x^i\hat{x}_{i} connected to x^j\hat{x}_{j}. Two nodes are incident if they are connected by an edge. A simplex is the convex hull of three edges that form a triangle. Once the hexagonal patch is created, the data consists of:

  • the nodes x^0,,x^6\hat{x}_{0},\dots,\hat{x}_{6};

  • the “boundary edges” {x^1,x^2},,{x^5,x^6},{x^6,x^1}\{\hat{x}_{1},\hat{x}_{2}\},\dots,\{\hat{x}_{5},\hat{x}_{6}\},\{\hat{x}_{6},\hat{x}_{1}\};

  • the six simplices formed by triangles with x^0\hat{x}_{0} as one of the nodes.

Two simplices are adjacent if they share an edge. An internal simplex is a simplex that is adjacent to three other simplices, and it is a boundary simplex otherwise. Therefore, the six simplices of the initial patch are all considered boundary since they are adjacent to exactly two others. Similarly, an edge of a simplex can be declared boundary or internal; internal edges are those that are shared with another simplex, and boundary edges are not. A boundary node is any node on a boundary edge, a frontal node is a boundary node on an edge shared by two simplices, and an internal node is a node that is not a boundary node.

Refer to caption
Figure 2: The hexagonal patch with normal vector indicated by an arrow. In this initial configuration, every node except for x^0\hat{x}_{0} is a boundary node. The boundary edges are blue.

Let xx be a chosen frontal node (see Section 2.2.3 for a general discussion on frontal node selection). By construction, xx is part of (at least) three distinct edges, two of which connect to boundary nodes, and at least one that connects to an internal node. The algorithm to grow a simplex is as follows. See Figure 3 for a visualization.

  1. 1.

    Compute an orthonormal basis for the tangent space 𝒯x\mathcal{T}_{x}\mathcal{M}.

  2. 2.

    Compute the (average) gap complement direction. Let x1,,xmx^{\circ}_{1},\dots,x^{\circ}_{m} denote the list of internal nodes nodes incident to xx. The gap complement direction is yc=x+1mi=1mxiy_{c}=-x+\frac{1}{m}\sum_{i=1}^{m}x_{i}^{\circ}.

  3. 3.

    Let x1x_{1} and x2x_{2} denote the boundary nodes incident to xx; form the edge directions y1=x1xy_{1}=x_{1}-x and y2=x2xy_{2}=x_{2}-x.

  4. 4.

    Orthogonally project ycy_{c}, y1y_{1} and y2y_{2} onto the tangent plane 𝒯x\mathcal{T}_{x}\mathcal{M}. Let these projections be denoted PycPy_{c}, Py1Py_{1} and Py2Py_{2}. Introduce a two-dimensional coordinate system on 𝒯x\mathcal{T}_{x}\mathcal{M} by way of a “unitary” (see Section 2.2) transformation to 2\mathbb{R}^{2}. Let P~yc\tilde{P}y_{c}, P~y1\tilde{P}y_{1} and P~y2\tilde{P}y_{2} denote the representatives in 2\mathbb{R}^{2}.

  5. 5.

    In the local two-dimensional coordinate system, compute the counter-clockwise (positive) angle θ1\theta_{1} required to complete a rotation from P~yc\tilde{P}y_{c} to P~y1\tilde{P}y_{1}, and the counter-clockwise (positive) angle θ2\theta_{2} required for rotation from P~yc\tilde{P}y_{c} to P~y2\tilde{P}y_{2}. The gap angle γ\gamma and orientation ρ\rho is

    γ=max{θ1θ2,θ2θ1},ρ={1,θ2>θ12θ2<θ1\gamma=\max\{\theta_{1}-\theta_{2},\theta_{2}-\theta_{1}\},\hskip 28.45274pt\rho=\left\{\begin{array}[]{ll}1,&\theta_{2}>\theta_{1}\\ 2&\theta_{2}<\theta_{1}\end{array}\right.
  6. 6.

    If γ<π6\gamma<\frac{\pi}{6} then close the gap: add the the triangle formed by the nodes {x,x1,x2}\{x,x_{1},x_{2}\} to the list of simplices, flag it as a boundary simplex, flag xx as an internal node, and conclude the growth step. Otherwise, proceed to step 7.

  7. 7.

    Generate the predictor “fan” in 2\mathbb{R}^{2}: let k=min{1,3θ/π}k=\min\left\{1,\lfloor 3\theta/\pi\rfloor\right\} and define the predictors

    P~y~j=R(jγ/k)P~yρ,j=1,,k,\tilde{P}\tilde{y}_{j}=R(j\gamma/k)\tilde{P}y_{\rho},\hskip 28.45274ptj=1,\dots,k,

    for R(θ)R(\theta) the 2×22\times 2 counterclockwise rotation matrix through angle θ\theta.

  8. 8.

    Invert the unitary transformation and map P~y~j\tilde{P}\tilde{y}_{j} into the tangent plane 𝒯x\mathcal{T}_{x}\mathcal{M}; let the result be the vectors Py~jP\tilde{y}_{j}, j=1,,kj=1,\dots,k.

  9. 9.

    Define predictors x^j=x+σPy~j\hat{x}_{j}=x+\sigma P\tilde{y}_{j} for j=1,,kj=1,\dots,k, where σ\sigma is a user-specified step size. Refine them using Newton’s method applied to (9), where Φ^1\hat{\Phi}_{1} and Φ^2\hat{\Phi}_{2} are now the orthonormal basis for 𝒯x\mathcal{T}_{x}\mathcal{M}.

  10. 10.

    Add the triangles formed by nodes {x,x^1,x^2}\{x,\hat{x}_{1},\hat{x}_{2}\}, {x,x^2,x^3}\{x,\hat{x}_{2},\hat{x}_{3}\}, \dots, {x,x^k1,x^k}\{x,\hat{x}_{k-1},\hat{x}_{k}\}. To the list of simplices, flag them as boundary simplices, and flag xx as an internal node. Do the same with the triangles formed by {x,x^1,}\{x,\hat{x}_{1},*\} and {x,x^k,}\{x,\hat{x}_{k},*\}, where * denotes ones of x1x_{1} and x2x_{2}, depending on orientation ρ\rho.

Refer to caption
Figure 3: Cartoon diagram of the simplex growth phase. For visualization purposes, we think of x1x_{1} and x2x_{2} as being in the tangent space 𝒯x\mathcal{T}_{x}\mathcal{M}; in reality, these should be close to the tangent space but not strictly contained in it. The boundary edges {x,x1}\{x,x_{1}\} and {x,x2}\{x,x_{2}\} are blue, and these form a simplices with the interior edge {x,x}\{x,x^{\circ}\} that generates the gap complement direction. The predictor “fan”, after being mapped back to the tangent space, consists of the vertices x^1\hat{x}_{1}, x^2\hat{x}_{2} and x^3\hat{x}_{3}. The four new simplices that can be formed, namely {x,x^1,x^2}\{x,\hat{x}_{1},\hat{x}_{2}\}, {x,x^2,x^3}\{x,\hat{x}_{2},\hat{x}_{3}\}, {x,x^1,x2}\{x,\hat{x}_{1},x_{2}\} and {x,x^3,x1}\{x,\hat{x}_{3},x_{1}\}, share the same angle at the xx vertex.

At the end of the simplex growth algorithm, one or more boundary simplices is added to the dictionary, and one additional node will be flagged as internal. The structure of this algorithm ensures that each simplex added to the dictionary will be adjacent to exactly two others, indicating that our convention of internal vs. boundary simplices is effective. Once the growth algorithm is complete, another frontal node is selected and the growth phase is repeated. This continues until a sufficient portion of the manifold has been computed (i.e. a user-specified exit condition is reached), or until a Newton’s method correction fails.

2.2 Comments on implementation

Here we collect some remarks concerning implementation of the finite-dimensional continuation. These comments may be applicable for general continuation problems, but we will often emphasize our specific situation which is continuation of periodic orbits in delay differential equations.

2.2.1 Complex rotations for the kernel of D𝒢(x)D\mathcal{G}(x)

First, generating the kernel elements {Φ^1,Φ^2}\{\hat{\Phi}_{1},\hat{\Phi}_{2}\} of D𝒢(x^)D\mathcal{G}(\hat{x}) for an approximate x^\hat{x} must be handled in a way that is appropriate to the space 𝒳\mathcal{X}. In our problem, 𝒳\mathcal{X} has additional structure: it is a subset of a complexified real vector space equipped with a discrete symmetry. However, in our implementation (that is, in the environment of MATLAB) we work on a generic complex vector space without this symmetry, so when we compute kernel elements using QR decomposition, the computed vectors are not necessarily in 𝒳\mathcal{X}. We fix this by performing a complex rotation to put the kernel back into 𝒳\mathcal{X}. This can always be done because Φ^1\hat{\Phi}_{1} and Φ^2\hat{\Phi}_{2}, as computed using QR, are always \mathbb{C}-linearly independent.

2.2.2 Orthonormal basis for the tangent space

The next point concerns the “orthonormal basis” of 𝒯x^\mathcal{T}_{\hat{x}}\mathcal{M}. Let us be a bit more precise. In our problem, 𝒳\mathcal{X} is a product of the form n×𝒱\mathbb{R}^{n}\times\mathcal{V}, where 𝒱2M+1\mathcal{V}\subset\mathbb{C}^{2M+1} is characterized by v=(vM,vM+1,,vM1,vM)𝒱v=(v_{-M},v_{-M+1},\dots,v_{M-1},v_{M})\in\mathcal{V} satisfies v=Svv=Sv, where Sv=(vM¯,vM1¯,,vM+1¯,vM¯)Sv=(\overline{v_{M}},\overline{v_{M-1}},\dots,\overline{v_{-M+1}},\overline{v_{-M}}). Consider the standard inner product ,\langle\cdot,\cdot\rangle on n+2M+1\mathbb{C}^{n+2M+1}. Once a basis {Φ^1,Φ^2}\{\hat{\Phi}_{1},\hat{\Phi}_{2}\} of 𝒯x^\mathcal{T}_{\hat{x}}\mathcal{M} has been computed, these basis vectors can be interpreted as being elements of n+2M+1\mathbb{C}^{n+2M+1}, and we say that they are orthogonal if Φ^1,Φ^2=0\langle\hat{\Phi}_{1},\hat{\Phi}_{2}\rangle=0. It is straightforward to verify that the Gram-Schmidt process applied to this basis produces yet another basis of 𝒯x^\mathcal{T}_{\hat{x}}\mathcal{M} (that is, it does not break the symmetry), and that the new basis is orthonormal with respect to ,\langle\cdot,\cdot\rangle.

2.2.3 Node prioritization for the simplex growth algorithm

A suitable selection of a frontal node for simplex growth might not be obvious. First, nodes that are more re-entrant (i.e. have many edges incident to them) are generally given higher priority. This is because such nodes are more likely to have the simplex growth algorithm perform the close the gap sub-routine. We want to avoid having thin simplices, so prioritizing the closing off of re-entrant nodes takes priority. After this, typically grow from the “oldest” boundary node.

2.2.4 Local coordinate system in 2\mathbb{R}^{2} for the tangent space

Finally, we must discuss the generation of a two-dimensional coordinate system for 𝒯x\mathcal{T}_{x}\mathcal{M}. If Φ^1\hat{\Phi}_{1} and Φ^2\hat{\Phi}_{2} are an orthonormal basis for the tangent space, then the map

yLy=(Φ^1y,Φ^2y)y\mapsto Ly=(\hat{\Phi}_{1}^{*}y,\hat{\Phi}_{2}^{*}y)

is invertible, where Φ^i\hat{\Phi}_{i}^{*} denotes the conjugate transpose of Φi\Phi_{i}. Writing Ly=(u,v)2Ly=(u,v)\in\mathbb{C}^{2} for y𝒳y\in\mathcal{X}, one can verify that (u,v)(u,v) is the unique solution of

y=Φ^1u+Φ^2v.y=\hat{\Phi}_{1}u+\hat{\Phi}_{2}v.

In particular, the range of this map is 2\mathbb{R}^{2} (that is, each of uu and vv is real) for our specific problem, where the space 𝒳\mathcal{X} is n×𝒱\mathbb{R}^{n}\times\mathcal{V}, with 𝒱\mathcal{V} having the symmetry described two paragraphs prior. The inverse map is

(u,v)L1(u,v)=Φ^1u+Φ^2v.(u,v)\mapsto L^{-1}(u,v)=\hat{\Phi}_{1}u+\hat{\Phi}_{2}v.

Moreover, this map is unitary in the sense that Ly,Lw=y,w\langle Ly,Lw\rangle=\langle y,w\rangle for y,w𝒳y,w\in\mathcal{X}, where on the left-hand side we take the standard inner product on 2\mathbb{R}^{2}. It is for these reasons that rotations on 2\mathbb{R}^{2} performed after the action of LL are consistent with rotations in the tangent space relative to the gap complement direction.

2.3 Continuation in a Banach space

Two-parameter continuation can be introduced generally in a Banach space, but here we take the perspective that 𝒢:𝒳𝒴\mathcal{G}:\mathcal{X}\rightarrow\mathcal{Y} of Section 2.1 is a projection of a nonlinear map G:XYG:X\rightarrow Y, and we are actually interested in computing the zeroes of GG, rather than those of 𝒢\mathcal{G}. That is to say, there exist projection operators π𝒳:X𝒳\pi_{\mathcal{X}}:X\rightarrow\mathcal{X}, π𝒴:Y𝒴\pi_{\mathcal{Y}}:Y\rightarrow\mathcal{Y}, and associated embeddings i𝒳:𝒳Xi_{\mathcal{X}}:\mathcal{X}\hookrightarrow X, i𝒴:𝒴Yi_{\mathcal{Y}}:\mathcal{Y}\hookrightarrow Y, such that 𝒢=π𝒴Gi𝒳\mathcal{G}=\pi_{\mathcal{Y}}G\circ i_{\mathcal{X}}.

While the zeroes of GG are what we want, we will still use the approximate zeroes of 𝒢\mathcal{G} to rigorously enclose them. Abstractly, let x^0,x^1,x^2𝒳\hat{x}_{0},\hat{x}_{1},\hat{x}_{2}\in\mathcal{X} be the three nodes of a simplex. Then we may introduce a coordinate system on this simplex as follows: write each element x^s\hat{x}_{s} of this simplex as a unique linear combination

x^s=x^0+s1(x^1x^0)+s2(x^2x^0)\displaystyle\hat{x}_{s}=\hat{x}_{0}+s_{1}(\hat{x}_{1}-\hat{x}_{0})+s_{2}(\hat{x}_{2}-\hat{x}_{0}) (10)

for s=(s1,s2)Δ={(a,b)2:0a,b1,0a+b1}s=(s_{1},s_{2})\in\Delta=\{(a,b)\in\mathbb{R}^{2}:0\leq a,b\leq 1,\hskip 2.84526pt0\leq a+b\leq 1\}. Let {Φ^j,1,Φ^j,2}\{\hat{\Phi}_{j,1},\hat{\Phi}_{j,2}\} for j=0,1,2j=0,1,2 denote an orthonormal basis for the kernel of D𝒢(x^j)D\mathcal{G}(\hat{x}_{j}). We can then form the interpolated kernels

Φ^s,i=Φ^0,i+s1(Φ^1,iΦ^0,i)+s2(Φ^2,iΦ^0,i)\hat{\Phi}_{s,i}=\hat{\Phi}_{0,i}+s_{1}(\hat{\Phi}_{1,i}-\hat{\Phi}_{0,i})+s_{2}(\hat{\Phi}_{2,i}-\hat{\Phi}_{0,i})

for i=1,2i=1,2. We introduce a nonlinear map Gs:XY×2G_{s}:X\rightarrow Y\times\mathbb{R}^{2},

Gs(x)=(G(x)Φ^s,1,π𝒳xx^sΦ^s,2,π𝒳xx^s).\displaystyle G_{s}(x)=\left(\begin{array}[]{c}G(x)\\ \langle\hat{\Phi}_{s,1},\pi_{\mathcal{X}}x-\hat{x}_{s}\rangle\\ \langle\hat{\Phi}_{s,2},\pi_{\mathcal{X}}x-\hat{x}_{s}\rangle\end{array}\right). (14)

The objective is to prove that GsG_{s} has a unique zero close to x^s\hat{x}_{s} for each sΔs\in\Delta. If this can be proven, and in fact GsG_{s} is CkC^{k} for some k1k\geq 1, then one can prove [10] that the the zero set of GG is a CkC^{k}, two-dimensional manifold. If this same can be proven for a collection of simplices, then the zero set is globally (i.e. on the union of the cobordant simplicial patches) a CkC^{k} manifold over all patches that can be proven; that is, the transition maps between patches are CkC^{k}.

Remark 1.

There is a subtle point concerning the relative orientations of the individual kernel elements Φ^0,i\hat{\Phi}_{0,i}, Φ^1,i\hat{\Phi}_{1,i} and Φ^2,i\hat{\Phi}_{2,i} that are used to define the interpolation Φ^s,i\hat{\Phi}_{s,i}. The validated continuation is unstable (and can even fail outright) if the interpolated kernels Φ^s,i\hat{\Phi}_{s,i} vary too much over sΔs\in\Delta. If these vectors all lived in the exact same two-dimensional tangent space, this would be very straightforward; we could simply (real) rotate and/or reflect each basis {Φ^1,i,Φ^2,i}\{\hat{\Phi}_{1,i},\hat{\Phi}_{2,i}\} for i=1,2i=1,2 so that they matched {Φ^1,0,Φ^2,0}\{\hat{\Phi}_{1,0},\hat{\Phi}_{2,0}\} exactly. However, these vectors live in different tangent spaces, so it is not as easy. The more consistent differential-geometric way to solve the problem would be to parallel transport the tangent basis {Φ^1,0,Φ^2,0}\{\hat{\Phi}_{1,0},\hat{\Phi}_{2,0}\} to the other two nodes and use these as bases for the tangent spaces there. We do nothing so sophisticated. We merely perform (real) rotations or reflections of the orthonormal bases {Φ^1,i,Φ^2,i}\{\hat{\Phi}_{1,i},\hat{\Phi}_{2,i}\} in their respective tangent spaces (two-dimensional) for i=1,2i=1,2 in such a way that, in norm, these are as close as possible to {Φ^1,0,Φ^2,0}\{\hat{\Phi}_{1,0},\hat{\Phi}_{2,0}\}. In practice, this has the effect of promoting enough alignment of the bases that proofs are feasible. We emphasize that this alignment process is only needed for the validated continuation; using misaligned tangent bases is not a problem for the steps described in Section 2.1.

2.4 Validated continuation and the radii polynomial approach

The radii polynomial approach is essentially a Newton-Kantorovich theorem with an approximate derivative, approximate inverse, and domain parametrization. It will be used to connect the approximate zeroes of 𝒢\mathcal{G} with exact zeroes of GG from the previous section. We state it generally for a family of sΔs\in\Delta -dependent maps Fs:X1X2F_{s}:X_{1}\rightarrow X_{2}. We include a short proof for completeness.

Theorem 1.

Suppose Fs:X1X2F_{s}:X_{1}\rightarrow X_{2} is differentiable for each sΔs\in\Delta, where X1X_{1} and X2X_{2} are Banach spaces. Let x^sX1\hat{x}_{s}\in X_{1} for all sΔs\in\Delta. Suppose there exist for each sΔs\in\Delta a bounded linear operator As:X1X2A_{s}^{\dagger}:X_{1}\rightarrow X_{2}, a bounded and injective linear operator As:X2X1A_{s}:X_{2}\rightarrow X_{1}, and non-negative reals Y0Y_{0}, Z0Z_{0}, Z1Z_{1} and Z2=Z2(r)Z_{2}=Z_{2}(r) such that

AsFs(x^s)\displaystyle||A_{s}F_{s}(\hat{x}_{s})|| Y0\displaystyle\leq Y_{0} (15)
IAsAsB(X1,X1)\displaystyle||I-A_{s}A_{s}^{\dagger}||_{B(X_{1},X_{1})} Z0\displaystyle\leq Z_{0} (16)
As(DFs(x^s)As)B(X1,X1)\displaystyle||A_{s}(DF_{s}(\hat{x}_{s})-A_{s}^{\dagger})||_{B(X_{1},X_{1})} Z1\displaystyle\leq Z_{1} (17)
As(DFs(x^s+δ)DFs(x^s))B(X1,X1)\displaystyle||A_{s}(DF_{s}(\hat{x}_{s}+\delta)-DF_{s}(\hat{x}_{s}))||_{B(X_{1},X_{1})} Z2(r), for all δBr(x^s)¯X,\displaystyle\leq Z_{2}(r),\hskip 14.22636pt\text{ for all }\delta\in\overline{B_{r}(\hat{x}_{s})}\subset X, (18)

for all sΔs\in\Delta, where Br(x^s)B_{r}(\hat{x}_{s}) is the closed ball of radius rr centered at x^s\hat{x}_{s}, and ||||B(X1,X1)||\cdot||_{B(X_{1},X_{1})} denotes the induced operator norm on X1X_{1}. Suppose there exists r0>0r_{0}>0 such that the radii polynomial

p(r)=rZ2(r)+(Z1+Z01)r+Y0p(r)=rZ_{2}(r)+(Z_{1}+Z_{0}-1)r+Y_{0}

satisfies p(r0)<0p(r_{0})<0. Then for each sΔs\in\Delta, there is a unique xsBr0(x^s)x_{s}\in B_{r_{0}}(\hat{x}_{s}) such that Fs(xs)=0F_{s}(x_{s})=0. If (s,x)Fs(x)(s,x)\mapsto F_{s}(x) and sAss\mapsto A_{s} are CkC^{k}, then the same is true of sxss\mapsto x_{s}.

Proof.

Define the Newton-like operator Ts(x)=xAsFs(x)T_{s}(x)=x-A_{s}F_{s}(x). We will show that TsT_{s} is a contraction on Br0(x^s)¯\overline{B_{r_{0}}(\hat{x}_{s})}, uniformly for sΔs\in\Delta. First, write xBr0(x^s)¯x\in\overline{B_{r_{0}}(\hat{x}_{s})} in the form x=x^s+δx=\hat{x}_{s}+\delta for some δ𝒳r||\delta||_{\mathcal{X}}\leq r. Then

DTs(x)\displaystyle||DT_{s}(x)|| =IAsDFs(x)\displaystyle=||I-A_{s}DF_{s}(x)||
IAsAs+As(AsDFs(x^s))+As(DFs(x^s)DFs(x^s+δ))\displaystyle\leq||I-A_{s}A_{s}^{\dagger}||+||A_{s}(A_{s}^{\dagger}-DF_{s}(\hat{x}_{s}))||+||A_{s}(DF_{s}(\hat{x}_{s})-DF_{s}(\hat{x}_{s}+\delta))||
Z0+Z1+Z2(r),\displaystyle\leq Z_{0}+Z_{1}+Z_{2}(r),

Now, using the triangle inequality and the mean-value inequality,

Ts(x)x^s\displaystyle||T_{s}(x)-\hat{x}_{s}|| Ts(x^s+δ)Ts(x^s)+Ts(x^s)x^s\displaystyle\leq||T_{s}(\hat{x}_{s}+\delta)-T_{s}(\hat{x}_{s})||+||T_{s}(\hat{x}_{s})-\hat{x}_{s}||
rsupt[0,1]DTs(x^s+tδ)B(X1,X1)+AsFs(x^s)\displaystyle\leq r\sup_{t\in[0,1]}||DT_{s}(\hat{x}_{s}+t\delta)||_{B(X_{1},X_{1})}+||A_{s}F_{s}(\hat{x}_{s})||
(Z0+Z1+Z2(r))r+Y0.\displaystyle\leq(Z_{0}+Z_{1}+Z_{2}(r))r+Y_{0}.

Choosing r=r0r=r_{0}, the radii polynomial implies that Ts(x)x^s<r||T_{s}(x)-\hat{x}_{s}||<r, so TsT_{s} is a self-map on Br0(x^s)¯\overline{B_{r_{0}}(\hat{x}_{s})} with its range in the interior. Moreover, since p(r0)<0p(r_{0})<0, we get that Z0+Z1+Z2(r0)<1Z_{0}+Z_{1}+Z_{2}(r_{0})<1, which proves that Ts:Br0(x^s)¯Br0(x^s)T_{s}:\overline{B_{r_{0}}(\hat{x}_{s})}\rightarrow B_{r_{0}}(\hat{x}_{s}) is a contraction (uniformly in sΔs\in\Delta). By the Banach fixed point theorem, TsT_{s} has a unique fixed point xsBr0(x^s)x_{s}\in B_{r_{0}}(\hat{x}_{s}) for each sΔs\in\Delta, and sxss\mapsto x_{s} is CkC^{k} provided the same is true of (s,x)Ts(x)(s,x)\mapsto T_{s}(x). Since AsA_{s} is injective, FsF_{s} has a unique zero in Br0(x^s)B_{r_{0}}(\hat{x}_{s}) if and only if TsT_{s} has a unique fixed point there. ∎

For our problem, injectivity of AsA_{s} will always follow from a successful verification of p(r0)<0p(r_{0})<0. See Lemma 4.

Let x^s\hat{x}_{s} be the convex combination defined by (10) for the simplex nodes x^0\hat{x}_{0}, x^1\hat{x}_{1} and x^2\hat{x}_{2}. We will say this simplex has been validated if we successfully find a radius r0r_{0} such that the conditions of the radii polynomial theorem are successful for the nonlinear map Gs:XY×2G_{s}:X\rightarrow Y\times\mathbb{R}^{2}.

In our implementation, we generate simplices “offline” first. This allows the workload to be distributed across several computers, since the validation step can be restricted to only a subset of the computed simplices. We do not implement a typical refinement procedure, where simplices that fail to validated are split, with more nodes added and corrected with Newton. Rather, we implement an adaptive refinement step, which can help with the validation if failure is primarily a result of interval over-estimation. See Section 4.2.1.

Remark 2.

The operators AsA_{s} and AsA_{s}^{\dagger} have standard interpretations in terms of the nonlinear map FsF_{s}. The operator As:X1X2A_{s}^{\dagger}:X_{1}\rightarrow X_{2} is expected to be an approximation of DFs(x^s)DF_{s}(\hat{x}_{s}), which means that Z1Z_{1} is a measure of the quality of the approximation. Conversely, As:X2X1A_{s}:X_{2}\rightarrow X_{1} is expected to be an approximation of the inverse of AsA_{s}^{\dagger}, so that Z0Z_{0} measures the quality of this approximation. Indirectly, AsA_{s} acts as an appoximation of DFs(x^s)1DF_{s}(\hat{x}_{s})^{-1}.

2.5 Globalizing the manifold

Theorem 1 guarantees that the map from the standard simplex to the zero set of Gs()G_{s}(\cdot) is C1C^{1}. There is then a natural question as to the smoothness of the manifold obtained by gluing together the images of the C1C^{1} maps. This is answered in the affirmative in [10], and is primarily a consequence of the numerical data being equal on cobordant simplices.

3 Continuation of periodic orbits through (degenerate) Hopf bifurcations

In this section, we construct a nonlinear map whose zeroes will encode periodic solutions of a delay differential equation

y˙(t)=f(y(t+μ1),,y(t+μJ),α,β)\displaystyle\dot{y}(t)=f(y(t+\mu_{1}),\dots,y(t+\mu_{J}),\alpha,\beta) (19)

for f:(n)J+1×2nf:(\mathbb{R}^{n})^{J+1}\times\mathbb{R}^{2}\rightarrow\mathbb{R}^{n}, with some (positive, negative or zero) constant delays μ1,,μJ\mu_{1},\dots,\mu_{J}, and distinguished systems parameters α,β\alpha,\beta. We will briefly consider ordinary differential equations in Section 5 as a special case. We assume that ff is sufficiently smooth to permit further partial derivative computations.

Following [30], we use the desingularization approach to isolate periodic orbits from (potentially) nearby steady states. This approach allows us to put a large distance (in the sense of a suitable Banach space) between steady states and periodic orbits that arise from Hopf bifurcations. This is exposited in Section 3.1, where we also discuss some details concerning non-polynomial nonlinearities.

The next Section 3.2 is devoted to the development of a nonlinear map whose zeroes encode periodic orbits of our delay differential equation. In this map, periodic orbits are isolated from fixed points. We present the map abstractly at the level of a function space, and then with respect to a more concrete sequence space.

In Section 3.3, we lift the map of the previous section into the scope of two-parameter continuation. We develop an abstract template for the map on a relevant Banach space, define an approximate Fréchet derivative near a candidate zero of this map, and investigate some properties of the Newton-like operator. Specifically, we verify that numerical data corresponding to an approximate real periodic orbit, under conditional contraction of the Newton-like operator, will converge to a real periodic orbit.

3.1 Desingularization, polynomial embedding and phase isolation

We begin by doing a “blowup” around the periodic orbit. Write y=x+azy=x+az, for xx a candidate equilibrium point. Then we get the rescaled vector field

f~(z1,,zJ,x,a,α,β)={a1(f(x+az1,,x+azJ,α,β)f(x,,x,α,β)),a0j=0Jdyjf(x,,x,α,β)zj,a=0.\displaystyle\tilde{f}(z_{1},\dots,z_{J},x,a,\alpha,\beta)=\left\{\begin{array}[]{ll}a^{-1}(f(x+az_{1},\dots,x+az_{J},\alpha,\beta)-f(x,\dots,x,\alpha,\beta)),&a\neq 0\\ \sum_{j=0}^{J}d_{y_{j}}f(x,\dots,x,\alpha,\beta)z_{j},&a=0.\end{array}\right.

The interpretation is that z=1||z||=1, so that aa behaves like the relative norm-amplitude of the periodic orbit. The vector field above is Ck1C^{k-1} provided the original function ff is CkC^{k} and xx is an equilibrium point; that is, f(x,,x,α,β)=0f(x,\dots,x,\alpha,\beta)=0. At this stage we can summarize by saying that our goal is to find a pair (x,z)(x,z) such

f(x,,x,α,β)\displaystyle f(x,\dots,x,\alpha,\beta) =0\displaystyle=0
z˙(t)\displaystyle\dot{z}(t) =f~(z(t+μ1),,z(t+μJ),x,a,α,β),\displaystyle=\tilde{f}(z(t+\mu_{1}),\dots,z(t+\mu_{J}),x,a,\alpha,\beta),
z\displaystyle||z|| =1\displaystyle=1

where zz is ω\omega-periodic for an unknown period ω\omega; equivalently, the frequency of zz is ψ=2πω\psi=\frac{2\pi}{\omega}.

Remark 3.

It is a common strategy in rigorous numerics, especially for periodic orbits, that time is re-scaled so that the period appears as a parameter in the differential equation. We do not do that here, since this would have the effect of dividing every delay μj\mu_{j} by the period. This causes its own set of problems.

In what follows, it will be beneficial for the vector field f~\tilde{f} to be polynomial. This is because we will make use of a Fourier spectral method, and polynomial nonlinearities in the function space translate directly to convolution-type nonlinearities on the sequence space of Fourier coefficients. While we can make due with non-polynomial nonlinearities, it is greatly simplifies the computer-assisted proof if they are polynomial. To fix this, we generally advocate the use of the polynomial embedding technique. The idea is that many analytic, non-polynomial functions are themselves solutions of polynomial ordinary differential equations. The reader may consult [van den Berg, Groothedde, Lessard] for a brief survey of this idea in the context of delay differential equations. See Section 7.2 for a specific example.

Applying the polynomial embedding procedure always introduces additional scalar differential equations. If we need to introduce mm extra scalar equations to get a polynomial vector field, this will also introduce mm natural boundary conditions that fix the initial conditions of the new components. As a consequence, we need to bring in mm unfolding parameters to balance the system. This is accomplished in a problem-specific way; see [van den Berg, Groothedde, Lessard] for some general guidelines and a discussion for the need of these extra unfolding parameters. By an abuse of notation, we assume f~\tilde{f} is polynomial (that is, the embedding has already been performed), and we write it as

f~(z1,,zJ,x,a,α,β,η),\tilde{f}(z_{1},\dots,z_{J},x,a,\alpha,\beta,\eta),

where ηm\eta\in\mathbb{R}^{m} is a vector representing the unfolding parameter, and we now interpret f~:(n+m)J+1×2n+m\tilde{f}:(\mathbb{R}^{n+m})^{J+1}\times\mathbb{R}^{2}\rightarrow\mathbb{R}^{n+m}. We then write the natural boundary condition corresponding to the polynomial embedding as

θBC(z(0),x,a,α,β,η)=0m.\theta_{BC}(z(0),x,a,\alpha,\beta,\eta)=0\in\mathbb{R}^{m}.

It can also be useful to eliminate non-polynomial parameter dependence from the vector field, especially if the latter has high-order polynomial terms with respect to the state variable zz. This can often be accomplished by introducing extra scalar variables. For example, if f~\tilde{f} is

αepxz1βz22+η1\alpha e^{-px}z_{1}-\beta z_{2}^{2}+\eta_{1}

and we want to eliminate the non-polynomial term epxe^{-px} from the vector field, then we can introduce a new variable η2\eta_{2} and impose the equality constraint 0=η2epx0=\eta_{2}-e^{-px}. The result is that the vector field becomes

αη2z1βz22+η1.\alpha\eta_{2}z_{1}-\beta z_{2}^{2}+\eta_{1}.

Since this operation introduces new variables and additional constraints, we incorporate the constraints as extra components in the natural boundary condition function θBC\theta_{BC} of the polynomial embedding. Since this type of operation will introduce an equal number of additional scalar variables and boundary conditions, we will neglect them from the dimension counting.

Remark 4.

If we want to formalize the embedding process for parameters, we can introduce differential equations for them. Indeed, in the example above, we have η˙2=0\dot{\eta}_{2}=0, and this differential equation can be added to the list of differential equations that result from polynomial embeddings of the original state variable zz. In this way, we can understand mm as the total embedding dimension. It should be remarked, however, that in numerical implementation, objects like η2\eta_{2} really are treated as scalar quantities.

The final thing we need to take into account is that every periodic orbit is equivalent to a one-dimensional continuum by way of phase shifts. Since our computer-assisted approach to proving periodic orbits is based on Newton’s method and contraction maps, we need to handle this lack of isolation. This can be done by including a phase condition. In this paper we will make use of an anchor condition; we select a periodic function z^\hat{z} having the same period as zz, and we require that

z(s),z^(s)𝑑s=0.\int\langle z(s),\hat{z}^{\prime}(s)\rangle ds=0.

3.2 Zero-finding problem

We are ready to write down a zero-finding problem for our rescaled periodic orbits. First, combining the work of the previous sections, we must simultaneously solve the equations

{z˙=f~(z(t+μ1),,z(t+μJ),x,a,α,β,η),(delay differential equations)z=1,(amplitude condition of scaled orbit)z(s),z^(s)𝑑s=0,(anchor condition)f(x,,x,α,β)=0,(x is a steady state)θBC(z(0),x,a,α,β,η)=0.(embedding boundary condition)\begin{cases}\dot{z}=\tilde{f}(z(t+\mu_{1}),\dots,z(t+\mu_{J}),x,a,\alpha,\beta,\eta),&\quad\text{(delay differential equations)}\\ \|z\|=1,&\quad\text{(amplitude condition of scaled orbit)}\\ \int\langle z(s),\hat{z}^{\prime}(s)\rangle ds=0,&\quad\text{(anchor condition)}\\ f(x,\dots,x,\alpha,\beta)=0,&\quad\text{($x$ is a steady state)}\\ \theta_{BC}(z(0),x,a,\alpha,\beta,\eta)=0.&\quad\text{(embedding boundary condition)}\end{cases} (20)

At this stage, the period ω\omega of the periodic orbit is implicit. In passing to the spectral representation, we will make it explicit. Define the frequency ψ=2πω\psi=\frac{2\pi}{\omega} and expand zz in Fourier series:

z(t)\displaystyle z(t) =kzkeikψt.\displaystyle=\sum_{k\in\mathbb{Z}}z_{k}e^{ik\psi t}. (21)

Recall that for a real (as opposed to complex-valued) periodic orbit, zkn+mz_{k}\in\mathbb{C}^{n+m} will satisfy the symmetry zk=zk¯z_{k}=\overline{z_{-k}}. To substitute (21) into the differential equation in (20), we must examine how time delays transform under Fourier series. Observe

z(t+μ)=kzkeikψμeikψt,z(t+\mu)=\sum_{k\in\mathbb{Z}}z_{k}e^{ik\psi\mu}e^{ik\psi t},

which means that at the level of Fourier coefficients, a delay of μ\mu corresponds to a complex rotation

zk(ζμ(ψ)z)k=defeikψμzk.\displaystyle z_{k}\mapsto(\zeta_{\mu}(\psi)z)_{k}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,e^{ik\psi\mu}z_{k}. (22)

Note that this operator is linear on Cn+mC^{\mathbb{Z}}_{n+m} and bounded on ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}). To densify our notation somewhat, we define ζ(ψ):ν1(n+m)ν1(n+m)J\zeta(\psi):\ell_{\nu}^{1}(\mathbb{C}^{n+m})\rightarrow\ell_{\nu}^{1}(\mathbb{C}^{n+m})^{J} by

ζ(ψ)z=(ζμ1(ψ)z,,ζμJ(ψ)z).\zeta(\psi)z=(\zeta_{\mu_{1}}(\psi)z,\dots,\zeta_{\mu_{J}}(\psi)z).

As for the derivative, we make use of the operator (Kz)k=kzk(Kz)_{k}=kz_{k}. Since f~\tilde{f} is polynomial, subsituting (21) into the first equation of (20) will result in an equation of the form

ψiKz=𝐟(ζ(ψ)z,x,a,α,β,η)\psi iKz=\mathbf{f}(\zeta(\psi)z,x,a,\alpha,\beta,\eta)

for a function 𝐟:(n+m)J×n××2×m(n+m)\mathbf{f}:(\mathbb{C}^{\mathbb{Z}}_{n+m})^{J}\times\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}^{2}\times\mathbb{R}^{m}\rightarrow(\mathbb{C}^{\mathbb{Z}}_{n+m}) being a (formal) vector polynomial with respect to Fourier convolution, in the arguments (n+m)J(\mathbb{C}^{\mathbb{Z}}_{n+m})^{J}. Observe that we have abused notation and identified the function zz in (21) with its sequence of Fourier coefficients. As an example, the nonlinearity zz(t)2z(t+μ)z\mapsto z(t)^{2}z(t+\mu) is transformed in Fourier to the nonlinearity

z(zz)(ζμ(ψ)z).z\mapsto(z*z)*(\zeta_{\mu}(\psi)z).

Now, let z^\hat{z} be an approximate periodic orbit. We can define new amplitude and phase conditions as functions of zz and the numerical data z^\hat{z}; see [30]. Then, define a map G:X×2UG:X\times\mathbb{R}^{2}\rightarrow U, with X=ν1(n+m)×n×××mX=\ell_{\nu}^{1}(\mathbb{C}^{n+m})\times\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}\times\mathbb{R}^{m}, and U=Kν(n+m)×n×××mU=K_{\nu}(\mathbb{C}^{n+m})\times\mathbb{R}^{n}\times\mathbb{C}\times\mathbb{C}\times\mathbb{C}^{m}, by

G(z,x,a,ψ,η,(α,β))\displaystyle G(z,x,a,\psi,\eta,(\alpha,\beta)) =(ψiKz+𝐟(ζ(ψ)z,x,a,α,β,η)f(x,,x,α,β)z,K2z^1z,iKz^θBC(kzk,x,a,α,β,η)).\displaystyle=\left(\begin{array}[]{c}-\psi iKz+\mathbf{f}(\zeta(\psi)z,x,a,\alpha,\beta,\eta)\\ f(x,\dots,x,\alpha,\beta)\\ \langle z,K^{2}\hat{z}\rangle-1\\ \langle z,iK\hat{z}\rangle\\ \theta_{BC}(\sum_{k}z_{k},x,a,\alpha,\beta,\eta)\end{array}\right). (28)

Let XX be equipped with the norm

(z,x,a,ψ,η)=max{zν,|x|,|a|,|ψ|,|η|},\displaystyle||(z,x,a,\psi,\eta)||=\max\{||z||_{\nu},|x|,|a|,|\psi|,|\eta|\}, (29)

where all Euclidean space norms are selected a priori and could be distinct. That is, we allow for the possibility of a refined weighting111It becomes notationally cubmersome to include references to weights, or to explicitly label the different norms, so we will refrain from doing so unless absolutely necessary. In these cases, footnotes will be used to emphasize any particular subtleties. Weighted max norms can make it easier to obtain computer assisted proofs, especially in continuation problem, which is why we have included this option. of the norms being used. Then XX is a Banach space, and the same is true for UU when equipped with an analogous norm, replacing ||||ν||\cdot||_{\nu} with ||||ν,K||\cdot||_{\nu,K}. If the vector field ff is twice continuously differentiable, then FF is continuously differentiable. This is because the polynomial embedding results in polynomial ordinary differential equations, while the blow-up procedure only results in one derivaive being lost. However, FF is generally not twice continuously differentiable; see later Section 4.3.

It will sometimes be convenient to compute norms on the n+m+4\mathbb{R}^{n+m+4}-projection of XX. In this case, if u=(z,y)Xu=(z,y)\in X and yn+m+4y\in\mathbb{R}^{n+m+4} is represented (ismorphically) as y=(x,a,ψ,η)n×××my=(x,a,\psi,\eta)\in\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}\times\mathbb{R}^{m}, then we define y=max{|x|,|a|,|ψ|,|η|}||y||=\max\{|x|,|a|,|\psi|,|\eta|\}, where any weighting is, again, implicit. Then (z,y)=max{zν,y}||(z,y)||=\max\{||z||_{\nu},||y||\}.

Introduce V=Symm(ν1(n+m))×n×××mV=\mbox{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}\times\mathbb{R}^{m}. Any zero of FF in the space VV uniquely corresponds to a real periodic orbit of (19) by way of the Fourier expansion (21) and the blow-up coordinates y=x+azy=x+az. Moreover, the restriction of FF to VV has range in W=Symm(Kν(n+m))×n×××mW=\operatorname{Symm}(K_{\nu}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}\times\mathbb{R}^{m}. Each of VV and WW are Banach spaces over the reals, and so from this point on we work with the restriction F:VWF:V\rightarrow W.

3.3 Finite-dimensional projection

To set up the rigorous numerics and the continuation, we need to define projections of VV and WW onto suitable finite-dimensional vector spaces. Let πM:ν1(n+m)ν1(n+m)\pi^{M}:\ell_{\nu}^{1}(\mathbb{C}^{n+m})\rightarrow\ell_{\nu}^{1}(\mathbb{C}^{n+m}) denote the projection operator

(πMz)k={zk,|k|M0|k|>M,(\pi^{M}z)_{k}=\left\{\begin{array}[]{ll}z_{k},&|k|\leq M\\ 0&|k|>M,\end{array}\right.

and π=Iν1(n+m)πM\pi^{\infty}=I_{\ell_{\nu}^{1}(\mathbb{C}^{n+m})}-\pi^{M} its complementary projector. Consider the finite-dimensional vector space

VM=defπM(Symm(n+m))×n×××m,V^{M}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\pi^{M}(\operatorname{Symm}(\mathbb{C}^{\mathbb{Z}}_{n+m}))\times\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}\times\mathbb{R}^{m},

and extend the projection to a map πM:VVM\pi^{M}:V\rightarrow V^{M} as follows:

πM(z,x,a,ψ,η)=(πMz,x,a,ψ,η).\pi^{M}(z,x,a,\psi,\eta)=(\pi^{M}z,x,a,\psi,\eta).

Now define a map 𝒢:VM×2VM\mathcal{G}:V^{M}\times\mathbb{R}^{2}\rightarrow V^{M}

𝒢(z,x,a,ψ,η,(α,β))\displaystyle\mathcal{G}(z,x,a,\psi,\eta,(\alpha,\beta)) =πMG(z,x,a,ψ,η,(α,β)).\displaystyle=\pi^{M}G(z,x,a,\psi,\eta,(\alpha,\beta)). (30)

𝒢\mathcal{G} is well-defined and smooth, and we have 𝒢=πMGiVM\mathcal{G}=\pi^{M}G\circ i_{V^{M}}, where iVM:VMVi_{V^{M}}:V^{M}\hookrightarrow V is the natural inclusion map. Therefore, this definition of the projection of GG is consistent with the abstract set-up of Section 2.3.

3.4 A reformulation of the continuation map

It will be convenient to identify V×2V\times\mathbb{R}^{2} and VM×2V^{M}\times\mathbb{R}^{2} respectively with isomorphic spaces

V×2\displaystyle V\times\mathbb{R}^{2} πM(Symm(ν1(n+m)))×n+m+4×π(Symm(ν1(n+m)))=defΩ\displaystyle\sim\pi^{M}(\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})))\times\mathbb{R}^{n+m+4}\times\pi^{\infty}(\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})))\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\Omega
VM×2\displaystyle V^{M}\times\mathbb{R}^{2} πM(Symm(ν1(n+m)))×n+m+4=defΩM\displaystyle\sim\pi^{M}(\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})))\times\mathbb{R}^{n+m+4}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\Omega^{M}

There is then the natural embedding (zM,ρ)(zM,ρ,0)(z^{M},\rho)\mapsto(z^{M},\rho,0) of ΩM\Omega^{M} into Ω\Omega. The isomorphism of VV with Ω\Omega is given by

(z,x,a,ψ,η,(α,β))(πMz,(x,a,ψ,η,α,β),πz).(z,x,a,\psi,\eta,(\alpha,\beta))\mapsto(\pi^{M}z,(x,a,\psi,\eta,\alpha,\beta),\pi^{\infty}z).

ΩM\Omega^{M} is finite-dimensional, and as such our language will sometimes reinforce this by describing matrices whose columns are elements of ΩM\Omega^{M}. This should be understood “up to isomorphism”. The purpose of the isomorphism of V×2V\times\mathbb{R}^{2} with Ω\Omega is to symbolically group all of the finite-dimensional objects together.

Given u^j=(z^j,ρ^j)ΩM\hat{u}_{j}=(\hat{z}_{j},\hat{\rho}_{j})\in\Omega^{M} for j=0,1,2j=0,1,2, let Φ^j\hat{\Phi}_{j} be a matrix whose columns are a basis for the kernel of DFM(z^j,ρ^j)DF^{M}(\hat{z}_{j},\hat{\rho}_{j}), and are therefore elements of VM×2ΩMV^{M}\times\mathbb{R}^{2}\sim\Omega^{M}. For sΔs\in\Delta, let u^s\hat{u}_{s} and Φ^s\hat{\Phi}_{s} be the usual interpolations of the elements u^j\hat{u}_{j} and bases Φ^j\hat{\Phi}_{j} for j=0,1,2j=0,1,2.

The continuation map GsG_{s} of (14) could now be defined for our periodic orbit function GG. However, it will be convenient in our subsequent discussions concerning the radii polynomial approach to re-interpret the codomain of GsG_{s} as being

Ω~=defπMSymm(Kν(n+m))×n+m+4×πSymm(Kν(n+m)).\tilde{\Omega}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\pi^{M}\mbox{Symm}(K_{\nu}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n+m+4}\times\pi^{\infty}\mbox{Symm}(K_{\nu}(\mathbb{C}^{n+m})).

Specifically, this will make it a bit easier to define an approximate inverse of DGs(u^s)DG_{s}(\hat{u}_{s}). The codomain of GsG_{s} is

W×2\displaystyle W\times\mathbb{R}^{2} =(Symm(Kν(n+m))×n×××m)×2\displaystyle=(\mbox{Symm}(K_{\nu}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n}\times\mathbb{R}\times\mathbb{R}\times\mathbb{R}^{m})\times\mathbb{R}^{2}
Symm(Kν(n+m))×n+m+4\displaystyle\sim\mbox{Symm}(K_{\nu}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n+m+4}
Ω~,\displaystyle\sim\tilde{\Omega},

where the isomorphisms can be realized by permuting the relevant components of GsG_{s} and splitting the Fourier space into direct sums. For (zM,ρ,z)Ω(z^{M},\rho,z^{\infty})\in\Omega, a suitable isomorphic representation of GsG_{s} is given by s:ΩΩ~\mathcal{F}_{s}:\Omega\rightarrow\tilde{\Omega},

s(zM,ρ,z)\displaystyle\mathcal{F}_{s}(z^{M},\rho,z^{\infty}) =(ψiKzM+πM𝐟(ζ(ψ)(zM+z),ρ)𝐉s(zM,ρ,z)ψiKz+π𝐟(ζ(ψ)(zM+z),ρ))=def(s(1)s(2)s(3)),\displaystyle=\left(\begin{array}[]{c}-\psi iKz^{M}+\pi^{M}\mathbf{f}(\zeta(\psi)(z^{M}+z^{\infty}),\rho)\\ \mathbf{J}_{s}(z^{M},\rho,z^{\infty})\\ -\psi iKz^{\infty}+\pi^{\infty}\mathbf{f}(\zeta(\psi)(z^{M}+z^{\infty}),\rho)\end{array}\right)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\left(\begin{array}[]{c}{\mathcal{F}}_{s}^{(1)}\\ {\mathcal{F}}_{s}^{(2)}\\ {\mathcal{F}}_{s}^{(3)}\end{array}\right), (37)
𝐉s(zM,ρ,z)\displaystyle\mathbf{J}_{s}(z^{M},\rho,z^{\infty}) =(f(x,,x,α,β)zM,K2u^s1zM,iKu^sθBC(k(zkM+zk),x,a,α,β,η)Φ^s,1,uMc^s,1Φ^s,2,uMc^s,2),uM=(zMρ)ΩM,\displaystyle=\left(\begin{array}[]{c}f(x,\dots,x,\alpha,\beta)\\ \langle z^{M},K^{2}\hat{u}_{s}\rangle-1\\ \langle z^{M},iK\hat{u}_{s}\rangle\\ \theta_{BC}(\sum_{k}(z^{M}_{k}+z^{\infty}_{k}),x,a,\alpha,\beta,\eta)\\ \langle\hat{\Phi}_{s,1},u^{M}\rangle-\hat{c}_{s,1}\\ \langle\hat{\Phi}_{s,2},u^{M}\rangle-\hat{c}_{s,2}\end{array}\right),\hskip 5.69054ptu^{M}=\left(\begin{array}[]{c}z^{M}\\ \rho\end{array}\right)\in\Omega^{M}, (46)

where ρ=(x,a,ψ,η,α,β)\rho=(x,a,\psi,\eta,\alpha,\beta), and the c^s,j\hat{c}_{s,j} for j=1,2j=1,2 are interpolations

c^s,j=Φ^0,j,u^0+s1(Φ^1,j,u^1Φ^0,j,u^0)+s2(Φ^2,j,u^2Φ^0,j,u^0).\displaystyle\hat{c}_{s,j}=\langle\hat{\Phi}_{0,j},\hat{u}_{0}\rangle+s_{1}(\langle\hat{\Phi}_{1,j},\hat{u}_{1}\rangle-\langle\hat{\Phi}_{0,j},\hat{u}_{0}\rangle)+s_{2}(\langle\hat{\Phi}_{2,j},\hat{u}_{2}\rangle-\langle\hat{\Phi}_{0,j},\hat{u}_{0}\rangle). (47)
Remark 5.

Strictly speaking, the final two components of 𝐉s\mathbf{J}_{s} are not the “standard” ones from (14). The inner product Φs,j,π𝒳xx^s\langle\Phi_{s,j},\pi_{\mathcal{X}}x-\hat{x}_{s}\rangle in the latter results in quadratic terms in ss, while the ones in (46) are ss-linear. The impact of this change is that, theoretically, the Newton correction is not strictly in the direction orthogonal to the interpolation of the tangent planes. This change has no theoretical bearing on the validated continuation, and is done only foe ease of computation: it is easier to compute derivatives of linear functions than nonlinear ones.

Remark 6.

We have abused notation somewhat, since now we interpret 𝐟\mathbf{f} as a map

𝐟:Symm(ν1(n+m))×n+m+4Symm(Kν(n+m)).\mathbf{f}:\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n+m+4}\rightarrow\operatorname{Symm}(K_{\nu}(\mathbb{C}^{n+m})).

It acts trivially with respect to the variable ψ\psi. Also, we emphasize that GsG_{s} depends on the numerical interpolants u^s\hat{u}_{s} and Φ^s\hat{\Phi}_{s}.

Since Theorem 1 is stated with respect to general Banach spaces, the validated continuation approach applies equally to the representation s:ΩΩ~\mathcal{F}_{s}:\Omega\rightarrow\tilde{\Omega} of Gs:VWG_{s}:V\rightarrow W. The only thing we need to do is specify a compatible norm on Ω\Omega. This is straightforward: for (zM,ρ,z)Ω(z^{M},\rho,z^{\infty})\in\Omega and ρ=(x,a,ψ,η,α,β)\rho=(x,a,\psi,\eta,\alpha,\beta), a suitable norm is

(zM,ρ,z)Ω=max{zM+zν,|x|,|a|,|ψ|,|η|,|α|,|β|},||(z^{M},\rho,z^{\infty})||_{\Omega}=\max\{||z^{M}+z^{\infty}||_{\nu},|x|,|a|,|\psi|,|\eta|,|\alpha|,|\beta|\},

where the Euclidean norms on the components of ρ\rho are the same222Including any weighting, which we recall is implicit. as the ones appearing in (29). With this choice, ||||Ω||\cdot||_{\Omega} is equivalent to the induced max norm on X×2X\times\mathbb{R}^{2}, with XX equipped with the norm (29) and 2\mathbb{R}^{2} the \infty-norm.

3.5 Construction of AsA_{s}^{\dagger} and AsA_{s}

Write u^s=(z^s,x^s,ψ^s,a^s,α^s,β^s)\hat{u}_{s}=(\hat{z}_{s},\hat{x}_{s},\hat{\psi}_{s},\hat{a}_{s},\hat{\alpha}_{s},\hat{\beta}_{s}), for sΔs\in\Delta. Denote the three vertices of Δ\Delta as s0=(0,0)s_{0}=(0,0), s1=(1,0)s_{1}=(1,0) and s2=(0,1)s_{2}=(0,1). Introduce an approximation of s\mathcal{F}_{s} as follows:

~s(zM,ρ,z)\displaystyle\tilde{\mathcal{F}}_{s}(z^{M},\rho,z^{\infty}) =(ψiKzM+πM𝐟(ζ(ψ)zM,ρ)𝐉s(zM,ρ,z)ψiKz)=def(~s(1)~s(2)~s(3)).\displaystyle=\left(\begin{array}[]{c}-\psi iKz^{M}+\pi^{M}\mathbf{f}(\zeta(\psi)z^{M},\rho)\\ \mathbf{J}_{s}(z^{M},\rho,z^{\infty})\\ -\psi iKz^{\infty}\end{array}\right)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\left(\begin{array}[]{c}\tilde{\mathcal{F}}_{s}^{(1)}\\ \tilde{\mathcal{F}}_{s}^{(2)}\\ \tilde{\mathcal{F}}_{s}^{(3)}\end{array}\right). (54)

Formally, ~s\tilde{\mathcal{F}}_{s} approximates s\mathcal{F}_{s} in the Fourier tail by neglecting the nonlinear terms, leaving only the part coming from the differentiation operator.

Proposition 2.

D~s(u^s)D\tilde{\mathcal{F}}_{s}(\hat{u}_{s}) has the representation

D~s(u^s)\displaystyle D\tilde{\mathcal{F}}_{s}(\hat{u}_{s}) =(D1~s(1)(u^s)D2~s(1)(u^s)0D1~s(2)(u^s)D2~s(2)(u^s)D3~s(2)(u^s)00D3~s(3)(u^s))=defAs.\displaystyle=\left(\begin{array}[]{ccc}D_{1}\tilde{\mathcal{F}}_{s}^{(1)}(\hat{u}_{s})&D_{2}\tilde{\mathcal{F}}_{s}^{(1)}(\hat{u}_{s})&0\\ D_{1}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})&D_{2}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})&D_{3}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})\\ 0&0&D_{3}\tilde{\mathcal{F}}_{s}^{(3)}(\hat{u}_{s})\end{array}\right)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,A_{s}^{\dagger}. (58)
Proof.

Since s(1)\mathcal{F}_{s}^{(1)} does not depend on zz^{\infty}, the upper-right block of DS(u^s)D\mathcal{F}_{S}(\hat{u}_{s}) is the zero map. Similarly, s(3)\mathcal{F}_{s}^{(3)} does not depend on either zMz^{M} or ρ\rho, so the finite-dimensional blocks in the zz^{\infty} (bottom) row are zero. ∎

The upper left 2×22\times 2 block is equivalent to a finite-dimensional matrix operator. In particular, D2~s(2)(u^s)D_{2}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s}) is real. Also,

D3~s(3)(u^s)=ψ^siKπD_{3}\tilde{\mathcal{F}}_{s}^{(3)}(\hat{u}_{s})=-\hat{\psi}_{s}iK\pi^{\infty}

is invertible, with

D3~s(3)(u^s)1=i(ψ^s)1(Kπ)1,D_{3}\tilde{\mathcal{F}}_{s}^{(3)}(\hat{u}_{s})^{-1}=i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1},

where ((Kπ)1z)k=def1kzk((K\pi^{\infty})^{-1}z)_{k}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\frac{1}{k}z_{k} for |k|M+1|k|\geq M+1. Suppose we can explicitly compute Sj(m+n+4)×(m+n+4)S_{j}\in\mathbb{R}^{(m+n+4)\times(m+n+4)} and Pj,Qj,RjP_{j},Q_{j},R_{j} complex matrices for j=0,1,2j=0,1,2 such that

(PjQjRjSj)(D1~sj(1)(u^sj)D2~sj(1)(u^sj)D1~sj(2)(u^sj)D2~sj(2)(u^sj))IΩM.\displaystyle\left(\begin{array}[]{cc}P_{j}&Q_{j}\\ R_{j}&S_{j}\end{array}\right)\left(\begin{array}[]{cc}D_{1}\tilde{\mathcal{F}}_{s_{j}}^{(1)}(\hat{u}_{s_{j}})&D_{2}\tilde{\mathcal{F}}_{s_{j}}^{(1)}(\hat{u}_{s_{j}})\\ D_{1}\tilde{\mathcal{F}}_{s_{j}}^{(2)}(\hat{u}_{s_{j}})&D_{2}\tilde{\mathcal{F}}_{s_{j}}^{(2)}(\hat{u}_{s_{j}})\end{array}\right)\approx I_{\Omega^{M}}. (63)

We can then prove the following lemma.

Lemma 3.

For s=(s(1),s(2))Δs=(s^{(1)},s^{(2)})\in\Delta, define matrix interpolants Ps=P1+s(1)(P2P1)+s(2)(P3P1)P_{s}=P_{1}+s^{(1)}(P_{2}-P_{1})+s^{(2)}(P_{3}-P_{1}), and analogously define interpolants QsQ_{s}, RsR_{s} and SsS_{s}. Introduce a family of operators AsA_{s} as follows:

As\displaystyle A_{s} =(PsQsQsD3~s(2)(u^s)i(ψ^s)1(Kπ)1RsSsSsD3~s(2)(u^s)i(ψ^s)1(Kπ)100i(ψ^s)1(Kπ)1).\displaystyle=\left(\begin{array}[]{ccc}P_{s}&Q_{s}&-Q_{s}D_{3}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1}\\ R_{s}&S_{s}&-S_{s}D_{3}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1}\\ 0&0&i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1}\end{array}\right). (67)

Suppose for j=1,2,3j=1,2,3, SjS_{j} is real and, as maps, PjP_{j}, QjQ_{j} and RjR_{j} are equivalent to

Pj\displaystyle P_{j} :πM(Symm(n+m))πM(Symm(n+m))\displaystyle:\pi^{M}(\operatorname{Symm}(\mathbb{C}^{\mathbb{Z}}_{n+m}))\rightarrow\pi^{M}(\operatorname{Symm}(\mathbb{C}^{\mathbb{Z}}_{n+m}))
Qj\displaystyle Q_{j} :m+n+4πM(Symm(n+m))\displaystyle:\mathbb{R}^{m+n+4}\rightarrow\pi^{M}(\operatorname{Symm}(\mathbb{C}^{\mathbb{Z}}_{n+m}))
Rj\displaystyle R_{j} :πM(Symm(n+m))n+m+4.\displaystyle:\pi^{M}(\operatorname{Symm}(\mathbb{C}^{\mathbb{Z}}_{n+m}))\rightarrow\mathbb{R}^{n+m+4}.

Then As:Ω~ΩA_{s}:\tilde{\Omega}\rightarrow\Omega is well-defined.

Proof.

One can show As:Ω~ΩA_{s}:\tilde{\Omega}\rightarrow\Omega is well-defined using the fact that each of PsP_{s}, QsQ_{s} and RsR_{s} is a real convex combination of maps to/from appropriate symmetric spaces, and noticing that iKπiK\pi^{\infty} (and hence its inverse) satisfy the symmetry (iKπz)k=(iKπz)k¯(iK\pi^{\infty}z)_{k}=\overline{(iK\pi^{\infty}z)_{-k}}

The point here is that, due to (63), we have

AsjD~sj(u^sj)1(Asj)1A_{s_{j}}\approx D\tilde{\mathcal{F}}_{s_{j}}(\hat{u}_{s_{j}})^{-1}\approx(A^{\dagger}_{s_{j}})^{-1}

for j=0,1,2j=0,1,2, and if the interpolation points u^j\hat{u}_{j} are close together, we expect D~s(u^sj)D\tilde{\mathcal{F}}_{s}(\hat{u}_{s_{j}}) to be invertible for all sΔs\in\Delta, and D~s(u^sj)1AsD\tilde{\mathcal{F}}_{s}(\hat{u}_{s_{j}})^{-1}\approx A_{s}.

Remark 7.

Checking the conditions of Lemma 3 amounts to verifying conjugate symmetries of the matrices PjP_{j}, QjQ_{j} and RjR_{j}. Numerical rounding makes this a nontrivial task, so we generally post-process the numerically computed matrices to impose these symmetry conditions.

4 Technical bounds for validated continuation of periodic orbits

Based on the previous section, we define a Newton-like operator Ts:ΩΩT_{s}:\Omega\rightarrow\Omega

Ts(u)=uAss(u),\displaystyle T_{s}(u)=u-A_{s}\mathcal{F}_{s}(u), (68)

for sΔs\in\Delta. As described in Section 2.4, our goal is to prove that TsT_{s} is a uniform (for sΔs\in\Delta) contraction in a closed ball centered at u^s\hat{u}_{s} using Theorem 1. If AsA_{s} can be proven (uniformly in ss) injective, this will prove the existence of a unique zero of s()\mathcal{F}_{s}(\cdot) close to u^s\hat{u}_{s}, thereby validating the simplex and proving the smooth patch of our manifold.

In Section 4.1, we will demonstrate how the bound Z0Z_{0} of the radii polynomial approach can be used to obtain a proof of uniform (in ss) injectivity of the operator AsA_{s}. We then provide some detailed discussion concerning general-purpose implementation of the bounds YY and ZZ in Section 4.2 through to Section 4.5.

4.1 Injectivity of AsA_{s}

The injectivity of AsA_{s} is a consequence of the successful identification of bounds Z0Z_{0} of Theorem 1 and the negativity of the radii polynomial. In particular,

Lemma 4.

Suppose IAsAsB(Ω,Ω)Z0||I-A_{s}A_{s}^{\dagger}||_{B(\Omega,\Omega)}\leq Z_{0} for all sΔs\in\Delta, with the operators AsA_{s} and AsA_{s}^{\dagger} of Section 3.5. If Z0<1Z_{0}<1, then AsA_{s} is injective for sΔs\in\Delta.

Proof.

First, observe AsA_{s} has non-trivial kernel if and only if there exists uΩMu\in\Omega^{M} such that Asu=0A_{s}u=0. This is a consequence of the structure of the operator and injectivity of (Kπ)1(K\pi^{\infty})^{-1}. Therefore, it suffices to verify the injectivity of the restriction 𝐀s=As|ΩM\mathbf{A}_{s}=A_{s}|_{\Omega^{M}}. Define 𝐀s=As|ΩM\mathbf{A}_{s}^{\dagger}=A_{s}^{\dagger}|_{\Omega^{M}}. By definition of the norm on Ω\Omega, we have I𝐀s𝐀sB(ΩM,ΩM)Z0<1||I-\mathbf{A}_{s}\mathbf{A}_{s}^{\dagger}||_{B(\Omega^{M},\Omega^{M})}\leq Z_{0}<1 for all sΔs\in\Delta. By Neumann series, it follows that 𝐀s𝐀s\mathbf{A}_{s}\mathbf{A}_{s}^{\dagger} is boundedly invertible, which implies 𝐀s\mathbf{A}_{s} is surjective. Since ΩM\Omega^{M} is finite-dimensional, 𝐀s\mathbf{A}_{s} is also injective. ∎

Corollary 5.

If the radii polynomial satisfies p(r0)<0p(r_{0})<0 for some r0>0r_{0}>0, then AsA_{s} is injective for sΔs\in\Delta.

4.2 The bound Y0Y_{0}

To begin, expand the product Ass(u^s)A_{s}\mathcal{F}_{s}(\hat{u}_{s}). We get

Ass(u^s)\displaystyle A_{s}\mathcal{F}_{s}(\hat{u}_{s}) =(Pss(1)(u^s)+Qss(2)(u^s)QsD3~s(2)(u^s)i(ψ^s)1(Kπ)1s(3)(u^s)Rss(1)(u^s)+Sss(2)(u^s)SsD3~s(2)(u^s)i(ψ^s)1(Kπ)1s(3)(u^s)i(ψ^s)1(Kπ)1s(3)(u^s))\displaystyle=\left(\begin{array}[]{c}P_{s}\mathcal{F}_{s}^{(1)}(\hat{u}_{s})+Q_{s}\mathcal{F}_{s}^{(2)}(\hat{u}_{s})-Q_{s}D_{3}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1}\mathcal{F}_{s}^{(3)}(\hat{u}_{s})\\ R_{s}\mathcal{F}_{s}^{(1)}(\hat{u}_{s})+S_{s}\mathcal{F}_{s}^{(2)}(\hat{u}_{s})-S_{s}D_{3}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1}\mathcal{F}_{s}^{(3)}(\hat{u}_{s})\\ i(\hat{\psi}_{s})^{-1}(K\pi^{\infty})^{-1}\mathcal{F}_{s}^{(3)}(\hat{u}_{s})\\ \end{array}\right)

Remark that s(3)(u^s)\mathcal{F}_{s}^{(3)}(\hat{u}_{s}) has range in a finite-dimensional subspace of Ω\Omega; specifically, it will be in the part of Ω\Omega such that the components in n+m\mathbb{C}^{\mathbb{Z}}_{n+m} with index (in absolute value) greater than Md+1Md+1 are zero, where dd is the maximum degree of the (convolution) polynomial 𝐟\mathbf{f}. As such, Ass(u^s)A_{s}\mathcal{F}_{s}(\hat{u}_{s}) is explicitly computable.

In practice, we must compute an enclosure of the norm Ass(u^s)||A_{s}\mathcal{F}_{s}(\hat{u}_{s})|| for all sΔs\in\Delta. This is slightly less trivial. We accomplish this using a first-order Taylor expansion with remainder. For the function (s,u)s(u)(s,u)\mapsto\mathcal{F}_{s}(u), denote by s\partial_{s} the Fréchet derivative with respect to ss, and DD the derivative with respect to uu. Given the interpolants u^s=(v^s,ρ^s)\hat{u}_{s}=(\hat{v}_{s},\hat{\rho}_{s}) and Φ^s,j\hat{\Phi}_{s,j}, let u^\hat{u}^{\prime} and Φ^j\hat{\Phi}_{j}^{\prime} denote their Fréchet derivatives with respect to ss. Also, denote c^s,j=sc^s,j\hat{c}_{s,j}^{\prime}=\partial_{s}\hat{c}_{s,j}, for c^s,j\hat{c}_{s,j} defined in (47). Note that these derivatives are constant, since the interpolants are linear in ss. Then

s(u^s)\displaystyle\mathcal{F}_{s}(\hat{u}_{s}) =s0(u^0)+(ss0(u^0)+Ds0(u^0)u^)s+12,\displaystyle=\mathcal{F}_{s_{0}}(\hat{u}_{0})+\big{(}\partial_{s}\mathcal{F}_{s_{0}}(\hat{u}_{0})+D\mathcal{F}_{s_{0}}(\hat{u}_{0})\hat{u}^{\prime}\big{)}s+\frac{1}{2}\mathcal{R},

where the remainder term \mathcal{R} will be elaborated upon momentarily. The product of the linear-order terms with AsA_{s} results in a function that is componentwise quadratic in sΔs\in\Delta, for which we can efficiently compute an upper bound on the norm. The derivative Ds(u^s)D\mathcal{F}_{s}(\hat{u}_{s}) is implementable, and will be further discussed in Section 4.3. As for the s\partial_{s} term,

ss(u)\displaystyle\partial_{s}\mathcal{F}_{s}(u) =(0s𝐉s(u)0),s𝐉s(u)=(0z,K2z^z,iKz^0Φ^1,uc^s,1Φ^2,uc^s,2),\displaystyle=\left(\begin{array}[]{c}0\\ \partial_{s}\mathbf{J}_{s}(u)\\ 0\end{array}\right),\hskip 5.69054pt\partial_{s}\mathbf{J}_{s}(u)=\left(\begin{array}[]{c}0\\ \langle z,K^{2}\hat{z}^{\prime}\rangle\\ \langle z,iK\hat{z}^{\prime}\rangle\\ 0\\ \langle\hat{\Phi}_{1}^{\prime},u\rangle-\hat{c}_{s,1}^{\prime}\\ \langle\hat{\Phi}_{2}^{\prime},u\rangle-\hat{c}_{s,2}^{\prime}\end{array}\right),

where u=(z,ρ)u=(z,\rho) and u^s=(z^s,ρ^s)\hat{u}_{s}=(\hat{z}_{s},\hat{\rho}_{s}).

As for the remainder \mathcal{R}, it is bounded by the norm of the second Fréchet derivative of ss(u^s)s\mapsto\mathcal{F}_{s}(\hat{u}_{s}), uniformly over sΔs\in\Delta. For each sΔs\in\Delta, let this second derivative be denoted 𝐃2s(u^s)\mathbf{D}^{2}\mathcal{F}_{s}(\hat{u}_{s}). Then

𝐃2s(u^s)\displaystyle\mathbf{D}^{2}\mathcal{F}_{s}(\hat{u}_{s}) =s2s(u^s)[e1,e2]+sDs(u^s)[u^e1,e2]+Dss(u^s)[e1,u^e2]\displaystyle=\partial_{s}^{2}\mathcal{F}_{s}(\hat{u}_{s})[e_{1},e_{2}]+\partial_{s}D\mathcal{F}_{s}(\hat{u}_{s})[\hat{u}^{\prime}e_{1},e_{2}]+D\partial_{s}\mathcal{F}_{s}(\hat{u}_{s})[e_{1},\hat{u}^{\prime}e_{2}] (69)
+D2s(u^s)[u^e1,u^e2].\displaystyle\phantom{=}+D^{2}\mathcal{F}_{s}(\hat{u}_{s})[\hat{u}^{\prime}e_{1},\hat{u}^{\prime}e_{2}].

where e1e_{1} and e2e_{2} denote the first and second (factor) projection maps on 2×2\mathbb{R}^{2}\times\mathbb{R}^{2}. The derivative D2sD^{2}\mathcal{F}_{s} will be discussed in Section 4.3. At this stage, we need only mention that it acts bilinearly on u^\hat{u}, and the latter is proportional in norm to the step size σ\sigma of the continuation scheme, so the D2sD^{2}\mathcal{F}_{s} term will be order σ2\sigma^{2}. As for the derivatives involving s\partial_{s}, most of the components are zero as evidenced by the previous calculation of ss(z,ρ)\partial_{s}\mathcal{F}_{s}(z,\rho), and it suffices to compute the relevant derivatives of 𝐉s\mathbf{J}_{s}. We have

s2𝐉s(u^s)[t1,t2]\displaystyle\partial_{s}^{2}\mathbf{J}_{s}(\hat{u}_{s})[t_{1},t_{2}] =0,Ds𝐉s(u^s)[t,h]=(0hz,K2z^thz,iKz^t0Φ^1t,hΦ^2t,h),\displaystyle=0,\hskip 5.69054ptD\partial_{s}\mathbf{J}_{s}(\hat{u}_{s})[t,h]=\left(\begin{array}[]{c}0\\ \langle h_{z},K^{2}\hat{z}^{\prime}t\rangle\\ \langle h_{z},iK\hat{z}^{\prime}t\rangle\\ 0\\ \langle\hat{\Phi}_{1}^{\prime}t,h\rangle\\ \langle\hat{\Phi}_{2}^{\prime}t,h\rangle\end{array}\right),

with (t,h)2×Ω(t,h)\in\mathbb{R}^{2}\times\Omega, h=(hz,hρ)h=(h_{z},h_{\rho}). In terms of the step size, s2𝐉s(u^s)\partial_{s}^{2}\mathbf{J}_{s}(\hat{u}_{s}) is order σ2\sigma^{2}, while Dss(u^s)D\partial_{s}\mathcal{F}_{s}(\hat{u}_{s}) is order σ\sigma. However, in (69), the latter term is multiplied by z^=O(σ)\hat{z}^{\prime}=O(\sigma). Therefore, as expected, the remainder \mathcal{R} is quadratic with respect to step size. We therefore compute

Y0As(s(u^s)21)+12As.Y_{0}\geq||A_{s}(\mathcal{F}_{s}(\hat{u}_{s})-2^{-1}\mathcal{R})||+\frac{1}{2}||A_{s}\mathcal{R}||.

Since R=O(σ2)||R||=O(\sigma^{2}), the bound can be tempered quadratically by reducing the step size. The caveat is that if As||A_{s}|| and/or s\mathcal{F}_{s} has large quadratic terms, it might still be necessary to take small steps.

Remark 8.

Directly computing the norm As||A_{s}\mathcal{R}|| would require a general-purpose implementation of the second derivative D2s(u^s)D^{2}\mathcal{F}_{s}(\hat{u}_{s}); see (69). As we have stated previously, such an implementation would be rather complicated. Therefore, in practice, we perform another level of splitting; namely, we use the bound

AsAs(𝐃2s(u^s)D2s(u^s)[u^e1,u^e2])+AsD2s(u^s)[u^e1,u^e2].||A_{s}\mathcal{R}||\leq||A_{s}\left(\mathbf{D}^{2}\mathcal{F}_{s}(\hat{u}_{s})-D^{2}\mathcal{F}_{s}(\hat{u}_{s})[\hat{u}^{\prime}e_{1},\hat{u}^{\prime}e_{2}]\right)||+||A_{s}D^{2}\mathcal{F}_{s}(\hat{u}_{s})[\hat{u}^{\prime}e_{1},\hat{u}^{\prime}e_{2}]||.

The first term on the right of the inequality is explicitly implementable using only the derivatives of 𝐉s\mathbf{J}_{s} and the finite blocks of AsA_{s}. For the second term, we use the bound AsD2s(u^s)[u^e1,u^e2]||A_{s}||\cdot||D^{2}\mathcal{F}_{s}(\hat{u}_{s})[\hat{u}^{\prime}e_{1},\hat{u}^{\prime}e_{2}]|| which, while not optimal, is implementable and good enough for our purposes.

4.2.1 Adaptive refinement

In continuation, the size of the Y0Y_{0} bound is severely limited by the step size. To distribute computations, we often want to compute the manifold first and then validate patches a posteriori. However, once the manifold has been computed, adjusting and re-computing patches of the manifold with smaller step sizes becomes complicated due to the need to ensure that cobordant simplices have matched data, as discussed in Section 2.5. When the Y0Y_{0} bound is too large due to interval arithmetic over-estimation, we can circumvent this by using adaptive refinement on the relevant simplex. Formally, we subdivide the simplex into four, using the interpolated zeroes at the nodes of the original simplex to define the data at the nodes of the four new ones. The result is that cobordant data still matches, allowing for globalization of the manifold. The advantage of this approach is that it can be safely done in a distributed manner; adaptively refining one simplex does not require re-validating any of its neighbours.

4.2.2 The bound Z0Z_{0}

The product AsAsA_{s}A_{s}^{\dagger} is block diagonal. Indeed, (IAsAs)|πΩ=0(I-A_{s}A_{s}^{\dagger})|_{\pi^{\infty}\Omega}=0, whereas

(IAsAs)|ΩM=IΩM(PsQsRsSs)(D1~s(1)(u^s)D2~s(1)(u^s)D1~s(2)(u^s)D2~s(2)(u^s)).\displaystyle(I-A_{s}A_{s}^{\dagger})|_{\Omega^{M}}=I_{\Omega^{M}}-\left(\begin{array}[]{cc}P_{s}&Q_{s}\\ R_{s}&S_{s}\end{array}\right)\left(\begin{array}[]{cc}D_{1}\tilde{\mathcal{F}}_{s}^{(1)}(\hat{u}_{s})&D_{2}\tilde{\mathcal{F}}_{s}^{(1)}(\hat{u}_{s})\\ D_{1}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})&D_{2}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s})\end{array}\right).

Therefore, to compute Z0Z_{0} it suffices to find an upper bound, uniformly in ss, for the norm of the above expression as a linear map from ΩMΩM\Omega^{M}\rightarrow\Omega^{M}. Interpreted as matrix operators, PsP_{s}, QsQ_{s}, RsR_{s} and SsS_{s} are interpolants of other explicit matrices. However, the derivatives Di~s(j)(u^s)D_{i}\tilde{\mathcal{F}}_{s}^{(j)}(\hat{u}_{s}), while evaluated at interpolants, are themselves “nonlinear” in ss.

Remark 9.

The implementation of ||(IAsAs)|ΩM||||(I-A_{s}A_{s}^{\dagger})|_{\Omega^{M}}|| is influenced by the way in which the dependence on ss is handled. For example, ss can be treated as a vector interval and the norm can be computed “in one step” using interval arithmeic, then we take the interval supremum to get a bound. This can result in some wrapping (over-estimation). One way to control the wrapping is to cover Δ\Delta in a mesh of balls, compute the norm ||(IAsAs)|ΩM||||(I-A_{s}A_{s}^{\dagger})|_{\Omega^{M}}|| for ss replaced with each of these interval balls, and take the maximum. Still another way is to carefully compute a Taylor expansion with respect to ss, although this task has a few technical issues due to the fact that s()\mathcal{F}_{s}(\cdot) is generally only C1C^{1}. We therefore only consider the “in one step” approach in our implementation.

4.3 A detour: derivatives of convolution-type delay polynomials

In our implementation of the Z1Z_{1} and Z2Z_{2} bounds, we will need to represent various partial derivatives of the map (z,ρ)𝐟(ζ(ψ)z,ρ)(z,\rho)\mapsto\mathbf{f}(\zeta(\psi)z,\rho), where ρ=(x,a,ψ,η,α,β)\rho=(x,a,\psi,\eta,\alpha,\beta). This can quickly become notationally cumbersome. Therefore, in this section we will elaborate on the structure of the derivatives of convolution terms of the following Kν()K_{\nu}(\mathbb{C})-valued map:

Θ(z,ρ)=defρ𝐦p=1d(eiψ(ρ)μjpKzcp)\displaystyle\Theta(z,\rho)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\rho^{\mathbf{m}}\prod_{p=1}^{d}\left(e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}}\right) (70)

for zν1(n+m)z\in\ell_{\nu}^{1}(\mathbb{C}^{n+m}) and ρm+n+4\rho\in\mathbb{R}^{m+n+4}. That is Θ:ν1(n+m)×n+m+4Kν()\Theta:\ell_{\nu}^{1}(\mathbb{C}^{n+m})\times\mathbb{R}^{n+m+4}\rightarrow K_{\nu}(\mathbb{C}). Here, ψ(ρ)\psi(\rho) denotes the frequency component of ρ\rho. In this section, zz will be indexed with the convention z=(z1,,zn+m)z=(z_{1},\dots,z_{n+m}), where zqν1()z_{q}\in\ell_{\nu}^{1}(\mathbb{C}) for q=1,,n+mq=1,\dots,n+m. The objects (70) can be interpreted as individual terms of 𝐟1\mathbf{f}_{1} through 𝐟n+m\mathbf{f}_{n+m}, which each have codomain Kν()K_{\nu}(\mathbb{C}). The product symbol indicates iterated convolution, while (eiψμKz)k=defeiψμkzk(e^{i\psi\mu K}z)_{k}\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,e^{i\psi\mu k}z_{k}. Here, dd\in\mathbb{N} is the (polynomial power) order of the term, the indices cp{1,,n+m}c_{p}\in\{1,\dots,n+m\} specify which factors of ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) are involved in the multiplication, while jp{1,,J}j_{p}\in\{1,\dots,J\} indicates which delays are associated to each of them. Finally, there is a multi-index 𝐦\mathbf{m} for multi-index power ρ𝐦\rho^{\mathbf{m}}. Importantly, the multi-index 𝐦\mathbf{m} is trivial in the frequency (ψ\psi) component, and the latter only enters 𝐟\mathbf{f} in the form of the delay mapping ζ(ψ)\zeta(\psi). The following lemma can then be proven by means of some rather tedious bookkeeping and the commutativity of the convolution product.

Lemma 6.

If q{1,,n+m}q\in\{1,\dots,n+m\}, then for zν1(n+m)z\in\ell_{\nu}^{1}(\mathbb{C}^{n+m}) and hν1()h\in\ell_{\nu}^{1}(\mathbb{C}),

ddzqΘ(z,ρ)h\displaystyle\frac{d}{dz_{q}}\Theta(z,\rho)h =ρ𝐦r=1cr=qd(eiψ(ρ)μjrKh)p=1prdeiψ(ρ)μjpKzcp\displaystyle=\rho^{\mathbf{m}}\sum_{\begin{subarray}{c}r=1\\ c_{r}=q\end{subarray}}^{d}(e^{i\psi(\rho)\mu_{j_{r}}K}h)*\prod_{\begin{subarray}{c}p=1\\ p\neq r\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}}
ddψ(ρ)Θ(z,ρ)\displaystyle\frac{d}{d\psi(\rho)}\Theta(z,\rho) =ρ𝐦r=1d(p=1prdeiψ(ρ)μjpKzcp)(iKμjreiψ(ρ)μjrKzcr)\displaystyle=\rho^{\mathbf{m}}\sum_{r=1}^{d}\left(\prod_{\begin{subarray}{c}p=1\\ p\neq r\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}}\right)*\left(iK\mu_{j_{r}}e^{i\psi(\rho)\mu_{j_{r}}K}z_{c_{r}}\right)
ddzq[ddψ(ρ)Θ(z,ρ)]h\displaystyle\frac{d}{dz_{q}}\left[\frac{d}{d\psi(\rho)}\Theta(z,\rho)\right]h =ρ𝐦r=1cr=qd(iKμjreiψ(ρ)μjrKh)p=1prdeiψ(ρ)μjpKzcp\displaystyle=\rho^{\mathbf{m}}\sum_{\begin{subarray}{c}r=1\\ c_{r}=q\end{subarray}}^{d}(iK\mu_{j_{r}}e^{i\psi(\rho)\mu_{j_{r}}K}h)*\prod_{\begin{subarray}{c}p=1\\ p\neq r\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}}
+ρ𝐦r=1cr=qd(eiψ(ρ)μjrKh)p=1prd(ξ=1ξr,pdeiψ(ρ)μjξKzcξ)(iKμjpeiψ(ρ)μjpKzcp)\displaystyle\phantom{=}+\rho^{\mathbf{m}}\sum_{\begin{subarray}{c}r=1\\ c_{r}=q\end{subarray}}^{d}(e^{i\psi(\rho)\mu_{j_{r}}K}h)*\sum_{\begin{subarray}{c}p=1\\ p\neq r\end{subarray}}^{d}\left(\prod_{\begin{subarray}{c}\xi=1\\ \xi\neq r,p\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{\xi}}K}z_{c_{\xi}}\right)*(iK\mu_{j_{p}}e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}})

Also, ddψ(ρ)[ddzqΘ(z,p)h]=ddzq[ddψ(ρ)Θ(z,ρ)]h\frac{d}{d\psi(\rho)}\left[\frac{d}{dz_{q}}\Theta(z,p)h\right]=\frac{d}{dz_{q}}\left[\frac{d}{d\psi(\rho)}\Theta(z,\rho)\right]h. If zn+mz\in\mathbb{C}_{\mathbb{Z}}^{n+m} is band-limited, d2dψ(ρ)2Θ(z,ρ)\frac{d^{2}}{d\psi(\rho)^{2}}\Theta(z,\rho) exists and

d2dψ(ρ)2Θ(z,ρ)\displaystyle\frac{d^{2}}{d\psi(\rho)^{2}}\Theta(z,\rho) =ρ𝐦r=1d(p=1prd(q=1qp,rdeiψ(ρ)μjqKzcq)(iKμjpeiψ(ρ)μjpKzcp)(iKμjreiψ(ρ)μjrKzcr))\displaystyle=\rho^{\mathbf{m}}\sum_{r=1}^{d}\left(\sum_{\begin{subarray}{c}p=1\\ p\neq r\end{subarray}}^{d}\left(\prod_{\begin{subarray}{c}q=1\\ q\neq p,r\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{q}}K}z_{c_{q}}\right)*(iK\mu_{j_{p}}e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}})*(iK\mu_{j_{r}}e^{i\psi(\rho)\mu_{j_{r}}K}z_{c_{r}})\right)
+ρ𝐦r=1d(p=1prdeiψ(ρ)μjpKzcp)(K2μjr2eiψ(ρ)μjrKzcr).\displaystyle\phantom{=}+\rho^{\mathbf{m}}\sum_{r=1}^{d}\left(\prod_{\begin{subarray}{c}p=1\\ p\neq r\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{p}}K}z_{c_{p}}\right)*\left(-K^{2}\mu_{j_{r}}^{2}e^{i\psi(\rho)\mu_{j_{r}}K}z_{c_{r}}\right).

Finally, if q1,q2{1,,n+m}q_{1},q_{2}\in\{1,\dots,n+m\} and h1,h2ν1()h_{1},h_{2}\in\ell_{\nu}^{1}(\mathbb{C}), then

d2dzq2dzq1Θ(z,p)[h1,h2]\displaystyle\frac{d^{2}}{dz_{q_{2}}dz_{q_{1}}}\Theta(z,p)[h_{1},h_{2}] =ρ𝐦r=1cr=q1d(eiψ(ρ)μjrKh1)(p=1cp=q2prd(eiψ(ρ)μjpKh2)ξ=1ξr,pdeiψ(ρ)μjξKzcξ)\displaystyle=\rho^{\mathbf{m}}\sum_{\begin{subarray}{c}r=1\\ c_{r}=q_{1}\end{subarray}}^{d}(e^{i\psi(\rho)\mu_{j_{r}}K}h_{1})*\left(\sum_{\begin{subarray}{c}p=1\\ c_{p}=q_{2}\\ p\neq r\end{subarray}}^{d}(e^{i\psi(\rho)\mu_{j_{p}}K}h_{2})*\prod_{\begin{subarray}{c}\xi=1\\ \xi\neq r,p\end{subarray}}^{d}e^{i\psi(\rho)\mu_{j_{\xi}}K}z_{c_{\xi}}\right)
Remark 10.

The requirement that zz be band-limited really is necessary for the existence of d2dψ(ρ)2Θ(z,p)\frac{d^{2}}{d\psi(\rho)^{2}}\Theta(z,p). Indeed, if zν1(n+m)z\in\ell_{\nu}^{1}(\mathbb{C}^{n+m}), then KzjKν()Kz_{j}\in K_{\nu}(\mathbb{C}) but K2zK^{2}z is not, and the latter terms appear in the second derivative. This is precisely why s\mathcal{F}_{s} is not twice differentiable.

4.4 The bound Z1Z_{1}

To have a hope at deriving a Z1Z_{1} bound, we will first determine the structure of the operator Ds(u^s)AsD\mathcal{F}_{s}(\hat{u}_{s})-A_{s}^{\dagger}. Represented as an “infinite block matrix”, most blocks are zero. One can verify that

Ds(u^s)As\displaystyle D\mathcal{F}_{s}(\hat{u}_{s})-A_{s}^{\dagger} =(00𝐙1(1,3)000𝐙1(3,1)𝐙1(3,2)𝐙1(3,3)),\displaystyle=\left(\begin{array}[]{ccc}0&0&\mathbf{Z}_{1}^{(1,3)}\\ 0&0&0\\ \mathbf{Z}_{1}^{(3,1)}&\mathbf{Z}_{1}^{(3,2)}&\mathbf{Z}_{1}^{(3,3)}\end{array}\right), (74)

with the individual terms 𝐙1(i,j)\mathbf{Z}_{1}^{(i,j)} being the operators

𝐙1(1,3)\displaystyle\mathbf{Z}_{1}^{(1,3)} =πMD1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)π\displaystyle=\pi^{M}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})\pi^{\infty}
𝐙1(3,1)\displaystyle\mathbf{Z}_{1}^{(3,1)} =πD1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)πM\displaystyle=\pi^{\infty}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})\pi^{M}
𝐙1(3,2)\displaystyle\mathbf{Z}_{1}^{(3,2)} =πD1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)z^sψ()+πD2𝐟(ζ(ψ^s)z^s,ρ^s)\displaystyle=\pi^{\infty}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta^{\prime}(\hat{\psi}_{s})\hat{z}_{s}\psi(\cdot)+\pi^{\infty}D_{2}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})
𝐙1(3,3)\displaystyle\mathbf{Z}_{1}^{(3,3)} =πD1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)π,\displaystyle=\pi^{\infty}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})\pi^{\infty},

and ψ:n+m+4\psi:\mathbb{R}^{n+m+4}\rightarrow\mathbb{R} once again denoting the frequency component projection. Computing Z1Z_{1} requires precomposing (74) with AsA_{s}, which has the structure (67).

4.4.1 Virtual padding

For hΩh\in\Omega with hΩ1||h||_{\Omega}\leq 1, denote

g=As(Ds(u^s)As)h,g=(gM,gρ,g).\displaystyle g=A_{s}(D\mathcal{F}_{s}(\hat{u}_{s})-A_{s}^{\dagger})h,\hskip 5.69054ptg=(g^{M},g^{\rho},g^{\infty}). (75)

Computing Z1Z_{1} is equivalent to a uniform (in hh) bound for gΩ||g||_{\Omega}. The tightness of this bound is determined by two levels of computation:

  • some finite norm computations that are done on the computer;

  • theoretical bounds, which are inversely proportional to the dimension of the object on which the finite norm computations are done.

By default, the size of the finite computation is linear in MM, the number of modes. This might suggest that explicitly increasing MM – that is, padding our solution with extra zeros and re-computing everything with more modes – is the only way to improve the bounds. Thankfully, this is not the case.

Intuitively, As(Ds(u^sAs)A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}-A_{s}^{\dagger}) is an “infinite matrix”, for which we have a canonical numerical center determined by the pre- and post- composition with the projection operator onto πMSymm(ν1(n+m))×n+m+4\pi^{M}\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n+m+4}. This is determined by the number of modes MM we have specified in our numerical zero. However, there is no reason to only compute this much of the infinite matrix explicitly; we could instead choose 𝐌M\mathbf{M}\geq M and compute the representation of this operator on π𝐌Symm(ν1(n+m))×n+m+4\pi^{\mathbf{M}}\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n+m+4}. The result is that a larger portion of As(Ds(u^sAs)A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}-A_{s}^{\dagger}) is stored in the computer’s memory. The advantage of doing this is that the explicit computations of norms are generally much tighter than theoretically guaranteed estimates, while the theoretical bounds related to the tail will be proportional to 1𝐌+1\frac{1}{\mathbf{M}+1}. See Figure 4 for a visual representation.

Refer to caption
Figure 4: Visual depiction of the virtual padding process as it applies to the operators DsD\mathcal{F}_{s} and AsA_{s}, left and right. When we do not do virtual padding (𝐌=M\mathbf{M}=M), the size of the objects we store in memory on the computer for matrix operations are strictly determined by the dimension of the numerical zero. These objects are the inner boxes in the infinite matrices above, and are depicted in pale yellow. The part outside of this box is controlled analytically. When we use nontrivial virtual padding (𝐌>M(\mathbf{M}>M), we store a larger amount of information in the computer for matrix operations; this is the outer dashed line box. Once again, the region outside the box is controlled analytically. Regions in white represent zeroes, while dark green represents the infinite part of the matrix, and light green a finite part of the matrix that is analytically controlled if there is no virtual padding.

To exploit these observations, we decompose ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) (and hence the symmetric subspace) as a direct sum

π𝐌(ν1(n+m))π𝐌+1,(ν1(n+m)),\pi^{\mathbf{M}}(\ell_{\nu}^{1}(\mathbb{C}^{n+m}))\oplus\pi^{\mathbf{M}+1,\infty}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})),

where 𝐌M\mathbf{M}\geq M, and π𝐌+1,\pi^{\mathbf{M}+1,\infty} is now interpreted as the complementary projector to π𝐌\pi^{\mathbf{M}}. In the subsections that follow, 𝐌\mathbf{M} will be the virtual padding dimension. We therefore re-interpret h=(h𝐌,hρ,h)h=(h^{\mathbf{M}},h^{\rho},h^{\infty}) and g=(g𝐌,gρ,g)g=(g^{\mathbf{M}},g^{\rho},g^{\infty}) as being in the product space

π𝐌Symm(ν1(n+m))×n+m+4×π𝐌+1,Symm(ν1(n+m)),\pi^{\mathbf{M}}\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m}))\times\mathbb{R}^{n+m+4}\times\pi^{\mathbf{M}+1,\infty}\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})),

which remains (isometrically) isomorphic to Ω\Omega. Remark that the finite blocks Ps,QsP_{s},Q_{s} and RsR_{s} coming from AsA_{s} must now be re-interpreted as linear maps involving π𝐌Symm(ν1(n+m))\pi^{\mathbf{M}}\operatorname{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})), as appropriate. Similarly, the 𝐁\mathbf{B} blocks can be replaced by

𝐙1(1,3)\displaystyle\mathbf{Z}_{1}^{(1,3)} =π𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)π𝐌+1,\displaystyle=\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})\pi^{\mathbf{M}+1,\infty}
𝐙1(3,1)\displaystyle\mathbf{Z}_{1}^{(3,1)} =π𝐌+1,D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)π𝐌\displaystyle=\pi^{\mathbf{M}+1,\infty}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})\pi^{\mathbf{M}}
𝐙1(3,2)\displaystyle\mathbf{Z}_{1}^{(3,2)} =π𝐌+1,D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)z^sψ()+π𝐌+1,D2𝐟(ζ(ψ^s)z^s,ρ^s)\displaystyle=\pi^{\mathbf{M}+1,\infty}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta^{\prime}(\hat{\psi}_{s})\hat{z}_{s}\psi(\cdot)+\pi^{\mathbf{M}+1,\infty}D_{2}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})
𝐙1(3,3)\displaystyle\mathbf{Z}_{1}^{(3,3)} =π𝐌+1,D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)π𝐌+1,,\displaystyle=\pi^{\mathbf{M}+1,\infty}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})\pi^{\mathbf{M}+1,\infty},

In our implementation, this virtual padding is implemented automatically on an as-needed basis.

4.4.2 Some notation

Let πj1,j2:ν1(n+m)ν1(n+m)\pi^{j_{1},j_{2}}:\ell_{\nu}^{1}(\mathbb{C}^{n+m})\rightarrow\ell_{\nu}^{1}(\mathbb{C}^{n+m}) be defined according to

(πj1,j2z)k={zk,j1|k|j20otherwise.(\pi^{j_{1},j_{2}}z)_{k}=\left\{\begin{array}[]{ll}z_{k},&j_{1}\leq|k|\leq j_{2}\\ 0&\mbox{otherwise.}\end{array}\right.

Let 𝟏j1,j2ν1(n+m)\mathbf{1}_{j_{1},j_{2}}\subset\ell_{\nu}^{1}(\mathbb{C}^{n+m}) denote the interval sequence

(𝟏j1,j2)k\displaystyle(\mathbf{1}_{j_{1},j_{2}})_{k} ={[𝟏]n+mν|k|,j1|k|j20otherwise,\displaystyle=\left\{\begin{array}[]{ll}[\mathbf{1}_{\mathbb{C}}]^{n+m}\nu^{-|k|},&j_{1}\leq|k|\leq j_{2}\\ 0&\mbox{otherwise},\end{array}\right.

where [𝟏]n+mn+m[\mathbf{1}_{\mathbb{C}}]^{n+m}\subset\mathbb{C}^{n+m} is the unit interval vector333Subject to any weighting of norms; that is, the components of [𝟏]n+m[\mathbf{1}_{\mathbb{C}}]^{n+m} should be scaled so that its norm is equal to 1 in n+m\mathbb{C}^{n+m}. in n+m\mathbb{C}^{n+m}. Let [𝟏]n+m+4[\mathbf{1}_{\mathbb{R}}]^{n+m+4} denote the unit interval vector444In case the euclidean norms associated to the scalar variables are not homogeneous and different weighting is used, each component of [𝟏]n+m+4[\mathbf{1}_{\mathbb{R}}]^{n+m+4} should be scaled so that the norm of [𝟏]n+m+4[\mathbf{1}_{\mathbb{R}}]^{n+m+4} is equal to 1 in the induced max norm. in n+m+4\mathbb{R}^{n+m+4}. Finally, we will occasionally require the use of

T(u,w)\displaystyle T(u,w) =(0m00D1θBC(kzk,x,a,α,β,η)jwj00),\displaystyle=\left(\begin{array}[]{c}0_{\mathbb{R}^{m}}\\ 0_{\mathbb{R}}\\ 0_{\mathbb{R}}\\ D_{1}\theta_{BC}(\sum_{k}z_{k},x,a,\alpha,\beta,\eta)\sum_{j}w_{j}\\ 0_{\mathbb{R}}\\ 0_{\mathbb{R}}\end{array}\right),

for u=(z,(x,a,α,β,η))ν1(n+m)×n+m+4u=(z,(x,a,\alpha,\beta,\eta))\in\ell_{\nu}^{1}(\mathbb{C}^{n+m})\times\mathbb{R}^{n+m+4}. Also, given wν1(n+m)w\in\ell_{\nu}^{1}(\mathbb{C}^{n+m}), define 𝐒(w)ν1(n+m)\mathbf{S}(w)\in\ell_{\nu}^{1}(\mathbb{C}^{n+m}) by 𝐒(w)j,0=kwj,k\mathbf{S}(w)_{j,0}=\sum_{k}w_{j,k}, and 𝐒(w),k=0\mathbf{S}(w)_{\cdot,k}=0 for k0k\neq 0. By construction, T(u,w)=T(u,𝐒(w))T(u,w)=T(u,\mathbf{S}(w)).

4.4.3 The norm g𝐌+g||g^{\mathbf{M}}+g^{\infty}||

In this section, the norm ||||||\cdot|| on Symm(ν1(n+m))\mbox{Symm}(\ell_{\nu}^{1}(\mathbb{C}^{n+m})) is identified with (xM,x)(xM,0,x)Ω(x^{M},x^{\infty})\mapsto||(x^{M},0,x^{\infty})||_{\Omega}.

Lemma 7.

Let z𝐟(z,)z\mapsto\mathbf{f}(z,\cdot) be a convolution polynomial of degree qq. Let555The absolute value of ψ𝟏\psi_{\mathbf{1}} is precisely the weight attributed to the frequency component, relative to the absolute value norm on \mathbb{R}. ψ𝟏=[1,1](0,(0,0,1,0,0))ΩM\psi_{\mathbf{1}}=[-1,1]\cdot||(0,(0,0,1,0,0))||_{\Omega^{M}}. Then

g𝐌+g\displaystyle||g^{\mathbf{M}}+g^{\infty}|| Psπ𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)𝟏𝐌+1,q𝐌+1𝐌+1|ψ^s|1IQsT(u^s,𝐯s+𝐰s),\displaystyle\leq||P_{s}\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\mathbf{1}_{\mathbf{M}+1,q\mathbf{M}}||+\frac{1}{\mathbf{M}+1}|\hat{\psi}_{s}|^{-1}||I-Q_{s}T(\hat{u}_{s},\mathbf{v}_{s}+\mathbf{w}_{s})||,

where the finitely-supported interval sequences 𝐯s,𝐰sν1(n+m)\mathbf{v}_{s},\mathbf{w}_{s}\subset\ell_{\nu}^{1}(\mathbb{C}^{n+m}) are defined according to

𝐰s\displaystyle\mathbf{w}_{s} =π𝐌+1,qM(D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)z^sψ𝟏1+D2𝐟(ζ(ψ^s)z^s,ρ^s)[𝟏]n+m+4)\displaystyle=\pi^{\mathbf{M}+1,qM}\Big{(}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta^{\prime}(\hat{\psi}_{s})\hat{z}_{s}\psi_{\mathbf{1}}^{-1}+D_{2}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})[\mathbf{1}_{\mathbb{R}}]^{n+m+4}\Big{)}
(𝐯s)j,k\displaystyle(\mathbf{v}_{s})_{j,k} ={([𝟏]n+m)jD1𝐟j(ζ(ψ^s)z^s,ρ^s)L(ν1(n+m),ν1()),j=1,,m+nk=00j=1,,m+nk0\displaystyle=\left\{\begin{array}[]{ll}([\mathbf{1}_{\mathbb{C}}]^{n+m})_{j}||D_{1}\mathbf{f}_{j}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})||_{L(\ell_{\nu}^{1}(\mathbb{C}^{n+m}),\ell_{\nu}^{1}(\mathbb{C}))},&\mbox{$j=1,\dots,m+n$, $k=0$}\\ 0&\mbox{$j=1,\dots,m+n$, $k\neq 0$}\end{array}\right.
Proof.

Explicitly, the sum g𝐌+gg^{\mathbf{M}}+g^{\infty} is

g𝐌+g\displaystyle g^{\mathbf{M}}+g^{\infty} =Ps𝐙11,3h+iψ^s1(IQsD3~s(2)(u^s))(Kπ𝐌+1:)1(𝐙13,1h𝐌+𝐙13,2hρ+𝐙13,3h)\displaystyle=P_{s}\mathbf{Z}_{1}^{1,3}h^{\infty}+i\hat{\psi}_{s}^{-1}(I-Q_{s}D_{3}\tilde{\mathcal{F}}_{s}^{(2)}(\hat{u}_{s}))(K\pi^{\mathbf{M}+1:\infty})^{-1}(\mathbf{Z}_{1}^{3,1}h^{\mathbf{M}}+\mathbf{Z}_{1}^{3,2}h^{\rho}+\mathbf{Z}_{1}^{3,3}h^{\infty})
=Ps𝐙11,3h+iψ^s1(IQsT(u^s,))(Kπ𝐌+1:)1(𝐙13,1h𝐌+𝐙13,2hρ+𝐙13,3h)\displaystyle=P_{s}\mathbf{Z}_{1}^{1,3}h^{\infty}+i\hat{\psi}_{s}^{-1}(I-Q_{s}T(\hat{u}_{s},\cdot))(K\pi^{\mathbf{M}+1:\infty})^{-1}(\mathbf{Z}_{1}^{3,1}h^{\mathbf{M}}+\mathbf{Z}_{1}^{3,2}h^{\rho}+\mathbf{Z}_{1}^{3,3}h^{\infty})

Since z𝐟(z,)z\mapsto\mathbf{f}(z,\cdot) is a convolution polynomial of degree qq, we have (D1𝐟(ζ(ψ^s)z^s,ρ^s)ek)j=0(D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})e_{k})_{j}=0 for all |j|𝐌|j|\leq\mathbf{M} whenever |k|>q𝐌|k|>q\mathbf{M}. It follows that the support of wπ𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)ww\mapsto\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})w is contained in the range of π0,q𝐌\pi^{0,q\mathbf{M}}. Restricting to the range of π𝐌+1,\pi^{\mathbf{M}+1,\infty},

Ps𝐙11,3hPsπ𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)𝟏𝐌+1,q𝐌.||P_{s}\mathbf{Z}_{1}^{1,3}h^{\infty}||\leq\left|\left|P_{s}\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\mathbf{1}_{\mathbf{M}+1,q\mathbf{M}}\right|\right|.

Next, the range of hρ𝐙1(3,2)hρh^{\rho}\mapsto\mathbf{Z}_{1}^{(3,2)}h^{\rho} is contained in that of π0,qM\pi^{0,qM} (recall that while we have used a different padding dimension 𝐌\mathbf{M} for the computation of norms, our data z^s\hat{z}_{s} is still band-limited to MM modes). We have

𝐙13,1h𝐌+𝐙13,3h=D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)(h𝐌+h)\displaystyle||\mathbf{Z}_{1}^{3,1}h^{\mathbf{M}}+\mathbf{Z}_{1}^{3,3}h^{\infty}||=||D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta(\hat{\psi}_{s})(h^{\mathbf{M}}+h^{\infty})|| 𝐯s,\displaystyle\leq||\mathbf{v}_{s}||,
𝐙13,2hρ=π𝐌+1,(D1𝐟(ζ(ψ^s)z^s,ρ^s)ζ(ψ^s)z^sψ()+D2𝐟(ζ(ψ^s)z^s,ρ^s))hρ\displaystyle\mathbf{Z}_{1}^{3,2}h^{\rho}=\pi^{\mathbf{M}+1,\infty}\big{(}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\zeta^{\prime}(\hat{\psi}_{s})\hat{z}_{s}\psi(\cdot)+D_{2}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\big{)}h^{\rho} 𝐰s\displaystyle\in\mathbf{w}_{s}

Using these bounds/inclusions, the fact that 𝐯s\mathbf{v}_{s} is supported on the zeroth Fourier index, and the properties of the map TT, we get

(IQsT(u^s,))(Kπ𝐌+1:)1(𝐙13,1h𝐌+𝐙13,2hρ+𝐙13,3h)\displaystyle||(I-Q_{s}T(\hat{u}_{s},\cdot))(K\pi^{\mathbf{M}+1:\infty})^{-1}(\mathbf{Z}_{1}^{3,1}h^{\mathbf{M}}+\mathbf{Z}_{1}^{3,2}h^{\rho}+\mathbf{Z}_{1}^{3,3}h^{\infty})|| IQsT(u^s,𝐯s+𝐰s)𝐌+1.\displaystyle\leq\frac{||I-Q_{s}T(\hat{u}_{s},\mathbf{v}_{s}+\mathbf{w}_{s})||}{\mathbf{M}+1}.

Combining the previous bounds, we get the result claimed in the lemma. ∎

Remark 11.

While perhaps symbolically intimidating, all of the quantities appearing in the statement of Lemma 7 are explicitly machine-computable, so a rigorous uniform (in ss and hh) bound for g𝐌+g||g^{\mathbf{M}}+g^{\infty}|| can indeed be computed.

4.4.4 The norm gρ||g^{\rho}||

In this section, the norm ||||||\cdot|| on n+m+4\mathbb{R}^{n+m+4} is identified with x||0,x,0)||Ωx\mapsto||0,x,0)||_{\Omega}. The bound for gρ||g^{\rho}|| bears a lot of similarity to the one for g𝐌+g||g^{\mathbf{M}}+g^{\infty}||, and its derivation is similar. We omit the proof of the following lemma.

Lemma 8.

With the same notation as in Lemma 7, we have the bound

gρ\displaystyle||g^{\rho}|| Rsπ𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)𝟏𝐌+1,q𝐌+1𝐌+1|ψ^s|1SsT(u^s,𝐯s+𝐰s).\displaystyle\leq||R_{s}\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\mathbf{1}_{\mathbf{M}+1,q\mathbf{M}}||+\frac{1}{\mathbf{M}+1}|\hat{\psi}_{s}|^{-1}||S_{s}T(\hat{u}_{s},\mathbf{v}_{s}+\mathbf{w}_{s})||.

4.4.5 Summary of the Z1Z_{1} bound

Combining the results of Lemma 7 and Lemma 8, it follows that if we choose Z1Z_{1} such that

Z1\displaystyle Z_{1} max{||Psπ𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)𝟏𝐌+1,q𝐌||+1𝐌+1|ψ^s|1||IQsT(u^s,𝐯s+𝐰s)||,\displaystyle\geq\max\left\{||P_{s}\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\mathbf{1}_{\mathbf{M}+1,q\mathbf{M}}||+\frac{1}{\mathbf{M}+1}|\hat{\psi}_{s}|^{-1}||I-Q_{s}T(\hat{u}_{s},\mathbf{v}_{s}+\mathbf{w}_{s})||\hskip 2.84526pt,\right.
||Rsπ𝐌D1𝐟(ζ(ψ^s)z^s,ρ^s)𝟏𝐌+1,q𝐌||+1𝐌+1|ψ^s|1||SsT(u^s,𝐯s+𝐰s)||}\displaystyle\phantom{\geq\max\big{\}}\}}\left.||R_{s}\pi^{\mathbf{M}}D_{1}\mathbf{f}(\zeta(\hat{\psi}_{s})\hat{z}_{s},\hat{\rho}_{s})\mathbf{1}_{\mathbf{M}+1,q\mathbf{M}}||+\frac{1}{\mathbf{M}+1}|\hat{\psi}_{s}|^{-1}||S_{s}T(\hat{u}_{s},\mathbf{v}_{s}+\mathbf{w}_{s})||\right\}

for all sΔs\in\Delta, then Z1Z_{1} will satisfy the bound (17) for our validated continuation problem. In the above bound, one must remember that the norms ||||||\cdot|| are actually restrictions of the norm ||||Ω||\cdot||_{\Omega} to various subspaces; see the preamble before Lemma 7 and Lemma 8.

4.5 The bound Z2Z_{2}

It is beneficial to perform a further splitting of the expression that defines the Z2Z_{2} bound. Recall that Z2(r)Z_{2}(r) should satisfy

As(Ds(u^s+δ)Ds(u^s))B(Ω,Ω)Z2(r)\displaystyle||A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}+\delta)-D\mathcal{F}_{s}(\hat{u}_{s}))||_{B(\Omega,\Omega)}\leq Z_{2}(r) (76)

uniformly for sΔs\in\Delta and δBr(u^s)¯\delta\in\overline{B_{r}(\hat{u}_{s})}. From Remark 10, Ds()D\mathcal{F}_{s}(\cdot) is not itself differentiable, so appealing to a “typical” second derivative estimate for this Z2Z_{2} bound is not going to work. To handle this, we use the strategy of [28] and split the norm to be computed using a triangle inequality. Given δBr(u^s)¯\delta\in\overline{B_{r}(\hat{u}_{s})}, split it as δ=δ1+δ2\delta=\delta_{1}+\delta_{2}, where the components of δ1\delta_{1} are all zero except for the frequency component, ψ\psi, and δ2\delta_{2} has zero for its frequency component. This decomposition is uniquely defined. Then, consider two new bounds to be computed:

As(Ds(u^s+δ1)Ds(u^s))B(Ω,Ω)\displaystyle||A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1})-D\mathcal{F}_{s}(\hat{u}_{s}))||_{B(\Omega,\Omega)} Z2,1(r)\displaystyle\leq Z_{2,1}(r)
As(Ds(u^s+δ1+δ2)Ds(u^s+δ1))B(Ω,Ω)\displaystyle||A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+\delta_{2})-D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}))||_{B(\Omega,\Omega)} Z2,1(r)\displaystyle\leq Z_{2,1}(r)

uniformly for sΔs\in\Delta and δBr(u^s)¯\delta\in\overline{B_{r}(\hat{u}_{s})}. If we define Z2(r)=defZ2,1(r)+Z2,2(r)Z_{2}(r)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,Z_{2,1}(r)+Z_{2,2}(r), then (76) will be satisfied. It turns out that this decomposition is effective. We will elaborate on this now.

4.5.1 The bound Z2,1Z_{2,1}

The partial derivative ddψ(ρ)Ds(z^sM,ρ^s,0)\frac{d}{d\psi(\rho)}D\mathcal{F}_{s}(\hat{z}^{M}_{s},\hat{\rho}_{s},0) exists and is continuous. This is a consequence of Lemma 6 (n.b. the inputs are band-limited) and the structure of s\mathcal{F}_{s}; see (37). As AsA_{s} is bounded, we can use the fundamental theorem of calculus (in Banach space) to obtain

As(Ds(u^s+δ1)Ds(u^s))hΩ\displaystyle||A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1})-D\mathcal{F}_{s}(\hat{u}_{s}))h||_{\Omega} supt[0,1]As(ddψ(ρ)Ds(u^s+tδ1)h)ψ(δ1)Ω\displaystyle\leq\sup_{t\in[0,1]}\left|\left|A_{s}\left(\frac{d}{d\psi(\rho)}D\mathcal{F}_{s}(\hat{u}_{s}+t\delta_{1})h\right)\psi(\delta_{1})\right|\right|_{\Omega} (77)

for any hΩh\in\Omega, hΩ1||h||_{\Omega}\leq 1. Due to the structure of s\mathcal{F}_{s}, the derivative term in the large parentheses will have

  • in the ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) components: convolution polynomials of the form (70) involving Fourier components coming from the set

    {ζ(ψ^s+ψ(δ1))z^s}{iζ(ψ^s+ψ(δ1))Kz^s}{ζ(ψ^s+ψ(δ1))K2z^s}{h}\{\zeta(\hat{\psi}_{s}+\psi(\delta_{1}))\hat{z}_{s}\}\cup\{i\zeta(\hat{\psi}_{s}+\psi(\delta_{1}))K\hat{z}_{s}\}\cup\{-\zeta(\hat{\psi}_{s}+\psi(\delta_{1}))K^{2}\hat{z}_{s}\}\cup\{h\}

    with coefficients in ρ^s\hat{\rho}_{s};

  • in the scalar components: finite-dimensional range functions of |k|M(z^s)k\sum_{|k|\leq M}(\hat{z}_{s})_{k} and δ(ψ1)\delta(\psi_{1}).

To obtain the tightest possible upper bound for (77), one would want to exploit as much structure of DsD\mathcal{F}_{s} (and the directional derivative in the frequency direction) as possible. However, to get a general-purpose implementable bound, we can apply the following operations to ddψ(ρ)Ds(u^s+tδ1)hψ(δ1)\frac{d}{d\psi(\rho)}D\mathcal{F}_{s}(\hat{u}_{s}+t\delta_{1})h\psi(\delta_{1}).

  1. 1.

    Replace all instances of ζ(ψ^s+ψ(δ1))\zeta(\hat{\psi}_{s}+\psi(\delta_{1})) with the interval [1,1][-1,1];

  2. 2.

    replace the component of hh in n+m+4\mathbb{R}^{n+m+4} with the unit interval vector666Note that the components of this “unit” interval vector might be uneven due to weighting. in ×n+m+4\times\mathbb{R}^{n+m+4};

  3. 3.

    replace all other variables777Note that if the norm on n+m\mathbb{C}^{n+m} associated to ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) is itself weighted, this must be taken into account. described in the bulleted list above with bounds for their norms, multiplied by the zero-supported basis element of the relevant space (e.g. for ν1\ell_{\nu}^{1} elements, the identity e0e_{0} of the Banach algebra; for ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}), the vectorized version of e0e_{0}), multiplied by the interval [1,1][-1,1];

  4. 4.

    compute operator norms of the blocks of AsA_{s} in (67);

  5. 5.

    complete block multiplications, taking the norm of the result, followed by an interval supremum (note, t[0,1]t\in[0,1] is also replaced by an interval).

We emphasize in the algorithm above, every object is a finite-dimensional quantity, so operations can indeed be performed in a suitable complex vector space on the computer. This strategy produces a true bound for (77) largely because of the Banach algebra and the interval arithmetic. For example, consider the impact of this computation on the quantity

z^sh+(eiK(ψ^s+ψ(δ1))z^s)(iKeiK(ψ^s+ψ(δ1)z^s)h-\hat{z}_{s}*h+(e^{iK(\hat{\psi}_{s}+\psi(\delta_{1}))}\hat{z}_{s})*(iKe^{iK(\hat{\psi}_{s}+\psi(\delta_{1})}\hat{z}_{s})*h

in ν1\ell_{\nu}^{1}. From the Banach algebra, if hν1||h||_{\nu}\leq 1, then a triangle inequality produces

z^s+eiK(ψ^s+ψ(δ1)z^sνiKeiK(ψ^s+ψ(δ1)z^sν\displaystyle||\hat{z}_{s}||+||e^{iK(\hat{\psi}_{s}+\psi(\delta_{1})}\hat{z}_{s}||_{\nu}\cdot||iKe^{iK(\hat{\psi}_{s}+\psi(\delta_{1})}\hat{z}_{s}||_{\nu} =(ae0)(ce0)ν+(ae0)(be0)(ce0)\displaystyle=||(ae_{0})*(ce_{0})||_{\nu}+||(ae_{0})*(be_{0})*(ce_{0})||
=ac+abc,\displaystyle=ac+abc,

where a=z^sνa=||\hat{z}_{s}||_{\nu}, b=iKz^sνb=||iK\hat{z}_{s}||_{\nu}, c=hν=1c=||h||_{\nu}=1, and e0e_{0} is the identity element in the Banach algebra ν1\ell_{\nu}^{1}. Now, define 𝟏=[1,1]\mathbf{1}=[-1,1] and compare the result to

(𝟏ae0)(𝟏ce0)+(𝟏ae0)(𝟏be0)(𝟏ce0).\displaystyle-(\mathbf{1}ae_{0})*(\mathbf{1}ce_{0})+(\mathbf{1}ae_{0})*(\mathbf{1}be_{0})*(\mathbf{1}ce_{0}).

The support of this sequence is the zeroth index, and we can explicitly compute the resulting interval. It is precisely 𝟏(ac+abc)e0\mathbf{1}\cdot(ac+abc)e_{0}. The supremum of this interval is ac+abcac+abc, which matches the analytical triangle inequality / Banach algebra bound computed previously.

Remark 12.

The bound obtained by applying the strategy above is incredibly crude; in effect, we use the triangle inequality for everything. However, producing fully general code to compute the second derivatives for this class of problems (polynomial delay differential equations with arbitrary numbers of delays) would be a messy programming task. Even with the present implementation, where we need only compute second derivatives evaluated at sequences that are supported on the zeroth Fourier mode, the implementation was far from trivial. Long term, it would be beneficial to implement second derivatives, as this would also allow for improvements to the Y0Y_{0} bound; see Section 4.2. The good news is that, theoretically, the computed bound will be O(r)O(r) for rr small enough, due to the linear multiplication of ψ(δ1)=O(r)\psi(\delta_{1})=O(r) appearing in (77).

4.5.2 The bound Z2,2Z_{2,2}

Let dδ2d_{\delta_{2}} denote the Gateaux derivative of Ds()D\mathcal{F}_{s}(\cdot) in the direction δ2\delta_{2}. We claim

dδ2Ds(u^s+δ1+tδ2)d_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})

exists and is continuous for t[0,1]t\in[0,1]. To see why, observe that that Ds(2)D\mathcal{F}_{s}^{(2)} is continuously differentiable, so we need only worry about the part in ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}), namely the components Ds(1)D\mathcal{F}_{s}^{(1)} and Ds(3)D\mathcal{F}_{s}^{(3)} that come from the vector field. The result now follows from Lemma 6. Indeed, a band-limited input is only required for the double derivative with respect to the frequency variable, and δ2\delta_{2} has zero for its frequency component. By the fundamental theorem of calculus,

As(Ds(u^s+δ1+δ2)Ds(u^s+δ1))hΩ\displaystyle||A_{s}(D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+\delta_{2})-D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}))h||_{\Omega} supt[0,1]As(dδ2Ds(u^s+δ1+tδ2)h)Ω\displaystyle\leq\sup_{t\in[0,1]}\left|\left|A_{s}\left(d_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h\right)\right|\right|_{\Omega} (78)

where we take hΩh\in\Omega, hΩ1||h||_{\Omega}\leq 1.

To bound (78), we make use of a very similar strategy to the one from Section 4.5.1 for Z2,1Z_{2,1}. The difference here is that the convolution polynomials in the ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) part of the Gateaux derivative have a more difficult structure. The main problem is in the mixed derivatives with respect to frequency and Fourier space; see the derivatives ddzqddψ(ρ)\frac{d}{dz_{q}}\frac{d}{d\psi(\rho)} in Lemma 6. These derivatives involve the action of the derivative operator KK on the sequence part of hh, which is not necessarily band-limited. If h=(hz,hρ)h=(h_{z},h_{\rho}) for hzν1h_{z}\in\ell_{\nu}^{1}, and this sequence is not band-limited, then generally KhzKh_{z} will not have a finite ν1\ell_{\nu}^{1}-norm. To combat this problem, we remark that only one such term can appear in any given convolution polynomial. This can be exploited as follows.

To begin, we factor AsA_{s} as follows.

As=((M+1)PsQsVs(1)(M+1)RsSsVs(2)00Vs(3))(I1M+1000I000(Kπ)1)=def𝐀s𝐊1.A_{s}=\left(\begin{array}[]{ccc}(M+1)P_{s}&Q_{s}&V_{s}^{(1)}\\ (M+1)R_{s}&S_{s}&V_{s}^{(2)}\\ 0&0&V_{s}^{(3)}\end{array}\right)\left(\begin{array}[]{ccc}I\frac{1}{M+1}&0&0\\ 0&I&0\\ 0&0&(K\pi^{\infty})^{-1}\end{array}\right)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\mathbf{A}_{s}\mathbf{K}^{-1}.

Note that 𝐀s\mathbf{A}_{s} is obtained from AsA_{s} by multiplying (on the right) the first “column” by (M+1)(M+1), and the last “column” by KπK\pi^{\infty}. The following lemma is a specific case of a lemma of van den Berg, Groothedde and Lessard [28].

Lemma 9.

Define an operator K~1=1M+1πM+K1π\tilde{K}^{-1}=\frac{1}{M+1}\pi^{M}+K^{-1}\pi^{\infty} on ν1\ell_{\nu}^{1}. Let u,v,ν1u,v,\in\ell_{\nu}^{1}. Then

K~1(Kuv)ν\displaystyle||\tilde{K}^{-1}(Ku*v)||_{\nu} Cνuνvν,Cν={ν2M+2elogν2M+2,ν2M+2<e1,ν2M+2e.\displaystyle\leq C_{\nu}||u||_{\nu}||v||_{\nu},\hskip 28.45274ptC_{\nu}=\left\{\begin{array}[]{ll}\frac{\nu^{2M+2}}{e\log\nu^{2M+2}},&\nu^{2M+2}<e\\ 1,&\nu^{2M+2}\geq e.\end{array}\right.

Now consider the product 𝐊1dδ2Ds(u^s+δ1+tδ2)h\mathbf{K}^{-1}d_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h. The part in ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) of the Gateaux derivative dδ2Ds(u^s+δ1+tδ2)hd_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h will be multiplied by the diagonal operator (K~1,,K~1)(\tilde{K}^{-1},\dots,\tilde{K}^{-1}). So consider how we might bound a term of the form

K~1(αiKhzjeiKψ(ρ^s+ψ(δ1))μj(z^s+tz(δ2))j)ν1()\displaystyle\tilde{K}^{-1}\left(\alpha iKh_{z}*\prod_{j}e^{iK\psi(\hat{\rho}_{s}+\psi(\delta_{1}))\mu_{j}}(\hat{z}_{s}+tz(\delta_{2}))_{j}\right)\in\ell_{\nu}^{1}(\mathbb{C}) (79)

in a given convolution polynomial that contains a factor KhzKh_{z}, for hzν1h_{z}\in\ell_{\nu}^{1} and hzν1||h_{z}||_{\nu}\leq 1. Here, α\alpha is a scalar that could depend on ρ^s\hat{\rho}_{s}, the scalar components of tδ2t\delta_{2}, and the delays, while z(δ2)z(\delta_{2}) is the part of δ2\delta_{2} in ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}). Note that by permutation invariance of the convolution, we may assume that KhzKh_{z} appears as the first term on the left, as we have done here. By Lemma 9, the above is bounded above by

CναjeiKψ(ρ^s+ψ(δ1))μj(z^s+tz(δ2))jν.\displaystyle C_{\nu}\left|\left|\alpha\prod_{j}e^{iK\psi(\hat{\rho}_{s}+\psi(\delta_{1}))\mu_{j}}(\hat{z}_{s}+tz(\delta_{2}))_{j}\right|\right|_{\nu}.

We can achieve the same effect by replacing KhzKh_{z} in (79) with Cνe0C_{\nu}e_{0}, for e0e_{0} being the identity in the convolution algebra, and taking the ν1\ell_{\nu}^{1}-norm. When K~1\tilde{K}^{-1} is multiplied by a polynomial term that does not contain a factor KhzKh_{z}, then a straightforward calculation shows that K~1B(ν1,ν1)=(M+1)1||\tilde{K}^{-1}||_{B(\ell_{\nu}^{1},\ell_{\nu}^{1})}=(M+1)^{-1}. The net result is that the impact on the norm of the polynomial term is a scaling by (M+1)1(M+1)^{-1}. Therefore, adapting the algorithm from the Z2,1Z_{2,1} bound calculation, it suffices to do perform the following operations to dδ2Ds(u^s+δ1+tδ2)hd_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h.

  1. 1.

    Multiply the part of dδ2Ds(u^s+δ1+tδ2)hd_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h in ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}) by 1M+1\frac{1}{M+1};

  2. 2.

    replace all instances of ζ(ψ^s+ψ(δ1))\zeta(\hat{\psi}_{s}+\psi(\delta_{1})) with the interval [1,1][-1,1];

  3. 3.

    replace all (remaining) instances of KK with (M+1)Cνe0[1,1](M+1)C_{\nu}e_{0}\cdot[-1,1];

  4. 4.

    replace the component of hh in n+m+4\mathbb{R}^{n+m+4} with the unit interval vector888Note that the components of this “unit” interval vector might be uneven due to weighting. in n+m+4\mathbb{R}^{n+m+4};

  5. 5.

    replace all other variables (z^s\hat{z}_{s}, ρ^s\hat{\rho}_{s}, δ1\delta_{1}, δ2\delta_{2}, the part of hh in ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}), and their relevant projections999See footnote 7 concerning the projections of hh in ν1(n+m)\ell_{\nu}^{1}(\mathbb{C}^{n+m}).) with bounds for their norms, multiplied by the zero-supported basis element of the relevant space, multiplied by the interval [1,1][-1,1];

  6. 6.

    compute operator norms of the blocks of 𝐀s\mathbf{A}_{s};

  7. 7.

    complete block multiplication of 𝐀s\mathbf{A}_{s} on the left, take the Ω\Omega-norm of the result, followed by an interval supremum (note: t[0,1]t\in[0,1] is also replaced by an interval).

The result is a (crude) upper bound of supt[0,1]Asdδ2Ds(u^s+δ1+tδ2)hΩ\sup_{t\in[0,1]}||A_{s}d_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h||_{\Omega} for hΩ1||h||_{\Omega}\leq 1. This bound is expected to be O(r)O(r) for rr small, since the Gateaux derivative dδ2Ds(u^s+δ1+tδ2)hd_{\delta_{2}}D\mathcal{F}_{s}(\hat{u}_{s}+\delta_{1}+t\delta_{2})h is O(δ2)O(||\delta_{2}||) for δ2r||\delta_{2}||\leq r small.

Remark 13.

In step 3, the variable replacement negates (by multilinearity of the convolution) the multiplication of the relevant polynomial term by (M+1)1(M+1)^{-1} that was done in step 1, while propagating forward the correct bound CνC_{\nu} that results from the interaction between K~1\tilde{K}^{-1} and the Kzhj()Kz_{h}*\prod_{j}(\cdots) polynomial term. This somewhat roundabout way of introducing the bound CνC_{\nu} in the correct locations is done in order to make the process implementable in generality.

5 Specification to ordinary differential equations

In Section 3, we demonstrated how rigorous two-parameter continuation of periodic orbits in delay differential equations can be accomplished in such a way that the continuation can pass through degenerate Hopf bifurcations. As ordinary differential equations are a special case – that is, where any delays are identically zero – the theory of the previous sections naturally applies equally to them. However, the formulation of the map (28) can be greatly simplified. Indeed, the reader familiar with numerical methods for periodic orbits has likely noticed that we have not performed the “usual” scaling out by the frequency, so that the period can be abstractly considered as 2π2\pi. With delay differential equations, this is not beneficial because it merely moves the frequency dependence into the delayed variables. Additionally, it is the presence of the delays that requires a more subtle analysis of the Z2Z_{2} bound; see Section 4.5.

To compare, with ordinary differential equations without delays, the technicalities with the Z2Z_{2} bound are absent. At the level of implementation, computing second and even third derivatives of the vector field in the Fourier representation is also much easier. In this section, we will present an alternative version of the zero-finding problem of Section 3.2 and analogous map GG from (28). However, we will not discuss the general implementation of the technical bounds YY and ZZ, since they are both simpler than the ones we have previously presented for the delay differential equations case, and can be obtained by fairly minor modifications of the bounds from [30]. We have implemented them in general in the codes at [4].

5.1 Desingularization, polynomial embedding and phase isolation

The set-up now starts with an ordinary differential equation

y˙(t)=f(y(t),α,β)\displaystyle\dot{y}(t)=f(y(t),\alpha,\beta) (80)

again depending on two real parameters α\alpha and β\beta. Performing the same blowup procecure as we did for the delay equations, we define

f~(z,x,a,α,β)={a1(f(x+az,α,β)f(x,α,β)),a0dyf(x,α,β)z,a=0.\displaystyle\tilde{f}(z,x,a,\alpha,\beta)=\left\{\begin{array}[]{ll}a^{-1}(f(x+az,\alpha,\beta)-f(x,\alpha,\beta)),&a\neq 0\\ d_{y}f(x,\alpha,\beta)z,&a=0.\end{array}\right.

The goal is therefore to find a pair (x,z)(x,z) such that

f(x,,x,α,β)\displaystyle f(x,\dots,x,\alpha,\beta) =0\displaystyle=0
z˙(t)\displaystyle\dot{z}(t) =f~(z(t),x,a,α,β),\displaystyle=\tilde{f}(z(t),x,a,\alpha,\beta),
z\displaystyle||z|| =1\displaystyle=1

where zz is ω\omega-periodic for an unknown period ω\omega; equivalently, the frequency of zz is ψ=2πω\psi=\frac{2\pi}{\omega}. Let μ=ψ1\mu=\psi^{-1} denote the reciprocal frequency. Define z~(t)=z(tμ)\tilde{z}(t)=z(t\mu). Then z~(t)\tilde{z}(t) is 2π2\pi-periodic. Substituting this into the differential equation above and dropping the tildes, we obtain the modified vector field

z˙(t)\displaystyle\dot{z}(t) =μf~(z(t),x,a,α,β),\displaystyle=\mu\tilde{f}(z(t),x,a,\alpha,\beta),

where now the scope is that zz should be 2π2\pi-periodic.

If a polynomial embedding is necessary to eliminate non-polynomial nonlinearities, we allow the inclusion of mm additional scalar equations that must be simultaneously solved, where we introduce an appropriate number mm of unfolding parameters, ηm\eta\in\mathbb{R}^{m}, to balance them. We again use the symbol θBC\theta_{BC} to function that defines these boundary conditions, with the equation being θBC(z(0),x,a,α,β,η)=0\theta_{BC}(z(0),x,a,\alpha,\beta,\eta)=0. We use the same anchor condition to handle the lack of isolation from phase shifts.

5.2 Zero-finding problem

Abusing notation and assuming now that f~\tilde{f} is polynomial after the embedding has been taken into account, the zero-finding problem is

{z˙=μf~(z(t),x,a,α,β,η),(differential equations)z=1,(amplitude condition of scaled orbit)z(s),z^(s)𝑑s=0,(anchor condition)f(x,α,β)=0,(x is a steady state)θBC(z(0),x,a,α,β,η)=0.(embedding boundary condition)\begin{cases}\dot{z}=\mu\tilde{f}(z(t),x,a,\alpha,\beta,\eta),&\quad\text{(differential equations)}\\ \|z\|=1,&\quad\text{(amplitude condition of scaled orbit)}\\ \int\langle z(s),\hat{z}^{\prime}(s)\rangle ds=0,&\quad\text{(anchor condition)}\\ f(x,\alpha,\beta)=0,&\quad\text{($x$ is a steady state)}\\ \theta_{BC}(z(0),x,a,\alpha,\beta,\eta)=0.&\quad\text{(embedding boundary condition)}\end{cases} (81)

where z^\hat{z} is understood to be a candidate numerical solution. Expanding zz as a Fourier series with period 2π2\pi, we have z(t)=kzkeiktz(t)=\sum_{k\in\mathbb{Z}}z_{k}e^{ikt} for some complex vectors zkz_{k} obeying the symmetry zk=zk¯z_{k}=\overline{z_{-k}}. If we now allow 𝐟(z,x,a,α,β,η)\mathbf{f}(z,x,a,\alpha,\beta,\eta) to be the representation of f~\tilde{f} as a convolution polynomial in the coefficients z=(zk)kz=(z_{k})_{k\in\mathbb{Z}}, then an analogous derivativation to the delay differential equations case produces the map

G(z,x,a,ψ,η,(α,β))\displaystyle G(z,x,a,\psi,\eta,(\alpha,\beta)) =(iKz+μ𝐟(z,x,a,α,β,η)f(x,α,β)z,K2z^1z,iKz^θBC(kzk,x,a,α,β,η))\displaystyle=\left(\begin{array}[]{c}-iKz+\mu\mathbf{f}(z,x,a,\alpha,\beta,\eta)\\ f(x,\alpha,\beta)\\ \langle z,K^{2}\hat{z}\rangle-1\\ \langle z,iK\hat{z}\rangle\\ \theta_{BC}(\sum_{k}z_{k},x,a,\alpha,\beta,\eta)\end{array}\right)

defined on the same Banach spaces, with the same codomain. However, this time, one can show that if ff is CkC^{k}, then GG really is Ck1C^{k-1}.

6 Proving bifurcations and bubbles

Modulo non-resonance conditions, we would generically expect Hopf bifurcations to occur on the level curve at amplitude zero of the computed 2-manifolds of Section 3. Hopf bubbles are equivalent to curves in the manifold that intersect the amplitude zero surface at two distinct points. As for bubble bifurcations, we can describe these in terms of a the existence of a relative local minimum of a projection of the computed 2-manifold, represented as the graph over amplitude and a distinguished parameter. We develop these points in this section, demonstrating how post-processing of data from the computer-assisted proofs can be used to prove the existence of Hopf bifurcations, bubbles, and degenerate bifurcations.

We should emphasize that to prove the existence of a single Hopf bubble, it suffices to identify a curve connecting two points at amplitude zero in the projection of the proven manifold in α×β×amplitude\alpha\times\beta\times\mbox{amplitude} space, for one of α\alpha or β\beta being fixed. While this does require verifying two Hopf bifurcations, the rest of the proof of an isolated bubble is fairly trivial, requiring only determining a sequence of simplices that enclose the curve. Therefore, we will only discuss how to prove bubble bifurcations in this section, since this can require a bit more analysis. We will not analyze Bautin bifurcations.

To simplify the presentation, we will assume throughout this section that the norm on Ω\Omega is such that (0,(0,a,0,0,0),0)=|a|||(0,(0,a,0,0,0),0)||=|a|. That is, the amplitude component is unit weighted relative to the absolute value.

6.1 Hopf bifurcation curves

In delay differential equations, the sufficient conditions for the existence of a Hopf bifurcation include the non-resonance check, which involves counting all eigenvalues all eigenvalues of the lienarization on the imaginary axis. This is somewhat beyond the scope of our work here, although there is literature on how this can be done using computer-assisted proofs [5, 24]. In this paper, we will consider the following related notion.

Definition 1.

The delay differential equation (3) has a Hopf bifurcation curve ΘHΔ\Theta_{H}\in\Delta if there exists a C1C^{1} parametrization Δs(α(s),β(s),x(s),y(s))\Delta\ni s\mapsto(\alpha(s),\beta(s),x(s),y(s)) such that x(s)x(s) is a steady state solution and y(s)y(s) is a periodic orbit at parameters (α(s),β(s))(\alpha(s),\beta(s)), such that:

  • ΘH(0)\Theta_{H}(0) and ΘH(1)\Theta_{H}(1) are on the boundary of Δ\Delta, and ΘH(t)\Theta_{H}(t) is in the interior of Δ\Delta for t(0,1)t\in(0,1).

  • xΘH=yΘHx\circ\Theta_{H}=y\circ\Theta_{H}; in other words, yy is a steady state on restriction to the image of ΘH\Theta_{H}.

  • For sΔ{ΘH(t):t[0,1]}s\in\Delta\setminus\{\Theta_{H}(t):t\in[0,1]\}, y(s)y(s) is not a steady state.

Contrary to the usual Hopf bifurcation, we do not reference the direction of the bifurcation, even along one-dimensional curves in the manifold of periodic orbits.

The existence of a Hopf bifurcation curve can be proven using post-processing of a validated contiuation on given simplex on which one expects a Hopf bifurcation to occur. The idea is that on a simplex that encloses a Hopf bifurcation, the amplitude in the blown-up variables (see Section 3.1) is expected to cross through zero. Since we are working in two-parameter continuation, such a crossing point should persist as a curve. A geometric construction based on the partial derivatives of the amplitude with respect to the simplex paramaterization can be used to construct this curve. See Figure 5 for a visualization.

To establish a more constructive proof, we need to introduce a few projection maps. Let u=(zM,ρ)ΩMu=(z^{M},\rho)\in\Omega^{M}, ρ=(x,a,ψ,η,α,β)\rho=(x,a,\psi,\eta,\alpha,\beta). Denote πau=a\pi^{a}u=a the projection to the amplitude component, and π(α,β)u=(α,β)\pi^{(\alpha,\beta)}u=(\alpha,\beta) the projection onto the parameters. Similarly, write πaρ=a\pi^{a}\rho=a. Denote the three vertices of the standard simplex Δ\Delta as s0=(0,0)s_{0}=(0,0), s1=(1,0)s_{1}=(1,0) and s2=(0,1)s_{2}=(0,1).

Refer to caption
Figure 5: Briefly, the partial derivatives of the amplitude of the family of periodic orbits, with respect to the simplex parametrization variables, have the same sign. From the assumptions of the proposition, the sign of the amplitude along the edge connecting s1s_{1} to s2s_{2} is opposite to that of the amplitude at s0s_{0}. Forming a ray (dashed-dotted line) connecting s0s_{0} to some point s¯\overline{s} on the edge connecting s1s_{1} to s2s_{2}, the mean-value theorem guarantees the existence of a unique point along the ray at which the amplitude is zero. Parameterizing over all s¯\overline{s} on the edge s1s_{1} to s2s_{2}, the Hopf curve ΘH\Theta_{H} (blue) is constructed.
Proposition 10.

Suppose a simplex containing the interpolated numerical data u^s\hat{u}_{s} has been validated, for sΔ.s\in\Delta. Let r>0r>0 be an associated validation radius from the radii polynomial. Let c=Z0+Z1+Z2(r)c=Z_{0}+Z_{1}+Z_{2}(r). Suppose the following are satisfied.

  • For k=1,2k=1,2,

    [πa(u^s0)r,πa(u^s0)+r][πa(u^sk)r,πa(u^sk)+r]<0,[\pi^{a}(\hat{u}_{s_{0}})-r,\pi^{a}(\hat{u}_{s_{0}})+r]\cdot[\pi^{a}(\hat{u}_{s_{k}})-r,\pi^{a}(\hat{u}_{s_{k}})+r]<0,

    where the multiplication is in the sense of interval arithmetic.

  • The sign of the interval [πa(u^s)r,πa(u^s)+r][\pi^{a}(\hat{u}_{s})-r,\pi^{a}(\hat{u}_{s})+r] is constant for ss on the edge of Δ\Delta not incident with s0s_{0}; that is, on the edge connecting s1s_{1} with s2s_{2}.

  • Let s\partial_{s} denote the Fréchet derivative operator with respect to the variable sΔs\in\Delta. For h2h\in\mathbb{R}^{2}, define

    ~sa(s)h\displaystyle\tilde{\partial}_{s}a(s)h =defπa(Sss[𝐉s](u^s+δr)h),~su(s)=defAss[s](u^s+δr),\displaystyle\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\pi^{a}\left(-S_{s}\partial_{s}[\mathbf{J}_{s}](\hat{u}_{s}+\delta_{r})h\right),\hskip 28.45274pt\tilde{\partial}_{s}u(s)\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,-A_{s}\partial_{s}[\mathcal{F}_{s}](\hat{u}_{s}+\delta_{r}), (82)
    γ\displaystyle\gamma =c1csupsΔ~su(s),\displaystyle=\frac{c}{1-c}\sup_{s\in\Delta}||\tilde{\partial}_{s}u(s)||,

    where δr=Br(0)¯ΩM\delta_{r}=\overline{B_{r}(0)}\subset\Omega^{M}. That is, γ\gamma is a real interval and ~sa(s)\tilde{\partial}_{s}a(s) is an inteval in (2)(\mathbb{R}^{2})^{*}. Let γ^=supγ\hat{\gamma}=\sup\gamma. The components of ~sa(s)+γ^[1,1]2\tilde{\partial}_{s}a(s)+\hat{\gamma}[-1,1]^{2} have the same sign for sΔs\in\Delta.

Then there is a unique (up to parameterizaztion) Hopf bifurcation curve ΘH:[0,1]Δ\Theta_{H}:[0,1]\rightarrow\Delta, and each of ΘH(0)\Theta_{H}(0) and ΘH(1)\Theta_{H}(1) lie on the edges of Δ\Delta that are incident with s0s_{0}.

Proof.

Suppose the radii polynomial proves the existence of usBr(u^s)u_{s}\in B_{r}(\hat{u}_{s}) such that s(us)=0\mathcal{F}_{s}(u_{s})=0. The C1C^{1} parametrization of the steady state (x(s)x(s)), periodic orbit (y(s)y(s)) and amplitude parametrization is a consequence of the computer-assisted proof. sus\partial_{s}u_{s} exists, and ss(us)+Ds(us)sus=0\partial_{s}\mathcal{F}_{s}(u_{s})+D\mathcal{F}_{s}(u_{s})\partial_{s}u_{s}=0. We claim Ds(y)D\mathcal{F}_{s}(y) is boundedly invertible for yBr(u^s)¯y\in\overline{B_{r}(\hat{u}_{s})}. First,

IAsDs(y)\displaystyle I-A_{s}D\mathcal{F}_{s}(y) =(IAsAs)+As(AsDs(u^s))+As(Ds(u^s)Ds(y))\displaystyle=(I-A_{s}A_{s}^{\dagger})+A_{s}(A_{s}^{\dagger}-D\mathcal{F}_{s}(\hat{u}_{s}))+A_{s}(D\mathcal{F}_{s}(\hat{u}_{s})-D\mathcal{F}_{s}(y))

whose norm is bounded above by Z0+Z1+Z2(r)=cZ_{0}+Z_{1}+Z_{2}(r)=c. Since the radii polynomial has validated u^s\hat{u}_{s}, it follows that c<1c<1. By the Neumann series, AsDs(y)A_{s}D\mathcal{F}_{s}(y) is boundedly invertible, with norm bounded by (1c)1(1-c)^{-1}. Since AsA_{s} is boundedly invertible by its construction, the same is true of Ds(y)D\mathcal{F}_{s}(y).

It follows that sus=Ds(us)1ss(us)\partial_{s}u_{s}=-D\mathcal{F}_{s}(u_{s})^{-1}\partial_{s}\mathcal{F}_{s}(u_{s}). Observe that

(AsDs(us)1)ss(u^s+δr)\displaystyle(A_{s}-D\mathcal{F}_{s}(u_{s})^{-1})\partial_{s}\mathcal{F}_{s}(\hat{u}_{s}+\delta_{r}) =(ADs(us))1(ADs(us)I)Asss(u^s+δr)\displaystyle=(AD\mathcal{F}_{s}(u_{s}))^{-1}(AD\mathcal{F}_{s}(u_{s})-I)A_{s}\partial_{s}\mathcal{F}_{s}(\hat{u}_{s}+\delta_{r})

By previous estimates, the above is bounded in norm by γ\gamma. Since s[s](us)s[s](u^s+δr)\partial_{s}[\mathcal{F}_{s}](u_{s})\in\partial_{s}[\mathcal{F}_{s}](\hat{u}_{s}+\delta_{r}) — see Section 4.2 — we have

susAss(u^s+δr)+Bγ(0)¯.\displaystyle\partial_{s}u_{s}\in-A_{s}\partial_{s}\mathcal{F}(\hat{u}_{s}+\delta_{r})+\overline{B_{\gamma}(0)}. (83)

Note that ~sa(t)h\tilde{\partial}_{s}a(t)h is precisely the amplitude component of s(ush)|s=t\partial_{s}(u_{s}h)|_{s=t}.

Consider the line [0,1]ts1+t(s¯s1)[0,1]\mapsto t\mapsto s_{1}+t(\overline{s}-s_{1}), where s¯\overline{s} is any point on the edge connecting s1s_{1} to s2s_{2} in Δ\Delta. Consider the function g(t)=πaus1+t(s¯s1)g(t)=\pi^{a}u_{s_{1}+t(\overline{s}-s_{1})}. By the assumptions of the proposition and the inclusion (83), we have g(0)g(1)<0g(0)g(1)<0 and ddtg(t)0\frac{d}{dt}g(t)\neq 0 for t(0,1)t\in(0,1). By the mean-value theorem, there is a unique t(0,1)t^{*}\in(0,1) such that g(t)=0g(t^{*})=0. Moreover, t=t(s¯)t^{*}=t^{*}(\overline{s}) depends continuously on s¯\overline{s}. Letting p:[0,1]Δp:[0,1]\rightarrow\Delta be a C1C^{1} parametrization of the edge connecting s1s_{1} to s2s_{2}, it follows by the definition of the map s\mathcal{F}_{s} that ΘH=tp\Theta_{H}=t^{*}\circ p is a Hopf bifurcation curve. ∎

Remark 14.

Proposition 10 is stated in such a way that we assume s0s_{0} and the edge connecting s1s_{1} to s2s_{2} lie on the opposite sides of Hopf curve ΘH\Theta_{H}. This is done almost without loss of generality. First, if the Hopf curve should bisect the simplex in such a way that s1s_{1} or s2s_{2} is separated from its opposing edge, a suitable re-labeling of the nodes transforms the problem into the form of the proposition. Second, a problem can arise if the Hopf curve intersects one of the vertices sjs_{j}, j=0,1,2j=0,1,2. Similarly, numerical difficulties can arise with the method of computer-assisted proof if the intersections of ΘH\Theta_{H} with the edges of Δ\Delta occur very close to the vertices. These problems can be alleviated by a careful selection of numerical data.

Remark 15.

We emphasize that most of the sufficient conditions of Proposition 10 can be checked using the output of the validation of a simplex. We also mention specifically that in the second point, the sign of [πa(u^s)r,πa(u^s)+r][\pi^{a}(\hat{u}_{s})-r,\pi^{a}(\hat{u}_{s})+r] need only be verified to match at s=s1s=s_{1} and s=s2s=s_{2}, due to linearity of u^s\hat{u}_{s}. The exception is the computation of ~sa\tilde{\partial}_{s}a and γ\gamma. The former is a finite computation, and the latter can be bounded in a straightforward manner; see Section 4.2 for details concerning the structure of ss\partial_{s}\mathcal{F}_{s}.

Remark 16.

A Hopf bifurcation curve necessarily generates continua of Hopf bifurcations. These can be constructed by selecting curves in Δ\Delta that are transversal to ΘH\Theta_{H}.

6.2 Degenerate Hopf bifurcations

In this section, we consider how one might prove the existence of a degenerate Hopf bifurcation, be it a bubble bifurcation or otherwise. For a first definition, we consider the bubble bifurcation.

Definition 2.

A bubble bifurcation (with quadratic fold) occurs at (α,β)(\alpha^{*},\beta^{*}) in (19) if there exists a Hopf curve ΘH\Theta_{H} with (παuΘH(t),πβuΘH(t))=(α,β)(\pi^{\alpha}u_{\Theta_{H}(t^{*})},\pi^{\beta}u_{\Theta_{H}(t^{*})})=(\alpha^{*},\beta^{*}) for some t(0,1)t^{*}\in(0,1), such that

  • the projection of ΘH\Theta_{H} into the (α,β)(\alpha,\beta) plane can be realized as a C2C^{2} graph β=β(α)\beta=\beta(\alpha).

  • β(α)=β\beta(\alpha^{*})=\beta^{*}, β(α)=0\beta^{\prime}(\alpha^{*})=0, and β′′(α)0\beta^{\prime\prime}(\alpha^{*})\neq 0.

  • there is a C2C^{2} diffeomorphism h:Dh(D)h:D\rightarrow h(D) defined on a neighbourhood DD of ΘH(t)Δ\Theta_{H}(t^{*})\in\Delta, such that h(ΘH(t))=(α,0)h(\Theta_{H}(t^{*}))=(\alpha^{*},0), and the periodic orbit yy (see Definition 1) is a steady state if and only if π2h(s)=0\pi_{2}h(s)=0, where π2\pi_{2} is the projection onto the second factor. Also, the projection π1\pi_{1} onto the first factor satisfies π1h(s)=πaus\pi_{1}h(s)=\pi^{a}u_{s} for all sDs\in D.

  • (α,0)(\alpha^{*},0) is a strict local extremum of the map (α,a)πβuh1(α,a)(\alpha,a)\mapsto\pi^{\beta}u_{h^{-1}(\alpha,a)}.

Our perspective is that a bubble bifurcation is characterized by the existence of a manifold of periodic orbits, parameterized in terms of amplitude and a parameter, such that at an isolated critical point, the projection of the manifold onto the other parameter has a local extremum. The parametrization of β\beta near α\alpha^{*} and the local minimum condition reflects the observation that a fold in the curve of Hopf bifurcations is sufficient condition for the birth of the bubble. That is, we have a family of bubbles (loops of Hopf bifurcations) that can be parameterized by β\beta in an interval of the form (βδ,β](\beta^{*}-\delta,\beta^{*}] or [β,β+δ)[\beta^{*},\beta^{*}+\delta), for δ>0\delta>0 small. This is a consequence of the (α,a)(\alpha,a)-parametrization of the simplex.

Remark 17.

We include the adjective “quadratic” to describe the fold in the Hopf curve to contrast with a related definition in Section 6.2.3, where the condition β′′(α)0\beta^{\prime\prime}(\alpha^{*})\neq 0 will be weakened.

6.2.1 Preparations

It will facilitate the presentation of the bubble bifurcation if we are able to construct the first and second derivatives of suss\mapsto u_{s}, the zeroes of (37), with respect to the simplex parameter ss. This was partially done in the proof of Proposition 10, but we will elaborate further here.

To compute the derivatives, the idea is to formally differentiate ss(us)s\mapsto\mathcal{F}_{s}(u_{s}) with respect to ss twice, allowing for the introduction of sus\partial_{s}u_{s} and s2us\partial_{s}^{2}u_{s}. The result is a system of three nonlinear equations, and solving for (us,sus,s2us)(u_{s},\partial_{s}u_{s},\partial_{s}^{2}u_{s}) amounts to computing zeroes of a map

Ω3(us,sus)(s(us),s[1](us,sus),s[2](us,sus,s2us))\displaystyle\Omega^{3}\ni(u_{s},\partial_{s}u_{s})\mapsto(\mathcal{F}_{s}(u_{s}),\mathcal{F}_{s}^{[1]}(u_{s},\partial_{s}u_{s}),\mathcal{F}_{s}^{[2]}(u_{s},\partial_{s}u_{s},\partial_{s}^{2}u_{s})) (84)
Remark 18.

It may be unclear whether the second derivative of ss(us)s\mapsto\mathcal{F}_{s}(u_{s}) can be given meaning. Indeed, as we have previously explained in Remark 10, the second Fréchet derivative of us(u)u\mapsto\mathcal{F}_{s}(u) does not, in general, exist. The (formal) second derivative of both sides of s(us)=0\mathcal{F}_{s}(u_{s})=0 produces

s2s(us)+2s[Ds](us)sus+D2s(us)[sus,sus]=0,\partial_{s}^{2}\mathcal{F}_{s}(u_{s})+2\partial_{s}[D\mathcal{F}_{s}](u_{s})\partial_{s}u_{s}+D^{2}\mathcal{F}_{s}(u_{s})[\partial_{s}u_{s},\partial_{s}u_{s}]=0,

and we can see that it is in fact possible to interpret D2s(us)[sus,sus]D^{2}\mathcal{F}_{s}(u_{s})[\partial_{s}u_{s},\partial_{s}u_{s}] as the Gateaux derivative of wDs(w)susw\mapsto D\mathcal{F}_{s}(w)\partial_{s}u_{s}, at w=usw=u_{s} and in the direction sus\partial_{s}u_{s}. However, even this may not exist as an element of Kν(n+m)K_{\nu}(\mathbb{C}^{n+m}), since the Fourier components of usu_{s} are generally not band-limited. It is therefore necessary to specify the codomain of (84) a bit more carefully. We will not elaborate on this subtlety here; the ramifications of this will be the topic of some of our future work. However, in the case of ordinary differential equations, there is no major complication. When there are no delays (or when they are all zero), s\mathcal{F}_{s} is twice continuously differentiable provided the same is true of ff.

Regardless how the partial derivatives are enclosed, the following equivalent version of Proposition 10 is available.

Proposition 11.

Assume the existence of a family of zeroes of (84) parameterized by sΔs\in\Delta and close to a numerical interpolant 𝐮^s=(u^s,su^s,s2u^s)\hat{\mathbf{u}}_{s}=(\hat{u}_{s},\partial_{s}\hat{u}_{s},\partial_{s}^{2}\hat{u}_{s}) has been proven, in the sense that we have identified r>0r>0 such that there is a unique zero of (84), for each sΔs\in\Delta, in the ball Br(𝐮^s)B_{r}(\hat{\mathbf{u}}_{s}). Assume the topolgy on this ball is such that the components a^s=πau^s\hat{a}_{s}=\pi^{a}\hat{u}_{s} satisfy

(a^s[1]as[1])hra(1)|h|,||(\hat{a}_{s}^{[1]}-a_{s}^{[1]})h||\leq r_{a}^{(1)}|h|,

for h2h\in\mathbb{R}^{2}, and a^s[0]as[0]ra(0)||\hat{a}_{s}^{[0]}-a_{s}^{[0]}||\leq r_{a}^{(0)}. Suppose

  • For j=1,2j=1,2, [a^s0[0]ra(0),a^s0[0]+ra(0)][a^sj[0]ra(0),a^sj[0]+ra(0)]<0[\hat{a}^{[0]}_{s_{0}}-r_{a}^{(0)},\hat{a}^{[0]}_{s_{0}}+r_{a}^{(0)}]\cdot[\hat{a}^{[0]}_{s_{j}}-r_{a}^{(0)},\hat{a}^{[0]}_{s_{j}}+r_{a}^{(0)}]<0, where the multiplication is in the sense of interval arithmetic.

  • The sign of [a^s[0]ra(0),a^s[0]+ra(0)][\hat{a}^{[0]}_{s}-r_{a}^{(0)},\hat{a}^{[0]}_{s}+r_{a}^{(0)}] is constant for ss on the edge of Δ\Delta not incident with s0s_{0}; that is, on the edge connecting s1s_{1} to s2s_{2}.

  • The components of the interval vector a^s[1]+ra(1)[1,1]2\hat{a}_{s}^{[1]}+r_{a}^{(1)}[-1,1]^{2} in 2\mathbb{R}^{2*} has the same sign for sΔs\in\Delta.

Then there exists a Hopf bifurcation curve ΘH:[0,1]Δ\Theta_{H}:[0,1]\rightarrow\Delta, each of ΘH(0)\Theta_{H}(0) and ΘH(1)\Theta_{H}(1) lie on the edges of Δ\Delta that are incident with s0s_{0}, and ΘH\Theta_{H} is C2C^{2}.

Comments analogous to those appearing in Remark 14 apply here as well. Note that Proposition 11 requires only the first derivatives, which can be enclosed using (83). However, if the existence of the second derivatives are in question, then ΘH\Theta_{H} can at best be proven to be C1C^{1}.

Definition 3.

A triple of line segments (v¯1,v¯,v¯2)(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) in Δ\Delta is s0s_{0}-oriented if there exist points t1,t,t2t_{1},t_{*},t_{2} on the edge (s1,s2)(s_{1},s_{2}), such that

  • v¯1\overline{v}_{1} is a subset of the line connecting s0s_{0} to t1t_{1};

  • v¯\overline{v}_{*} is a subset of the line connecting s0s_{0} to tt_{*};

  • v¯2\overline{v}_{2} is a subset of the line connecting s0s_{0} to t2t_{2};

  • under the total order on the edge (s1,s2)(s_{1},s_{2}) defined by abs1a2s1b2a\leq b\Leftrightarrow||s_{1}-a||_{2}\leq||s_{1}-b||_{2}, we have v¯1v¯v¯2\overline{v}_{1}\leq\overline{v}_{*}\leq\overline{v}_{2}.

In this case the associated wedge cover is the simplex in Δ\Delta with vertices s0,t1s_{0},t_{1} and t2t_{2}, and it is denoted W(v¯1,v¯,v¯2)W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}).

These s0s_{0}-oriented line segments will be used to enclose potential bubble bifurcation points. Some finer control can be specified with the following.

Definition 4.

Let (v¯1,v¯,v¯2)(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) be an s0s_{0}-oriented triple in Δ\Delta. Let ΘH\Theta_{H} be a Hopf curve in Δ\Delta associated to a family usu_{s} of zeroes of s\mathcal{F}_{s}. A Hopf-bounding trapezoid in W(v¯1,v¯,v¯2)W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) is a trapezoid contained in W(v¯1,v¯,v¯2)W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) whose edges include both v¯1\overline{v}_{1} and v¯2\overline{v}_{2}, and such that the other edges, denoted w¯+\overline{w}_{+} and w¯\overline{w}_{-}, satisfy πaus<0\pi^{a}u_{s}<0 for sw¯s\in\overline{w}_{-}, and πaus>0\pi^{a}u_{s}>0 for sw¯+s\in\overline{w}_{+}.

It is straightforward to verify that, under the assumptions of Proposition 11 , a trapezoid in W(v¯1,v¯,v¯2)W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) is Hopf-bounding provided two of its edges match v¯1\overline{v}_{1} and v¯2\overline{v}_{2}, while

a^s[0]+ra(0)\displaystyle\hat{a}_{s}^{[0]}+r_{a}^{(0)} <0,sw¯\displaystyle<0,\quad s\in\overline{w}_{-}
a^s[0]ra(0)\displaystyle\hat{a}_{s}^{[0]}-r_{a}^{(0)} >0,sw¯+.\displaystyle>0,\quad s\in\overline{w}_{+}.

6.2.2 Enclosure of a bubble bifurcation

The following proposition provides verifiable conditions for the existence of a bubble bifurcation. Some are more explicit than others.

Proposition 12.

Let the hypotheses of Proposition 11 be satisfied. Let (v¯1,v¯,v¯2)(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) be an s0s_{0}-oriented triple of line segments in Δ\Delta. Suppose the topology on Br(𝐮^s)B_{r}(\hat{\mathbf{u}}_{s}) is such that the components παu^s[k]=α^s[k]\pi^{\alpha}\hat{u}_{s}^{[k]}=\hat{\alpha}_{s}^{[k]}, πβu^s[k]=β^s[k]\pi^{\beta}\hat{u}_{s}^{[k]}=\hat{\beta}_{s}^{[k]}, πau^s[k]=a^s[k]\pi^{a}\hat{u}_{s}^{[k]}=\hat{a}_{s}^{[k]} satisfy

(α^s[k]αs[k])[h1,,hk]\displaystyle||(\hat{\alpha}^{[k]}_{s}-\alpha^{[k]}_{s})[h_{1},\dots,h_{k}]|| rα(k)|h1||hk|\displaystyle\leq r_{\alpha}^{(k)}|h_{1}|\cdots|h_{k}|
(β^s[k]βs[k])[h1,,hk]\displaystyle||(\hat{\beta}^{[k]}_{s}-\beta^{[k]}_{s})[h_{1},\dots,h_{k}]|| rβ(k)|h1||hk|\displaystyle\leq r_{\beta}^{(k)}|h_{1}|\cdots|h_{k}|
(a^s[k]as[k])[h1,,hk]\displaystyle||(\hat{a}^{[k]}_{s}-a^{[k]}_{s})[h_{1},\dots,h_{k}]|| ra(k)|h1||hk|\displaystyle\leq r_{a}^{(k)}|h_{1}|\cdots|h_{k}|

for kk-tuples h1,,hk2h_{1},\dots,h_{k}\in\mathbb{R}^{2}, and k=0,1,2k=0,1,2. With the empty tuple (k=0k=0), the norm reduces to absolute value on \mathbb{R}. Let 𝐫κ\mathbf{r}_{\kappa}, for κ{α,β,a}\kappa\in\{\alpha,\beta,a\} be interval vectors in 2\mathbb{R}^{2} such that (𝐫κ)j=[1,1]rκ(1)(\mathbf{r}_{\kappa})_{j}=[-1,1]r_{\kappa}^{(1)}, for j=1,2j=1,2. Finally, given sΔs\in\Delta, denote by VsV_{s} the set

Vs={v2:(a^s[1]+𝐫a)v=0,v2=1}.V_{s}=\{v\in\mathbb{R}^{2}:(\hat{a}_{s}^{[1]}+\mathbf{r}_{a})\cdot v=0,\hskip 5.69054pt||v||_{2}=1\}.

Suppose the following are satisfied.

  1. 1.

    (α^s[1]+𝐫α)vs(\hat{\alpha}_{s}^{[1]}+\mathbf{r}_{\alpha})\cdot v_{s} is bounded away from zero, for all sΔs\in\Delta and all vsVsv_{s}\in V_{s}.

  2. 2.

    infa^v¯i[0]+ra(0)<0<supa^v¯i[0]ra(0)\inf\hat{a}_{\overline{v}_{i}}^{[0]}+r_{a}^{(0)}<0<\sup\hat{a}_{\overline{v}_{i}}^{[0]}-r_{a}^{(0)} for i=1,2i=1,2.

  3. 3.

    infa^v¯[0]+ra(0)<0<supa^v¯[0]ra(0)\inf\hat{a}_{\overline{v}_{*}}^{[0]}+r_{a}^{(0)}<0<\sup\hat{a}_{\overline{v}_{*}}^{[0]}-r_{a}^{(0)}.

  4. 4.

    β^v¯[0]+rβ(0)<β^v¯i[0]rβ(0)\hat{\beta}^{[0]}_{\overline{v}_{*}}+r_{\beta}^{(0)}<\hat{\beta}^{[0]}_{\overline{v}_{i}}-r_{\beta}^{(0)} for i=1,2i=1,2.

  5. 5.

    The components of a^s[1]+𝐫a\hat{a}_{s}^{[1]}+\mathbf{r}_{a} are bounded away from zero for all sΔs\in\Delta.

  6. 6.

    The determinant of the 2×22\times 2 interval matrix

    ((α^s[1])1+[1,1]rα(1)(α^s[1])2+[1,1]rα(1)(a^s[1])1+[1,1]ra(1)(a^s[1])2+[1,1]ra(1))\displaystyle\left(\begin{array}[]{cc}(\hat{\alpha}^{[1]}_{s})_{1}+[-1,1]r_{\alpha}^{(1)}&(\hat{\alpha}^{[1]}_{s})_{2}+[-1,1]r_{\alpha}^{(1)}\\ (\hat{a}^{[1]}_{s})_{1}+[-1,1]r_{a}^{(1)}&(\hat{a}^{[1]}_{s})_{2}+[-1,1]r_{a}^{(1)}\end{array}\right) (87)

    is bounded away from zero for sΔs\in\Delta.

  7. 7.

    Defining

    cs=1a^s[1]+𝐫a2(β^s[1]+𝐫β)(a^s[1]+𝐫a),\displaystyle c_{s}=\frac{1}{||\hat{a}^{[1]}_{s}+\mathbf{r}_{a}||^{2}}(\hat{\beta}^{[1]}_{s}+\mathbf{r}_{\beta})\cdot(\hat{a}^{[1]}_{s}+\mathbf{r}_{a}), (88)

    it holds that the interval β^s[2][vs,vs]csa^s[2][vs,vs]+(rβ(2)+cs2ra(2))[1,1]\hat{\beta}^{[2]}_{s}[v_{s},v_{s}]-c_{s}\hat{a}^{[2]}_{s}[v_{s},v_{s}]+(r_{\beta}^{(2)}+||c_{s}||_{2}r_{a}^{(2)})[-1,1] is bounded away from zero for all sΔs\in\Delta and vsVsv_{s}\in V_{s}.

  8. 8.

    The matrix

    (βs[2][αs,αs]+βs[1]α2sβs[2][as,αs]+βs[1]aαsβs[2][as,αs]+βs[1]aαsβs[2][as,as]+βs[1]a2s)\displaystyle\left(\begin{array}[]{cc}\beta_{s}^{[2]}[\partial_{\alpha}s,\partial_{\alpha}s]+\beta_{s}^{[1]}\partial_{\alpha}^{2}s&\beta_{s}^{[2]}[\partial_{a}s,\partial_{\alpha}s]+\beta_{s}^{[1]}\partial_{a}\partial_{\alpha}s\\ \beta_{s}^{[2]}[\partial_{a}s,\partial_{\alpha}s]+\beta_{s}^{[1]}\partial_{a}\partial_{\alpha}s&\beta_{s}^{[2]}[\partial_{a}s,\partial_{a}s]+\beta_{s}^{[1]}\partial_{a}^{2}s\end{array}\right) (91)

    is uniformly (for sΔs\in\Delta) definite, where defining h:Δ2h:\Delta\rightarrow\mathbb{R}^{2} by h(s)=(αs,as)h(s)=(\alpha_{s},a_{s}), the partial derivatives in the matrix above are

    [αsas]\displaystyle\left[\begin{array}[]{cc}\partial_{\alpha}s&\partial_{a}s\end{array}\right] =Dh(s)1\displaystyle=Dh(s)^{-1}
    α2s\displaystyle\partial_{\alpha}^{2}s =Dh(s)1D2h(s)[αs,αs]\displaystyle=-Dh(s)^{-1}D^{2}h(s)[\partial_{\alpha}s,\partial_{\alpha}s]
    aαs\displaystyle\partial_{a}\partial_{\alpha}s =Dh(s)1D2h(s)[as,αs]\displaystyle=-Dh(s)^{-1}D^{2}h(s)[\partial_{a}s,\partial_{\alpha}s]
    a2s\displaystyle\partial_{a}^{2}s =Dh(s)1D2h(s)[as,as].\displaystyle=-Dh(s)^{-1}D^{2}h(s)[\partial_{a}s,\partial_{a}s].

Then, there exists a bubble bifurcation with quadratic fold at some (α,β)(\alpha^{*},\beta^{*}) in the projection of ΘHW(v¯1,v¯,v¯2)\Theta_{H}\cap W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) onto the (α,β)(\alpha,\beta) plane.

Proof.

By a suitable reparametrization, we may assume that ΘH(t)=1||\Theta_{H}^{\prime}(t)||=1 for all t(0,1)t\in(0,1). Denote a(s)=πausa(s)=\pi^{a}u_{s}. The definition of the Hopf curve is that a(ΘH(t))=0a(\Theta_{H}(t))=0 or all t[0,1]t\in[0,1]. As consequence, (sa)ΘH=0(\partial_{s}a)\Theta_{H}^{\prime}=0, so that ΘH\Theta_{H}^{\prime} is dual to the orthogonal complement of sa\partial_{s}a. That is, ΘHsVs\Theta_{H}^{\prime}\in\cup_{s}V_{s}. Now,

ddtαΘH(t)[0]=αΘH(t)[1]ΘHsΔ(vsVs(α^s[1]+𝐫α)vs),\frac{d}{dt}\alpha^{[0]}_{\Theta_{H}(t)}=\alpha_{\Theta_{H}(t)}^{[1]}\Theta_{H}^{\prime}\in\bigcup_{s\in\Delta}\left(\bigcup_{v_{s}\in V_{s}}(\hat{\alpha}_{s}^{[1]}+\mathbf{r}_{\alpha})\cdot v_{s}\right),

which is bounded away from zero by assumption (condition 1 of the proposition). It follows that tαΘH(t)[0]t\mapsto\alpha_{\Theta_{H}(t)}^{[0]} is monotone, so ΘH\Theta_{H} can be parameterized by α\alpha, for α\alpha in the monotone range of tαΘH(t)[0]t\mapsto\alpha_{\Theta_{H}(t)}^{[0]}. Consequently, the projection of the Hopf curve ΘH\Theta_{H} in the (α,β)(\alpha,\beta) plane can be represented as a graph β=β(α)\beta=\beta(\alpha).

Conditions 2–4 of the proposition guarantee that each of the segments s¯1\overline{s}_{1}, s¯2\overline{s}_{2} and t¯\overline{t} enjoy the following properties:

  • as one-dimensional manifolds, they enclose an intersection with the Hopf curve;

  • the β\beta-components of v¯1\overline{v}_{1} and v¯2\overline{v}_{2} are strictly greater than those of v¯\overline{v}_{*}.

As consequence, β=β(α)\beta=\beta(\alpha) possesses an internal (to its domain, relative to the previously-computed range) critical point which is a global minimizer. Let this point be α\alpha^{*}, so β(α)=defβ\beta(\alpha^{*})\,\stackrel{{\scriptstyle\mbox{\tiny{\raisebox{0.0pt}[0.0pt][0.0pt]{def}}}}}{{=}}\,\beta^{*}. Let the associated point on the simplex be ΘH(t)\Theta_{H}(t^{*}).

We wish to show that β\beta^{*} is a strict, isolated minimum of β\beta. We will do that by proving β′′(α)0\beta^{\prime\prime}(\alpha^{*})\neq 0. It is enough to prove that d2dt2βΘH(t)[0]0\frac{d^{2}}{dt^{2}}\beta^{[0]}_{\Theta_{H}(t)}\neq 0 whenever ddtβΘH(t)[0]=0\frac{d}{dt}\beta^{[0]}_{\Theta_{H}(t)}=0. If the latter is satisfied, then we have simultaneously

aΘH(t)[1]ΘH\displaystyle a_{\Theta_{H}(t)}^{[1]}\Theta_{H}^{\prime} =0\displaystyle=0
βΘH(t)[1]ΘH\displaystyle\beta_{\Theta_{H}(t)}^{[1]}\Theta_{H}^{\prime} =0.\displaystyle=0.

Since ΘH0\Theta_{H}^{\prime}\neq 0, it must be the case that αΘH(t)[1]\alpha^{[1]}_{\Theta_{H}(t)} and βΘH(t)[1]\beta_{\Theta_{H}(t)}^{[1]} are colinear. By assumption 5, αΘH(t)[1]\alpha^{[1]}_{\Theta_{H}(t)} is bounded away from zero, so the quantity csc_{s} of (88) is well-defined and

βΘH(t)[1]cΘH(t)aΘH(t)[1].\displaystyle\beta_{\Theta_{H}(t)}^{[1]}\in c_{\Theta_{H}(t)}a_{\Theta_{H}(t)}^{[1]}. (92)

On its own, (92) might not seem particular useful. However, consider that

d2dt2βΘH(t)[0]\displaystyle\frac{d^{2}}{dt^{2}}\beta^{[0]}_{\Theta_{H}(t)} =βΘH(t)[2][ΘH,ΘH]+βΘH(t)[1]ΘH′′\displaystyle=\beta_{\Theta_{H}(t)}^{[2]}[\Theta_{H}^{\prime},\Theta_{H}^{\prime}]+\beta_{\Theta_{H}(t)}^{[1]}\Theta_{H}^{\prime\prime}
0\displaystyle 0 =aΘH(t)[2][ΘH,ΘH]+aΘH(t)[1]ΘH′′,\displaystyle=a_{\Theta_{H}(t)}^{[2]}[\Theta_{H}^{\prime},\Theta_{H}^{\prime}]+a_{\Theta_{H}(t)}^{[1]}\Theta_{H}^{\prime\prime},

where the second equation comes from taking a second derivative of the definition a(ΘH(t))=0a(\Theta_{H}(t))=0 of the Hopf curve. Using the second equation together with (92), we can get the inclusion

βΘH(t)[1]ΘH′′cΘH(t)aΘH(t)[2][ΘH,ΘH],\beta_{\Theta_{H}(t)}^{[1]}\Theta_{H}^{\prime\prime}\in-c_{\Theta_{H}(t)}a_{\Theta_{H}(t)}^{[2]}[\Theta_{H}^{\prime},\Theta_{H}^{\prime}],

thereby removing the dependence on the second derivative ΘH′′\Theta_{H}^{\prime\prime} of the Hopf curve. Substituting into the expression for the second derivative of tβΘH(t)t\mapsto\beta_{\Theta_{H}(t)}, we get

d2dt2βΘH(t)[0]\displaystyle\frac{d^{2}}{dt^{2}}\beta^{[0]}_{\Theta_{H}(t)} βΘH(t)[2][ΘH,ΘH]cΘH(t)aΘH(t)[2][ΘH,ΘH]\displaystyle\in\beta_{\Theta_{H}(t)}^{[2]}[\Theta_{H}^{\prime},\Theta_{H}^{\prime}]-c_{\Theta_{H}(t)}a_{\Theta_{H}(t)}^{[2]}[\Theta_{H}^{\prime},\Theta_{H}^{\prime}]
sΔ(vsVsβ^s[2][vs,vs]csa^s[2][vs,vs]+(rβ(2)+cs2ra(2))[1,1]),\displaystyle\subset\bigcup_{s\in\Delta}\left(\bigcup_{v_{s}\in V_{s}}\hat{\beta}^{[2]}_{s}[v_{s},v_{s}]-c_{s}\hat{a}^{[2]}_{s}[v_{s},v_{s}]+(r_{\beta}^{(2)}+||c_{s}||_{2}r_{a}^{(2)})[-1,1]\right),

which is bounded away from zero by condition 7. Therefore, β′′(α)0\beta^{\prime\prime}(\alpha^{*})\neq 0.

Next, we need to verify the local parametrization of the simplex near ΘH(t)\Theta_{H}(t^{*}) in terms of (α,a)(\alpha,a). This is a fairly direct consequence of the inverse function theorem, using the condition 6 of the proposition. This shows that the function hh defined in condition 8 of the proposition defines a local diffeomorphism near ΘH(t)\Theta_{H}(t^{*}). The Hessian of Γ:(α,a)βh1(α,a)\Gamma:(\alpha,a)\mapsto\beta_{h^{-1}(\alpha,a)} can be computed by implicit differentiation; if h(s)=(α,a)h(s)=(\alpha,a), then the Hessian is precisely the 2×22\times 2 matrix of condition 8. Since this matrix is uniformly definite on the simplex Δ\Delta, every critical point must be a local extremum. We already know that αΓ(α,0)=β(α)=0\partial_{\alpha}\Gamma(\alpha^{*},0)=\beta^{\prime}(\alpha^{*})=0. The other partial derivative aΓ(α,0)\partial_{a}\Gamma(\alpha^{*},0) is zero due to the amplitude symmetry of periodic orbits. Therefore, (α,0)(\alpha^{*},0) is a strict local extremum of the map (α,a)πβuh1(α,a)(\alpha,a)\mapsto\pi^{\beta}u_{h^{-1}(\alpha,a)}. ∎

Remark 19.

Checking conditions 2–6 of the proposition is fairly routine. Conditions 1 and 7, however, deserve some extra attention. If the step size is small, we should expect the derivatives of the solution suss\mapsto u_{s} to be close to constant. In this way, the set VsV_{s} should not vary too much (in a Hausdorff sense). VsV_{s} geometrically consists of two arcs on the unit circle, and we expect the angles defining these arcs to be nearly constant over the simplex. Consequently, the interval computations of conditions 1 and 7 are, indeed, implementable, but the feasibility of the checks — that these intervals are bounded away from zero — will be strongly influenced by the size of the radius, rr, and any weighting in the norm. As for condition 8, we have not included all of the associated radii, but since all of the intermediary derivatives appearing in the matrix (91) admit (by assumption) rigorous enclosures, the matrix is implementable. Therefore, condition 8 can be checked using a suitable package to compute eigenvalues of interval matrices.

Remark 20.

Condition 4 is formulated in such a way that β\beta^{*} is a local minimum of the curve β=β(α)\beta=\beta(\alpha). This condition can be reformulated in a straightforward way to allow it instead to be a local maximum. Also, comments analogous to those appearing in Remark 14 apply here as well.

Corollary 13.

It suffices to verify the conditions 6,7,8 of Proposition 12 for all ss in a given Hopf-bounding trapezoid in W(v¯1,v¯,v¯2)W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}).

6.2.3 A degenerate Hopf bifurcation without second derivatives

In Remark 18, we pointed out that, unfortunately, the lack of second-differentiability of the map s\mathcal{F}_{s} is a serious obstruction to computing second derivatives s2us\partial_{s}^{2}u_{s} and, consequently, checking all the conditions of Proposition 12. While it is not a problem when all of the delays are zero (i.e. an ODE), we would like to provide a constructive result for delay equations. Along these lines, let us slightly weaken Definition 2.

Definition 5.

A degenerate Hopf bifurcation occurs at (α,β)(\alpha^{*},\beta^{*}) in (19) if there exists a Hopf curve ΘH\Theta_{H} with (παuΘH(t),πβuΘH(t))=(α,β)(\pi^{\alpha}u_{\Theta_{H}(t^{*})},\pi^{\beta}u_{\Theta_{H}(t^{*})})=(\alpha^{*},\beta^{*}) for some t(0,1)t^{*}\in(0,1), such that

  • the projection of ΘH\Theta_{H} into the (α,β)(\alpha,\beta) plane can be realized as a C1C^{1} graph β=β(α)\beta=\beta(\alpha);

  • β(α)=β\beta(\alpha^{*})=\beta^{*} and β(α)=0\beta^{\prime}(\alpha^{*})=0;

  • there is a C1C^{1} diffeomorphism h:Dh(D)h:D\rightarrow h(D) defined on a neighbourhood DD of ΘH(t)Δ\Theta_{H}(t^{*})\in\Delta, such that h(ΘH(t))=(α,0)h(\Theta_{H}(t^{*}))=(\alpha^{*},0), and the periodic orbit zz (see Definition 1) is a steady state if and only if π2h(s)=0\pi_{2}h(s)=0, where π2\pi_{2} is the projection onto the second factor. Also, the projection π1\pi_{1} onto the first factor satisfies and π1h(s)=πaus\pi_{1}h(s)=\pi^{a}u_{s} for all sDs\in D.

The main difference between the above and Definition 2 is we no longer require that β\beta^{*} is an extremum of β\beta. We also do not make any specifications concerning the geometry of the implicit map (α,a)β(\alpha,a)\mapsto\beta near the Hopf bifurcation curve. Clearly, a bubble bifurcation with quadratic fold satisfies the conditions of the above definition. However, the new definition permits other types of degenerate Hopf bifurcations, including Bautin bifurcations. The following proposition can now be proven using the same ideas as Proposition 12.

Proposition 14.

Assume a family of zeroes usu_{s} of the map s\mathcal{F}_{s} has been proven, in addition to the first derivatives sus\partial_{s}u_{s}, close to a numerical interpolant 𝐮^s=(u^s,su^s)\hat{\mathbf{u}}_{s}=(\hat{u}_{s},\partial_{s}\hat{u}_{s}), in the sense that we have identified r>0r>0 such that there is a unique zero of (84), for each sΔs\in\Delta, in the ball Br(𝐮^s)B_{r}(\hat{\mathbf{u}}_{s}). Suppose the topology on Br(𝐮^s)B_{r}(\hat{\mathbf{u}}_{s}) is such that the components παu^s[k]=α^s[k]\pi^{\alpha}\hat{u}_{s}^{[k]}=\hat{\alpha}_{s}^{[k]}, πβu^s[k]=β^s[k]\pi^{\beta}\hat{u}_{s}^{[k]}=\hat{\beta}_{s}^{[k]}, πau^s[k]=a^s[k]\pi^{a}\hat{u}_{s}^{[k]}=\hat{a}_{s}^{[k]} satisfy

(α^s[k]αs[k])[h1,,hk]\displaystyle||(\hat{\alpha}^{[k]}_{s}-\alpha^{[k]}_{s})[h_{1},\dots,h_{k}]|| rα(k)|h1||hk|\displaystyle\leq r_{\alpha}^{(k)}|h_{1}|\cdots|h_{k}|
(β^s[k]βs[k])[h1,,hk]\displaystyle||(\hat{\beta}^{[k]}_{s}-\beta^{[k]}_{s})[h_{1},\dots,h_{k}]|| rβ(k)|h1||hk|\displaystyle\leq r_{\beta}^{(k)}|h_{1}|\cdots|h_{k}|
(a^s[k]as[k])[h1,,hk]\displaystyle||(\hat{a}^{[k]}_{s}-a^{[k]}_{s})[h_{1},\dots,h_{k}]|| ra(k)|h1||hk|\displaystyle\leq r_{a}^{(k)}|h_{1}|\cdots|h_{k}|

for kk-tuples h1,,hk2h_{1},\dots,h_{k}\in\mathbb{R}^{2}, and k=0,1k=0,1. With the empty tuple (k=0k=0), the norm reduces to absolute value on \mathbb{R}. Let (v¯1,v¯,v¯2)(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) be an s0s_{0}-oriented triple of line segments in Δ\Delta. Suppose conditions 1–3, 5 and 6 of Proposition 12 are satisfied. Then, there exists a degenerate Hopf bifurcation at some (α,β)(\alpha^{*},\beta^{*}) in the projection of ΘHW(v¯1,v¯,v¯2)\Theta_{H}\cap W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}) onto the (α,β)(\alpha,\beta) plane.

Corollary 15.

It suffices to verify condition 6 of Proposition 14 for all ss in a Hopf-bounding trapezoid in W(v¯1,v¯,v¯2)W(\overline{v}_{1},\overline{v}_{*},\overline{v}_{2}).

7 Examples

7.1 Extended Lorenz-84 system

The extended Lorenz-84 system is the following system of four coupled ODEs:

u˙1\displaystyle\dot{u}_{1} =u22u32au1afbu42\displaystyle=-u_{2}^{2}-u_{3}^{2}-au_{1}-af-bu_{4}^{2}
u˙2\displaystyle\dot{u}_{2} =u1u2cu1u3u2+d\displaystyle=u_{1}u_{2}-cu_{1}u_{3}-u_{2}+d
u˙3\displaystyle\dot{u}_{3} =cu1u2+u1u3u3\displaystyle=cu_{1}u_{2}+u_{1}u_{3}-u_{3}
u˙4\displaystyle\dot{u}_{4} =eu4+bu4u1+μ\displaystyle=-eu_{4}+bu_{4}u_{1}+\mu

We consider the parameters a=0.25a=0.25, b=0.987b=0.987, d=0.25d=0.25, e=1.04e=1.04, f=2f=2 to be fixed, while μ\mu and cc are treated as parameters.

We started the continuation at c=1c=1, close to a Hopf bifurcation. Using a step size of 0.020.02, we generated an approximate triangulation of the manifold with 11928 simplices (including simplices created by adaptive refinement needed for proofs). We used N=5N=5 Fourier modes. To capture a “wider” section of the manifold, we restricted the simplex growth to amplitude in the interval [0.1,0.3][-0.1,0.3]. Figure 6 is a plot of the triangulation, projected into amplitude and parameter space, while we restricted to the zero amplitude plane in Figure 7 to generate a plot of the Hopf bifurcation curve. The former figure allows for visualization of the traditional square-root amplitude curvature near the Hopf bifurcation curve. Interesting, far from the bifurcation curve, there appears to be a near-circular “hole” in the manifold. We have not studied its structure in detail, and have no insight into its significance. The restriction of the amplitude to [0.1,0.3][-0.1,0.3] results in the top and bottom edges appearing “ragged”, since simplices can not organically grow to produce hexagon patches. The latter Figure 7 indicates the presence of three bubble bifurcations.

Refer to caption
Figure 6: The manifold of (proven) periodic orbits in the extended Lorenz-84 system, in the projection of amplitude and the parameters c,μc,\mu. Note the square-root curvature of amplitude near μ0.04\mu\approx 0.04 (right side of plot), indicative of Hopf bubbles. There are over 10,000 simplices, so to provide a clean figure, we have not plotted the edges.
Refer to caption
Figure 7: Intersection of the surface in Figure 6 with the amplitude zero plane, which corresponds to the Hopf bifurcation curve. Hopf bubbles are present in the former figure near μ0.04\mu\approx 0.04, so the fold in the present figure likely corresponds to a bubble bifurcation with quadratic fold. There is a symmetric fold near μ0.04\mu\approx-0.04, and with respect to the other variable near μ0\mu\approx 0, for a total of three bubble bifurcations.

7.2 Time-delay SI model

We consider the time-delay SI model

y˙(t)\displaystyle\dot{y}(t) =y(t)+R0epy(tτ)y(t)(1y(t)),\displaystyle=-y(t)+R_{0}e^{-py(t-\tau)}y(t)(1-y(t)),

for which there is an analytical proof of a bubble bifurcation [17] near (R0,p)(2.1474,1.6617)(R_{0},p)\approx(2.1474,1.6617), for delay τ=10\tau=10. We will replicate this analysis using our rigorous two-parameter continuation.

7.2.1 Set-up

We begin by desingularizing the vector field. Writing y=x+ay~y=x+a\tilde{y}, for xx being a steady state, we get, dropping the tilde,

y˙=y+g(a,yτp)R0epxx(1x)+R0epxeapyτ(ay2+y(12x)),\displaystyle\dot{y}=-y+g(a,y_{\tau}p)R_{0}e^{-px}x(1-x)+R_{0}e^{-px}e^{-apy_{\tau}}(-ay^{2}+y(1-2x)),

where yτ=y(tτ)y_{\tau}=y(t-\tau), and gg is defined by

g(a,y)\displaystyle g(a,y) ={a1(eay1),a0y,a=0.\displaystyle=\left\{\begin{array}[]{ll}a^{-1}(e^{-ay}-1),&a\neq 0\\ -y,&a=0.\end{array}\right.

Observe, yg(a,y)=eay\partial_{y}g(a,y)=-e^{-ay} and

g(a,y)=yk=11k!(ay)k1.g(a,y)=-y\sum_{k=1}^{\infty}\frac{1}{k!}(-ay)^{k-1}.

gg is indeed analytic. Whenever gg (or its derivatives) must be rigorously evaluated, we construct Taylor polynomials of sufficiently high order and propagate error from the remainder accordingly.

Now we polynomialize. Define z2=eapyz_{2}=e^{-apy} and z1=g(a,yp)z_{1}=g(a,yp). Then

z˙1\displaystyle\dot{z}_{1} =pz2(y+R0epxz1(tτ)x(1x)+R0epxz2(tτ)(ay2+y(12x)))\displaystyle=-pz_{2}\big{(}-y+R_{0}e^{-px}z_{1}(t-\tau)x(1-x)+R_{0}e^{-px}z_{2}(t-\tau)(-ay^{2}+y(1-2x))\big{)}
z˙2\displaystyle\dot{z}_{2} =apz2(y+R0epxz1(tτ)x(1x)+R0epxz2(tτ)(ay2+y(12x)))\displaystyle=-apz_{2}\big{(}-y+R_{0}e^{-px}z_{1}(t-\tau)x(1-x)+R_{0}e^{-px}z_{2}(t-\tau)(-ay^{2}+y(1-2x))\big{)}

They can be more compactly written as z˙1=z2z˙0\dot{z}_{1}=-z_{2}\dot{z}_{0} and z˙2=apz2z˙0\dot{z}_{2}=-apz_{2}\dot{z}_{0}. We also have the implied boundary conditions

z1(0)\displaystyle z_{1}(0) =g(a,z0(0)p)\displaystyle=g(a,z_{0}(0)p)
z2(0)\displaystyle z_{2}(0) =eapz0(0),\displaystyle=e^{-apz_{0}(0)},

where z0=yz_{0}=y. We need two unfolding parameters to compensate for the two extra boundary conditions.

Lemma 16.

If zz is a periodic solution of

z˙0(t)\displaystyle\dot{z}_{0}(t) =z0(t)+R0μz1(tτ)x(1x)+R0μz2(tτ)(az0(t)2+z0(t)(12x))\displaystyle=-z_{0}(t)+R_{0}\mu z_{1}(t-\tau)x(1-x)+R_{0}\mu z_{2}(t-\tau)(-az_{0}(t)^{2}+z_{0}(t)(1-2x))
z˙1(t)\displaystyle\dot{z}_{1}(t) =pz2(t)z˙0(t)+η1\displaystyle=-pz_{2}(t)\dot{z}_{0}(t)+\eta_{1}
z˙2(t)\displaystyle\dot{z}_{2}(t) =apz2(t)z˙0(t)+η2\displaystyle=-apz_{2}(t)\dot{z}_{0}(t)+\eta_{2}

for some η1,η2\eta_{1},\eta_{2}\in\mathbb{R}, and zz satisfies z1(0)=g(a,z0(0)p)z_{1}(0)=g(a,z_{0}(0)p) and z2(0)=eapz0(0)z_{2}(0)=e^{-apz_{0}(0)}, then η1=η2=0\eta_{1}=\eta_{2}=0.

Proof.

First, suppose η20\eta_{2}\neq 0. Then necessarily, z2z_{2} has constant sign because the differential equation for z2z_{2} is affine-linear and η20\eta_{2}\neq 0. Since z2(0)>0z_{2}(0)>0, we must have z2>0z_{2}>0. But this means that

z˙2(t)z2(t)+apz˙0(t)=η2z2(t),\frac{\dot{z}_{2}(t)}{z_{2}(t)}+ap\dot{z}_{0}(t)=\frac{\eta_{2}}{z_{2}(t)},

a contradiction, since the left side is periodic and η20\eta_{2}\neq 0. We may therefore assume that η2=0\eta_{2}=0. Then z˙2(t)=apz2(t)z˙0(t)\dot{z}_{2}(t)=apz_{2}(t)\dot{z}_{0}(t), and it follows again that z2>0z_{2}>0. But this means

z˙1(t)z2(t)+pz˙0(t)=η1z2(t),\frac{\dot{z}_{1}(t)}{z_{2}(t)}+p\dot{z}_{0}(t)=\frac{\eta_{1}}{z_{2}(t)},

and since the left-hand side is periodic, it follows that η1=0\eta_{1}=0. ∎

To complete the polynomial embedding, we further polynomialize the parameters. This is done to ensure compatibility with the numerical implementation. We define μ=R0epx\mu=R_{0}e^{-px}, so that the complete polynomialized vector field is

z˙0(t)\displaystyle\dot{z}_{0}(t) =z0(t)+μz1(tτ)x(1x)+μz2(tτ)(az0(t)2+z0(t)(12x))\displaystyle=-z_{0}(t)+\mu z_{1}(t-\tau)x(1-x)+\mu z_{2}(t-\tau)(-az_{0}(t)^{2}+z_{0}(t)(1-2x))
z˙1(t)\displaystyle\dot{z}_{1}(t) =pz2(t)z˙0(t)+η1\displaystyle=-pz_{2}(t)\dot{z}_{0}(t)+\eta_{1}
z˙2(t)\displaystyle\dot{z}_{2}(t) =apz2(t)z˙0(t)+η2.\displaystyle=-apz_{2}(t)\dot{z}_{0}(t)+\eta_{2}.

The complete set of boundary conditions is

0\displaystyle 0 =z1(0)g(a,z0(0)p)\displaystyle=z_{1}(0)-g(a,z_{0}(0)p)
0\displaystyle 0 =z2(0)eapz0(0)\displaystyle=z_{2}(0)-e^{-apz_{0}(0)}
0\displaystyle 0 =μR0epx\displaystyle=\mu-R_{0}e^{-px}

In the terminology of Remark 4, the embedding dimension is m=3m=3. The steady-state equation is scalar, and is given by

0=x+μx(1x).0=-x+\mu x(1-x).

7.2.2 Results

We validated a patch of manifold initially with 406 simplices at a step size of 5×1065\times 10^{-6}. In the validation of nearly every simplex, three layers of adaptive refinement were needed to keep the Y0Y_{0} bound under control. We have plotted the manifold in Figure 8 without the refinements included. The projection into the (R0,p)(R_{0},p) plane is provided in Figure 9. The results are consistent with the analysis of Leblanc [17].

Refer to caption
Figure 8: The manifold of (proven) periodic orbits in the time delay SI model. A very small step size was necessary to get proofs without too many adaptive refinement steps. The Y0Y_{0} bound is the clear bottleneck.
Refer to caption
Figure 9: Intersection of the surface from Figure 8 with the amplitude zero, which corresponds to a Hopf bifurcation curve in the time-delay SI model. The validation radius is 2×1052\times 10^{-5} over the entire manifold, so the location R02.1474R_{0}\approx 2.1474 of the bubble bifurcation is consistent with the analysis of Leblanc. A tighter validation radius could be obtained with a smaller step size. Because of how the curve is plotted, skew simplices from them 2-manifold have the effect of making the curve appear “thicker” in some parts than others.

7.3 FitzHugh-Nagumo equation

The FitzHugh-Nagumo ODE is

u˙\displaystyle\dot{u} =u(uα)(1u)w+I\displaystyle=u(u-\alpha)(1-u)-w+I
w˙\displaystyle\dot{w} =ϵ(uγw)\displaystyle=\epsilon(u-\gamma w)

for scalar parameters α,ϵ,γ,I\alpha,\epsilon,\gamma,I. It is a cubic vector field in the state variables, and numerical simulations suggest the existence of Hopf bubbles (see Section 5.8 of [7]). We fix α=0.1,γ=1\alpha=0.1,\,\gamma=1, while leaving ϵ\epsilon and II as parameters for the continuation. We start the continuation near (ϵ,I)=(0.3,0.3)(\epsilon,I)=(0.3,0.3) and compute a triangulation of the manifold with 9006 simplices (including those needed for adaptive refinement) at step size 0.01. For this example, we used N=7N=7 Fourier modes. A plot of the proven simplices from the manifold is provided in Figure 10. The Hopf bifurcation curve appears in Figure 11.

Refer to caption
Refer to caption
Figure 10: Left: Projection of the manifold of periodic orbits for the FitzHugh-Nagumo equation into ϵ×I×amplitude\epsilon\times I\times\mbox{amplitude} space. At this level of scaling, the curvature of the manifold is not easily visible. Right: zoomed-in portion near the bubble bifurcation. Here, the curvature is more easily seen. There are over 9000 simplices in the left figure, so we have not plotted the edges.
Refer to caption
Figure 11: Intersection of the surface from the left pane of Figure 10 with the amplitude zero plane. This corresponds to the Hopf bifurcation curve for the Fitzhugh-Nagumo system. The curve resembles a parabola, with a bubble bifurcation at its vertex.

7.4 An ODE with a periodic orbit 2-manifold resembling a fish

Consider the three-dimensional ODE system

y˙1\displaystyle\dot{y}_{1} =βy1y2y1(y12+y22+y32+α2)\displaystyle=\beta y_{1}-y_{2}-y_{1}(y_{1}^{2}+y_{2}^{2}+y_{3}^{2}+\alpha^{2}) (93)
y˙2\displaystyle\dot{y}_{2} =y1+βy2y2(y12+y22+ϵy32+α2)\displaystyle=y_{1}+\beta y_{2}-y_{2}(y_{1}^{2}+y_{2}^{2}+\epsilon y_{3}^{2}+\alpha^{2}) (94)
y˙3\displaystyle\dot{y}_{3} =y35+3y330.01y3+0.1α+0.01(y12+y22),\displaystyle=-y_{3}^{5}+3y_{3}^{3}-0.01y_{3}+0.1\alpha+0.01(y_{1}^{2}+y_{2}^{2}), (95)

for parameters α,β\alpha,\beta, and a real control parameter ϵ\epsilon. When ϵ=1\epsilon=1, a change of variables to cylindrical coordinates shows that periodic orbits are in one-to-one correspondence with solutions (r,z)(r,z) of the set of algebraic equations

0\displaystyle 0 =βr2α2z2\displaystyle=\beta-r^{2}-\alpha^{2}-z^{2}
0\displaystyle 0 =z5+3z30.01z+0.1α+0.01r2.\displaystyle=-z^{5}+3z^{3}-0.01z+0.1\alpha+0.01r^{2}.

When ϵ1\epsilon\neq 1, the radial symmetry in (y1,y2)(y_{1},y_{2}) is broken and this change of variables is no longer informative. We set ϵ=0.8\epsilon=0.8 in (93)–(95) and used our validated continuation scheme to rigorously compute a 2-manifold of periodic orbits. In the projection of amplitude and parameters (α,β)(\alpha,\beta), the result is a figure that qualitatively resembles an angelfish. See Figure 12. In this projection, the manifold has several folds and appears to exhibit a singularity where it pinches onto a single point. Plotting the manifold in a different projection more clearly allows us to see that this singularity is merely an artifact of the projection; see Figure 13. The Hopf bifurcation curve is plotted in Figure 14. For this example, we used N=9N=9 Fourier modes and a step size 0.01. We computed a comparatively small portion of the manifold, since the interesting geometry was localized close to (α,β)=(0,0)(\alpha,\beta)=(0,0). We computed and validated 1007 simplices. This example did not require any adaptive refinement.

Refer to caption
Refer to caption
Figure 12: Left: Projection of the manifold of periodic orbits for the ODEs (93)–(95) into amplitude×α×β\mbox{amplitude}\times\alpha\times\beta space, viewed from the α\alpha axis. Note the “pinching” of the manifold, at which this projection becomes singular, near β103\beta\approx 10^{-3}. Right: an illustration of an angelfish, for comparison.
Refer to caption
Refer to caption
Figure 13: Left: Projection of the manifold of periodic orbits for the ODEs (93)–(95) into amplitude×α×β\mbox{amplitude}\times\alpha\times\beta space. There are Hopf bubbles prior to the pinching phenomenon that happens near β103\beta\approx 10^{-3}. Right: projection into α×β×y3(0)\alpha\times\beta\times y_{3}(0) space, where y3y_{3} denotes the third component of the periodic orbit in the blown-up coordinates. The pinching point in the left figure projection is caused by a pair of simultaneous folds, clearly visible in the right figure projection.
Refer to caption
Figure 14: Intersection of the surface from the left pane of Figure 13 with the amplitude zero, which corresponds to a Hopf bifurcation curve for the ODE system (93)–(95). There is a bubble bifurcation at (α,β)=(0,0)(\alpha,\beta)=(0,0). The apparent self-intersection of the Hopf curve is a consequence of the projection, and does not represent a bifurcation point. Because of how the curve is plotted, skew simplices from them 2-manifold have the effect of making the curve appear “thicker” in some parts than others.

8 Discussion

We have proposed validated continuation as an alternative way of exploring degenerate Hopf bifurcations. In combination with rigorous numerics and additional a posteriori post-processing, one can prove the existence of Hopf bifurcation curves and bubble bifurcations. The library BiValVe is rather flexible, and with the additions of the present paper, can handle multiparameter continuation problems for periodic orbits in ordinary and delay differential equations, near and far from Hopf bifurcations.

Without access to second derivatives of solutions of the zero-finding problem (37), it is difficult to prove bubble bifurcations with quadratic folds. That is, we are only able to prove the weaker characterization of Definition 5. This is a major barrier in applying the method to delay equations. We believe that a suitable re-formulation of the zero-finding problem, taking into account the additional unbounded operators that result from derivatives with respect to frequency, could resolve the issue.

Another way to prove the “quadratic fold” part of the bubble bifurcation, would be to compute the second derivatives of Hopf bifurcation curves without using the machinery of sequences spaces. Along these lines, it would be interesting to use pseudo-arclength continuation to continue Hopf bifurcation curves directly. Computer-assisted proofs of isolated Hopf bifurcations in delay differential equations are completed in [5], and with minimal changes, pseudo-arclength continuation could be used to do continuation of Hopf bifurcations. The map from [5] is finite-dimensional and as smooth as the delay vector field, so the derivatives of the Hopf curve could be rigorously computed that way instead. However, this trick can not be used to prove that (α,a)πβuh1(α,a)(\alpha,a)\mapsto\pi^{\beta}u_{h^{-1}(\alpha,a)} has a strict local extremum at the bifurcation point. Indeed, the latter map is defined in terms of the periodic orbits themselves, rather than the algebraic properties of the vector field and the eigenvalues of the linearization.

As remarked in [28], the Z2Z_{2} bound associated to delay periodic orbit validation suffers from a fundamental limitation: it scales linearly with respect to the number of Fourier modes. Therefore, while we have not needed to use many Fourier modes in our examples, it would be very costly (or infeasible) to do continuation of a periodic orbit that required many modes to represent. This is because the Y0Y_{0} bound is naturally dependent on step size, so even if an isolated solution has an exceptionally good numerical defect, a very small step size might be needed to hedge against a large Z2Z_{2}. In this way, while we can compute manifolds of periodic orbits with delay near (degenerate) Hopf bifurcations, we expect that in large-amplitude regimes or for complicated orbits, completing a validation would be difficult. To compare, the situation is far better for ordinary differential equations. The Z2Z_{2} bound is generally unharmed by having many Fourier modes, and second derivatives of the solutions can be computed by solving an auxiliary zero-finding problem using similar techniques from rigorous numerics.

There are other codimension-2+ bifurcations that could be studied from the point of view of validated multi-parameter continuation. For example, the cusp bifurcation should be amenable to this type of analysis, and is simpler than the present work because it involves only bifurcations of fixed points rather than periodic orbits. There is also the Bautin bifurcation, for which the analysis of Section 6 could be replicated. In fact, our continuation scheme is able to validate manifolds of periodic orbits passing through Bautin points. As a very brief final example, recall the normal form (3)–(4), which has a Bautin bifurcation at (x,y)=(0,0)(x,y)=(0,0) at the parameters (α,β)=(0,0)(\alpha,\beta)=(0,0). Periodic orbits in this ODE are equivalent (by a change of variables to polar coordinates) to scalar solutions rr of

0=r(β+αr2r4).\displaystyle 0=r(\beta+\alpha r^{2}-r^{4}). (96)

Agnostic to this particular representation of the zeroes, our code is able to validate a large section of the manifold of periodic orbits directly from the ODEs. See Figure 15. As expected, we were able to validated this manifold using very few Fourier modes: three, in this case.

Refer to caption
Figure 15: Proven section of the manifold of periodic orbits associated to the Bautin normal form. Note that the amplitude is in fact equal to rr from equation (96).

Hopf bubbles have been observed in the Mackey-Glass equation [15] at the classical parameters, and some of our preliminary investigations suggest that the equation possesses a bubble bifurcation. It would be interesting to use a combination of polynomial embedding and blow-up to investigate this bifurcation. However, the added complexity of using both blow-up and polynomial embedding presents a challenge; the resulting (polynomial) delay vector field ends up being high-order with dozens of distinct nonlinear terms.

Acknowledgments

Kevin E. M. Church is supported by a CRM-Simons Postdoctoral Fellowship. Elena Queirolo is supported by Walter Benjamin Programme DFG-Antrag QU579/1-1.

References

  • [1] Peter Ashwin, Stephen Coombes, and Rachel Nicks. Mathematical Frameworks for Oscillatory Network Dynamics in Neuroscience. The Journal of Mathematical Neuroscience, 6(1):2, dec 2016.
  • [2] N. Bautin. Behaviour of dynamical systems near the boundaries of stability regions. OGIZ GOSTEXIZDAT. Leningrad. In Russian. 1949.
  • [3] Peter A. Braza. The Bifurcation Structure of the Holling–Tanner Model for Predator-Prey Interactions Using Two-Timing. SIAM Journal on Applied Mathematics, 63(3):889–904, jan 2003.
  • [4] Kevin E.M. Church and Elena Queirolo. BiValVe: Bifurcation Validation Venture. https://github.com/elenaquei/bubbles/releases/tag/v1
  • [5] Kevin E.M. Church and Jean-Philippe Lessard. Rigorous verification of Hopf bifurcations in functional differential equations of mixed type. Physica D: Nonlinear Phenomena, 429:133072, jan 2022.
  • [6] Michael G. Crandall and Paul H. Rabinowitz. The Hopf Bifurcation Theorem in infinite dimensions. Archive for Rational Mechanics and Analysis, 67(1):53–72, 1977.
  • [7] Geneviève Dupont, Martin Falcke, Vivien Kirk, and James Sneyd. Models of Calcium Signalling, volume 43 of Interdisciplinary Applied Mathematics. Springer International Publishing, Cham, 2016.
  • [8] Hassan A. El-Morshedy and Alfonso Ruiz-Herrera. Asymptotic convergence in delay differential equations arising in epidemiology and physiology. SIAM Journal on Applied Mathematics, 81(4):1781–1798, 2021.
  • [9] Teresa Faria and Luis T. Magalhães. Normal forms for retarded functional differential equations with parameters and applications to hopf bifurcation, 1995.
  • [10] Marcio Gameiro, Jean Philippe Lessard, and Alessandro Pugliese. Computation of Smooth Manifolds Via Rigorous Multi-parameter Continuation in Infinite Dimensions. Foundations of Computational Mathematics, 16(2), 2016.
  • [11] Martin Golubitsky and William F. Langford. Classification and unfoldings of degenerate Hopf bifurcations. Journal of Differential Equations, 41(3):375–415, 1981.
  • [12] Brian Hassard and Katie Jiang. Degenerate Hopf bifurcation and isolas of periodic solutions in an enzyme-catalyzed reaction model. Journal of Mathematical Analysis and Applications, 177:170–189, 1993.
  • [13] Weihua Jiang, Qi An, and Junping Shi. Formulation of the normal form of Turing-Hopf bifurcation in partial functional differential equations. Journal of Differential Equations, 268(10):6067–6102, 2020.
  • [14] Yuri A Kuznetsov. Elements of Applied Bifurcation Theory. Springer New York, 2004.
  • [15] Tibor Krisztin and Eduardo Liz. Bubbles for a Class of Delay Differential Equations. Qualitative Theory of Dynamical Systems, 10(2):169–196, oct 2011.
  • [16] Joe Latulippe, Derek Lotito, and Donovan Murby. A mathematical model for the effects of amyloid beta on intracellular calcium. PLOS ONE, 13(8):e0202503, aug 2018.
  • [17] Victor G. LeBlanc. A Degenerate Hopf Bifurcation in Retarded Functional Differential Equations, and Applications to Endemic Bubbles. Journal of Nonlinear Science, 26(1):1–25, 2016.
  • [18] Maoxing Liu, Eduardo Liz, and Gergely Röst. Endemic bubbles generated by delayed behavioral response: Global stability and bifurcation switches in an sis model. SIAM Journal on Applied Mathematics, 75(1):75–91, 2015.
  • [19] Zhihua Liu, Pierre Magal, and Shigui Ruan. Hopf bifurcation for non-densely defined Cauchy problems. Zeitschrift fur Angewandte Mathematik und Physik, 62(2):191–222, 2011.
  • [20] Stephen B. Margolis and Bernard J. Matkowsky. New Modes of Quasi-Periodic Combustion Near a Degenerate Hopf Bifurcation Point. SIAM Journal on Applied Mathematics, 48(4):828–853, aug 1988.
  • [21] J. E. Marsden and M. McCracken. The Hopf Bifurcation and Its Applications, volume 19 of Applied Mathematical Sciences. Springer New York, New York, NY, 1976.
  • [22] Joseph Minicucci, Molly Alfond, Angelo Demuro, David Gerberry, and Joe Latulippe. Quantifying the dose-dependent impact of intracellular amyloid beta in a mathematical model of calcium regulation in xenopus oocyte. PLOS ONE, 16(1):e0246116, jan 2021.
  • [23] Richmond Opoku-sarkodie and Ferenc A Bartha. Dynamics of an SIRWS model with waning of immunity and varying immune boosting period. pages 1–26, 2022.
  • [24] J. P. Lessard and J. D. Mireles James. A functional analytic approach to validated numerics for eigenvalues of delay equations. Journal of Computational Dynamics, 7(1):123–158, 2020.
  • [25] Aldo Rustichini. Hopf bifurcation for functional differential equations of mixed type. Journal of Dynamics and Differential Equations, 1(2):145–177, apr 1989.
  • [26] N. Sherborne, K. B. Blyuss, and I. Z. Kiss. Bursting endemic bubbles in an adaptive network. Physical Review E, 97(4):042306, apr 2018.
  • [27] James Sneyd, Shawn Means, Di Zhu, John Rugis, Jong Hak Won, and David I. Yule. Modeling calcium waves in an anatomically accurate three-dimensional parotid acinar cell. Journal of Theoretical Biology, 419:383–393, apr 2017.
  • [28] Jan Bouwe van den Berg, Chris Groothedde, and Jean Philippe Lessard. A General Method for Computer-Assisted Proofs of Periodic Solutions in Delay Differential Problems. Journal of Dynamics and Differential Equations, 2020.
  • [29] Jan Bouwe van den Berg Elena Queirolo. A general framework for validated continuation of periodic orbits in systems of polynomial ODEs. Journal of Computational Dynamics. 8(1):59, 2021.
  • [30] Jan Bouwe van den Berg, Jean-Philippe Lessard, and Elena Queirolo. Rigorous Verification of Hopf Bifurcations via Desingularization and Continuation. SIAM Journal on Applied Dynamical Systems, 20(2):573–607, jan 2021.
  • [31] Xiaoli Wang, Junping Shi, and Guohong Zhang. Interaction between water and plants: Rich dynamics in a simple model. Discrete & Continuous Dynamical Systems - B, 22(7):2971–3006, 2017.