This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Symplectic algorithms for stable manifolds in control theory

Guoyuan Chen School of Data Sciences, Zhejiang University of Finance & Economics, Hangzhou 310018, Zhejiang, P. R. China gychen@zufe.edu.cn  and  Gaosheng Zhu School of Mathematics, Tianjin University, Tiajin 300072, P. R. China gaozsc@163.com
Abstract.

In this note, we propose a symplectic algorithm for the stable manifolds of the Hamilton-Jacobi equations combined with an iterative procedure in [Sakamoto-van der Schaft, IEEE Transactions on Automatic Control, 2008]. Our algorithm includes two key aspects. The first one is to prove a precise estimate for radius of convergence and the errors of local approximate stable manifolds. The second one is to extend the local approximate stable manifolds to larger ones by symplectic algorithms which have better long-time behaviors than general-purpose schemes. Our approach avoids the case of divergence of the iterative sequence of approximate stable manifolds, and reduces the computation cost. We illustrate the effectiveness of the algorithm by an optimal control problem with exponential nonlinearity.

The authors are grateful to the anonymous referees for useful comments and suggestions. G. Chen is supported by NSFC (No.11771386) and First Class Discipline of Zhejiang - A (Zhejiang University of Finance and Economics- Statistics). G. Zhu is supported by NSFC (No.11871356).

1. Introduction

It is well known that an optimal feedback control can be given by solving an associated Hamilton-Jacobi (HJ) equation (see e.g. [23]) and HH^{\infty} feedback control can be obtained from solutions of one or two Hamilton-Jacobi equations (see e.g. [37, 20, 38, 6]). Unfortunately, the Hamilton-Jacobi equation in general can not be solved analytically. Hence numerical method becomes important. Seeking approximate solutions of Hamilton-Jacobi equations from control theory has been studied extensively. There are several approaches: Taylor series method, Galerkin method, state-dependent Riccati equation method, algebraic method, etc. See e.g. [3, 24, 21, 8, 7, 27, 25, 9, 4, 5, 26, 28, 19, 29, 2] and the references therein. These methods may have good performance for concrete control systems. However, in general, they may have various disadvantages such as heavy computation cost for higher dimensional state spaces, restriction on simple nonlinearity of the systems, etc.

For the stationary Hamilton-Jacobi equations which are related to infinite horizon optimal control and HH^{\infty} control problems, [32] developed an iterative procedure to construct an approximate sequence that converges to the exact solution of the associated Hamiltonian system on the stable manifold. It is based on the fact that the stabilizing solutions of stationary Hamilton-Jacobi equations correspond to the generating functions of the stable manifolds (Lagrangian) of the associated Hamiltonian systems at certain equilibriums (cf. e.g. [25, 30, 32]). This approach has better performances for various nonlinear feedback control systems, especially for the ones with more complicated nonlinearities, see e.g. [31, 18, 17].

We should note that the computation approach in [32] (as well as [31, 18, 17]) depends essentially on the radius of convergence of the iterative procedure which is not estimated analytically. Moreover, since the errors of less iterative steps are tremendous when we enlarge the local approximate stable manifolds in unstable direction, to obtain a stable manifold with proper size for applications, the number of iterative steps should be large. This may make the computation time-consuming.

In this note, inspired by [32], we construct a sequence of local approximate stable manifolds of stationary Hamilton-Jacobi equations by iterative procedure with a precise estimate of radius of convergence, and enlarge the local manifolds to larger ones by symplectic algorithms. To be more precise, as in [32] we firstly generate a sequence of local approximate stable manifolds near the equilibrium by an iterative procedure to solve the associated Hamiltonian system of the Hamilton-Jacobi equation. Then we extend the local approximate stable manifolds to large ones by symplectic algorithms for solving initial value problem of the associated Hamiltonian system (Section 4 below). We emphasize that in our approach, the radius of convergence of the iterative sequences are estimated precisely and the error of the local approximate stable manifold can be controlled as small as possible (Theorem 3.1 below). Therefore the significant point is how to extend the local stable manifolds. There are various numerical methods for general ODEs, e.g. Runge-Kutta methods of various orders. For our applications in Hamiltonian systems, we will use symplectic algorithms. The symplectic structure plays a significant role in design of numerical methods for Hamiltonian systems. Generally speaking, the symplectic algorithms are designed to preserve the symplectic structure of the Hamiltonian systems. Hence, for Hamiltonian systems, symplectic algorithm improves qualitative behaviours, and gives a more accurate long-time integration comparing with general-purpose methods such as Runge-Kutta schemes. Various kinds of symplectic algorithms, e.g., symplectic Euler, Störmer-Verlet, symplectic Runge-Kutta, were constructed since 1950s. A detailed history of symplectic methods and related topics can be found in [14]. We refer the readers to the books [16, 13, 10, 34] for a complete presentation of various symplectic algorithms for Hamiltonian systems.

The note is organized as follows. In Section 2, basic notations for the stable manifolds of Hamilton-Jacobi equations in control theory are given. Section 3 is devoted to constructing the iterative procedure, and proves the precise estimate of radius of convergence as well as the error of the approximate solutions. The symplectic scheme which extends the local approximate stable manifold is described in Section 4. In Section 5, the effectiveness of our algorithm is illustrated by application to an optimal control problem. The Appendix gives the details of the proof of Theorem 3.1.

2. The Hamilton-Jacobi equation and the stable manifolds

In this section, for the convenience of the reader we recall the relevant material without proofs, thus making our exposition self-contained. For more details, see [32] and [31].

Let Ωd\Omega\subset\mathbb{R}^{d} be a domain containing 0. We consider the following Hamilton-Jacobi equation

H(x,p)=pTf(x)12pTR(x)p+q(x)=0,\displaystyle H(x,p)=p^{T}f(x)-\frac{1}{2}p^{T}R(x)p+q(x)=0, (2.1)

where p:=Vp:=\nabla V for some unknown function VV, f:Ωdf:\Omega\to\mathbb{R}^{d}, R:Ωd×dR:\Omega\to\mathbb{R}^{d\times d}, q:Ωq:\Omega\to\mathbb{R} are CC^{\infty} and R(x)R(x) is symmetric matrix for all xΩx\in\Omega. Furthermore, we assume that f(0)=0f(0)=0 and q(0)=0q(0)=0, qx(0)=0\frac{\partial q}{\partial x}(0)=0. Hence for xx near 0, f(x)=Ax+O(|x|2),q(x)=12xTQx+O(|x|3),f(x)=Ax+O(|x|^{2}),\,q(x)=\frac{1}{2}x^{T}Qx+O(|x|^{3}), where A=fx(0)A=\frac{\partial f}{\partial x}(0), Q=2qx2(0)Q=\frac{\partial^{2}q}{\partial x^{2}}(0) is the Hessian of qq at 0. We say that a solution VV of (2.1) is stabilizing if p(0)=0p(0)=0 and 0 is an asymptotically stable equilibrium of the vector field f(x)R(x)p(x)f(x)-R(x)p(x) where p(x)=V(x)p(x)=\nabla V(x).

This kind of problem arises from infinite horizon optimal control. For example, consider the optimal control problem

x˙=f(x)+g(x)u,in Ω,\displaystyle\dot{x}=f(x)+g(x)u,\quad\mbox{in }\Omega, (2.2)

where gg is a smooth d×md\times m matrix-valued function, uu is an mm-dimensional feedback control function of column form. Assume the instantaneous cost function L(x,u)=12(q(x)+ruTu),L(x,u)=\frac{1}{2}(q(x)+ru^{T}u), where r>0r>0, q(x)0q(x)\geq 0 is a smooth function. Define the cost functional by J(x,u)=0+L(x(t),u(t))𝑑t.J(x,u)=\int_{0}^{+\infty}L(x(t),u(t))dt. Then the corresponding Hamilton-Jacobi equation is of form (2.1) with R(x):=r1g(x)g(x)TR(x):=r^{-1}g(x)g(x)^{T}. The optimal controller is

u(x)=r1g(x)Tp(x),\displaystyle u(x)=-r^{-1}g(x)^{T}p(x), (2.3)

where p(x)=V(x)p(x)=\nabla V(x).

From the symplectic geometry point of view, a stabilizing solution VV of (2.1) corresponds to a stable (Lagrangian) submanifold. That is, ΛV:={(x,p)|p=V}\Lambda_{V}:=\{(x,p)\,|\,p=\nabla V\} is a stable (Lagrangian) manifold which is invariant under the flow of the associated Hamiltonian system of (2.1):

{x˙=Hp(x,p)=f(x)R(x)p,p˙=Hx(x,p)=(fx)Tp+12(pTR(x)p)x(qx)T.\displaystyle\left\{\begin{array}[]{l}\dot{x}=H_{p}(x,p)=f(x)-R(x)p,\\ \dot{p}=-H_{x}(x,p)=-(\frac{\partial f}{\partial x})^{T}p+\frac{1}{2}\frac{\partial(p^{T}R(x)p)}{\partial x}-(\frac{\partial q}{\partial x})^{T}.\end{array}\right. (2.6)

Conversely, if a dd-dimensional manifold Λ\Lambda in (x,p)(x,p)-space is invariant with respect to the flow (2.6), and at some point (x0,p0)(x_{0},p_{0}), the projection of Λ\Lambda to the xx-space is surjective, then Λ\Lambda is a Lagrangian submanifold in a neighborhood of (x0,p0)(x_{0},p_{0}) and there is a solution VV of (2.1) in a neighborhood of x0x_{0} such that ΛV=Λ\Lambda_{V}=\Lambda. Therefore, if we get the stable manifold, then the optimal controller can be obtained from (2.3). See e.g. [32].

Denote z=(x,p)z=(x,p). Let JJ be the standard symplectic matrix [0IdId0]\left[\begin{array}[]{cc}0&I_{d}\\ -I_{d}&0\end{array}\right], where IdI_{d} denotes the identity matrix of dd-dimensional. Then the vector field on left side of (2.6) is J1H(z)J^{-1}\nabla H(z). Let XH(z)=J1H(z)X_{H}(z)=J^{-1}\nabla H(z) be the Hamiltonian vector field of HH. Note that z0=(0,0)z_{0}=(0,0) is an equilibrium of XHX_{H}. Then the derivative of the Hamiltonian vector field at z0z_{0} is a Hamiltonian matrix, i.e., (JDXH(z0))T=JDXH(z0)(JDX_{H}(z_{0}))^{T}=JDX_{H}(z_{0}). We say that the equilibrium z0z_{0} is hyperbolic if DXH(z0)DX_{H}(z_{0}) has no imaginary eigenvalues. It is well known that if DXH(z0)DX_{H}(z_{0}) is hyperbolic, then its eigenvalues are symmetric with respect to the imaginary axis. From the Stable Manifold Theorem, there exists a global stable manifold 𝒮z0\mathcal{S}_{z_{0}} of z0z_{0}. Moreover, 𝒮z0\mathcal{S}_{z_{0}} is a Lagrangian submanifold of (2d,ω)(\mathbb{R}^{2d},\omega) where ω\omega is the standard symplectic structure (see e.g. [1]). Near z0z_{0}, the Hamiltonian system (2.6) can be rewritten as

z˙=DXH(z0)z+N(z),\displaystyle\dot{z}=DX_{H}(z_{0})z+N(z), (2.7)

where N(z)N(z) is the nonlinear term.

A sufficient condition for the existence local stabilizing solution for (2.1) is obtained by van der Schaft [37] based on an observation on the Riccati equation. Assume (x0,p0)=(0,0)(x_{0},p_{0})=(0,0) without loss of generality. Let R:=R(0)R:=R(0). The Riccati equation

PA+ATPPRP+Q=0\displaystyle PA+A^{T}P-PRP+Q=0 (2.8)

is the linearization of (2.1) at the origin. A symmetric matrix PP is said to be the stabilizing solution of (2.8) if it is a solution of (2.8) and ARPA-RP is stable. The Riccati equation (2.8) has a stabilizing solution if and only if the following two conditions hold: (1) DXH(z0)DX_{H}(z_{0}) is hyperbolic; (2) the generalized eigenspace EE_{-} for dd stable eigenvalues satisfies EIm(0,Id)T=2d.E_{-}\oplus{\rm Im}(0,I_{d})^{T}=\mathbb{R}^{2d}. See e.g. [22]. If (2.8) has a stabilizing solution PP, then there exists a local stabilizing solution VV of (2.1) around the origin such that 2Vx2(0)=P\frac{\partial^{2}V}{\partial x^{2}}(0)=P. This yields a local solution of Hamilton-Jacobi solution ([32]). Consider the Lyapunov equation

(ARP)S+S(ARP)T=R.\displaystyle(A-RP)S+S(A-RP)^{T}=R. (2.9)

Some direct computations yield the following result ([32]).

Lemma 2.1.

Assume that SS is a solution of (2.9). Then

T1DXH(z0)T=[B00BT],\displaystyle T^{-1}DX_{H}(z_{0})T=\left[\begin{array}[]{cc}B&0\\ 0&-B^{T}\\ \end{array}\right], (2.12)

where B=ARPB=A-RP and T=[IdSPPS+Id].T=\left[\begin{array}[]{cc}I_{d}&S\\ P&PS+I_{d}\\ \end{array}\right].

From Lemma 2.1, by coordinates transformation [x¯p¯]=T1[xp]\left[\begin{array}[]{c}\bar{x}\\ \bar{p}\\ \end{array}\right]=T^{-1}\left[\begin{array}[]{c}x\\ p\\ \end{array}\right], system (2.7) becomes a separated form

{x¯˙=Bx¯+Ns(x¯,p¯)p¯˙=BTp¯+Nu(x¯,p¯),\displaystyle\left\{\begin{array}[]{l}\dot{\bar{x}}=B\bar{x}+N_{s}(\bar{x},\bar{p})\\ \dot{\bar{p}}=-B^{T}\bar{p}+N_{u}(\bar{x},\bar{p})\end{array}\right., (2.15)

where Ns(x¯,p¯)N_{s}(\bar{x},\bar{p}) and Nu(x¯,p¯)N_{u}(\bar{x},\bar{p}) are nonlinear terms corresponding to N(z)N(z) after the coordinates transformation.

3. The local approximate stable manifolds: iteration

In this section, we shall give an iterative procedure to construct a sequence of local approximate stable manifolds near equilibrium for the Hamiltonian system in the form

{x˙=Bx+ns(x,y),y˙=BTy+nu(x,y).\displaystyle\left\{\begin{array}[]{l}\dot{x}=Bx+n_{s}(x,y),\\ \dot{y}=-B^{T}y+n_{u}(x,y).\end{array}\right. (3.3)

Assumption 1: BB has eigenvalues with negative real parts. It follows that there exist positive constants a,ba,b such that eBtaebt\|e^{Bt}\|\leq ae^{-bt} for t0t\geq 0.

Assumption 2: ns,nu:d×ddn_{s},n_{u}:\mathbb{R}^{d}\times\mathbb{R}^{d}\to\mathbb{R}^{d} are continuous and satisfy the following conditions: For all L>0L>0, 0<lL0<l\leq L, |x|+|y|l|x|+|y|\leq l and |x|+|y|l|x^{\prime}|+|y^{\prime}|\leq l,

|ns(x,y)ns(x,y)|M(L)l(|xx|+|yy|),\displaystyle|n_{s}(x,y)-n_{s}(x^{\prime},y^{\prime})|\leq M(L)l(|x-x^{\prime}|+|y-y^{\prime}|),
|nu(x,y)nu(x,y)|M(L)l(|xx|+|yy|),\displaystyle|n_{u}(x,y)-n_{u}(x^{\prime},y^{\prime})|\leq M(L)l(|x-x^{\prime}|+|y-y^{\prime}|),

where M(L)M(L) is increasing with respect to LL.

As in [32], to solve equation (3.3), we define the following iterative sequence for k=1,2,k=1,2,\cdots,

{xk+1=eBtξ+0teB(ts)ns(xk(s),yk(s))𝑑syk+1=teBT(ts)nu(xk(s),yk(s))𝑑s\displaystyle\left\{\begin{array}[]{l}x_{k+1}=e^{Bt}\xi+\int_{0}^{t}e^{B(t-s)}n_{s}(x_{k}(s),y_{k}(s))ds\\ y_{k+1}=-\int_{t}^{\infty}e^{B^{T}(t-s)}n_{u}(x_{k}(s),y_{k}(s))ds\end{array}\right. (3.6)

and x0=eBtξx_{0}=e^{Bt}\xi, y0=0y_{0}=0 for an arbitrary ξd\xi\in\mathbb{R}^{d}. Equivalent to (3.6), we consider the following ODE:

{x˙k+1=Bxk+1+ns(xk(t),yk(t))y˙k+1=BTyk+1+nu(xk(t),yk(t))\displaystyle\left\{\begin{array}[]{l}\dot{x}_{k+1}=Bx_{k+1}+n_{s}(x_{k}(t),y_{k}(t))\\ \dot{y}_{k+1}=-B^{T}y_{k+1}+n_{u}(x_{k}(t),y_{k}(t))\end{array}\right. (3.9)

with boundary conditions xk+1(0)=ξ,yk+1(+)=0x_{k+1}(0)=\xi,\,y_{k+1}(+\infty)=0 and x0=eBtξx_{0}=e^{Bt}\xi, y0=0y_{0}=0, t0t\geq 0. This form is more convenient to apply numerical methods to ODEs.

Inspired by [32, Theorem 5], we can prove the following convergence result whose proof is included in Appendix.

Theorem 3.1.

Assume that system (3.3) satisfies Assumption 1-2, M(L)L3b8aM(L)L\geq\frac{3b}{8a} and |ξ|3b16a2M|\xi|\leq\frac{3b}{16a^{2}M}, where M=M(L)M=M(L) is the constant depending on LL given by Assumption 2. Let {xk}\{x_{k}\} and {yk}\{y_{k}\} be the sequences defined by (3.6). Then xk(t)0x_{k}(t)\to 0, yk(t)0y_{k}(t)\to 0 as t+t\to+\infty, and there exist functions xx and yy such that {xk}\{x_{k}\} and {yk}\{y_{k}\} uniformly converge to xx and yy respectively, (x,y)(x,y) is solution of (3.3), and for all t0t\geq 0,

|xk(t)x(t)|C(a,b,M)|ξ|22k1ebt,\displaystyle|x_{k}(t)-x(t)|\leq C(a,b,M)\frac{|\xi|^{2}}{2^{k-1}}e^{-bt},
|yk(t)y(t)|C(a,b,M)|ξ|22k1e2bt,\displaystyle|y_{k}(t)-y(t)|\leq C(a,b,M)\frac{|\xi|^{2}}{2^{k-1}}e^{-2bt}, (3.10)

where C(a,b,M)>0C(a,b,M)>0 is a constant depending only on a,b,Ma,b,M.

Remark 3.1.

Compared to [32, Theorem 5], we improve the result in two aspects: (1) a sufficient estimate of |ξ||\xi| is given; (2) the error of iteration is estimated precisely. The constant C(a,b,M)C(a,b,M) also can be calculated explicitly (see the proof of Theorem 3.1 in the Appendix below).

4. Extension of the local stable manifolds by symplectic algorithms

In this section, the local stable manifolds will be enlarged by symplectic algorithms.

4.1. Structure of the approximate stable manifold

By Lemma 2.1 and Theorem 3.1, we obtain a sequence of local approximate stable manifolds of (2.6) near equilibrium (0,0)(0,0). Let 𝕊ρ={ξd||ξ|=ρ},\mathbb{S}_{\rho}=\{\xi\in\mathbb{R}^{d}\,|\,|\xi|=\rho\}, where ρ\rho is the radius of convergence chosen by Theorem 3.1. Denote the local approximate stable manifold by

Λk={(xk(t,ξ),pk(t,ξ))|t0,ξ𝕊ρ},\Lambda_{k}=\{(x_{k}(t,\xi),p_{k}(t,\xi))\,|\,t\geq 0,\,\xi\in\mathbb{S}_{\rho}\},

and denote the boundary of Λk\Lambda_{k} by Λk\partial\Lambda_{k}. Then Λk={(ξ,pk(0,ξ))|ξ𝕊ρ}\partial\Lambda_{k}=\{(\xi,p_{k}(0,\xi))\,|\,\xi\in\mathbb{S}_{\rho}\}. As kk\to\infty, Λk\Lambda_{k} tends to the exact stable manifold near equilibrium (0,0)(0,0) parameterized by (t,ξ)(t,\xi)

Λ:={(x(t,ξ),p(t,ξ))|t0,ξ𝕊ρ}.\displaystyle\Lambda:=\{(x(t,\xi),p(t,\xi))\,|\,t\geq 0,\,\xi\in\mathbb{S}_{\rho}\}. (4.1)

Its boundary Λ={(ξ,p(0,ξ))|ξ𝕊ρ}\partial\Lambda=\{(\xi,p(0,\xi))\,|\,\xi\in\mathbb{S}_{\rho}\}. Consider the initial value problem of Hamiltonian system

{x˙=Hp(x,p)p˙=Hx(x,p), for t0 with (x(0),p(0))Λ.\displaystyle\left\{\begin{array}[]{l}\dot{x}=H_{p}(x,p)\\ \dot{p}=-H_{x}(x,p)\end{array}\right.,\mbox{ for }t\leq 0\mbox{ with }(x(0),p(0))\in\partial\Lambda. (4.4)

Then by the invariance of the stable manifold,

Λg\displaystyle\Lambda_{\rm g} :=\displaystyle:= {(x(t),p(t))|t0,(x(0),p(0))Λ}Λ\displaystyle\{(x(t),p(t))\,|\,t\leq 0,\,(x(0),p(0))\in\partial\Lambda\}\cup\Lambda
=\displaystyle= {(x(t,ξ),p(t,ξ))|t,ξ𝕊ρ},\displaystyle\{(x(t,\xi),p(t,\xi))\,|\,t\in\mathbb{R},\,\xi\in\mathbb{S}_{\rho}\},

is the global stable manifold of (0,0)(0,0). Hence we extend local stable manifold Λ\Lambda to the global one. Moreover, we should emphasize that each solution curve of (4.4), (x(t),p(t))(x(t),p(t)), with (x(0),p(0))Λ(x(0),p(0))\in\partial\Lambda lies in Λg\Lambda_{\rm g}. Numerically, we compute the approximations by the following initial value problem

{x˙=Hp(x,p)p˙=Hx(x,p)for t0 with (xk(0),pk(0))Λk.\displaystyle\left\{\begin{array}[]{l}\dot{x}=H_{p}(x,p)\\ \dot{p}=-H_{x}(x,p)\end{array}\right.\mbox{for }t\leq 0\mbox{ with }(x_{k}(0),p_{k}(0))\in\partial\Lambda_{k}. (4.7)

Letting (xk(t),pk(t))(x_{k}(t),p_{k}(t)) be numerical solutions of (4.7), we obtain an approximate stable manifold

Λk,g:={(xk(t),pk(t))|t0,(xk(0),pk(0))Λk}Λk\displaystyle\Lambda_{k,\rm g}:=\{(x_{k}(t),p_{k}(t))\,|\,t\leq 0,\,(x_{k}(0),p_{k}(0))\in\partial\Lambda_{k}\}\cup\Lambda_{k}
={(xk(t,ξ),pk(t,ξ))|t,ξ𝕊ρ}.\displaystyle=\{(x_{k}(t,\xi),p_{k}(t,\xi))\,|\,t\in\mathbb{R},\,\xi\in\mathbb{S}_{\rho}\}. (4.8)

Therefore, the key point is to numerically solve the problem (4.7). For general ODEs, there are many kinds of numerical algorithms for the initial value problems. For example, Runge-Kutta of various orders. However, we should point out that (4.7) is a Hamiltonian system. Symplectic algorithms have better performance for such kind of systems.

4.2. Symplectic algorithms of Hamiltonian systems

For Hamiltonian systems, it is natural to use symplectic algorithms. A numerical one-step method yn+1=Φh(yn)y_{n+1}=\Phi_{h}(y_{n}) (with step size hh) is called symplectic if Φh\Phi_{h} is a symplectic map, that is, DΦh(y)TJDΦh(y)=J,D\Phi_{h}(y)^{T}JD\Phi_{h}(y)=J, where DΦh(y)D\Phi_{h}(y) is the tangent map of Φh\Phi_{h} at yy. Symplectic structure is an essential property of Hamiltonian system (cf., e.g., [16, Chapter VI.2]). Symplectic algorithms preserve this geometric structure for each step. Hence compared to general purpose numerical algorithms (e.g. Runge-Kutta methods), symplectic algorithms have much better long-time qualitative behaviours. There are many types of symplectic algorithms, e.g., symplectic Runge-Kutta of various orders, Störmer-Verlet methods, Nyström method, etc. We refer the readers to [16, 13] for more details of symplectic algorithms.

In what follows, we illustrate our procedure of extension of the local stable manifold by the Störmer-Verlet method which is a simple symplectic algorithm of 2-order. Other symplectic algorithms of higher orders may have better numerical results.

Let hh be a step size, and t0=0t_{0}=0 be the initial time, and let tn=nht_{n}=nh, tn+1/2=(n+1/2)ht_{n+1/2}=(n+1/2)h. Hence for problem (4.7), we should choose h<0h<0. Let pn=p(tn)p_{n}=p(t_{n}), pn+1/2=p(tn+1/2)p_{n+1/2}=p(t_{n+1/2}), xn=x(tn)x_{n}=x(t_{n}), xn+1/2=x(tn+1/2)x_{n+1/2}=x(t_{n+1/2}). Denote Hp(x,p)=Hp(x,p)H_{p}(x,p)=\frac{\partial H}{\partial p}(x,p), Hx(x,p)=Hx(x,p)H_{x}(x,p)=\frac{\partial H}{\partial x}(x,p). In our case, the Hamiltonian H(x,p)H(x,p) is given by (2.1) where HpH_{p} and HxH_{x} can be calculated. Then we have the following theorem.

Theorem 4.1.

Given (xn,pn)(x_{n},p_{n}), the Störmer-Verlet schemes

{pn+1/2=pnh2Hx(xn,pn+1/2)xn+1=xn+h2[Hp(xn,pn+1/2)+Hp(xn+1,pn+1/2)]pn+1=pn+1/2h2Hx(xn+1,pn+1/2), and \left\{\begin{array}[]{l}p_{n+1/2}=p_{n}-\frac{h}{2}H_{x}(x_{n},p_{n+1/2})\\ x_{n+1}=x_{n}+\frac{h}{2}\left[H_{p}(x_{n},p_{n+1/2})+H_{p}(x_{n+1},p_{n+1/2})\right]\\ p_{n+1}=p_{n+1/2}-\frac{h}{2}H_{x}(x_{n+1},p_{n+1/2}),\quad\mbox{ and }\end{array}\right. (4.9)
{xn+1/2=xn+h2Hp(xn+1/2,pn)pn+1=pnh2[Hx(xn+1/2,pn)+Hx(xn+1/2,pn+1)]xn+1=xn+1/2+h2Hp(xn+1/2,pn+1),\left\{\begin{array}[]{l}x_{n+1/2}=x_{n}+\frac{h}{2}H_{p}(x_{n+1/2},p_{n})\\ p_{n+1}=p_{n}-\frac{h}{2}\left[H_{x}(x_{n+1/2},p_{n})+H_{x}(x_{n+1/2},p_{n+1})\right]\\ x_{n+1}=x_{n+1/2}+\frac{h}{2}H_{p}(x_{n+1/2},p_{n+1}),\end{array}\right. (4.10)

are symplectic methods of order 2.

A complete proof of this Theorem and more details of the Störmer-Verlet method can be found in [15, 16]. Note that in general (4.9) and (4.10) are implicit equations which can be solved by Newton’s iteration method. For example, in (4.9), the first two equations are implicit. The third equation is explicit if pn+1/2,xn+1p_{n+1/2},x_{n+1} were found. Recall that the key point of Newton’s iteration method is to give a proper initial guess at the beginning of iteration. For the first two equations of (4.9), a good initial guess of (xn+1,pn+1/2)(x_{n+1},p_{n+1/2}) is (xn,pn)(x_{n},p_{n}) since usually the step size hh is small and (xn+1,pn+1/2)(x_{n+1},p_{n+1/2}) is close to (xn,pn)(x_{n},p_{n}). Hence, only a few times of iteration in Newton’s method should be applied to arrive at the accuracy needed. Therefore the computation cost is cheap for Newton’s iteration at this point.

For our case, from (2.6), the Störmer-Verlet scheme (4.9) becomes

{pn+1/2=pn+h2[[fx(xn)]Tpn+1/2+12(pn+1/2TRpn+1/2)x(xn)qx(xn)]xn+1=xn+h2[f(xn)+f(xn+1)(R(xn)+R(xn+1))pn+1/2]pn+1=pn+1/2+h2[[fx(xn+1)]Tpn+1/2+12(pn+1/2TRpn+1/2)x(xn+1)qx(xn+1)].\displaystyle\left\{\begin{array}[]{l}p_{n+1/2}=p_{n}+\frac{h}{2}\left[-\left[\frac{\partial f}{\partial x}(x_{n})\right]^{T}p_{n+1/2}\right.\\ \quad\quad\quad\quad+\left.\frac{1}{2}\frac{\partial(p_{n+1/2}^{T}Rp_{n+1/2})}{\partial x}(x_{n})-\frac{\partial q}{\partial x}(x_{n})\right]\\ x_{n+1}=x_{n}+\frac{h}{2}\left[f(x_{n})+f(x_{n+1})\right.\\ \quad\quad\quad\quad\left.-(R(x_{n})+R(x_{n+1}))p_{n+1/2}\right]\\ p_{n+1}=p_{n+1/2}+\frac{h}{2}\left[-\left[\frac{\partial f}{\partial x}(x_{n+1})\right]^{T}p_{n+1/2}\right.\\ \quad\quad\quad\quad+\left.\frac{1}{2}\frac{\partial(p_{n+1/2}^{T}Rp_{n+1/2})}{\partial x}(x_{n+1})-\frac{\partial q}{\partial x}(x_{n+1})\right].\\ \end{array}\right. (4.17)

In many applications, a special case is that R(x)R(x) is a constant matrix, then the first equation is explicit since (pn+1/2TRpn+1/2)x(xn)=0\frac{\partial(p_{n+1/2}^{T}Rp_{n+1/2})}{\partial x}(x_{n})=0. That is,

pn+1/2=[Id+h2[fx(xn)]T]1[pnh2qx(xn)].\displaystyle p_{n+1/2}=\left[I_{d}+\frac{h}{2}\left[\frac{\partial f}{\partial x}(x_{n})\right]^{T}\right]^{-1}\left[p_{n}-\frac{h}{2}\frac{\partial q}{\partial x}(x_{n})\right]. (4.18)

Note that Id+h2(fx(xn))TI_{d}+\frac{h}{2}(\frac{\partial f}{\partial x}(x_{n}))^{T} is invertible since hh is small. Hence in (4.17), the second equation is the only implicit equation.

Symplectic algorithms have favourable long term behaviours such as energy conservation. Assume that a Hamiltonian H:DH:D\to\mathbb{R} (D2dD\subset\mathbb{R}^{2d}) is analytic. Suppose that Φh(y)\Phi_{h}(y) is the Störmer-Verlet method with step size h>0h>0. If the numerical solution stays in some compact set KDK\subset D, then there exists h0h_{0} such that

H(xn,pn)=H(x0,p0)+O(h2),\displaystyle H(x_{n},p_{n})=H(x_{0},p_{0})+O(h^{2}), (4.19)

in exponential large interval 0<nheh0/(2h)0<nh\leq e^{h_{0}/(2h)}. For example, the Hamiltonian of the example in Section 5 is analytic. Moreover, in concrete problem the constant h0h_{0} can be computed explicitly [16, Section 8.1]. For more general symplectic algorithms of various orders, similar energy estimates hold. See e.g. [16, Section 8.1]. As pointed out in [32, 31], the value of the Hamiltonian (i.e. the energy) is usually used as a measure for the accuracy of the approximate solutions. That is, if H(xn,pn)H(x_{n},p_{n}) is not close to H(x0,p0)H(x_{0},p_{0}), then the solution trajectory leaves the stable manifold. The estimates (4.19) indicate that the steps of symplectic algorithm can be controlled by the value of Hamiltonian well. In practice, we set |H(xn,pn)H(x0,p0)||H(x_{n},p_{n})-H(x_{0},p_{0})| along the numerical trajectories to satisfy certain accuracy δ>0\delta>0. Once |H(xn,pn)H(x0,p0)|>δ|H(x_{n},p_{n})-H(x_{0},p_{0})|>\delta, the numerical computation will stop and record the time. Such technique is called Hamiltonian check as in [31].

4.3. Computation procedure

We are now in the position to summarize the computation procedure.

Step 1. Transform (2.6) into a system of form (2.15). To apply the iterative method, we transform the Hamiltonian system into the separated form (2.15) by a coordinates transformation as in Lemma 2.1.

Step 2. Compute the local approximate stable manifold by iteration. We give a precise estimate of the radius ρ\rho of ξ\xi which makes the sequences {xk}\{x_{k}\} and {pk}\{p_{k}\} convergent by Theorem 3.1. Using numerical methods (e.g., Runge-Kutta method), (3.9) is solved for different points ξ𝕊ρ\xi\in\mathbb{S}_{\rho}. Here the number of points ξ\xi can be properly chosen in concrete problems. Then we get a local approximate stable manifold Λk={(xk(t,ξ),pk(t,ξ))|t0,ξ𝕊ρ}\Lambda_{k}=\{(x_{k}(t,\xi),p_{k}(t,\xi))\,|\,t\geq 0,\,\xi\in\mathbb{S}_{\rho}\}. Note that the error can be controlled by kk and ρ\rho by Theorem 3.1.

Step 3. Extend the local approximate stable manifold by symplectic algorithm. Rewrite Λk\Lambda_{k} in the original coordinates by Λ^k=TΛk\hat{\Lambda}_{k}=T\Lambda_{k} where TT is the coordinates transform given by Lemma 2.1. Taking advantage of symplectic algorithm such as the Störmer-Verlet method, symplectic Runge-Kutta method of various orders, we solve the initial value problem (4.7). Then we find a larger approximate stable manifold. We shall use the Hamiltonian check to indicate that the trajectories stay close to the exact stable manifold.

Remark 4.1.

In practice, we can find an applicable approximate stable manifold as follows: Firstly, a proper set of ξ𝕊ρ\xi\in\mathbb{S}_{\rho} is chosen according to the concrete problem. From Theorem 3.1, a corresponding set of local curves in local approximate stable manifold is obtained. Then, extending the local curves by solving (4.7) numerically, we find a set of longer curves in the global approximate stable manifold. Using dd-dimensional interpolation method or dd-variable polynomial fitting (see e.g. [31]), we obtain an approximate function p~(x)\tilde{p}(x) whose graph is an approximate stable manifold. For a detailed example, see Section 5.2 below.

Remark 4.2.

In [32], [31], the local approximate stable manifold is extended by using negative tt in (3.9) (or, equivalently, (3.6)) and taking more iterative times kk. In our approach, we first generate a local approximate stable manifold by (3.9) with t0t\geq 0, then extend this local approximate stable manifold by solving the initial value problem (4.7) for negative tt. We should emphasize that in our approach, we do not use negative tt in the iterative procedure (3.9). This can avoid the divergent case of iterative sequence for negative tt as in [32], [31] and also reduce the computation cost.

5. Example

In this section, we apply the symplectic algorithm to a 2-dimensional optimal control problem with exponential nonlinearity. The existed methods may be difficult to applied to the systems with such kind of complicated nonlinearities as showed in [31, 18, 17].

Throughout this section, we use the following notations. kk: the iterative times for local approximate stable manifold as in (3.9); ξ\xi: the initial condition given in Theorem 3.1; hh_{-}: the step size of the numerical methods for t<0t<0 by (4.7). RK45 method is from scipy.integrate.RK45 in python environment. This algorithm is based on error control as 4-order Runge-Kutta method, and its steps are taken using the 5-order accurate formula. For more details this algorithm, see [11, 35]. We should mention that other numerical environment such as Matlab also includes a similar algorithm named by ode45. In what follows, we implement RK45 in iterative procedure (3.9) for the local stable manifold.

Let us consider the following optimal control problem with exponential nonlinearity:

{x˙1=ex21+u1,x˙2=(x1+13x13)+u2,\displaystyle\left\{\begin{array}[]{l}\dot{x}_{1}=e^{x_{2}}-1+u_{1},\\ \dot{x}_{2}=-(x_{1}+\frac{1}{3}x_{1}^{3})+u_{2},\end{array}\right. (5.3)

where u=(u1,u2)Tu=(u_{1},u_{2})^{T} is the feedback control function. Let

f(x)=[ex21,(x1+13x13)]T.f(x)=\left[e^{x_{2}}-1,-(x_{1}+\frac{1}{3}x_{1}^{3})\right]^{T}.

Define the cost function by

012[xT(t)x(t)+uT(t)u(t)]𝑑t.\int_{0}^{\infty}\frac{1}{2}[x^{T}(t)x(t)+u^{T}(t)u(t)]dt.

Then the corresponding Hamilton-Jacobi equation is

H(x,p)=pTf(x)12pTRp+12xTQx=0,H(x,p)=p^{T}f(x)-\frac{1}{2}p^{T}Rp+\frac{1}{2}x^{T}Qx=0,

where p=Vp=\nabla V is the gradient of the value function in column form, x=(x1,x2)Tx=(x_{1},x_{2})^{T}, R=I2R=I_{2}, Q=I2Q=I_{2}. Then by (2.3) the optimal controller is u=pu=-p. The associated Hamiltonian system is

{x˙=f(x)Rpp˙=[fx]TpQx,\displaystyle\left\{\begin{array}[]{l}\dot{x}=f(x)-Rp\\ \dot{p}=-\left[\frac{\partial f}{\partial x}\right]^{T}p-Qx,\end{array}\right. (5.6)

where [fx(x)]T=[01x12ex20]\left[\frac{\partial f}{\partial x}(x)\right]^{T}=\left[\begin{array}[]{cc}0&-1-x_{1}^{2}\\ e^{x_{2}}&0\\ \end{array}\right]. Then (0,0)(0,0) is an equilibrium and the Hamiltonian matrix is given by [ARQAT],\left[\begin{array}[]{cc}A&-R\\ -Q&-A^{T}\\ \end{array}\right], where A=[0110].A=\left[\begin{array}[]{cc}0&1\\ -1&0\\ \end{array}\right]. The stabilizing solution PP of the Riccati equation is I2I_{2}. From Lemma 2.1, we find a matrix T=[I20.5I2I20.5I2]T=\left[\begin{array}[]{cc}I_{2}&-0.5I_{2}\\ I_{2}&0.5I_{2}\\ \end{array}\right] such that the Hamiltonian system (5.6) becomes a separated form

[x¯˙p¯˙]=[B00BT][x¯p¯]+[ns(x¯,p¯)nu(x¯,p¯)],\displaystyle\left[\begin{array}[]{c}\dot{\bar{x}}\\ \dot{\bar{p}}\\ \end{array}\right]=\left[\begin{array}[]{cc}B&0\\ 0&-B^{T}\end{array}\right]\left[\begin{array}[]{c}\bar{x}\\ \bar{p}\\ \end{array}\right]+\left[\begin{array}[]{c}n_{s}(\bar{x},\bar{p})\\ n_{u}(\bar{x},\bar{p})\\ \end{array}\right], (5.15)

where PP is the stabilizing solution of the Riccati equation (2.8), B=ARPB=A-RP, x=x(x¯,p¯),p=p(x¯,p¯)x=x(\bar{x},\bar{p}),p=p(\bar{x},\bar{p}) are defined by [xp]=T[x¯p¯],\left[\begin{array}[]{c}x\\ p\\ \end{array}\right]=T\left[\begin{array}[]{c}\bar{x}\\ \bar{p}\\ \end{array}\right], and [ns(x¯,p¯)nu(x¯,p¯)]:=T1[f(x)Ax[fx]Tp+ATp].\left[\begin{array}[]{c}n_{s}(\bar{x},\bar{p})\\ n_{u}(\bar{x},\bar{p})\\ \end{array}\right]:=T^{-1}\left[\begin{array}[]{c}f(x)-Ax\\ -\left[\frac{\partial f}{\partial x}\right]^{T}p+A^{T}p\\ \end{array}\right]. Here the eigenvalues of BB are 1±i-1\pm i. Then we consider an iterative procedure as (3.9) with x¯0=eBtξ,p¯0=0\bar{x}_{0}=e^{Bt}\xi,\,\bar{p}_{0}=0, and x¯k(0)=ξ,p¯k(+)=0\bar{x}_{k}(0)=\xi,\,\bar{p}_{k}(+\infty)=0 for k=1,2,3,k=1,2,3,\cdots.

For (5.15) we can choose M(L)=(9/4)LM(L)=(9/4)L for L>2/3L>2/3 and M(L)=3/2M(L)=3/2 for L2/3L\leq 2/3. Moreover, a=1,b=1a=1,\,b=1. Hence by Theorem 3.1, if M(L)L3/8M(L)L\geq 3/8, then |x¯k|+|p¯k|<L|\bar{x}_{k}|+|\bar{p}_{k}|<L for all k=1,2,k=1,2,\cdots. That is, we can choose L1/4L\geq 1/4. Hence for |ξ|316M(L)0.125|\xi|\leq\frac{3}{16M(L)}\approx 0.125, {x¯k(t,ξ)}\{\bar{x}_{k}(t,\xi)\} and {p¯k(t,ξ)}\{\bar{p}_{k}(t,\xi)\} converge to exact solution x¯(t,ξ)\bar{x}(t,\xi) and p¯(t,ξ)\bar{p}(t,\xi) respectively. In what following, we shall take ξ\xi in the sphere 𝕊0.12\mathbb{S}_{0.12}. That yields a local approximate stable manifold Λk={(x¯k(t,ξ),p¯k(t,ξ))|t0,ξ𝕊0.12}\Lambda_{k}=\{(\bar{x}_{k}(t,\xi),\bar{p}_{k}(t,\xi))\,|\,t\geq 0,\,\xi\in\mathbb{S}_{0.12}\} and Λk={(ξ,p¯k(0,ξ))|ξ𝕊0.12}\partial\Lambda_{k}=\{(\xi,\bar{p}_{k}(0,\xi))\,|\,\xi\in\mathbb{S}_{0.12}\}.

Next, as in Section 4.3, the local stable manifold Λk\Lambda_{k} will be enlarged to a global one Λk,g\Lambda_{k,g} from (4.7). We solve the initial value problem (4.7) by the Störmer-Verlet scheme (4.17). In our problem (5.6), algorithm (4.17) becomes

{pn+1/2=[I2+h2[fx(xn)]T]1[pnhxn]xn+1=xn+h2[f(xn)+f(xn+1)2pn+1/2]pn+1=pn+1/2h2[[fx(xn+1)]Tpn+1/2+xn+1)].\displaystyle\left\{\begin{array}[]{l}p_{n+1/2}=\left[I_{2}+\frac{h}{2}\left[\frac{\partial f}{\partial x}(x_{n})\right]^{T}\right]^{-1}\left[p_{n}-hx_{n}\right]\\ x_{n+1}=x_{n}+\frac{h}{2}\left[f(x_{n})+f(x_{n+1})-2p_{n+1/2}\right]\\ p_{n+1}=p_{n+1/2}-\frac{h}{2}\left[\left[\frac{\partial f}{\partial x}(x_{n+1})\right]^{T}p_{n+1/2}+x_{n+1})\right].\\ \end{array}\right. (5.19)

Here the second equation is the only implicit equation which can be solved by Newton’s iteration method.

5.1. Comparison with other methods

To show the effectiveness of our approach, the extension result of the Störmer-Verlet scheme is compared with other two methods, i.e., the first one is solving extension initial problem (4.7) by the RK45 algorithm, and the second one is the method of iteration for negative time in problem (3.6) as in Sakamoto-van der Schaft [32, 31]. For the differences of these methods, see Remark 4.2 above. Figure 1 illustrates the values of the Hamiltonian function along the approximate curves from the three methods with ξ=(0.12×2/2,0.12×2/2)𝕊0.12\xi=(0.12\times\sqrt{2}/2,0.12\times\sqrt{2}/2)\in\mathbb{S}_{0.12}. It shows that the Störmer-Verlet method is much better than the RK45 extension method and the extension approach of Sakamoto-van der Schaft [32, 31].

Refer to caption
Refer to caption
Refer to caption
Figure 1. Hamiltonian values along extended curves with ξ=(0.12×2/2,0.12×2/2)\xi=(0.12\times\sqrt{2}/2,0.12\times\sqrt{2}/2). The first subfigure is the Störmer-Verlet scheme with k=3k=3, h=0.005h_{-}=-0.005, t[3.8,10]t\in[-3.8,10]; the second one corresponds to the RK45 extension method with k=3k=3, t[3.8,10]t\in[-3.8,10]; the third one is from the extension method of [32, 31] with k=50k=50, t[3.5,10]t\in[-3.5,10]. We should point out that in the third subfigure, the time interval is smaller since the Hamiltonian value from the extension method of [32, 31] blows up to infinite on [3.8,10][-3.8,10].

From (4.1), the approximate stable manifold Λk,g\Lambda_{k,g} is made up by all curves of solutions (xk(t,ξ),pk(t,ξ))(x_{k}(t,\xi),p_{k}(t,\xi)), ξ𝕊0.12\xi\in\mathbb{S}_{0.12}. To construct the optimal controller, we need to find numerical solutions (xk(t,ξ),pk(t,ξ))(x_{k}(t,\xi),p_{k}(t,\xi)) with Hamiltonian value under certain error tolerance. Projection of the approximate stable manifold (i.e. all extended curves) under certain error tolerance to xx-plane is a domain. In Figure 2, we illustrate the domains of projection of all extended curves with ξ𝕊0.12\xi\in\mathbb{S}_{0.12} to xx-plane by the Störmer-Verlet method, RK45 and the method of Sakamoto-van der Schaft ([32, 31]) under certain error tolerances. It is clear that the Störmer-Verlet method generates a much larger domain comparing with the other two extension methods.

We should point out that, as shown above, the method of Sakamoto-van der Schaft ([32, 31]) requires much more iterative times, hence the computation cost is more expensive.

Refer to caption
Figure 2. The domains generated by projections of extended approximate local stable manifold to xx-plane. The largest domain Ω\Omega (with green line boundary) is from the extension by the Störmer-Verlet scheme for k=3k=3, h=103h_{-}=-10^{-3} and the absolute value of Hamiltonian less than 10410^{-4}. The domain with red dashed line boundary (resp., blue dotted line boundary) corresponds to the RK45 extension method for k=3k=3 (resp., the extension method of [32, 31] for k=50k=50) with the absolute value of Hamiltonian less than 10310^{-3}.
Remark 5.1.

We should mention that the Störmer-Verlet scheme is just a two order symplectic algorithm. If we need a larger domain, or a smaller Hamiltonian value tolerance is required, then other symplectic algorithms such as symplectic Runge-Kutta schemes of higher orders (see e.g. [16]) can be implemented.

5.2. Construct optimal controller by the approximate stable manifold

To find the approximate optimal controller uu by the stable manifold, we use the polynomial fitting method as in [31]. To be more precise, we first choose 200200 ξ\xis according to the uniform distribution in 𝕊0.12\mathbb{S}_{0.12}. Using the computation procedure in Section 4.3 based on Störmer-Verlet scheme for time interval [3.5,10][-3.5,10], h=103h_{-}=-10^{-3} and k=3k=3, we find 200200 extended curves in the stable manifold with the Hamiltonian less than 10410^{-4}. The projections of these curves to xx-plane belong to the domain Ω2\Omega\subset\mathbb{R}^{2} (the largest domain in Figure 2). Then we select 10 points (including the point at time 3.5-3.5 and the other 9 points chosen randomly in (3.5,0](-3.5,0] according to the uniform distribution) on each curve and the point (x,p)=(0,0)(x,p)=(0,0). Hence we obtained a set 𝒲\mathcal{W} of sample points on the stable manifold. To approximate stable manifold, we assume a polynomial fitting of form

ppol(x)=[0i,j5Cij1x1ix2j,0i,j5Cij2x1ix2j]T,p_{\rm pol}(x)=\left[\sum_{0\leq i,j\leq 5}C^{1}_{ij}x_{1}^{i}x_{2}^{j},\sum_{0\leq i,j\leq 5}C^{2}_{ij}x_{1}^{i}x_{2}^{j}\right]^{T},

where the coefficients CijkC_{ij}^{k} are chosen by the least square technique based on the sample set 𝒲\mathcal{W}. That is, {Cijk|i,j=1,,5;k=1,2}\{C_{ij}^{k}\,|\,i,j=1,\cdots,5;k=1,2\} are the coefficients to minimize

(xl,pl)𝒲(ppol(xl)pl)2.\sum_{(x_{l},p_{l})\in\mathcal{W}}(p_{\rm pol}(x_{l})-p_{l})^{2}.

Then, using (2.3), the approximate optimal controller is given by u~(x)=ppol(x)\tilde{u}(x)=-p_{\rm pol}(x). Hence by system (5.3), the closed loop can be computed for any xΩx\in\Omega. Here we demonstrate the controlled trajectories from the controller u~\tilde{u} at some points in Figure 3. We can clearly see that the controller asymptotically stabilizes these trajectories.

Refer to caption
Refer to caption
Figure 3. The controlled trajectories for the initial point (4,3.6)(4,3.6) and (5,4)(-5,4).

6. Conclusion

In this note, using a symplectic algorithm for the stable manifolds of the Hamilton-Jacobi equations from infinite horizon optimal control, we construct a sequence of local approximate stable manifolds for Hamiltonian system at some hyperbolic equilibrium. In our approach, we first construct an iterative sequence of local approximate stable manifolds at equilibrium based on a precise estimates for the radius of convergence. Then, using symplectic algorithms, we enlarge the local approximate stable manifolds by solving initial value problem of the associated Hamiltonian system for t<0t<0. This avoids the divergent case of the iterative sequence as in [32] for unstable direction t<0t<0, and the computation cost can be reduced. Moreover, the symplectic algorithms have better long-time behaviours than general purpose numerical methods such as Runge-Kutta. The effectiveness of our algorithm is demostrated by an optimal control problem with exponential nonlinearity. Lastly, we should mention that the symplectic algorithm has potentials to be applied to more general Hamilton-Jacobi equations in control theory since such kind of equations have associated Hamiltonian systems as the characteristic systems.

.1. Proof of Theorem 3.1

Firstly, we prove that for |ξ|3b16a2M|\xi|\leq\frac{3b}{16a^{2}M},

|xk(t)|α¯ebt,|yk(t)|β¯e2bt,t0,\displaystyle|x_{k}(t)|\leq\underline{\alpha}e^{-bt},\quad|y_{k}(t)|\leq\underline{\beta}e^{-2bt},\quad\forall t\geq 0, (.1)

where

α¯=3a216|ξ|2g+g2a216|ξ|2+a|ξ|,β¯=a216|ξ|2g+g2a216|ξ|2.\displaystyle\underline{\alpha}=\frac{\frac{3a^{2}}{16}|\xi|^{2}}{g+\sqrt{g^{2}-\frac{a^{2}}{16}|\xi|^{2}}}+a|\xi|,\,\underline{\beta}=\frac{\frac{a^{2}}{16}|\xi|^{2}}{g+\sqrt{g^{2}-\frac{a^{2}}{16}|\xi|^{2}}}. (.2)

Here g=3b32aMa4|ξ|3b64aMg=\frac{3b}{32aM}-\frac{a}{4}|\xi|\geq\frac{3b}{64aM}.

In fact, By Assumption 1, |x0(t)|a|ξ|ebt,|y0(t)|=0|x_{0}(t)|\leq a|\xi|e^{-bt},|y_{0}(t)|=0, t0.t\geq 0. Hence α0=a|ξ|\alpha_{0}=a|\xi|, β0=0\beta_{0}=0. For k=1,2,k=1,2,\cdots, we prove by inductive method. Assume that the claim holds for kk. Then from (3) and (.1),

|xk+1(t)|a|ξ|ebt+aMebt0tebs(|xk(s)+yk(s)|2)𝑑s\displaystyle|x_{k+1}(t)|\leq a|\xi|e^{-bt}+aMe^{-bt}\int_{0}^{t}e^{bs}(|x_{k}(s)+y_{k}(s)|^{2})ds
[aMb(αk+βk)2+a|ξ|]ebt, and\displaystyle\leq\left[\frac{aM}{b}(\alpha_{k}+\beta_{k})^{2}+a|\xi|\right]e^{-bt},\quad\mbox{ and }
|yk+1(t)|\displaystyle|y_{k+1}(t)| \displaystyle\leq aMebttebs(|xk(s)+yk(s)|2)𝑑s\displaystyle aMe^{bt}\int_{t}^{\infty}e^{-bs}(|x_{k}(s)+y_{k}(s)|^{2})ds
\displaystyle\leq aM3b(αk+βk)2e2bt.\displaystyle\frac{aM}{3b}(\alpha_{k}+\beta_{k})^{2}e^{-2bt}.

Let c0=aM3bc_{0}=\frac{aM}{3b}. Define α0=a|ξ|,β0=0\alpha_{0}=a|\xi|,\,\beta_{0}=0, αk+1=3c0(αk+βk)2+a|ξ|,βk+1=c0(αk+βk)2,\alpha_{k+1}=3c_{0}(\alpha_{k}+\beta_{k})^{2}+a|\xi|,\,\beta_{k+1}=c_{0}(\alpha_{k}+\beta_{k})^{2}, for k=1,2,3,k=1,2,3,\cdots. We should point out that the definition of αk,βk\alpha_{k},\beta_{k} is different from that in [32]. Note first that α1>α0\alpha_{1}>\alpha_{0} and β1>β0\beta_{1}>\beta_{0}. By mathematical induction, we can prove that αk+1>αk\alpha_{k+1}>\alpha_{k} and βk+1>βk\beta_{k+1}>\beta_{k}. Next solving

{α¯=3c0(α¯+β¯)2+a|ξ|,β¯=c0(α¯+β¯)2,\displaystyle\left\{\begin{array}[]{l}\underline{\alpha}=3c_{0}(\underline{\alpha}+\underline{\beta})^{2}+a|\xi|,\\ \underline{\beta}=c_{0}(\underline{\alpha}+\underline{\beta})^{2},\end{array}\right. (.5)

we have solution (.2). Then (.1) follows.

Secondly, by a similar argument as in [32], we find that

|xk+1(t)xk(t)|γkebt,|yk+1(t)yk(t)|εke2bt\displaystyle|x_{k+1}(t)-x_{k}(t)|\leq\gamma_{k}e^{-bt},\,|y_{k+1}(t)-y_{k}(t)|\leq\varepsilon_{k}e^{-2bt} (.6)

where {γk}\{\gamma_{k}\} and {εk}\{\varepsilon_{k}\} satisfy γ1=a3M|ξ|2b,ε1=a3M|ξ|23b\gamma_{1}=\frac{a^{3}M|\xi|^{2}}{b},\,\varepsilon_{1}=\frac{a^{3}M|\xi|^{2}}{3b} and γk+1=a(α¯+β¯)Mb(γk+εk)\gamma_{k+1}=\frac{a(\underline{\alpha}+\underline{\beta})M}{b}(\gamma_{k}+\varepsilon_{k}), εk+1=a(α¯+β¯)M3b(γk+εk)\varepsilon_{k+1}=\frac{a(\underline{\alpha}+\underline{\beta})M}{3b}(\gamma_{k}+\varepsilon_{k}). Moreover, {γk}\{\gamma_{k}\} and {εk}\{\varepsilon_{k}\} are decreasing and limkγk=0\lim_{k\to\infty}\gamma_{k}=0, limkεk=0\lim_{k\to\infty}\varepsilon_{k}=0. Consequently, it holds that

γk+εk\displaystyle\gamma_{k}+\varepsilon_{k} \displaystyle\leq [43a(α¯+β¯)Mb]k14a3M|ξ|23b.\displaystyle\left[\frac{4}{3}\frac{a(\underline{\alpha}+\underline{\beta})M}{b}\right]^{k-1}\frac{4a^{3}M|\xi|^{2}}{3b}. (.7)

Thirdly, we prove that if M(L)L>3b8aM(L)L>\frac{3b}{8a}, it holds that for all kk\in\mathbb{N},

|xk(t)|+|yk(t)|L,t[0,).\displaystyle|x_{k}(t)|+|y_{k}(t)|\leq L,\quad\forall t\in[0,\infty). (.8)

In fact, from (.2) and |ξ|3b16a2M|\xi|\leq\frac{3b}{16a^{2}M}, it holds that α¯2164baM,β¯364baM.\underline{\alpha}\leq\frac{21}{64}\frac{b}{aM},\quad\underline{\beta}\leq\frac{3}{64}\frac{b}{aM}. Hence |xk(t)|2164baMebt.|x_{k}(t)|\leq\frac{21}{64}\frac{b}{aM}e^{-bt}. Similarly, |yk(t)|364baMe2bt.|y_{k}(t)|\leq\frac{3}{64}\frac{b}{aM}e^{-2bt}. Therefore, we obtain that |xk(t)|+|yk(t)|38baMebt.|x_{k}(t)|+|y_{k}(t)|\leq\frac{3}{8}\frac{b}{aM}e^{-bt}. Then since M(L)L>3b8aM(L)L>\frac{3b}{8a}, (.8) holds. Since M=M(L)M=M(L) is increasing with respect to LL, the conclusion holds.

Finally, let kk\in\mathbb{N} be any fixed number. From (.6) and (.7), for all jj\in\mathbb{N}, it holds that

|xk+j(t)xk(t)|\displaystyle|x_{k+j}(t)-x_{k}(t)|
\displaystyle\leq [43a(α¯+β¯)Mb]k14a3M|ξ|23(ba(α¯+β¯)M)ebt,\displaystyle\left[\frac{4}{3}\frac{a(\underline{\alpha}+\underline{\beta})M}{b}\right]^{k-1}\frac{4a^{3}M|\xi|^{2}}{3(b-a(\underline{\alpha}+\underline{\beta})M)}e^{-bt},
|yk+j(t)yk(t)|\displaystyle|y_{k+j}(t)-y_{k}(t)|
\displaystyle\leq [43a(α¯+β¯)Mb]k14a3M|ξ|23ba(α¯+β¯)Me2bt.\displaystyle\left[\frac{4}{3}\frac{a(\underline{\alpha}+\underline{\beta})M}{b}\right]^{k-1}\frac{4a^{3}M|\xi|^{2}}{3b-a(\underline{\alpha}+\underline{\beta})M}e^{-2bt}.

Here we used the fact that 43a(α¯+β¯)Mb1/2\frac{4}{3}\frac{a(\underline{\alpha}+\underline{\beta})M}{b}\leq 1/2. Rewriting (.1) in form (3.1), the conclusions of this theorem hold since jj is arbitrary. This completes the proof. \Box

References

  • [1] R. A. Abraham and J. E. Marsden. Foundation of mechanics. Benjamin/Cummings, Reading, MA, 2nd edition, 1978.
  • [2] C. O. Aguilar and A. J. Krener. Numerical solutions to the Bellman equation of optimal control. Journal of Optimization Theory and Applications, 160(2):527–552, 2014.
  • [3] E.G. Al’Brekht. On the optimal stabilization of nonlinear systems. Journal of Applied Mathematics and Mechanics, 25(5):1254–1266, 1961.
  • [4] M. D. S. Aliyu. An approach for solving the Hamilton–Jacobi–Isaacs equation (HJIE) in nonlinear H{H}_{\infty} control. Automatica, 39(5):877–884, 2003.
  • [5] M. D. S. Aliyu. A transformation approach for solving the Hamilton–Jacobi–Bellman equation in H2{H}_{2} deterministic and stochastic optimal control of affine nonlinear systems. Automatica, 39(7):1243–1249, 2003.
  • [6] J. A. Ball, J. W. Helton, and M. L. Walker. H{H}^{\infty} control for nonlinear systems with output feedback. IEEE Transactions on Automatic Control, 38(4):546–559, 1993.
  • [7] R. W. Beard. Successive Galerkin approximation algorithms for nonlinear optimal and robust control. International Journal of Control, 71(5):717–743, 1998.
  • [8] R. W. Beard, G. N. Saridis, and J.  T. Wen. Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation. Automatica, 33(12):2159–2177, 1997.
  • [9] S. C. Beeler, H. T. Tran, and H. T. Banks. Feedback control methodologies for nonlinear systems. Journal of Optimization Theory and Applications, 107(1):1–33, 2000.
  • [10] S. Blanes and F. Casas. A concise introduction to geometric numerical integration. Chapman and Hall/CRC, 2016.
  • [11] John R. Dormand and Peter J. Prince. A family of embedded Runge-Kutta formulae. Journal of computational and applied mathematics, 6(1):19–26, 1980.
  • [12] E. Faou. Geometric numerical integration and Schrödinger equations, volume 15. European Mathematical Society, 2012.
  • [13] K. Feng and M. Qin. Symplectic geometric algorithms for Hamiltonian systems. Springer, 2010.
  • [14] L. Gauckler, E. Hairer, and C. Lubich. Dynamics, numerical analysis, and some geometry. Proceedings of the International Congress of Mathematicians (ICM 2018), pages 453–485, 2019.
  • [15] E. Hairer, C. Lubich, and G. Wanner. Geometric numerical integration illustrated by the Störmer–Verlet method. Acta Numerica, 12:399–450, 2003.
  • [16] E. Hairer, C. Lubich, and G. Wanner. Geometric numerical integration: structure-preserving algorithms for ordinary differential equations, volume 31. Springer Science & Business Media, 2006.
  • [17] T. Horibe and N. Sakamoto. Nonlinear optimal control for swing up and stabilization of the acrobot via stable manifold approach: Theory and experiment. IEEE Transactions on Control Systems Technology, 27(6):2374–2387, Nov 2019.
  • [18] T. Horibe and N. Sakamoto. Optimal swing up and stabilization control for inverted pendulum via stable manifold method. IEEE Transactions on Control Systems Technology, 26(2):708–715, 2017.
  • [19] T. Hunt and A. J. Krener. Improved patchy solution to the Hamilton-Jacobi-Bellman equations. In 49th IEEE Conference on Decision and Control (CDC), pages 5835–5839. IEEE, 2010.
  • [20] A. Isidori and A. Astolfi. Disturbance attenuation and H{H}^{\infty}-control via measurement feedback in nonlinear systems. IEEE Transactions on Automatic Control, 37(9):1283–1293, 1992.
  • [21] G. Kreisselmeier and T. Birkholzer. Numerical nonlinear regulator design. IEEE Transactions on Automatic Control, 39(1):33–46, 1994.
  • [22] P. Lancaster and L. Rodman. Algebraic Riccati equations. Clarendon press, 1995.
  • [23] E. B. Lee and L. Markus. Foundations of Optimal Control Theory. New York: Wiley, 1967.
  • [24] D. L. Lukes. Optimal regulation of nonlinear dynamical systems. SIAM Journal on Control, 7(1):75–100, 1969.
  • [25] J. Markman and I. N. Katz. An iterative algorithm for solving Hamilton–Jacobi type equations. SIAM Journal on Scientific Computing, 22(1):312–329, 2000.
  • [26] W. M. McEneaney. A curse-of-dimensionality-free numerical method for solution of certain HJB PDEs. SIAM Journal on Control and Optimization, 46(4):1239–1276, 2007.
  • [27] C. P. Mracek and J. R. Cloutier. Control designs for the nonlinear benchmark problem via the state-dependent Riccati equation method. International Journal of Robust and Nonlinear Control, 8(4-5):401–433, 1998.
  • [28] C. Navasca and A. J. Krener. Patchy solutions of Hamilton-Jacobi-Bellman partial differential equations. In Modeling, Estimation and Control, pages 251–270. Springer, 2007.
  • [29] T. Ohtsuka. Solutions to the Hamilton-Jacobi equation with algebraic gradients. IEEE Transactions on Automatic Control, 56(8):1874–1885, 2010.
  • [30] N. Sakamoto. Analysis of the Hamilton–Jacobi equation in nonlinear control theory by symplectic geometry. SIAM Journal on Control and Optimization, 40(6):1924–1937, 2002.
  • [31] N. Sakamoto. Case studies on the application of the stable manifold approach for nonlinear optimal control design. Automatica, 49(2):568–576, 2013.
  • [32] N. Sakamoto and A. J. van der Schaft. Analytical approximation methods for the stabilizing solution of the Hamilton–Jacobi equation. IEEE Transactions on Automatic Control, 53(10):2335–2350, 2008.
  • [33] J. M. Sanz-Serna. Symplectic integrators for Hamiltonian problems: an overview. Acta numerica, 1:243–286, 1992.
  • [34] J. M. Sanz-Serna and M.-P. Calvo. Numerical Hamiltonian problems. Courier Dover Publications, 2018.
  • [35] Lawrence F. Shampine. Some practical Runge-Kutta formulas. Mathematics of Computation, 46: 135–150, 1986.
  • [36] Y. B. Suris. The problem of integrable discretization: Hamiltonian approach, volume 219. Birkhäuser, 2012.
  • [37] A. J. van der Schaft. On a state space approach to nonlinear H{H}_{\infty} control. Systems & Control Letters, 16(1):1–8, 1991.
  • [38] A. J. van der Schaft. L2{L}_{2}-gain analysis of nonlinear systems and nonlinear state feedback H{H}_{\infty} control. IEEE Transactions on Automatic Control, 37(6):770–784, 1992.