Towards Global Optimal Control via Koopman Lifts
Abstract
This paper introduces a framework for solving time-autonomous nonlinear infinite horizon optimal control problems, under the assumption that all minimizers satisfy Pontryagin’s necessary optimality conditions. In detail, we use methods from the field of symplectic geometry to analyze the eigenvalues of a Koopman operator that lifts Pontryagin’s differential equation into a suitably defined infinite dimensional symplectic space. This has the advantage that methods from the field of spectral analysis can be used to characterize globally optimal control laws. A numerical method for constructing optimal feedback laws for nonlinear systems is then obtained by computing the eigenvalues and eigenvectors of a matrix that is obtained by projecting the Pontryagin-Koopman operator onto a finite dimensional space. We illustrate the effectiveness of this approach by computing accurate approximations of the optimal nonlinear feedback law for a Van der Pol control system, which cannot be stabilized by a linear control law.
keywords:
Pontryagin’s maximum principle, Koopman operators, symplectic geometry, optimal feedback control, ,
1 Introduction
During the last decades algorithms and software for computing local solutions of nonlinear optimal control problems have reached a high level of maturity [4, 9, 35], enabling its deployment in industrial applications [29]. By now, such local solutions can be computed within the milli- and microsecond range. Moreover, auto-generated implementations of real-time local optimal control solvers run on embedded hardware systems [24, 18]. In contrast to these developments in local optimal control, algorithms for locating global minimizers of non-convex optimal control problems, can hardly be claimed to be ready for widespread industrial application. There are at least two reasons for this. Firstly, generic global optimization methods can often only be applied to problems of modest dimensions. And secondly, their run-time usually exceeds the run-time of local solvers by orders of magnitude. Nevertheless, in the last years there have been a number of promising developments in the field of global optimal control, which are reviewed next.
Dynamic Programming (DP) [3], which proceeds by approximately solving the Hamilton-Jacobi-Bellman (HJB) equation [13], is a historically important method able to find globally optimal feedback laws. In [23] and [15] tailored discretization grids for DP implementations have been developed that can successfully solve nonlinear optimal control problems, as long as the dimension of the state of the system is small. For higher dimensional state-spaces these methods are, however, not applicable as the complexity of storing and processing the value functions during the DP recursion grows exponentially with the number of states. Other deterministic methods for global optimization problems involving nonlinear differential equations are based on Branch-and-Bound (BB) [7, 11, 22, 32] and its variants, including -BB methods [27, 8]. These BB algorithms have in common that they can effectively solve problems with a small number of decision variables. Unfortunately, implementations of branch-and-bound search easily run out of memory in higher dimensional spaces due to exponentially growing search trees. In particular, as the discretization of control inputs typically leads to a large number of optimization variables, BB-methods are usually not suited for solving optimal control problems.
An alternative to Branch-and-Bound are the so-called Branch-and-Lift (BL) methods [16]. In contrast to BB, these methods never discretize the control inputs directly. Instead, these methods branch over orthogonal projections of the control function in lower dimensional subspaces of increasing dimension, until an -suboptimal global control input is found. A rigorous analysis of the mathematical properties of such BL methods can be found in [17]. However, in the context of developing practical implementations of Branch-and-Lift one faces numerical challenges. In particular, these methods require the computation of accurate enclosures of moment-constrained reachable sets of nonlinear differential equations. At the current state of research, the lack of generic methods for computing such moment-constrained reachable set enclosures limits the applicability of Branch-and-Lift—currently, some problems have been solved successfully with BL by using enclosure propagation methods that are tailored to particular applications [12, 16].
In order to mitigate the above-mentioned limitations of existing global optimal control methods, this paper proposes a completely new framework for constructing methods for solving nonlinear infinite horizon optimal control problems. Here, the main idea is to analyze the spectrum of a Koopman operator that is associated with Pontryagin’s differential equation—under the assumption that Pontryagin’s optimality condition [28] is satisfied at the optimal solution. In order to understand the contributions of this paper, it is important to first be aware of the fact that Koopman operators have originally been introduced almost years ago—in fact, by Koopman himself [19]. This theory has later been extended for more general nonlinear differential equations in [26], where the concept of Koopman mode analysis has been introduced. A more complete analysis of these Koopman mode-based methods from the field of ergodic theory has, however, only appeared much later [1]. Notice that a mathematical analysis of finite dimensional approximations of the Koopman operator can be found in [14] and a variety of related applications of approximate Koopman mode analysis methods can be found in [6]. Moreover, in [20] Koopman mode estimation heuristics for data driven control have been introduced.
Contribution.
The key idea of this paper is to introduce a Koopman operator that is associated with the Pontryagin differential equation of a nonlinear infinite-horizon optimal control problem—in the following we called a Pontryagin-Koopman operator—which can be used to construct globally optimal feedback laws. Section 2 introduces infinite horizon optimal control problems and briefly reviews Pontryagin’s necessary condition for optimality. This review section is followed by a presentation the main contributions of this article, which can be outlined as follows.
-
1.
Section 3 uses ideas from the field of symplectic geometry [2] to characterize the structural properties of Pontryagin-Koopman operators. Section 4 leverages on these structural properties in order to perform a symplectic spectral analysis that finally leads to a complete characterization of globally optimal feedback control laws (see Section 5, Theorem 18). Because this is the first time that the symplecticity of Pontryagin-Koopman operators is analyzed and used to derive practical characterizations of optimal control laws, Theorem 18 can be regarded as the main theoretical contribution of this paper.
-
2.
In order to avoid misunderstandings, it is mentioned clearly that this paper does not claim to solve all the numerical issues regarding the discretization of Pontryagin-Koopman operators that would have to be solved to develop a generic software for global optimal control. However, Section 6 illustrates the practical applicability of the proposed theoretical framework by designing an optimal regulator for a controlled Van der Pol oscillator. Although this implementation is only based on a naive Galerkin projection of the Pontryagin-Koopman operator onto a finite dimensional Legendre basis, this section successfully applies the proposed framework to construct accurate approximations of nonlinear globally optimal feedback laws—in this example, for a system that cannot be stabilized by a linear control law.
The above contributions are relevant for the future of optimal control algorithm development, as they pave the way towards the development of practical procedures for approximating globally optimal control laws of a very general class of nonlinear systems. Therefore, Section 7 does not only conclude the paper, but also outlines the potential and relevance of the proposed framework for future research in control.
Notation.
The distance of a point to a trajectory is denoted by
We denote with the set of (potentially complex-valued) Lebesgue measurable functions whose -th power of the absolute value is integrable on . Moreover, we use the notation to denote the Sobolev space of integrable functions on , whose weak derivatives up to order are all in . The symbol denotes the Hermitian transpose of the matrix . The symbol denotes the gradient operator. The associated second order derivative operator is denoted by . Last but not least, the support of a function is denoted by
2 Infinite horizon optimal control
This section introduces infinite horizon optimal control problems and briefly summarizes existing methods for analyzing the local stability properties of optimal periodic limit orbits (see Section 2.2). Moreover, Sections 2.3 and 2.4 review Pontryagin’s necessary optimal condition thereby introducing the notation that is used throughout the paper.
2.1 Problem formulation
This paper considers infinite horizon optimal control problems of the form
(1) | ||||||
Here, denotes the state trajectory and denotes the control input. The initial value is assumed to be given. Throughout this paper the following assumptions are imposed.
Assumption 1.
The function is twice continuously differentiable and globally Lipschitz continuous in .
Assumption 2.
The function is twice continuously differentiable.
The constant in (1) denotes the optimal average cost,
assuming that this limit exists. Notice that does not need to be known explicitly in order to solve (1), as adding constant offsets to the objective does not affect the solution of an optimization problem. In this paper, the constant is merely introduced for mathematical reasons, such that remains finite and well-defined on infinite horizons under mild regularity assumptions that shall be introduced further below.
2.2 Limit behavior and periodic orbits
In practical instances, one is often interested in whether optimal solutions for the state trajectory of (1) converge to an optimal steady-state or, in more generality, to an optimal periodic limit orbit. These optimal steady states or more general periodic orbits are defined as follows.
Definition 3.
The function is called an optimal periodic orbit, if there exists a period such that for all
-
1.
and ,
-
2.
for all , and
-
3.
.
For the special case that and are constant functions, this pair is called an optimal steady-state.
In order to prepare the following analysis, it is helpful to introduce the shorthands
(2) | ||||
assuming that is an optimal periodic orbit. Here, and denote the partial derivatives of with respect to and and an analogous notation is then also used for the mixed second order derivatives of . The above matrix-valued functions can be used to construct sufficient conditions under which an optimal periodic orbit is locally stabilizable. Here, one relies on the theory of periodic Riccati differential equations [5] of the form
(3) | ||||
It is well-known [5] that if Assumptions 1 and 2 are satisfied and if has full rank, then the existence of a periodic and positive definite function satisfying (3) is sufficient to ensure that the solution of (1) converges to the optimal periodic orbit as long as is sufficiently small—that is, if is in a small neighborhood of the optimal orbit . However, this statement is of a rather local nature. This means that, if we wish to understand the global behavior of system (1), an analysis of the periodic Riccati equation is not sufficient. The following sections focus on analyzing global solutions of (1), under the assumption that is not necessarily small.
2.3 Pontryagin’s differential equations
Pontryagin’s maximum principle [28] can be used to derive necessary conditions for the minimizers of (1). The first order variational optimality condition is summarized as follows. Let be the Hamiltonian function of (1),
(4) |
If Assumptions 1 and 2 hold, is, by construction, twice continuously differentiable in all variables. Next, let
(5) |
denote the associated parametric minimizer of . At this point, we introduce the following regularity assumption.
Assumption 4.
The parametric minimizer in (5) satisfies the second order sufficient condition, .
Notice that for the practically relevant special case that is affine in and strongly convex in , Assumption (4) always holds whenever Assumptions 1 and 2 are satisfied.
Next, if Assumptions 1, 2, and 4 are satisfied, it is well-known [28] that any minimizer of (1) necessarily satisfies Pontryagin’s differential equation
(6) | ||||
(7) |
for a co-state function . In the following, we summarize this differential equation in the form
(8) |
where denotes the stacked state and a stacked version of the right-hand side functions of (6) and (7).
2.4 Necessary boundary and limit conditions
Besides Pontryagin’s differential equation (8), optimal solutions of (1) satisfy necessary boundary conditions. First, of course, the initial condition must hold. Moreover, the co-state satisfies a necessary limit condition, which can be summarized as follows.
Proposition 5.
Let Assumptions 1, 2, and 4 hold. Moreover, let be a primal optimal solution of (1) converging to an optimal periodic orbit at which the periodic Riccati differential equation (3) admits a positive definite solution. Then, there exists a periodic function such that the associated co-state necessarily satisfies
Notice that Proposition 5 is—at least in very similar versions—well-known in the literature [28, 21]. However, as this result is important for understanding the developments in this paper, the following proof briefly recalls the main argument, why this proposition holds.
Proof. Since (3) admits a positive definite solution, the value function in (1) is well-defined and differentiable in a neighborhood of the optimal orbit . Thus, (1) can be replaced by an equivalent finite-horizon optimal control problem as long as is used as a terminal cost. Now, Pontryagin’s principle for optimal control problems with Mayer terms [28] yields the boundary condition
for any horizon length . Clearly, since converges to the optimal periodic orbit , must converge to the associated dual limit orbit . ∎
3 Symplectic Koopman operators
This section analyzes the symplectic structure of the flow associated with Pontryagin’s differential equation. These symplectic structures are needed to understand the properties of the Pontryagin Koopman operators. In fact, Section 3.2 introduces a symplectic test space—that is, a space of observables—in which the Pontryagin-Koopman operator inherits certain symplecticity properties of its underlying incompressible flow field. Moreover, Section 3.3 leverages on ideas from the field of symplectic geometry [2] in order to work out the symplectic dual of the Pontryagin-Koopman operator. Notice that these developments are the basis for the developments in Section 4, which uses a symplectic spectral analysis in order to characterize globally optimal control laws.
3.1 Symplectic Flows
Let denote the flow of Pontryagin’s differential equations such that
for all . If Assumptions 1, 2, and 4 hold, then is a well-defined, continuously differentiable function. In the mathematical literature [2] it is well-known that is a so-called symplectic flow. In order to reveal this symplectic structure it is helpful to introduce the block matrix
Now, the structural properties of can be summarized as follows.
Lemma 6.
Proof. Let us first recall that the right-hand side of Pontryagin’s differential equation is given by (6) and (7), which can also be summarized as
(14) |
This implies that the derivative of with respect can be written in the form
(17) |
Here, we have used Assumptions 1 and 2 to ensure that the second derivatives of with respect to and exist. Moreover, we have used Assumption 4, which implies that is invertible and, as a consequence, that the implicit function theorem holds. In particular, we have
In this form, it becomes clear that the matrix is symmetric and we arrive at the intermediate result
(18) |
In order to proceed, we write the first order variational differential equation for in the form
(19) |
Now, the main idea of this proof is to show that the function
(20) |
vanishes, . Here, we first have by construction, since . Moreover, the derivative of with respect to time is given by
Thus, in summary, we must have for all , which yields the statement of this lemma. ∎
The following corollary summarizes two immediate consequences of Lemma 6 which are both equivalent to stating that is an incompressible flow.
Corollary 7.
Proof. By taking the determinant on both sides of (9) and using that , we find that
Since and since is continuous, this is only possible if (21) holds. Next, by taking the logarithm on both sides of (21) and differentiating with respect to time, we find111Alternatively, the equation can also directly be found by substituting the explicit expression (17) from the proof of Lemma 6.
This completes the proof of the corollary. ∎
3.2 Pontryagin-Koopman operators
The developments from the previous section can be used to analyze the structural properties of the Pontryagin-Koopman operator, which are defined to be the Koopman operator that is associated with the flow of Pontryagin’s differential equation, or, more formally:
Definition 8 (Pontryagin-Koopman Operator).
We use the notation to denote the Pontryagin Koopman operator of , which is defined such that
with denoting the composition operator.
It is well known [19, 26] that is a linear operator satisfying for all . Moreover, satisfies , which follows by substituting in the previous equation and using that .
In order to introduce a notion of symplecticity in the space of observables of the Pontryagin-Koopman operator, we introduce the bilinear form
which is defined as
for all . Because we have , the bilinear form is skew-symmetric,
and is a symplectic space [2].
Theorem 9.
Proof. Assumptions 1, 2, and 4 imply that the function and its inverse are both continuously differentiable. This implies in particular that the equivalence
holds. Next, the definition of the operator and the chain rule for differentiation imply that
(22) |
for all . Furthermore, because Assumptions 1, 2, and 4 hold, it follows from Lemma 6 that
Let us multiply this equation with from the left and substitute . This yields
Next, we multiply the latter equation with from the right and use the relation once more, in order to arrive at the equation
(23) |
Thus, by substituting the previous equations, we get
In order to further transform this integral, we need to introduce the change of variables . Because Corollary 7 ensures that
this change of variables is volume preserving. Thus, this implies that
for all . Thus, is a symplectic operator. ∎
3.3 Symplectic duality
In order to prepare the analysis of the properties of the spectrum of symplectic Koopman operators, it is helpful to introduce a notion of duality. However, instead of defining duality in a Hilbert space, we propose to introduce the following notion of symplectic duality.
Definition 10.
Let be a linear operator. If there exists a linear operator with
for all , then we call the symplectic adjoint operator of .
In the following, we consider a linear differential operator that is given by
(24) |
for all . Notice that this operator is found by differentiating the definition of the Pontryagin-Koopman operator with respect to time, such that .
Lemma 11.
Proof. For any function , we have
(26) |
which follows by differentiating (24) on both sides. Next, we know from Theorem 9 that is a symplectic operator. Thus, we can differentiate the equation
on both sides with respect to , which yields
Now, if , we can expand the derivative on the right as
and, after changing variables, , recalling that , and resorting terms,
(27) |
Moreover, we recall that the equation
(28) |
also holds (see Equation (18)). Thus, in summary, we have
Because the latter equation holds for all , we find that is the symplectic adjoint of , as claimed by the statement of this lemma. ∎
Remark 12.
In many articles on Koopman operators, a so-called transport differential equation of the form
is considered. If one introduces suitable boundary conditions, this advection PDE can be interpreted as the derivative of a Perron-Frobenius operator that is dual to the Koopman operator with respect to the standard -scalar product [10]. Notice that in our case, the symplectic adjoint operator
happens to coincide with the right-hand operator of the above advection PDE. Thus, physically, one could interpret the operator as an advection operator of the incompressible flow field recalling that .
4 Spectral Analysis
In this section, we show that the symplectic structures of the Koopman operator, as analyzed in the previous section, have important consequences on its set of eigenvalues. Moreover, Section 5 uses these spectral properties of the Koopman operator to characterize global optimal control laws that are associated with the infinite horizon optimal control problem (1).
4.1 Eigenfunctions and Eigenvalues
As already mentioned in the introduction, the spectrum of general Koopman operators has been analyzed by many authors; for example in [1, 6, 26]. In the context of this paper, we call a weakly differentiable function an eigenfunction of the differential Pontryagin-Koopman operator , if the Lebesgue measure of the set is either unbounded, or, if it exists, is not equal to , and
for a potentially complex eigenvalue . The fact that the symplectic adjoint of the operator is given by has important consequences on its spectrum, which can be summarized as follows.
Theorem 13.
Proof. Because the functions are eigenfunctions of , they satisfy
(29) |
Thus, since is a bilinear form, we have
and, after resorting terms,
Thus, if , we must have . This is equivalent to the statement of the theorem. ∎
Remark 14.
The statement of Theorem 13 is formally not directly applicable for eigenfunctions of that are locally weakly twice differentiable but whose derivatives are not square integrable, such that are not elements of the Sobolev space . However, in practice, one is usually interested in constructing eigenfunctions of on a compact domain , in the following called the region of interest, such that
(30) |
This region of interest can, for example, model a-priori bounds on the optimal primal and dual trajectories of (1). Consequently, because we are simply not interested in how the flow and the associated eigenfunctions are defined outside of the set , these functions can simply be redefined arbitrarily for , for example, such that the desired eigenfunctions of satisfy by construction. Thus, for the purpose of this paper, it is not restrictive at all to assume that the derivatives of the eigenfunctions of are square integrable. Notice that this technique is also illustrated by our tutorial case study in Section 6, where we explain how to choose and how to discretize (30) on .
Let denote the spectrum of ; that is, the set of eigenvalues of the linear operator . In the following, we use the notation to denote an eigenfunction that is associated with an eigenvalue . Moreover, for a function , we use the shorthand notation
to denote the matrix that is obtained by evaluating the skew symmetric bilinear form for all possible combinations of the components of . We call a skew-orthogonal function in the symplectic space , if it satisfies . Notice that the construction of such skew orthogonal functions is straightforward by using the standard skew-symmetric variant of the Gram-Schmidt algorithm [2].
Definition 15.
The operator is said to admit a spectral decomposition with respect to a given function , if there exists a generalized function such that
(31) |
Notice that the above definition uses the “control engineering notation” for generalized functions, which means that we use as if it was a standard function, although this notation suppresses the distributional nature of . Thus, in mathematical terms, this notation has to be translated as “ represents a distribution of order ; that is, a linear operator on that is Lipschitz continuous with respect to the -norm”.222Notice that, for example, could be a Dirac distribution, which is not a function in the traditional sense but, by construction, a Lipschitz continuous linear operator. The following statement is a consequence of Theorem 13.
Corollary 16.
Proof. By substituting (31) in the equation , we find that the equation
(32) |
holds. Let us have a closer look at the terms in (32). Clearly, the unit matrix on the left has full rank. But, on the other side, we have an integral over the rank matrices . This integral term can only have full rank, if there are at least pairs of eigenvalues for which
But now Theorem 13 implies that all these pairs must be such that and, after sorting all eigenvalues with respect to their real-part, we find that there must be at least eigenvalues with non-negative real part and associated mirrored eigenvalues with non-positive real-part, as claimed by the statement of this corollary. ∎
Remark 17.
Notice that Corollary 16 makes a statement about the spectrum of under the assumption that this operator admits a spectral decomposition with respect to at least one skew orthogonal function. Thus, one should further ask the question under which assumptions the existence of such a spectral decomposition can be guaranteed for at least one such skew orthogonal function. In full generality, this question is difficult to answer, but sufficient conditions for the existence of (much more general and, in our context, sufficient) spectral decompositions can be found in [1, 6, 25, 26], which use ideas from the field of ergodic theory [33] as well as Yoshida’s theorem [34]. For example, if the monodromy matrix, , of a periodic optimal orbit with period length is diagonalizable, one can ensure that a spectral decomposition is possible. This sufficient condition follows by applying Proposition 3 in [25] to the Pontryagin differential equation. A more complete review of such results from the field of functional analysis would, however, go beyond the scope of the present paper.
5 Optimal Feedback Control Laws
The theoretical results from the previous sections can be used to derive optimality conditions, which, in turn, can be used to develop practical numerical algorithms for solving (1). Let us introduce the set
of unstable eigenvalues and its associated invariant manifold
(35) |
Because we assume that the eigenfunctions are weakly differentiable, we may assume without loss of generality that the functions are also continuous—otherwise there exists a continuous function for almost every and we can use instead of recalling that we work with Sobolev spaces, in which such arguments are indeed possible.
In the following, we say that is a regular optimal control law of (1), if the closed-loop trajectories,
are minimizers of (1) at which Pontryagin’s necessary optimality conditions are satisfied such that converges to an optimal periodic orbit at which the conditions of Proposition 5 are satisfied.
Theorem 18.
Proof. Since is a time-autonomous infinitesimal generator of , the eigenfunctions of also satisfy [26]
(36) |
for all . The remainder of the proof is divided into two parts. In the first part, we show that (36) implies that the optimal periodic limit orbit is a subset of the manifold and, in the second part, we use this property of the limit orbit to construct a globally optimal feedback law.
Part I: Let denote the optimal periodic limit orbit of (1) with period time . If we would have for a time , we must have
for at least one . Because the optimal periodic orbit satisfies Pontryagin’s differential equation, this implies in particular that
as has a strictly positive real part. But this is a contradiction, since is periodic. Thus, in summary, we must have
(37) |
Part II: Let denote an optimal solution of (1) with . Now, if we would have , then there would have to exist at least one for which
But if this would be the case, then we would also have
(38) |
as is strictly unstable. But this limit statement is in conflict with (37), since is continuous and we assume that converges to the optimal periodic limit orbit. Thus, in summary, there exists for every a with and the corresponding map from to can be denoted by . The associated optimal control input is given by
This is already sufficient to establish the statement of this theorem, as the optimal feedback law must be time-autonomous.∎
Notice that Theorem 18 can be used to systematically search for globally optimal solutions of (1). Here, one of the key observations is that if, in addition to the assumptions of Theorem 18, the assumptions of Corollary 16 are also satisfied and if none of the non-negative eigenvalues happens to be on the imaginary axis, then the parametric equation system
(39) |
consists of at least independent equations while the number of variables, , is equal to , too. Thus, this equation system can, in many practical instances, be expected to admit a finite number of parametric solutions only. This means that, if all the above regularity assumptions are satisfied, Theorem 18 singles out a finite number of candidate control laws —at least one of which must be globally optimal.
6 Numerical example
This section illustrates how Theorem 18 can be used to construct accurate approximations of globally optimal control laws of (1) for a Van der Pol oscillator system with
(42) | ||||
(43) |
The associated system, , has a linearly uncontrollable equilibrium at . Notice that for this particular example an explicit solution of the Hamilton-Jacobi-Bellman equation is known [30]: it turns out that the optimal value function and the globally optimal feedback law are given, respectively, by
(44) |
In the following, this explicit solution is, however, only used to assess the accuracy of the proposed method; that is, the numerical procedure below neither knows the above expression for nor for .
6.1 Galerkin discretization
In order to illustrate the practical applicability of the developments in Section 5, we introduce a simple Galerkin discretization of the operator . Let be orthogonal functions with respect to the standard -scalar product on . Next, if we compute the coefficients
for all , the matrix can be interpreted as a Galerkin approximation of the operator over the subspace spanned by . In our implementation, we set to the th multivariate Legendre polynomial on the -dimensional compact interval box and we set outside of this domain; that is, for . Consequently, our discretization can only be expected to be accurate inside our region of interest , but other choices for and for the basis functions would be possible, too.
Remark 19.
Standard Galerkin methods are, in general, numerically unstable when applied to advection operators and, consequently, although this method happens to yield reasonable approximations for our particular example, such naive discretization schemes cannot be recommended in general. More advanced discretization schemes for linear advection operators can be found in the modern PDE literature; see, for example [31]. A more complete discussion of such numerical discretization methods for Pontryagin-Koopman operators, is, however, beyond the scope of this paper.



6.2 Approximations of the optimal feedback law
The above Galerkin approximation of the operator can be used to construct approximate eigenfunctions. In detail, if is a left eigenvector of with eigenvalue , then
is an approximation of an eigenfunction of . The right plot in Figure 1 shows the spectrum of the matrix for different choices of (blue circles: , red squares: , and black circles with a cross: ). In order to understand the structure of this spectrum it is helpful to recall Corollary 16, which predicts that there exits at least eigenvalues of such that and are also eigenvalues of . Notice that such symmetric eigenvalue pairs are indeed present in the spectrum of , although is only a Galerkin approximation of .
In order to further illustrate how the above spectral analysis of can be used to construct approximations of the globally optimal control law , we can compute an approximation of the manifold by using the approximate eigenfunctions instead of the exact ones (see Theorem 18). For example, for , the matrix has the eigenvalues
(45) |
and the associated Galerkin approximation of the globally optimal control law is given by
(46) |
The left plot in Figure 1 shows and compares it to the optimal feedback law . In fact, the squared integral error over is approximately . Quite remarkably, this approximate feedback law can even be used to control the system for initial values outside of . The plot in the middle of Figure 1 shows the corresponding trajectories of the closed-loop system that are obtained by using the approximately optimal feedback law .
7 Conclusions
This paper has presented an analysis of infinite horizon nonlinear optimal control problems, whose minimizers satisfy Pontryagin’s necessary conditions of optimality. The proposed formalism is based on Pontryagin-Koopman operators, which have been shown to possess a symplectic structure, as revealed by Theorem 9. Moreover, Theorem 13 and Corollary 16 have established conditions under which the spectrum of the differential Pontryagin-Koopman operator contains at least mirrored eigenvalues. This spectral structure is used in Theorem 18 to characterize optimal control laws.
The theoretical findings of this paper have been applied to construct accurate approximations of a globally optimal control law for a Van der Pol oscillator, which illustrates the potential of the proposed Pontryagin-Koopman operator based framework for the design of global optimal control algorithms. Here, it needs to be highlighted that, in contrast to dynamic programming methods, which rely on the discretization of nonlinear Hamilton-Jacobi-Bellman PDEs, the proposed framework for global optimal control relies on the computation of eigenfunctions of a linear differential operator. This opens the door to an application of linear algebra methods and tailored discretization schemes for linear PDEs, which have never been considered for the computation of optimal control laws. Therefore, the development of tailored, structure-exploiting projection and linear algebra methods for symplectic Pontryagin-Koopman operators and their application to global optimal control can be regarded as a promising direction for future research.
References
- [1] H. Arbabi and I. Mezić. Ergodic theory, dynamic mode decomposition, and computation of spectral properties of the Koopman operator. SIAM Journal on Applied Dynamical Systems, 16(4), 2096–2126.
- [2] V.I. Arnold and S.P. Novikov. Symplectic geometry. Dynamical systems IV. Springer, 2001.
- [3] R. E. Bellman. Dynamic Programming. Princeton University Press, Princeton, 1957.
- [4] L.T. Biegler. An overview of simultaneous strategies for dynamic optimization. Chemical Engineering and Processing, 46:1043–1053, 2007.
- [5] S. Bittanti, P. Colaneri, and G. De Nicolao. The periodic Riccati equation. Springer, 1991.
- [6] M. Budišić, R. Mohr, and I. Mezić. Applied Koopmanism. Chaos, 22(4), 047510, 2012.
- [7] B. Chachuat, A.B. Singer, and P.I. Barton. Global methods for dynamic optimization and mixed-integer dynamic optimization. Ind. Eng. Chem. Res., 45(25):8373–8392, 2006.
- [8] H. Diedam and S. Sager. Global optimal control with the direct multiple shooting method. Optimal Control Applications and Methods, DOI. 10.1002/oca.2324, 2017.
- [9] M. Diehl, H.G. Bock, J.P. Schlöder, R. Findeisen, Z. Nagy, and F. Allgöwer. Real-time optimization and nonlinear model predictive control of processes governed by differential-algebraic equations. Journal of Process Control, 12(4):577–585, 2002.
- [10] J. Ding. The point spectrum of Frobenius-Perron and Koopman operators. Proceedings of the American Mathematical Society, 126(5), 1998.
- [11] W. R. Esposito and C. A. Floudas. Deterministic global optimization in nonlinear optimal control problems. J. Global Optim., 17:96–126, 2000.
- [12] X. Feng, M.E. Villanueva, B. Chachuat, and B. Houska. Branch-and-lift algorithm for obstacle avoidance control. In In Proceedings of the 56th IEEE Conference on Decision and Control, page 745–750, 2017.
- [13] H. Frankowska. Lower semicontinuous solutions of Hamilton-Jacobi-Bellman equations. SIAM Journal on Control and Optimization, 31(1):257–272.
- [14] N. Govindarajan, R. Mohr, S. Chandrasekaran, and I. Mezić. On the approximation of Koopman spectra for measure preserving transformations. SIAM Journal on Applied Dynamical Systems, 18(3):1454–1497, 2019.
- [15] L. Grüne and W. Semmler. Using dynamic programming with adaptive grid scheme to solve nonlinear dynamic models in economics. Computing in Economics and Finance 2002, (99), 2002.
- [16] B. Houska and B. Chachuat. Branch-and-lift algorithm for deterministic global optimization in nonlinear optimal control. Journal of Optimization Theory and Applications, 162(1):208–248, 2014.
- [17] B. Houska and B. Chachuat. Global optimization in Hilbert space. Mathematical Programming, Series A, 173:221–249, 2019.
- [18] B. Houska, H.J. Ferreau, and M. Diehl. An auto-generated real-time iteration algorithm for nonlinear MPC in the microsecond range. Automatica, 47(10):2279–2285, 2011.
- [19] B.O. Koopman. Hamiltonian systems and transformations in Hilbert space. Proceedings of the National Academy of Sciences of the United States of America, 17(5):315–318, 1931.
- [20] M. Korda and I. Mezić. Linear predictors for nonlinear dynamical systems: Koopman operator meets model predictive control. Automatica, 93:149–160, 2018.
- [21] D. Liberzon. Calculus of Variations and Optimal Control Theory: A Concise Introduction. Princeton University Press, 2012.
- [22] Y. Lin and M. A. Stadtherr. Deterministic global optimization of nonlinear dynamic systems. AIChE J., 53(4):866–875, 2007.
- [23] R. Luss. Optimal control by dynamic programming using systematic reduction in grid size. International Journal of Control, 51(5):995–1013, 1990.
- [24] J. Mattingley and S. Boyd. Automatic Code Generation for Real-Time Convex Optimization. Cambridge University Press, 2009.
- [25] I. Mauroy, A.and Mezić. Global stability analysis using the eigenfunctions of the Koopman operator. IEEE Transactions on Automatic Control, 61(11):3356–3369, 2016.
- [26] I. Mezić. Spectral properties of dynamical systems, model reduction and decompositions. Nonlinear Dynamics, 41(1–3):309–325, 2005.
- [27] I. Papamichail and C.S. Adjiman. A rigorous global optimization algorithm for problems with ordinary differential equations. J. Global Optim., 24:1–33, 2002.
- [28] L.S. Pontryagin, V.G. Boltyanskii, R.V. Gamkrelidze, and E.F. Mishchenko. The mathematical theory of optimal processes. John Wiley & Sons, New York, 1962.
- [29] S.J. Qin and T.A. Badgwell. A survey of industrial model predictive control technology. Control engineering practice, 11(7):733–764, 2003.
- [30] Luis Rodrigues, Didier Henrion, and Brian J Cantwell. Symmetries and analytical solutions of the hamilton–jacobi–bellman equation for a class of optimal control problems. Optimal Control Applications and Methods, 37(4):749–764, 2016.
- [31] Hans-Görg Roos, Martin Stynes, and Lutz Tobiska. Robust numerical methods for singularly perturbed differential equations: convection-diffusion-reaction and flow problems, volume 24. Springer Science & Business Media, 2008.
- [32] A. B. Singer and P. I. Barton. Global optimization with nonlinear ordinary differential equations. J. Global Optim., 34(2):159–190, 2006.
- [33] N. Wiener and A. Wintner. Harmonic analysis and ergodic theory. American Journal of Mathematics, 63(2):415–426, 1941.
- [34] K. Yosida. Functional analysis. Springer, Berlin, 1978.
- [35] V.M. Zavala and L.T. Biegler. The advanced step NMPC controller: optimality, stability and robustness. Automatica, 45:86–93, 2009.