Matrix Pontryagin principle approach to controllability metrics maximization under sparsity constraints
Abstract
Controllability maximization problem under sparsity constraints is a node selection problem that selects inputs that are effective for control in order to minimize the energy to control for desired state. In this paper we discuss the equivalence between the sparsity constrained controllability metrics maximization problems and their convex relaxation. The proof is based on the matrix-valued Pontryagin maximum principle applied to the controllability Lyapunov differential equation.
keywords:
Sparse optimal control, node selection problem, controllability maximization1 Introduction
Sparse optimal problems have attracted a lot of attention in the field of optimal control. Such an approach is useful to find a small number of essential information that is closely related to the control performance of interest, and it is applied widely, for example, Ikeda et al. (2021). This paper investigates the application of sparse optimization to controllability maximization problem, one of the control node selection problems. The problem is known as the optimization problem minimizing the energy to control for the desired state.
These problems are generally formulated as maximization of some metric of the controllability Gramian with constraints, but it is known that the problems include combinatorial structures. To circumvent this, relaxed problems, where the norms are replaced by the norms, are considered for its computational tractability. Then, the problem is how to prove the equivalence between the main problem and its relaxation. The paper Ikeda and Kashima (2018) proved the equivalence when the trace of controllability Gramian is adopted as the metric, but its usefulness as a metric is questionable since the designed Gramian may include the zero eigenvalue, so the trace metric does not automatically ensure the controllability. The paper Ikeda and Kashima (2022) considered the minimum eigenvalue and the determinant of the controllability Gramian which is useful as metrics, but it avoided the proof of equivalence because of the difficulty and treated approximation problems that are easy to prove the equivalence. In view of this, this paper newly proposes a method to prove the equivalence for general metrics of controllability. Specifically, we adopted the controllability Lyapunov differential equation. The controllability Lyapunov differential equation is a matrix-valued differential equation whose solution is the controllability Gramian. By considering the optimal control problem for this Lyapunov differential equation, we can strictly treat useful metrics that are related to the controllability Gramian.
The remainder of this paper is organized as follows. Section 2 provides mathematical preliminaries. Section 3 formulates our node scheduling problem using controllability Lyapunov differential equation, and gives a sufficient condition for the main problem to boil down to the relaxation problem. Section 4 offers concluding remarks.
Notation
We denote the set of all positive integers by and the set of all real numbers by . Let . We denote the zero matrix of size by and the identity matrix of size by . For any , we denote the Frobenius norm of by , and the inner product of and by . Let be a closed subset of and . A matrix is a proximal normal to the set at the point if and only if there exists a constant such that for all . The proximal normal cone to at is defined as the set of all such , which is denoted by . We denote the limiting normal cone to at by , i.e., . For other notations, see (Ikeda and Kashima, 2022, Section II).
2 Preliminary
Let us consider the following continuous-time linear system
(1) |
where is the state vector consisting of nodes, where is the state of the -th node at time ; is the exogenous control input that influences the network dynamics. Then the controllability Gramian for the system is defined by
(2) |
We next show why the controllability Gramian is used as the metric of the ease of control. We here recall the minimum-energy control problem:
(3) |
The minimum control energy is then given by (Verriest and Kailath (1983)). Based on this, recent works have been considered to make as large as possible. In this paper we design in order to maximize some metric of the controllability Gramian. As the constraints, we introduce and constraints on to take account of the upper bound of the total time length of node activation and the number of activated nodes at each time. We consider the following optimal problem that maximizes some metric of under sparsity constraints:
(4) |
where is a metric of the controllability Gramian, and and is constant.
Since the maximization problem in (4) is a combinatorial optimization problem, we consider the following relaxation problem:
(5) |
This problem is easier to treat than the main problem (especially if is concave, problem (5) is a convex optimization problem). We, however, have to consider the equivalence between the main problem and the corresponding relaxation problem. Ikeda and Kashima (2022) formulated alternative approximation problem instead of proving the equivalence. Then this paper proves the equivalence between the main problem and the relaxed one by using the controllability Lyapunov differential equation.
3 Proposed method
In this section, we formulate a controllability Lyapunov differential equation which holds the controllability Gramian as a solution, and then formulate an optimization problem for a system in which the state space representation is given by the derived differential equation. We provide an equivalence theorem between the main problem and the corresponding relaxation problem.
3.1 Problem formulation and relaxation
Controllability Lyapunov differential equation is given as follows:
(6) |
Then the controllability Gramian defined by (2) corresponds to the solution of (6) at . Here we consider the following optimal control problem.
Problem 1 (Main problem).
(7) |
Note that since , so we rewrite the controllability Lyapunov differential equation. Problem 1 is a combinatorial optimization problem, so we consider the following relaxed problem, where the norms are replaced by the norms, respectively.
Problem 2 (Relaxed problem).
(8) |
In what follows, we suppose that is continuously differentiable.
3.2 discreteness and equivalence
We define the set of feasible solutions of Problem 1 and Problem 2 by and , i.e.,
Note that , since for all and on for any measurable function with on . The inclusion is proper in general, since the / constraints do not automatically guarantee the / constraints and some functions in are not obviously binary. Then, we first show the discreteness of solutions of Problem 2, which guarantees that the optimal solutions of Problem 2 belongs to the set . For this purpose, we prepare lemmas.
Lemma 1 (Matrix Pontryagin principle).
Let us consider the following optimization problem
(9) |
where is continuously differentiable, is continuous, is continuous with respect to , , , , , , and . Note that is given. We define Hamiltonian function associated to problem (9) by
(10) |
Let the process be a local minimizer for the problem (9). Then there exists a matrix , and a scalar equal to 0 or 1 satisfying the following conditions:
-
•
the nontriviality condition:
(11) -
•
the transversality condition:
(12) -
•
the adjoint equation for almost every :
(13) -
•
the maximum condition for almost every :
(14)
Proof.
We define a mapping by
(15) |
where denotes the th column of a matrix . From Athans (1967), is a regular linear mapping (hence exists), and preserves the inner product. Then problem (9) is equivalent to
(16) |
where , , and corresponds to respectively. We define the Hamiltonian function associated to problem (16) by and denote the local minimizer for problem (16) by (, ). Then there exists an arc and a scalar equal to 0 or 1 satisfying the following conditions (Pontryagin’s Maximum Principle (Clarke (2013))):
-
•
the nontriviality condition:
(17) -
•
the transversality condition:
(18) -
•
the adjoint equation for almost every
(19) -
•
the maximum condition for almost every
(20)
Since exists, we obtain the Hamiltonian function associated to as follows:
(21) |
which satisfies (11), (12), (13), and (14), where , , . This completes the proof. ∎
Lemma 2.
Define a set
(22) |
and fix any , where is a set of positions of elements of for which inequality constraints are given. Then any satisfies
(23) | |||
(24) | |||
(25) |
Proof.
Theorem 1.
Proof.
We first reformulate Problem 2 into a form to which Lemma 1 is applicable. The value is equal to the final state of the system with . Define and matrices , , , by
Then, Problem 2 is equivalently expressed as follows:
(30) |
where , , . This is an optimal control problem to which Lemma 1 is applicable. We define the Hamiltonian function associated to problem (30) by
We define two matrices as follows:
(31) |
Then (, ) is the local minimizer of problem (30) because of the equivalence between Problem 2 and problem (30), and there exists a scalar equal to 0 or 1 and a matrix satisfying the conditions (11), (12), (13), (14). It follows from (13) that
which leads to
(32) |
where
(33) |
with and . Note that
by (12), then we have
(34) | |||
(35) | |||
(36) | |||
(37) | |||
(38) |
from Lemma 2. Substituting these into (32), we get
(39) |
Then, we have
It follows from (14) that
(40) |
We here claim that . Indeed, if , follows from (11), i.e., there exists some that satisfies
(41) |
Hence, from (40) and (41), we have for all , i.e., . This contradicts to (41). Thus, . From the assumption, it is easy to verify that
-
1)
we have almost everywhere for all ,
-
2)
there exists , such that
almost everywhere.
Hence, we find
(42) |
for almost every , where
This completes the proof. ∎
The following theorem is the main result, which shows the equivalence between Problem 1 and Problem 2.
Theorem 2 (equivalence).
Proof.
Denote any solution of Problem 2 by . It follows from Theorem 1 that almost everywhere. Note that the null set does not affect the cost, and hence we can adjust the variables so that on , without loss of the optimality. We have
for all . Since , we have and for all and . Thus, . Then,
(43) |
where the first relation follows from , the second relation follows from , and the last relation follows from . Hence, we have
(44) |
which implies . Hence, and is not empty.
Next, take any . Note that , since . In addition, it follows from (44) that . Therefore, , which implies . This gives . ∎
4 Conclusion
In this paper, we discussed the equivalence between the sparsity constrained controllability metrics maximization problems and their convex relaxation. The proof is based on the matrix-valued Pontryagin maximum principle applied to the controllability Lyapunov differential equation. The existence of optimal solutions and computational cost are currently under investigation.
References
- Athans (1967) Athans, M. (1967). The matrix minimum principle. Information and Control, 11, 592–606.
- Clarke (2013) Clarke, F. (2013). Functional Analysis, Calculus of Variations and Optimal Control, volume 264. Springer Science & Business Media.
- Ikeda and Kashima (2018) Ikeda, T. and Kashima, K. (2018). Sparsity-constrained controllability maximization with application to time-varying control node selection. IEEE Control Systems Letters, 2(3), 321–326.
- Ikeda and Kashima (2022) Ikeda, T. and Kashima, K. (2022). Sparse control node scheduling in networked systems based on approximate controllability metrics. IEEE Transactions on Control of Network Systems.
- Ikeda et al. (2021) Ikeda, T., Sakurama, K., and Kashima, K. (2021). Multiple sparsity constrained control node scheduling with application to rebalancing of mobility networks. IEEE Transactions on Automatic Control.
- Verriest and Kailath (1983) Verriest, E. and Kailath, T. (1983). On generalized balanced realizations. IEEE Transactions on Automatic Control, 28(8), 833–844.