Finsler geometries on strictly accretive matrices
Abstract
In this work we study the set of strictly accretive matrices, that is, the set of matrices with positive definite Hermitian part, and show that the set can be interpreted as a smooth manifold. Using the recently proposed symmetric polar decomposition for sectorial matrices, we show that this manifold is diffeomorphic to a direct product of the manifold of (Hermitian) positive definite matrices and the manifold of strictly accretive unitary matrices. Utilizing this decomposition, we introduce a family of Finsler metrics on the manifold and charaterize their geodesics and geodesic distance. Finally, we apply the geodesic distance to a matrix approximation problem, and also give some comments on the relation between the introduced geometry and the geometric mean for strictly accretive matrices as defined by S. Drury in [S. Drury, Linear Multilinear Algebra. 2015 63(2):296–301].
Key words: Accretive matrices; matrix manifolds; Finsler geometry; numerical range; geometric mean
1 Introduction
Given a complex number we can write it in its Cartesian form , where is the real part and is the imaginary part, or we can write it in its polar form as , where is the magnitude and is the phase. The standard metric on defines the (absolute) distance between and as which is efficiently computed using the Cartesian form as . However, sometimes a logarithmic (relative) distance between the numbers contains information that is more relevant for the problem at hand. One such distance is given by , and in this distance measure the point is as close to as it is to . This type of distances have wide application in engineering problems, e.g., as demonstrated in the use of Bode plots and Nichols charts in control theory [41]. Moreover, this type of logarithmic metric has been generalized to (Hermitian) positive definite matrices, with plenty of applications, for example in computing geometric means between such matrices [37], [11, Chp. 6], [29, Chp. XII]. This generalization can done by identifying the set of positive matrices as a smooth manifold and introducing a Riemannian or Finsler metric on it. Here, we follow a similar path and extend this type of logarithmic metrics to so called strictly accretive matrices. More specifically, the outline of the paper is as follows: in Section 2 we review relevant background material and set up the notation used in the paper. Section 3 is devoted to showing that the set of strictly accretive matrices can be interpreted as a smooth manifold, and that this manifold is diffeomorphic to a direct product of the smooth manifold of positive definite matrices and the smooth manifold of strictly accretive unitary matrices. The latter is done using the newly introduced symmetric polar decomposition for sectorial matrices [44]. In Section 4 we introduce a family of Finsler metrics on the manifold, by means of the decomposition from the previous section and so called (Minkowskian) product functions [39]. In particular, this allows us to characterize the corresponding geodesics and compute the geodesic distance. Finally, in Section 5 we given an application of the metric to a matrix approximation problem and also give some comments on the relation between the geodesic midpoint and the geometric mean between strictly accretive matrices as introduced in [16].
2 Background and notation
In the following section we introduce some background material needed for the rest of the paper. At the same time, this section is also be used to set up the notation used throughout. To this end, let denote the set of matrices over the filed of complex numbers. For , let denotes its complex conjugate transpose, let denote its Hermitian part, and let denote its skew-Hermitian part.111Note that this is equivalent to the Toeplitz decomposition since if , then and , see [23, p. 7]. Moreover, by we denote the identity matrix, and for by we denote its spectrum, i.e., , and by we denote it singular values, i.e., .
The following is a number of different sets of matrices that will be used throughout: denotes the set of invertible matrices, denotes the set of unitary matrices, denotes the set of Hermitian matrices, denotes the set of positive definite matrices, i.e., s.t. , denotes the set of skew-Hermitian matrices, and denotes the set of strictly accretive matrices, i.e., if and only if .222The naming used here is the same as in [28, p. 281], in contrast to [8].
Two matrices are said to be congruent if there exists a matrix such that . For matrices we define the inner product , which gives the Frobenius norm . By we denote the spectral norm, i.e., , the larges singular value of . Next, a function is called a symmetric gauge function if for all and all i) if , ii) , iii) , and iv) for all where is any permutation of [36], [34, Sec. 3.I.1]. For any unitary invariant norm, i.e., norms such that for all and all , there exists a symmetric gauge function such that [10, Thm IV.2.1 ], [34, Thm. 10.A.1]. For this reason we will henceforth denote such norms . Moreover, we will call a symmetric gauge function, and the corresponding norm, smooth if it is smooth outside of the origin, cf. [32, Thm. 8.5].
For a vector , by we denote the vector obtained by sorting the elements in in a nonincreasing order. More precisely, is obtained by permuting the elements of such that where . For two vectors , we say that is submajorized (weakly submajorized) by if for and [34, p. 12]. Submajorization (weak submajorization) is a preorder on , and we write . On the equivalence classes of vectors sorted in nonincreasing order it is a partial ordering [34, p. 19].
2.1 Sectorial matrices and the phases of a matrix
Given a matrix , we define the numerical range (field of values) as
Using the numerical range, we can define the set of so called sectorial matrices as
The name comes from the fact that the numerical range of a matrix is contained in a sector of opening angle less than . The latter can be seen from the Toeplitz-Hausdorff theorem, which states that for any matrix , is a convex set [45, Thm. 4.1], [20, Thm. 1.1-2]. Recently, sectorial matrices have received considerable attention in the literature, see, e.g., [8, 35, 17, 33, 46, 44, 12].
Sectorial matrices have several interesting properties. In particular, if is sectorial it is congruent to a unitary diagonal matrix , i.e., for some [21, 15, 19, 26, 24]. Although the decomposition is not unique, the elements in are unique up to permutation, and any such decomposition is called a sectorial decoposition [46]. Using this decomposition, we define the phases of , denoted , as the phases of the eigenvalues of [19, 44, 12];333In [19], these were called canonical angles. by convention we defined them to belong the an interval of length strictly less than . With this definition we have, e.g., that if and only if and , which is true if and only if . Note that the phases of a sectorial matrix is different from the angles of the eigenvalues, i.e., in general where . More precisely, equality holds for normal matrices. The phases have a number of desirable properties that the angles of the eigenvalues do not, see [44].
Another decomposition of sectorial matrices, which will in fact be central to this work, is the so called symmetric polar decomposition [44, Thm. 3.1]: for there is a unique decomposition given by
where and . The latter is the set of sectorial unitary matrices. The phases of are given by the phases of , which is in fact the phases of the eigenvalues of . Therefore, we have that if and only if it has a symmetric polar decomposition such that , i.e., the set of strictly accretive unitary matrices.
2.2 Riemannian and Finsler manifolds
Smooth manifolds are important mathematical objects which show up in such diverse fields as theoretical physics [40], robotics [38], and statistics and information theory [2]. Intuitively, they can be thought of as spaces that locally look like the Euclidean space, and on these spaces one can introduce geometric concepts such as curves and metrics. In particular, all smooth manifolds admit a so called Riemannian metric [27, Thm. 1.4.1], [30, Prop. 13.3], and Riemannian geometry is a well-studied subjects, see, e.g., one of the monographs [40, 29, 27, 30, 31]. An relaxation of Riemannian geometry leads to so called Finsler geometry [14]; loosely expressed it can be interpreted as chaining the tangent space from a Hilbert space to a Banach space. For an introduction to Finsler geometry, see, e.g., [9, 42, 13].
More specifically, given a smooth manifold , for we denote the tangent space by and the tangent bundle by . A Riemannian metric is induced by an inner product on the tangent space, , that varies smoothly with the base point . Using this inner product, one defines the norm , which in fact defines a smooth function on the slit tangent bundle . In this work we consider Finsler structures on smooth (matrix) manifolds, but we will limit the scope to Finsler structures , , where is a norm on which is not necessarily induced by an inner product, and such that is smooth on the slit tangent bundle . Given a piece-wise smooth curve , the arc length on the manifold is defined using this Finsler structure. More precisely, it is defined as
where is the derivative of with respect to . Using arc length, the geodesic distance between two points is defined as
and a minimizing curve (if one exists) is called a geodesic. A final notion we need is that of diffeomorphic manifolds. More precisely, two smooth manifolds and are said to be diffeomorphic if there exists a diffeomorphism , i.e., a function which is a smooth bijection with a smooth inverse. In this case we write .
Next, we summarize some results regarding two matrix manifolds, namely and , together with specific Finsler structures. These will be needed later.
2.2.1 A family of Finsler metrics on and their geodesics
Riemannian and Finsler geometry on is a well-studied subject, and we refer the reader to, e.g., [37], [11, Chp. 6] or [29, Chp. XII]. Here, we summarize some of the results we will need for later. To this end, note that the tangent space at is , and given , we can introduce the inner product on the tangent space as
(2.1) |
with corresponding norm . The geodesic between in the induced Riemannian metric is given by
(2.2) |
and the length of the curve, i.e., the Riemannian distance between and , is given by
Interestingly, if the norm on is changed to any other unitary invariant matrix norm
the expressions for a geodesic between two matrices remains unchanged and the corresponding distance is given by [11, Sec. 6.4]. However, the geodesic (2.2) might no longer be the unique shortest curve [11, p. 223].
An alternative expression for the geodesic (2.2) is given by the following proposition.
Proposition 2.1.
Let be any smooth unitarily invariant norm and consider the Finsler structure given by , . For , a geodesic between them can be written as where , is a simultaneous diagonalization by congruence of and , i.e., and is diagonal with positive elements on the diagonal. Moreover, the geodesic distance is
(2.3) |
Proof.
To show the second equality in (2.3), note that
where the second equality comes from that the singular values of a Hermitian matrix, i.e., the matrix , are the absolute values of the eigenvalues, the third equality follows since the symmetric gauge function is invariant under sign changes, the forth can be seen by using a unitary diagonalization of , and the fifth equality comes from that the spectrum is invariant under similarity. Next, since both , by [23, Thm. 7.6.4] we can simultaneously diagonalize and by congruence, i.e., there exists an such that and , where and where are all strictly larger than . In fact, are the eigenvalues of , which means that which in turn gives the last equality in (2.3). Finally, this also gives that ∎
2.2.2 A family of Finsler metrics on and their geodesics
The set of unitary matrices is a Lie group, and results related to Riemannian and Finsler geometry on can be found in, e.g., [3, 5, 4, 7, 6]. Again, we here summarize some of the results that we will need for later. To this end, note that the tangent space at is and given , we can introduce the inner product on the tangent space as
(2.4) |
with corresponding induced norm . The induced Riemannian metric have shortest curves between given by
(2.5) |
where , and where is such that . Moreover, the geodesic distance is
(2.6) |
and the geodesic is unique if . Similarly to the results for the smooth manifold , if the norm on is changed to any other unitary invariant matrix norm the expressions for a geodesic in (2.5) is unchanged and the expression for the geodesic distance (2.6) is only changed to using the corresponding norm [4, 7]. However, even if the geodesic (2.5) might not be unique in this case [7, Sec. 3.2].
3 The smooth manifold of strictly accretive matrices
In this section we prove that is a smooth manifold diffemorphic to . The latter is fundamental for the introduction of the Finsler structures in the following section. For improved readability, some of the technical results of this section are deferred to Appendix A. To this end, we start by proving the following.
Theorem 3.1.
is a connected smooth manifold, and at a point the tangent space is .
Next, we prove the following characterization of the manifold.
Theorem 3.2.
, where and are embedded submanifolds.
This theorem follows as a corollary to the following proposition.
Proposition 3.3.
For , let be the symmetric polar decomposition. Then the mapping is a diffeomorphism between the smooth manifolds and .
Proof.
Since and are smooth manifolds (see [11, Chp. 6] and Lemma A.4, respectively), is also a smooth manifold [30, Ex. 1.34]. Next, note that the matrix square root is a diffeomorphism of to itself, with the matrix square as the inverse, cf. [23, Thm. 7.2.6]. Therefore, it suffices to show that the mapping is a diffeomorphism between and . To this end, first observe that the latter is a bijection due to the existence and uniqueness of a symmetric polar decomposition [44, Thm. 3.1]. Moreover, that the inverse is smooth follows since the components in are polynomial in the components in and .
To show that and are smooth in , we note that since is strictly accretive, . This means that we can write
where is a normal matrix444To see this, note that a matrix is normal if and only if and commute, see, e.g., [45, Thm. 9.1]. (cf. [46, Proof of Cor. 2.5]) which by construction depends smoothly on . Now, let be the polar decomposition of . Since depends smoothly on , and since the polar decomposition is smooth in the matrix (Lemma A.5), and are smooth in . Moreover, since is normal, and commute [45, Thm. 9.1], and thus and commute. Therefore, , where all components depend smoothly on . Now, let and let be the polar decomposition of . Similar to before, and are thus both smooth in . Finally, we thus have that and since and are unitary so is . By the uniqueness of the symmetric polar decomposition [44, Thm. 3.1] it follows that and , which are both smooth in . ∎
Proof of Theorem 3.2.
Remark 3.4.
The results in Theorem 3.1 and 3.2 can easily be generalized to other subsets of sectorial matrices, namely any subset of all matrices such that there exists , , and , for which and . To see this, note that for we have that , i.e., that with a diffeomorphism between the components in the symmetric polar decomposition. Examples of such sets of matrices are and the set of strictly dissipative matrices, i.e., matrices such that [28, p. 279], and matrices such that .555Note that the latter has unfortunately also been termed dissipative in the literature, see [28, p. 279].
4 A family of Finsler metrics on and their geodesics
In this section we introduce a family of Finsler structures on and in particular we will characterize the geodesics and geodesic distances corresponding to these structures. To this end, by Theorem 3.2 we have that . Moreover, and are smooth manifolds that are well-studied in the literature, and since and are embedded submanifolds of a desired property would be that when restricted to any of the two embedded submanifold the introduced Finsler structure would yield the corresponding known Finsler structure. To this end, we first characterize the geodesics and the geodesic distance on .
Proposition 4.1.
Let be any smooth unitarily invariant norm and consider the Finsler structure given by , . A geodesic between and is given by
Moreover, the geodesic distance is given by
(4.1) |
Proof.
Let . The proposition follows if we can show that a geodesic on between and , given by (2.5), remains in for , and that in (2.5) and (2.6) takes the form . To show the latter, note that , where since both and are strictly accretive, see [43], [44, Thm. 6.2]. Therefore we can use the principle branch of the logarithm, which gives . Next, to show that for , note that . By [16, Prop. 3.1 and Thm. 3.4], is strictly accretive, and a repeated argument now gives that is strictly accretive for a dense set of . By continuity of the map , the result follows. ∎
Next, let and be two smooth symmetric gauge functions and consider the Finsler manifolds and as defined in Proposition 2.1 and Proposition 4.1, respectively. In the Riemannian case, there is a canonical way to introduce a metric on , namely the product metric [30, Ex. 13.2]. However, in the case of products of Finsler manifolds there is no canonical way to introduce a Finsler structure on a product space, cf. [9, Ex. 11.1.6], [39]. Here, we consider so called (Minkowskian) product manifolds [39] and to this end we next define so called (Minkowskian) product functions.
Definition 4.2 ([39]).
A function is called a product function if it satisfies the following conditions:
-
i)
if and only if ,
-
ii)
for all and all ,
-
iii)
is smooth on ,
-
iv)
on , for ,
-
v)
on .
For any product function , is a Finsler manifold [39], and we therefore define the Finsler manifold as follows.666In [39], the convention is that the Finsler structure is squared compared to the one in [9]. We follow the convention of the latter.777Note that functions fulfilling i)-v) are not necessarily symmetric gauge functions. As an example, consider [39, Rem. 6]; this is not a symmetric gauge function since in general . Conversely, symmetric gauge functions do not in general fulfill i)-v). As an example, consider [34, p. 138]; at any point such that , and hence condition iv) is not fulfilled for this function.
Definition 4.3.
One particular example of a product function is , which in [39] this was called “the Euclidean product”, and in the Riemannian case this leads to the canonical product manifold. Moreover, the geodesics and geodesic distance can be characterized for general product functions . This leads to the following result.
Theorem 4.4.
Let , and let and be the corresponding symmetric polar decompositions. On the Finsler manifold , a geodesic from to is given by
(4.2a) | |||
where | |||
(4.2b) | |||
(4.2c) |
Moreover, the geodesic distance from to is given by
(4.3) | ||||
where denotes the angles of the eigenvalues.
Proof.
That is a Finsler manifold follows from the discussion leading up to the theorem; see [39]. Moreover, that (4.2) is a geodesic follows (by construction) by using [39, Thm. 3] together with Proposition 2.1 and Proposition 4.1; this also gives the first two equalities in (4.3). To prove the last equality, first observe that is the geometric mean of and and hence positive definite [11, Thm. 4.1.3], [16, Sec. 3]. Therefore, is Hermitian and thus
Similarly, is unitary and thus is skew-Hermitian. Therefore, and hence
Finally, for unitary invariant norms we have that , and since for any symmetric gauge function , the last equality follows. ∎
Next, we derive some properties of the geodesic distance in Theorem 4.4.
Proposition 4.5.
For matrices we have that
-
1)
-
2)
-
3)
, and the geodesic midpoint between and is
-
4)
for any we have that
Proof.
To prove the statements, first note that , and that . To prove 1), we observe that
For the positive definite part, we have that
and since the symmetric gauge function is invariant under sign changes the distance corresponding to the positive definite part is equal. Similarly, for the strictly accretive unitary part we have that , where the first equality follows from that the absolute value of the angles of the eigenvalues of a unitary matrix are invariant under the operation of taking conjugate transpose, and the last equality follows since the angles of the eigenvalues are invariant under unitary congruence. Statement 2) follows by a similar argument. To prove statement 3),
where the second equity follows by an argument similar to previous ones, and the third equality follows from property ii) for symmetric gauge functions and property ii) for product functions. The second equality in 3) now follows from 1), and that follows by a direct calculation using (4.2). Finally, to prove 4), simply note that , i.e., the same unitary congruence transformation applied to and individually. A direct calculation, using the unitary invariance of eigenvalues, the matrix logarithm, and the norms, gives the result. ∎
Remark 4.6.
Note that is in general not a complete metric space. In particular, in the Riemannian case, i.e., with symmetric gauge functions , , and product function , is not a complete metric space since it is not geodesically complete [31, Thm. 6.19]. The latter is due to the fact that is not geodesically complete; for geodesics are not defined for all since they will reach the boundary. For an example of a Cauchy sequence that does not converge to an element in , consider the sequence , where for all . In this case, the geodesic distance between and is given by
since . Thus is a Cauchy sequence, however . Since the Hopf-Rinow theorem [31, Thm. 6.19] also carries over to Finsler geometry [9, Thm. 6.6.1], similar statements are true also in the general case.
Remark 4.7.
Using Remark 3.4, the above results can easily be generalized to the same subsets of sectorial matrices. In fact, a direct calculation shows that all the algebraic expressions in Theorem 4.4 still hold in this case. However, statements 1)-3) of Proposition 4.5 use the fact that if then . This is in general not true for other sets .
5 An application and some related results
By construction, on the Finsler manifold the question “given , which matrix is closest to ” have the answer “, where is the symmetric polar decomposition.” Similarly, the corresponding question “which matrix is closest to ” have the answer “”. In this section we consider an application of the distance on the Finsler manifold to another matrix approximation problem, namely finding the closest matrix of bounded log-rank, the definition of which is given in Section 5.1. Moreover, in Section 5.2 we consider the relation between the midpoint of geodesics on and the geometric mean of strictly accretive matrices as introduced in [16].
5.1 Closest matrix of bounded log-rank
For a positive definite matrix , we defined the log-rank as the rank of the matrix logarithm of . This is equivalent to the number of eigenvalues of that are different from . We denote this . Analogously, for a unitary matrix the log-rank can be defined as the rank of the matrix logarithm of , which is equivalent to the number of eigenvalues of with phase different from . We denote this . For strictly accretive matrices we define the log-rank as follows.
Definition 5.1.
For with symmetric polar decomposition , we define the log-rank of as
We now consider the log-rank approximation problem: given find , the latter with log-rank bounded by , that is closest to in the geodesic distance . This can be formulated as the optimization problem
(5.1a) | ||||
subject to | (5.1b) |
Let be the symmetric polar decomposition. By properties i) - iv) in the Definition 4.2 of product functions , for each such function the distance is nondecreasing in each argument separately. Therefore, by the form of the geodesic distance (4.3) and the definition of log-rank on , it follows that (5.1) splits into two separate problems over and , namely
(5.2a) | ||||
subject to | (5.2b) |
and
(5.3a) | ||||
subject to | (5.3b) |
In fact, this gives the following theorem.
Theorem 5.2.
In [47, Thm. 3] it was shown that (5.3) always has an optimal solution, and that it is the same for all symmetric gauge functions . More precisely, the optimal solution is obtained from by setting the phases of with smallest absolute value equal to . That is, let be a diagonalization of where and where are the phases of ordered so that . Then
is the optimal solution to (5.3). In the same spirit, the optimal solution to (5.2) can be charaterized as follows.
Proposition 5.3.
Let , and let be a diagonlization of where the eigenvalues are ordered so that . Then is a minimizer to (5.2) for all symmetric gauge functions .
Proof.
Clearly, and hence is feasible to (5.2). Next, . Moreover, since and hence is unitary diagonalizable and has positive eigenvalues, we have that . Now, to show that is the minimizer to (5.2) for all symmetric gauge functions , it is equivalent to show that for all such that ; see, e.g., [18, Thm. 4], [36, Thm. 1], [22, Sec. 3.5], [34, Prop. 4.B.6], [45, Thm. 10.35 ].
To this end, using [34, Thm. 9.H.1.f] (or [10, Cor III.4.6 ], [45, Thm. 10.30 ]) we have that ,888To see this, take and in [34, Thm. 9.H.1.f], and use the fact that the eigenvalues of are invariant under the similarity transform . and by [10, Ex 11.3.5] we therefore have that . Moreover, by a direct calculation it can be verified that holds. Therefore, if we can show that
(5.4) | ||||
for all for all such that | , |
we would have that for all such ,
and by transitivity of preorders the result follows. To show (5.4), we formulate the following equivalent optimization problem: let and consider
subject to | |||
where is minimizing with respect to the preordering . The solution to the latter is to take for the elements of with largest absolute value. ∎
5.2 On the geometric mean for strictly accretive matrices
The geometric mean of strictly accretive matrices, denoted by , was introduced in [16] as a generalization of the geometric mean for positive definite matrices [11, Chp. 4 and 6], [37]. In particular, in [16] it was shown that for there is a unique solution to the equation . This solution is given explicitly as
which is also the same algebraic expression as for the geometric mean of positive definite matrices.
The geometric mean for positive definite matrices can also be interpreted as the midpoint on the geodesic connecting the matrices [11, Sec. 6.1.7]. With the Finsler geometry , we can therefore get an alternative definition of the geometric mean between strictly accretive matrices as the geodesic midpoint. However, for we in general have that . This can be seen by the following simple example.
Example 5.4.
Let and let . Then we have that and . Thus, in general . In fact, equality holds in this case if and only if and commute.
Instead, the midpoint of the geodesic between two matrices can be expressed using the geometric mean as
(5.5) |
which follows directly from (4.2). Using this representation, we can characterize when . In order to do so, we first need the following two auxiliary result.
Lemma 5.5.
Let , and let be is polar decomposition, where and . is normal if and only if is the symmetric polar decomposition of .
Proof.
First, using [21, Lem. 9] we conclude that since is sectorial, is also sectorial. Now, is normal if and only if and commute [45, Thm. 9.1], which is true if and only if and commute. Hence is normal if and only if , and by the existence and uniqueness of the symmetric polar decomposition the result follows. ∎
Lemma 5.6.
Let and . For all , the unique strictly accretive solution to
is .
Proof.
That solves the equations is easily verified by simply plugging it in. Moreover, that is unique follows from the uniqueness of the geometric mean for strictly accretive matrices [16, Sec. 3] and the fact that for any we have that . ∎
Proposition 5.7.
For , let and be the corresponding symmetric polar decompositions. We have that if one of the following holds:
-
i)
,
-
ii)
,
-
iii)
and are commuting normal matrices.
Proof.
Using (5.5), the first statement follows immediately. To prove the second statement, let and . From (5.5) it therefor follows that . Using Lemma 5.6 with and , where therefore have that
To prove the third statement, by Lemma 5.5 we have that , , and , commute. Moreover, since commuting normal matrices are simultaneously unitarilty diagonalizable [23, Thm. 2.5.5], and since a unitary diagonlization is unique up to permutation of the eigenvalues and eigenvectors, it follows that and all commute. Using this together with (5.5), a direct calculation gives the result. ∎
As noted in the above proof, if and are normal and commute they are also simultaneously unitarily diagonalizable [23, Thm. 2.5.5], i.e., and for some . In this case, using Lemma 5.6 we have that , and the geometric mean between and can thus be interpreted as the (independent) geometric mean between the corresponding pairs of eigenvalues. In fact, the latter observation can be generalized to all pairs of matrices that can be simultaneously diagonalized by congruence, albeit that the elements of the diagonal matrices are not necessarily eigenvalues in this case (cf. Proposition 2.1).
Proposition 5.8.
Let and assume that and , where and where and are diagonal matrices. Then .
Proof.
Let and assume that and , where and where and are diagonal matrices. A direct application of Lemma 5.6, with and , gives the result. ∎
As a final remark, note that if and in Proposition 5.8 are unitary, then and are sectorial decompositions of and , respectively. Now, let be the polar decomposition of , with and . Hence we have that and , i.e., is the positive definite part and and are the strictly accretive unitary part in the symmetric polar decomposition of and , respectively. From Proposition 5.7.ii) we therefore have that in this case.
6 Conclusions
In this work we show that the set of strictly accretive matrices is a smooth manifold that is diffeomorphic to a direct product of the smooth manifold of positive definite matrices and the smooth manifold of strictly accretive unitary matrices. Using this decomposition, we introduced a family of Finsler metrics and studied their geodesics and geodesic distances. Finally, we consider the matrix approximation problem of finding the closest strictly accretive matrix of bounded log-rank, and also discuss the relation between the geodesic midpoint and the previously introduced geometric mean between accretive matrices.
There are several interesting ways in which these results can be extended. For example, in the case of positive definite matrices the geometric framework offered by the Riemannian manifold construction gives yet another interpretation of the geometric mean. In fact, the geometric mean between two positive definite matrices and is also the (unique) solution to the variational problem [11, Sec. 6.2.8], [37, Prop. 3.5], and this interpretation can be used to extend the geometric mean to a mean between several matrices [37]. In a similar way, a geometric mean between the strictly accretive matrices can be defined as the solution to
however such a generalization would need more investigation. For example, even in the case of the Riemannian metric on the manifold of positive definite matrices, analytically computing the geometric mean between several matrices is nontrivial [37, Prop. 3.4]. Nevertheless, there are efficient numerical algorithms for solving the latter problem, see, e.g., the survey [25] or the monograph [1] and references therein.
The idea of this work was to introduce a metric that separates the “magnitudes” and the “phases” of strictly accretive matrices. However, the similarities between the manifold of positive definite matrices and the manifold of unitary matrices raises a question about another potential geometry on that does not explicitly use the product structure. More precisely, note that since all strictly accretive matrices have a unique, strictly accretive square root, the inner product on the tangent space , given by (2.4), can be defined analogously to the one on , given by (2.1), namely as
for and . Based on the similarities between the inner products, and the corresponding geodesics and geodesic distances, we ask the following question: for and , if we define the inner product on as
what is the form of the geodesics and the geodesic distance?
Acknowledgments
The authors would like to thank Wei Chen, Dan Wang, Xin Mao, Di Zhao, and Chao Chen for valuable discussions.
Appendix A Technical results from Section 3
Lemma A.1.
is an open sets in .
Proof.
Lemma A.2.
is connected.
Proof.
By [30, Prop. 1.11], is connected if and only if it is path-connected. To show the latter, it suffices to show that any is path-connected to . To this end, let be the sectorial decomposition of . A piece-wise smooth path connecting and is given by
and hence is connected. ∎
Lemma A.3.
The tangent space at an is given by .
Proof.
This follows since is an open subset of . ∎
Lemma A.4.
is a connected smooth manifold and at a point the tangent space is .
Proof.
Lemma A.5 (Cf. [29, Prop. VII.2.5]).
For , let where and be the polar decomposition of . The mapping is a diffeomorphis between the manifolds and .
Proof.
First, since and are smooth manifolds so is [30, Ex. 1.34]. Next, for each the polar decomposition is unique [23, Thm. 7.3.1], [45, Prob. 3.2.20], and for each pair of matrices we have that ; thus the mapping is bijective. Now, is smooth in and since it is polynomial in the coefficients, i.e., the inverse mapping is smooth. To prove the converse, note that the components in the polar decomposition are given by and , cf. [23, p. 449], [45, p. 288]. Since , , and the matrix square root is a smooth function on , cf. [23, Thm. 7.2.6]. Therefore, since both and are compositions of smooth functions of , they both depend smoothly on the components of . ∎
References
- [1] Pierre-Antoine Absil, Robert Mahony, and Rodolphe Sepulchre. Optimization algorithms on matrix manifolds. Princeton University Press, Princeton, NJ, 2008.
- [2] Shun-ichi Amari and Hiroshi Nagaoka. Methods of information geometry. American Mathematical Society, Providence, RI, 2000.
- [3] Esteban Andruchow. Short geodesics of unitaries in the metric. Canadian Mathematical Bulletin, 48(3):340–354, 2005.
- [4] Esteban Andruchow, Gabriel Larotonda, and Lázaro A. Recht. Finsler geometry and actions of the -Schatten unitary groups. Transactions of the American Mathematical Society, 362(1):319–344, 2010.
- [5] Esteban Andruchow and Lázaro A. Recht. Geometry of unitaries in a finite algebra: variation formulas and convexity. International Journal of Mathematics, 19(10):1223–1246, 2008.
- [6] Jorge Antezana, Eduardo Ghiglioni, and Demetrio Stojanoff. Minimal curves in and with respect to the spectral and the trace norms. Journal of Mathematical Analysis and Applications, 483(2):123632, 2020.
- [7] Jorge Antezana, Gabriel Larotonda, and Alejandro Varela. Optimal paths for symmetric actions in the unitary group. Communications in Mathematical Physics, 328(2):481–497, 2014.
- [8] Charles S. Ballantine and Charles R. Johnson. Accretive matrix products. Linear and Multilinear Algebra, 3(3):169–185, 1975.
- [9] David Bao, Shiing-Shen Chern, and Zhongmin Shen. An introduction to Riemann-Finsler geometry. Springer, New York, NY, 2000.
- [10] Rajendra Bhatia. Matrix analysis. Springer, New York, NY, 1997.
- [11] Rajendra Bhatia. Positive definite matrices. Princeton university press, Princton, NJ, 2007.
- [12] Wei Chen, Dan Wang, Sei Zhen Khong, and Li Qiu. Phase analysis of MIMO LTI systems. In 2019 IEEE 58th Conference on Decision and Control (CDC), pages 6062–6067. IEEE, 2019.
- [13] Xinyue Cheng and Zhongmin Shen. Finsler geometry. Springer, Berlin, Heidelberg, 2012.
- [14] Shiing-Shen Chern. Finsler geometry is just Riemannian geometry without the quadratic restriction. Notices of the American Mathematical Society, 43(9):959–963, 1996.
- [15] Charles R. DePrima and Charles R. Johnson. The range of in GL(n, C). Linear Algebra and its Applications, 9:209–222, 1974.
- [16] Stephen Drury. Principal powers of matrices with positive definite real part. Linear and Multilinear Algebra, 63(2):296–301, 2015.
- [17] Stephen Drury and Minghua Lin. Singular value inequalities for matrices with numerical ranges in a sector. Operators and Matrices, 8(4):1143–1148, 2014.
- [18] Ky Fan. Maximum properties and inequalities for the eigenvalues of completely continuous operators. Proceedings of the National Academy of Sciences of the United States of America, 37(11):760, 1951.
- [19] Susana Furtado and Charles R. Johnson. Spectral variation under congruence. Linear and Multilinear Algebra, 49(3):243–259, 2001.
- [20] Karl E. Gustafson and Duggirala K.M. Rao. Numerical range. Springer, New York, NY, 1997.
- [21] Alfred Horn and Robert Steinberg. Eigenvalues of the unitary part of a matrix. Pacific Journal of Mathematics, 9(2):541–550, 1959.
- [22] Roger A. Horn and Charles R. Johnson. Topics in matrix analysis. Cambridge University Press, New York, NY, 1994.
- [23] Roger A. Horn and Charles R. Johnson. Matrix analysis. Cambridge university press, New York, NY, 2013.
- [24] Roger A. Horn and Vladimir V. Sergeichuk. Canonical forms for complex matrix congruence and ∗congruence. Linear algebra and its applications, 416(2-3):1010–1032, 2006.
- [25] Ben Jeuris, Raf Vandebril, and Bart Vandereycken. A survey and comparison of contemporary algorithms for computing the matrix geometric mean. Electronic Transactions on Numerical Analysis, 39:379–402, 2012.
- [26] Charles R. Johnson and Susana Furtado. A generalization of Sylvester’s law of inertia. Linear Algebra and its Applications, 338(1-3):287–290, 2001.
- [27] Jürgen Jost. Riemannian geometry and geometric analysis. Springer, Berlin, Heidelberg, 2008.
- [28] Tosio Kato. Perturbation theory for linear operators. Springer, Berlin, Heidelberg, 1995.
- [29] Serge Lang. Fundamentals of differential geometry. Springer, New York, NY, 1999.
- [30] John M. Lee. Introduction to Smooth Manifolds. Springer, New York, NY, 2013.
- [31] John M. Lee. Introduction to Riemannian Manifolds. Springer, Cham, 2018.
- [32] Adrian S. Lewis. Group invariance and convex matrix analysis. SIAM Journal on Matrix Analysis and Applications, 17(4):927–949, 1996.
- [33] Chi-Kwong Li and Nung-Sing Sze. Determinantal and eigenvalue inequalities for matrices with numerical ranges in a sector. Journal of Mathematical Analysis and Applications, 410(1):487–491, 2014.
- [34] Albert W Marshall, Ingram Olkin, and Barry C Arnold. Inequalities: theory of majorization and its applications. Springer, New York, NY, 2nd edition, 2011.
- [35] Roy Mathias. Matrices with positive definite Hermitian part: Inequalities and linear systems. SIAM journal on matrix analysis and applications, 13(2):640–654, 1992.
- [36] Leon Mirsky. Symmetric gauge functions and unitarily invariant norms. The quarterly journal of mathematics, 11(1):50–59, 1960.
- [37] Maher Moakher. A differential geometric approach to the geometric mean of symmetric positive-definite matrices. SIAM Journal on Matrix Analysis and Applications, 26(3):735–747, 2005.
- [38] Richard M. Murray, Zexiang Li, and S. Shankar Sastry. A mathematical introduction to robotic manipulation. CRC press, Boca Raton, FL, 1994.
- [39] Tsutomu Okada. Minkowskian product of Finsler spaces and Berwald connection. Journal of Mathematics of Kyoto University, 22(2):323–332, 1982.
- [40] Barrett O’Neill. Semi-Riemannian geometry with applications to relativity. Academic press, San Diego, CA, 1983.
- [41] Karl Johan Åström and Richard M. Murray. Feedback systems. Princeton university press, Princeton, NJ, 2008.
- [42] Hideo Shimada and Vasile Soriin Sabău. Finsler geometry. In P.L. Antonelli, editor, Finslerian Geometries: A Meeting of Minds, pages 15–24. Springer, Dordrecht, 2000.
- [43] Robert C. Thompson. On the eigenvalues of a product of unitary matrices I. Linear and Multilinear Algebra, 2(1):13–24, 1974.
- [44] Dan Wang, Wei Chen, Sei Zhen Khong, and Li Qiu. On the phases of a complex matrix. Linear Algebra and its Applications, 593:152–179, 2020.
- [45] Fuzhen Zhang. Matrix theory: basic results and techniques. Springer, New York, NY, 2011.
- [46] Fuzhen Zhang. A matrix decomposition and its applications. Linear and Multilinear Algebra, 63(10):2033–2042, 2015.
- [47] Di Zhao, Axel Ringh, Li Qiu, and Sei Zhen Khong. Low phase-rank approximation. Submitted, 2020.