Minimizing effects of the Kalman gain on Posterior covariance Eigenvalues, the characteristic polynomial and symmetric polynomials of Eigenvalues
Abstract
The Kalman gain is commonly derived as the minimizer of the trace of theposterior covariance. It is known that it also minimizes the determinant of the posterior covariance. I will show that it also minimizes the smallest Eigenvalue and the chracteristic polynomial on and is critical point to all symmetric polynomials of the Eigenvalues, minimizing some. This expands the range of uncertainty measures for which the Kalman Filter is optimal.
keywords:
Kalman Filter , Uncertainty measures , symetric Polynomials[label1]organization=Department of Mathematics, University of Tennessee Knoxville,city=Knoxville, postcode=37996, state=TN, country=USA
In a Kalman filter algorithm the Kalman gain is defined by
(1) |
where is the covariance matrix of prior, is the covariance matrix of the Likelihood and is the measurement operator. [1, 5] The posterior covariance matrix is defined through
(2) |
evaluated at , where is the identity matrix. Note that as covariance matrices and are symmetric and striclty positive, properties which inherits.
The Kalman gain defined in equation (1) is often derived as the minimizer of the total posterior variance, i.e. the trace of .[1, 5, 3, 6, 4]. In other words
(3) |
It was shown in [2] that also minimizes the posterior generalized variance, defined as the determinant of , i.e.
(4) |
Let be the characteristic Polynomial of and be its roots, the Eigenvalues of . Let .
Theorem 1.
Let and . The Kalman gain is a critical point of and, if
(5) |
Proof.
Let and let . Using matrix calculus it follows that
(6) |
The equaton is solved if and only if is solved. The latter however is only true for , implying that is a critical point of and hence of .
For the term is positive definite, meaning the minimizing properties of are preserved.
∎
Corollary 1.1.
minimizes the smallest Eigenvalue
(7) |
Proof.
Let . Suppose . Then . By continuity of these characteristic Polynomials there is , i.e. such that . This contradicts Theorem 1. ∎
Corollary 1.2.
Let . Then is critical point of the map for all and minimizer for even :
(8) |
Proof.
For the claim holds by applying Theorem 1 at . For let . Pick such that . Then let . This function has the same critical points and extrema as in Theorem 1. Since the first part of the claim follows.
The second part follows if for even all . Hence let now be an even number. Since all Eigenvalues of are strictly positive and the results from expanding the product , he signs of the are alternating, meaning for all . Since we ultimately care about I will assume WLOG that for even . Hence and . Therefore has a root, which we chose to be , between and . ∎
Remark 1 (Elementary symmetric Polynomials).
The elementary symmetric Polynomials in variables are defined as
(9) |
Examples are and . They are invariant under permutation of their entries and appear naturally as the coefficients of the characteristic Polynomial:
(10) |
The previous theorem thus also applies to these elementary symmetrical polynomials evaluated at Eigenvalues of .
Corollary 1.3.
The Kalman gain is a critical point of the map , where is an arbirtray symmetric Polynomial, i.e. , where is a permutation.
Proof.
The non-leading coefficients of the characteristic polynomial are the elementary symmetric polynonials evaluated at the Eigenvalues of . From the previous corollary we showed that is a critical point for these. By the fundamental theorem of symmetric polynomials there is a polynomial such that
(11) |
The claim follows by chain rule. ∎
Remark 2 (Special cases).
For we can see that . The Kalman gain minimizes this function for all . For it is found that the sum of all , i.e. the sum of all elementary symmetric Polynomials in Eigenvalues of are minimized by . At this evaluates to reproducing the result from [2] . It is easy to see that as . Since minimizes this along the limiting process the minimizing of is rediscovered.
References
- [1] Mark Asch, Marc Bocquet, and Maëlle Nodet. Data Assimilation: Methods, Algorithms, and Applications. Fundamentals of Algorithms. Society for Industrial and Applied Mathematics.
- [2] Eviatar Bach. Proof that the kalman gain minimizes the generalized variance, 2021.
- [3] Andrew H. Jazwinski. Stochastic Processes and Filtering Theory. Academic Press, Inc.
- [4] Sören Laue, Matthias Mitterreiter, and Joachim Giesen. MatrixCalculus.org – Computing Derivatives of Matrix and Tensor Expressions. In Ulf Brefeld, Elisa Fromont, Andreas Hotho, Arno Knobbe, Marloes Maathuis, and Céline Robardet, editors, Machine Learning and Knowledge Discovery in Databases, Lecture Notes in Computer Science, pages 769–772. Springer International Publishing.
- [5] Simo Sarkka. Bayesian Filtering and Smoothing. Cambridge University Press, Cambridge, 2013.
- [6] Ricardo Todling. Estimation Theory and Foundations of Atmospheric Data Assimilation.