[1,3]\fnmStephen S.-T. \surYau
1]\orgdivDepartment of Mathematical Sciences, \orgnameTsinghua University, \orgaddress\cityBeijing, \postcode100084, \countryChina
2]\orgdivSchool of Mathematics, \orgnameRenmin University of China, \orgaddress\cityBeijing, \postcode100872, \countryChina
3]\orgnameBeijing Institute of Mathematical Sciences and Applications, \orgaddress\cityBeijing, \postcode101408, \countryChina
On the Convergence Analysis of Yau-Yau Nonlinear Filtering Algorithm: from a Probabilistic Perspective
Abstract
At the beginning of this century, a real time solution of the nonlinear filtering problem without memory was proposed in [1, 2] by the third author and his collaborator, and it is later on referred to as Yau-Yau algorithm. During the last two decades, a great many nonlinear filtering algorithms have been put forward and studied based on this framework. In this paper, we will generalize the results in the original works and conduct a novel convergence analysis of Yau-Yau algorithm from a probabilistic perspective. Instead of considering a particular trajectory, we estimate the expectation of the approximation error, and show that commonly-used statistics of the conditional distribution (such as conditional mean and covariance matrix) can be accurately approximated with arbitrary precision by Yau-Yau algorithm, for general nonlinear filtering systems with very liberal assumptions. This novel probabilistic version of convergence analysis is more compatible with the development of modern stochastic control theory, and will provide a more valuable theoretical guidance for practical implementations of Yau-Yau algorithm.
keywords:
nonlinear filtering, DMZ equation, Yau-Yau algorithm, convergence analysis, stochastic partial differential equationpacs:
[MSC Classification]60G35, 93E11, 60H15, 65M12
1 Introduction
Filtering is an important subject in the field of modern control theory, and has wide applications in various scenarios such as signal processing [3][4], weather forecast [5][6], aerospace industrial [7][8] and so on. The core objective of filtering problem is pursuing accurate estimation or prediction to the state of a given stochastic dynamical system based on a series of noisy observations [9][10]. For practical implementations, it is also necessary that the estimation or prediction to the state can be computed in a recursive and real-time manner.
In the filtering problems we consider in this paper, the evolution of state processes, as well as the noisy observations, is governed by the following system of stochastic differential equations,
(1) |
in the filtered probability space , where is a fixed termination; is the state process we would like to track; is the noisy observation to the state process ; and are mutually independent, -adapted, -dimensional standard Brownian motions; is a -valued random variable with probability density function , which is independent of and ; , and are sufficiently smooth vector- or matrix- valued functions.
Mathematically, for a given test function , the optimal estimation of based on the historical observations up to time is the conditional expectation , where the -algebra generated by historical observations. Such estimation is ‘optimal’ in the sense that,
(2) |
Therefore, the main task of filtering problem can be specified into finding efficient algorithms to numerically compute the conditional expectation , or equivalently, the conditional probability distribution or the conditional probability density function (if exists).
With some regularity assumptions on the coefficients in the system (1), the conditional probability measure is absolutely continuous with respect to the Lebesgue measure in , and the conditional probability density function , (without considering a normalization constant), can be described by the well-known DMZ equation, which is named after three researchers: T. E. Duncan [11], R. E. Mortensen [12] and M. Zakai [13], who derived the equation independently in the late 1960s.
The DMZ equation satisfied by the unnormalized conditional probability density function is a second-order stochastic partial differential equation, and does not possess an explicit-form solution in general. In their works [1] and [2], the third author and his collaborator propose a two-stage algorithm framework to compute the solution of the DMZ equation numerically in a memoryless and real-time manner. Later on, this algorithm is referred to as Yau-Yau algorithm.
The basic idea of Yau-Yau algorithm is that the heavy computation burden of numerically solving a Kolmogorov-type partial differential equation (PDE) can be done off-line, and in the meanwhile, the on-line procedure only consists of the basic computations such as multiplication by the exponential function of the observations. In the framework of Yau-Yau algorithm, various kinds of methods to solving the Kolmogorov-type PDEs, such as spectral methods [14][15], proper orthogonal decomposition [16], tensor training [17], etc, are proposed and applied in specific examples of nonlinear filtering problems. Numerical results in the previous works mentioned above show that Yau-Yau algorithm can provide accurate and real-time estimations to the state process of very general nonlinear filtering problems in low and medium-high dimensional space.
In the original works [1] and [2], the convergence analysis of Yau-Yau algorithm is conducted pathwisely. For regular paths of observations with some boundedness conditions, it is proved that the numerical solution provided by Yau-Yau algorithm will converge to the exact solution of DMZ equation both pointwisely and in -sense, as the size of time-discretization step tends to zero, while in the works on the practical implementations of Yau-Yau algorithm, such as [14] and [15], the convergence analysis mainly focuses on the capability of numerical methods to approximate the solution of Kolmogorov-type PDEs arising in Yau-Yau algorithms.
In this paper, we will revisit the convergence analysis of Yau-Yau algorithm from a probabilistic perspective. Instead of considering the convergence results pathwisely, we prove that the solution of Yau-Yau algorithm will converge to the exact solution of DMZ equation in expectation, and also, after the normalization procedure, the approximated solution to the filtering problem provided by Yau-Yau algorithm will converge to the conditional expectation .
The advantage of this probabilistic perspective is that for a theoretically rigorous convergence analysis, instead of regularity assumptions on observation paths, we only need to make assumptions on the coefficients of the filtering system (1) and the test function , which are verifiable off-line in advance for practitioners. In the meanwhile, as shown in the main results of this paper in Section 3, those assumptions we need are in fact quite general, and it is straightforward to check that the most commonly used test functions and , , corresponding to the conditional mean and conditional covariance matrix, as well as the linear Gaussian systems (with , , and , ), satisfy all the assumptions.
Moreover, to the best of the authors’ knowledge, most of the theoretical analysis of PDE-based filtering algorithm mainly deals with convergence results with respect to the DMZ equation. In such probabilistic perspective we consider here in this paper, however, it is natural and convenient to make a step forward and discuss the approximation capability of Yau-Yau algorithm to the normalized conditional expectation and conditional probability distribution. In this way, we will provide a thorough convergence analysis of the Yau-Yau algorithm for filtering problems.
The organization of this paper is as follows. Section 2 serves as preliminaries, in which we will summarize some basic concepts of filtering problems and the main procedure of Yau-Yau algorithm. The main theorems in this paper will be stated in Section 3, together with a sketch of the proofs. In the next four sections, we will provide the detailed proofs of the lemmas and theorems. We first focus on the properties of the exact solution of DMZ equation in Section 4 and Section 5, and then deal with the approximated solutions given by Yau-Yau algorithm in Section 6 and Section 7. Finally, Section 8 is a conclusion.
2 Preliminaries
In this section, we would like to briefly summarize the theory of nonlinear filtering, including the change-of-measure approach to deriving the DMZ equation, as well as the main idea and procedures of Yau-Yau algorithm.
In the change-of-measure approach to deriving the DMZ equation corresponding to the filtering system (1), we first introduce a series of reference probability measures , absolutely continuous to the original probability measure with Radon derivatives given by
(3) |
According to Girsanov’s theorem, as long as the process defined in (3) is a martingale, then under the reference probability measure , the observation process is a standard Brownian motion which is independent of the state process .
We also introduce the process , , to be the inverse of , which is also a Radon derivative and can be expressed by the stochastic integral with respect to as follows:
(4) |
Therefore, for any -measurable, integrable random variable , its expectation with respect to measure can be computed by
(5) |
where means the expectation is taken under the probability measure .
As an extension of Bayesian formula in the context of continuous-time stochastic processes, the following Kallianpur-Striebel formula allows us to express and calculate the solution of filtering problem, , by a ratio of conditional expectations under :
(6) |
Since the denominator in (6) is independent of the test function , people often refer to the nominator, , as the unnormalized conditional expectation of . The corresponding measure-valued stochastic process defined by
(7) |
is also referred to as unnormalized conditional probability measure, and we also denote the unnormalized conditional expectation by
(8) |
With sufficient regularity assumptions on the coefficients and test function , the evolution of is governed by the following well-known DMZ equation:
(9) |
where
(10) |
is a second-order elliptic operator with .
If the stochastic measures , , are almost surely absolutely continuous to the Lebesgue measure in , and the density functions (or the Radon derivatives) , as well as the derivatives of , is square-integrable, then is the solution to the following equation, (which is also referred to as the DMZ equation):
(11) |
which is a second-order stochastic partial differential equation with
(12) |
the adjoint operator of .
In this case, the unnormalized conditional expectation can be expressed as
(13) |
and the (normalized) conditional expectation can be calculated by
(14) |
Because the solution of (11) does not have a closed form for general nonlinear filtering systems, efficient numerical methods must be proposed, so that we can get a good approximation to the conditional expectation through the equation (14).
At the beginning of this century, the third author and his collaborator proposed a two-stage algorithm to numerically solve the DMZ equation (11) in a memoryless and real-time manner, which is often referred to as Yau-Yau algorithm. Here, we would like to briefly introduce the basic idea and main procedure of this algorithm.
Firstly, if we consider the exponential transformation
(15) |
then the function satisfies the robust DMZ equation
(16) |
where the stochastic differential terms in the original DMZ equation (11) are eliminated and
(17) | ||||
are stochastic functions that depend on the specific value of observation at time .
Instead of solving equation (11) directly, we will mainly focus on the robust DMZ equation (16), especially the corresponding initial-boundary value (IBV) problems in a closed ball , with a given radius .
(18) |
where is a function supported in and satisfies
(19) |
and for , so that the initial value is compatible with the boundary condition in (18). And from now on, we would like to drop the subscript, , in the notation for the simplicity of notations, and use to denote the solution to the IBV problem (19).
Let be a uniform partition of the time interval , with , . On each time interval , consider the IBV problem of the following parabolic equation
(20) |
with the value of coefficients and frozen at the left point and initial value .
With another exponential transformation given by
(21) |
the newly-constructed function satisfies
(22) |
After the two exponential transformations (15) and (21), the function we would like to use to approximate the unnormalized conditional probability density function at time is given by
(23) |
and the value for approximating the conditional expectation is given by
(24) |
The main idea of the Yau-Yau algorithm is that the problem of solving the DMZ equation satisfied by the unnormalized probability density function can be separated into two parts. The computationally expensive part of solving the (IBV) problems of parabolic equation (22) can be done off-line, because it is a deterministic Kolmogorov-type PDE which is independent of observations, and at least, the corresponding semi-group, , can be analyzed and approximated off-line. When new observation comes, the remaining task is only about calculating exponential transformations and numerical integrals. The framework of Yau-Yau algorithm is shown in Algorithm 1.
In the next section, we will give a mathematically rigorous interpretation of the approximation results (23) and (24) from a probabilistic perspective. In particular, we only need assumptions on the test function and the coefficients of the filtering systems to derive the convergence result. These assumptions are also easy to verify off-line before the observations come, and therefore, this convergence analysis will provide a guidance for practitioners to determine the parameters in the implementations of Yau-Yau algorithm for practical use.
3 Main Results
In this section, we would like to state the main result in this paper and also provide a sketch of the proof.
Firstly, besides the smoothness and regularity requirements which guarantee the existence of conditional expectation and the existence of the solution to the DMZ equation, let us further introduce four particular assumptions on the coefficients of the system, the initial distribution and the test function.
For the state equation in the filtering system (1), the drift term is assumed to be Lipschitz, and the the diffusion term , together with , is assumed to have bounded partial derivatives up to second order, i.e.,
Assumption (A1) guarantees that the state equation of (1) has a strong solution in , and especially, the state equation for linear filter satisfies this assumption.
Also, in order to conduct energy estimations for the (stochastic) partial differential equations, we would like to assume that the diffusion term in the state equation is nondegenerate, in the sense that
For the initial distribution , we would like to assume that it is smooth enough and possesses finite high-order moments:
Assumption (A3) is satisfied by commonly-used light-tailed distributions such as Gaussian distributions. In fact, in the following convergence analysis, we only require Assumption (A3) to hold for sufficiently large , rather than for all .
Finally, it is assumed that the test function is at most polynomial growth:
which is satisfied by most of the commonly-used test functions, such as those correspond to the conditional mean and covariance matrix.
Based on the above assumptions (A1) to (A4), the main result in this paper is stated as follows:
Theorem 1.
Fix a terminal time and the filtering system (1) with smooth coefficients. If Assumptions (A1) to (A4) hold, then for every , there exists , and , such that we can conduct the Yau-Yau algorithm in the closed ball and the uniform partition of : with for , and the numerical solution approximates the exact solution of the filtering problems at each time step well in the sense of mathematical expectation, i.e.,
(25) |
Here in this section, we provide a sketch of the proof of Theorem 25, in which the main idea of the proof is illustrated. The detailed proofs of those key estimations here will be given in order in the next four sections.
A Sketch of the Proof of Theorem 25.
Let be the Radon derivative defined in (4). And therefore, for every integrable, -measurable random variable , we have
(26) |
According to the properties of conditional expectations, the expectation of the approximation error of Yau-Yau algorithm can be estimated as follows:
where we use the fact that is -measurable and for integrable, -measurable random variable ,
(27) |
Therefore, the remaining task for us is to estimate the two error terms and , and to show that and can be arbitrarily small with sufficiently large and sufficiently small .
Firstly, for the estimation of
(28) |
we would like to utilize an intermediate function , , which is the solution of IBV problem of the DMZ equation (11) and will be introduced in (43) in Section 4. And we have
(29) |
For the estimation of
(30) |
since in the closed ball , ,
(31) |
and thus,
(32) | ||||
Combining (29) and (32), we have
(33) | ||||
According to Theorem 2 in Section 4, for every , there exists , which depends on , , , , , such that
(34) | ||||
Therefore, for every , with Assumption (A3) for the initial distribution , as long as we take , there exists , such that
(35) | ||||
Therefore, as long as , there exists , such that
(37) |
Let us choose , and for this particular , according to Theorem 5 in Section 7, there exists a time step , such that
(38) |
and thus,
(39) |
Take , and back to (33), and we obtain the desired result, that is, we have found and , such that
(40) |
∎
4 Estimation of the density outside the ball
In this section, we will provide an estimation of the value of the unnormalized conditional probability density outside a ball , with large enough.
Especially, we will show that almost all the density of is contained in the closed ball , and the estimations (34) in the proof of Theorem 25 in Section 3 holds with Assumptions (A1) to (A4).
Theorem 2.
With Assumptions (A1) to (A4), there exists a constant which only depends on , , , and , such that
(41) |
and
(42) |
holds for all .
Proof of Theorem 2.
We first consider the following IBV problem on the ball :
(43) |
where is the function defined in (19), such that the initial value is compatible with the boundary conditions.
Let and define
(44) |
Then, according to the IBV problem (43) satisfied by the function , we have
(45) | ||||
By the Gauss-Green formula, we have
(46) | ||||
where is the unit outward normal vector of , denotes the measure on ,
(47) |
and
(48) | ||||
Since , , on and
(49) |
Moreover, we have
(50) |
where is a continuous function on , because and .
Therefore,
(51) | ||||
where the last inequality holds because is positive semi-definite. Thus,
(52) |
Similarly,
(53) |
Therefore,
(54) |
where
(55) |
Since , then
(56) |
where is the Kronecker’s symbol, with , if , and otherwise.
Notice that
(57) |
With the assumption that , we have
(58) |
Because is Lipschitz continuous according to (A1).
(59) |
Therefore,
(60) | ||||
Let us denote by the above upper bound of :
(61) |
which is a constant that depends on , and , but does not depend on .
Take expectation with respect to the reference probability measure , we obtain
(62) |
Here we use the fact that is a Brownian motion with respect to .
According to the Gronwall’s inequality, we have
(63) | ||||
Let tends to infinity, and we have
(64) |
Therefore,
(65) | ||||
Moreover, with condition (A4),
(66) | ||||
∎
5 Approximation of by the IBV problem in
With the estimation in Theorem 2, because almost all the density of is contained in the closed ball for large enough, it is natural to think about approximating by the solution, , to the corresponding initial-boundary value (IBV) problem (43) of DMZ equation in the ball .
It will be rigorously proved in this section that, for large enough, can be approximated well by defined in (43), and in particular, the estimation (36) holds in the proof of Theorem 25 in Section 3.
The main result in this section is stated as follows:
Theorem 3.
With Assumptions (A1) to (A4), there exists a constant which only depends on , , and , such that
(67) |
holds for large enough (for example, ), where is the solution of the IBV problem (43).
Proof of Theorem 3.
For each , consider the auxiliary function
(68) |
and
(69) |
Define , . Then, according to the maximum principle for SPDEs (cf. [18], for example), we have , for all and a.s. . Let be the stochastic process defined by
(70) |
Since is the solution to the SPDE
(71) | ||||
the -valued stochastic process satisfies
(72) | ||||
According to the Gauss-Green formula, we have
where, as in the proof of Theorem 2,
(73) |
(74) | ||||
denotes the outward normal vector of the boundary and denotes the measure on .
Notice that and
(75) |
Moreover,
(76) |
and therefore,
(77) |
Hence,
(78) | ||||
Take expectation with respect to the probability measure , and we have
(79) | ||||
For , , , and together with the Lipschitz conditions for ,
(80) |
(81) | ||||
Also, according to direct computations,
(82) | ||||
where is the Kronecker’s symbol. Thus,
(83) |
We would like to remark that the estimation in (83) is quite rough. Each term on the right-hand side of (83) corresponds to one term on the right-hand side of (82), and the purpose is just to show the second-order derivatives are also bounded by a constant independent of .
Notice that . Together with the bounded condition for , we have
(84) |
where is a constant which depends on , but does not depend on .
According to Theorem 2, the integral
(85) |
which is also bounded by a constant independent of , thus,
(86) |
where is a constant which depends on .
By Gronwall’s inequality,
(87) |
where is a constant which depends on , , and .
6 Regularity of the Approximated Function
In this section, we will discuss the regularity of , , which is the solution of a series of coefficient-frozen equations (20).
The main purpose of this section is to show that under mild conditions, the recursively defined functions will not explode in the finite time interval , even if the time-discretization step , in the sense that the -norm of () is square integrable with respect to the probability measure , and the expectations, , are uniformly bounded for .
As shown in the next section, this following theorem is an essential intermediate result for the convergence analysis of this time-discretization scheme.
Theorem 4.
Let be the solution to the IBV problem of the coefficients-frozen equation (20). Then, with Assumptions (A1) to (A4), the -norm of is square-integrable with respect to the probability measure , and we have
(91) |
where is a constant that depends on , , , , but is uniform in .
In the proof of Theorem 4, we will consider another exponential transformation given by
(92) |
Direct computation implies that is the solution of
(93) |
and recursively, we can rewrite the initial value in (93) by
(94) |
Under the reference probability measure , is a Brownian motion and
(95) |
with the -dimensional identity matrix. We would like to study the regularity of first, utilizing the Markov property of , and then derive the regularity results for .
For the sake of discussing the regularity of in a recursive manner, we need the following lemma which describes the relationship between and from (94).
Lemma 1.
For , let , be the solution of (93). The end-point values and satisfy (94). Let us denote by the space of quartic-integrable functions in . Assume that , and the -norm, , is quartic integrable with respect to , i.e.,
(96) |
then , its -norm, is quartic integrable with respect to , and for sufficiently small time-discretization step size , we have
(97) |
where is a constant that depends on and .
Proof of Lemma 1.
According to the expression (94) and the definition of on , because of the Markov property of , is independent of .
Because the observation function is assumed to be smooth enough, and is a bounded domain in , there exists a constant , which may depend on , such that the maximum of the absolute value of , together with its partial derivatives up to order , is bounded above by .
Therefore, by Fubini’s theorem,
(98) | ||||
Next, let us estimate the expectations of functions of normal random variable arising in the above expressions, for small time-discretization step .
In fact, because , we have
(99) |
In the bounded domain ,
(100) | ||||
Therefore, for (for example ),
(101) |
Thus,
(102) |
∎
Now, we are ready to give the proof of Theorem 4.
Proof of Theorem 4.
The idea of this proof is to study the regularity of , recursively, and then obtain the regularity of based on the relationship (94).
In fact, according to the Cauchy-Schwartz inequality,
with , a constant depending only on .
Under the reference probability measure , is a standard -dimensional Brownian motion, and therefore, the expectation
is bounded.
Hence, it remains to show that there exists a constant , such that,
(103) |
holds uniformly for .
In the time interval , is the solution to (93). According to the regularity results of parabolic partial differential equations, we have
(104) |
where is a constant which depends on the coefficients of the filtering system. The techniques in the proof of (104) is standard, and the proof of a counterpart, in which -norm (instead of -norm) is considered, can be found in the textbook [19]. We also provide a detailed proof in the Appendix, for the readers’ convenience and in order to keep this paper self-contained.
7 Convergence Analysis of the Time Discretization Scheme
This section serves to show that the solution of the coefficient-frozen equations (20) can approximate the solution of the original robust DMZ equation (18) well, if the time-discretization step size is small enough.
Also, we will show in this section that, after the exponential transformation , the -norm of the difference between the unnormalized densities (defined by (43)) and (defined by (22)) still converges to zero, as . In particular, the estimation (38) holds in the proof of Theorem 25 in Section 3.
Theorem 5.
Proof of Theorem 5.
Since is globally Lipschitz, , and is a bounded domain, there exists a constant , such that the absolute value of each component in and , as well as there first and second order derivatives, are dominated by in the ball , i.e.,
(108) | ||||
Then, according to equations (18) and (20) satisfied by and in ,
(110) | ||||
Because on the boundary , and , we have and with the outward normal vector of and on . Thus, the first three terms on the right-hand side of (110) can be estimated by
(111) | ||||
where we use the fact that is positive semi-definite and the definition of and in (17); and are constants which depend only on ; and denotes the measure on .
Also, by the definition of and in (17), we have the following estimation of the differences
(112) | ||||
where and are constants which only depends on .
Hence,
(113) | ||||
holds for almost all and almost surely, where , and are constants which depend on the coefficients of the system.
Under the reference probability distribution , the observation process is a standard -dimensional Brownian motion, and therefore,
(114) |
Let be the event which represents the observation process is not severely abnormal, and is the indicator function of the set .
For a fixed , let us first take the expectation with respect to on the event for both sides of (113), and we have
Here, the second inequality holds because of the property of the event , the third inequality holds according to the Cauchy-Schwartz inequality and the last equality holds because is a normal distributed random vector.
On the event , the observation process is bounded. Therefore, according to the regularity results of parabolic partical differential equations (cf. [19], Section 7.1, Theorem 6), the integrals and are also bounded for almost every , as long as and . Thus,
(115) | ||||
where are constants which depend on .
Similarly, we also have the estimation for the integral on the set :
(116) | ||||
and thus
(117) | ||||
Therefore,
(118) | ||||
and
(119) | ||||
Notice that by definition. Inductively, we have
(120) | ||||
where is a constant which depends on .
Also, for the value we are concerned with in (107),
(121) | ||||
On the event , let
then,
(122) | ||||
where is a constant which is related to the volume of the -dimensional ball , and is the random variable given by
(123) |
and
(124) |
According to the Burkholder-Davis-Gundy inequality (cf. [20], Chapter 3, Theorem 3.28, for example), there exists , such that
(125) |
and also, because are normal random variables, the expectation of
is bounded.
For the value , because
(126) |
then
(127) | ||||
Notice that is the solution to the stochastic partial differential equation
(128) |
and the boundedness of
(129) |
follows from the regularity theory of stochastic partial differential equation.
In the monograph [21], the authors provided a similar regularity result, and proved that is bounded by the initial values. Here in our case, we will prove that there exists , such that
(130) |
The detailed proof of (130) can be found in the Appendix.
Therefore, we have
(131) |
where is a constant that does not depend on or .
Furthermore, as we have discussed in the previous section, is also bounded above, and thus, we have
(132) |
where is a constant which does not depend on or .
In summary, for each , there exists , such that
(133) |
and for this particular , there exists , such that
(134) |
Therefore, for every ,
(135) | ||||
∎
8 Conclusion
In this paper, we provide a novel convergence analysis of Yau-Yau algorithm from a probabilistic perspective. With very liberal assumptions only on the coefficients of the filtering systems and the initial distributions (without assumptions on particular paths of observations), we can prove that Yau-Yau algorithm can provide accurate approximations with arbitrary precision to a quite broad class of statistics for the conditional distribution of state process given the observations, which includes the most commonly used conditional mean and covariance matrix. Therefore, the capability of Yau-Yau algorithm to solve very general nonlinear filtering problems is theoretically verified in this paper.
In the process of deriving this probabilistic version of the convergence results, we study the properties of the exact solution, , to the DMZ equation and the approximated solution , given by Yau-Yau algorithm, respectively.
For the exact solution of the DMZ equation, we have shown in Section 4 and Section 5 that most of the density of will remain in the closed ball , and can be approximated well by the corresponding initial-boundary value problem of DMZ equation in . This result also implies that it is very unlikely for the state process to reach infinity within finite terminal time.
For the approximated solution given by Yau-Yau algorithm, we have first proved in Section 6 that , which evolves in a recursive manner, will not explode in finite time interval, even if the time-discretization step . And then, in Section 7, the convergence of is proved and the convergence rate is also estimated to be .
It is clear that the properties of exact solutions and approximated solutions, which we have proved in this paper, highly rely on the nice properties of Brownian motion and Gaussian distributions, especially the Markov and light-tail properties. On the one hand, Brownian motion and Gaussian distribution are up to now, among the most commonly used objects in the mathematical modeling of many areas of applications, and can describe most scenarios in practice. On the other hand, for those systems driven by non-Markov or heavy-tailed processes, minimum mean square criteria, together with the conditional expectations (if exist), may not result in a satisfactory estimation of the state process. In this case, the studies of estimations based on other criteria, such as maximum a posteriori (MAP) [22][23][24], will be a promising direction.
Finally, in this paper, we only consider filtering systems and conduct convergence analysis in time interval with a fixed finite terminal time . It is also interesting to study the behavior of the DMZ equation and the approximation capability of Yau-Yau algorithm in the case where the terminal time , especially for filtering systems with further stable assumptions. We will continue working on how to combine the existing studies on filter stability, such as [25][26], with our techniques developed in this paper, and hopefully, obtain some convergence results of Yau-Yau algorithm for the whole time line .
Declarations
Funding
This work is supported by National Natural Science Foundation of China (NSFC) grant (12201631) and Tsinghua University Education Foundation fund (042202008).
Conflict of interest/Competing interests
The authors have no competing interests to declare that are relevant to the content of this article.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Data, Materials and/or Code availability
Not applicable.
Author contribution
All authors contributed to the study conception and design. The first draft of the manuscript was written by Zeju Sun and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Appendix A Regularity Results of Parabolic Partial Differential Equation and Stochastic Evolution Equation
In this appendix, we will provide a detailed proof of the regularity results of the parabolic partial differential equation and the stochastic evolution equation.
For the purpose of deriving (104) and (130), the regularity results is slightly different from standard ones considered in square-integrable functional spaces.
Theorem 6.
Let be the solution of the following IBV problem:
(136) |
where is the ball in with radius ; , , are smooth enough functions. Assume that the matrix-valued function is uniformly positive definite, i.e., there exists , such that
(137) |
If the initial value is quartic-integrable in , then there exists a constant , which depends on the coefficients of the system, such that
(138) |
Remark 1.
In fact, Assumption (A2) in the main text will imply the coercivity condition (137). This is because the closed ball is a compact set of , and the continuous function in Assumption (A2) will map to a compact set. Therefore, there exists , such that , for all .
Proof.
Let us define
(139) |
Then the parabolic equation (136) can be written in a divergence form
(140) |
Hence,
(141) | ||||
In the bounded domain , there exists a constant , such that
(142) |
Thus,
(143) |
and by Gronwall’s inequality, we have
(144) |
∎
Theorem 7.
Consider the IBV problem of stochastic partial differential equation given by
(145) |
where is a standard -dimensional Brownian motion in the filtered probability space ; is the ball in with radius , and
(146) |
Assume that the coefficients , , are smooth enough and the Assumption (A2) holds for the matrix-valued function , which implies that is uniformly positive definite in , i.e., there exists , such that
(147) |
If the initial value is square-integrable in , then there exists a constant , which depends on , and the coefficients of the system, such that
(148) |
Proof.
Let us define
(149) |
Then the stochastic partial differential equation in (145) can be rewritten in divergence form:
(150) |
Let
(151) |
then according to Itô’s formula,
(152) |
and
(153) | ||||
After taking expectations, we have
(154) | ||||
Notice that
(155) | ||||
Hence,
(156) | ||||
In the bounded domain , there exists , such that
(157) |
Thus,
(158) |
According to Gronwall’s inequality,
(159) |
which is the desired result. ∎
References
- \bibcommenthead
- Yau and Yau [2000] Yau, S.-T., Yau, S.S.-T.: Real time solution of nonlinear filtering problem without memory i. Mathematical research letters 7(6), 671–693 (2000)
- Yau and Yau [2008] Yau, S.-T., Yau, S.S.-T.: Real time solution of the nonlinear filtering problem without memory ii. SIAM Journal on Control and Optimization 47(1), 163–195 (2008)
- Candy [2016] Candy, J.V.: Bayesian Signal Processing: Classical, Modern, and Particle Filtering Methods, (2016)
- Roth et al. [2017] Roth, M., Hendeby, G., Fritsche, C., Gustafsson, F.: The ensemble kalman filter: a signal processing perspective. EURASIP Journal on Advances in Signal Processing 2017, 1–16 (2017)
- Galanis et al. [2006] Galanis, G., Louka, P., Katsafados, P., Pytharoulis, I., Kallos, G.: Applications of kalman filters based on non-linear functions to numerical weather predictions. Ann. Geophys 24, 2451–2460 (2006)
- Chen and Yu [2014] Chen, K., Yu, J.: Short-term wind speed prediction using an unscented kalman filter based state-space support vector regression approach. Applied energy 113, 690–705 (2014)
- Ichard [2015] Ichard, C.: Random media and processes estimation using non-linear filtering techniques: application to ensemble weather forecast and aircraft trajectories. PhD thesis, Université de Toulouse, Université Toulouse III-Paul Sabatier (2015)
- Sun et al. [2019] Sun, J., Blom, H.A., Ellerbroek, J., Hoekstra, J.M.: Particle filter for aircraft mass estimation and uncertainty modeling. Transportation Research Part C: Emerging Technologies 105, 145–162 (2019)
- Jazwinski [2007] Jazwinski, A.H.: Stochastic Processes and Filtering Theory. Courier Corporation, (2007)
- Bain [2009] Bain, A.: Fundamentals of Stochastic Filtering, 1st ed. 2009. edn. Springer, New York, NY (2009)
- Duncan [1967] Duncan, T.E.: Probability densities for diffusion processes with applications to nonlinear filtering theory and detection theory. PhD thesis, Stanford University, Stanford, California (May 1967)
- Mortensen [1966] Mortensen, R.E.: Optimal control of continuous-time stochastic systems. Technical report, California Univ. Berkeley Electronics Research Lab (1966)
- Zakai [1969] Zakai, M.: On the optimal filtering of diffusion processes. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete 11(3), 230–243 (1969)
- Luo and Yau [2013] Luo, X., Yau, S.S.-T.: Hermite spectral method to 1-d forward kolmogorov equation and its application to nonlinear filtering problems. IEEE Transactions on Automatic Control 58(10), 2495–2507 (2013)
- Dong et al. [2020] Dong, W., Luo, X., Yau, S.S.-T.: Solving nonlinear filtering problems in real time by legendre galerkin spectral method. IEEE Transactions on Automatic Control 66(4), 1559–1572 (2020)
- Wang et al. [2019] Wang, Z., Luo, X., Yau, S.S.-T., Zhang, Z.: Proper orthogonal decomposition method to nonlinear filtering problems in medium-high dimension. IEEE Transactions on Automatic Control 65(4), 1613–1624 (2019)
- Li et al. [2022] Li, S., Wang, Z., Yau, S.S.-T., Zhang, Z.: Solving nonlinear filtering problems using a tensor train decomposition method. IEEE Transactions on Automatic Control (2022)
- Chekroun et al. [2016] Chekroun, M.D., Park, E., Temam, R.: The stampacchia maximum principle for stochastic partial differential equations and applications. Journal of Differential Equations 260(3), 2926–2972 (2016)
- Evans [2010] Evans, L.C.: Partial Differential Equations, 2nd ed. edn. American Mathematical Society, (2010)
- Karatzas [1998] Karatzas, I.: Brownian Motion and Stochastic Calculus, 2nd edn. New York: Springer, (1998)
- Rozovsky [2018] Rozovsky, B.L.: Stochastic Evolution Systems Linear Theory and Applications to Non-Linear Filtering, 2nd edn. Cham : Springer, (2018)
- Godsill et al. [2001] Godsill, S., Doucet, A., West, M.: Maximum a posteriori sequence estimation using monte carlo particle filters. Annals of the Institute of Statistical Mathematics 53, 82–96 (2001)
- Saha et al. [2012] Saha, S., Mandal, P.K., Bagchi, A., Boers, Y., Driessen, J.N.: Particle based smoothed marginal map estimation for general state space models. IEEE transactions on signal processing 61(2), 264–273 (2012)
- Kang et al. [2023] Kang, J., Salmon, A., Yau, S.S.-T.: Log-concave posterior densities arising in continuous filtering and a maximum a posteriori algorithm. SIAM Journal on Control and Optimization 61(4), 2407–2424 (2023)
- Atar [1998] Atar, R.: Exponential stability for nonlinear filtering of diffusion processes in a noncompact domain. Annals of Probability 26(4), 1552–1574 (1998)
- Ocone and Pardoux [1996] Ocone, D., Pardoux, E.: Asymptotic stability of the optimal filter with respect to its initial condition. SIAM Journal on Control and Optimization 34(1), 226–243 (1996)