This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bayesian Filtering for Nonlinear Stochastic Systems Using
Holonomic Gradient Method with Integral Transform

Tomoyuki Iori and Toshiyuki Ohtsuka This work was partly supported by JSPS KAKENHI Grant Numbers JP18J22093, JP21K21285, and JP15H02257.T. Iori is with the Department of Information and Physical Sciences, Graduate School of Information Science and Technology, Osaka University, 1-5 Yamadaoka, Suita, Osaka 565–0871, Japan [email protected]T. Ohtsuka is with the Department of Systems Science, Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo-ku, Kyoto 606–8501, Japan [email protected]©2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
Abstract

This paper proposes a symbolic-numeric Bayesian filtering method for a class of discrete-time nonlinear stochastic systems to achieve high accuracy with a relatively small online computational cost. The proposed method is based on the holonomic gradient method (HGM), which is a symbolic-numeric method to evaluate integrals efficiently depending on several parameters. By approximating the posterior probability density function (PDF) of the state as a Gaussian PDF, the update process of its mean and variance can be formulated as evaluations of several integrals that exactly take into account the nonlinearity of the system dynamics. An integral transform is used to evaluate these integrals more efficiently using the HGM compared to our previous method. Further, a numerical example is provided to demonstrate the efficiency of the proposed method over other existing methods.

I Introduction

In most engineering applications, knowledge of the state of a dynamical system is required for monitoring or control via feedback. The optimal filtering theory was developed to estimate the current state of a system using the history of its outputs that are observable but contaminated by random noise. Following the success of the Kalman filter (KF) [1, 2] for linear systems with Gaussian noise, optimal filtering theory has been extended to applications involving nonlinear systems as well as non-Gaussian noise [3, 4, 5, 6, 7].

Bayesian filtering is a general framework of optimal filtering. Although numerous problems can be formulated as the Bayesian filtering problems, they are difficult to implement as algorithms on computers, particularly for the real-time applications in systems and control. The Monte-Carlo scheme, which was first introduced for the particle filter (PF) [8, 9] and subsequently used in other filtering algorithms, such as [7], is a powerful tool to compute integrations accompanied by statistical operations, such as marginalizations and expectations in nonlinear Bayesian filtering problems. Moreover, the nonlinearity of the system dynamics can be exactly accounted for by updating the particles according to the dynamics. To reap these benefits of the Monte-Carlo scheme, however, the number of particles must be sufficiently large, which requires heavy computational resources and may be unacceptable for applications with short sampling intervals. Another approach involves extending the KF to nonlinear cases such as the extended KF (EKF) [10], unscented KF (UKF) [11], cubature KF (CKF) [12], and Gauss-Hermite KF (GHKF) [13]. In all these methods, the posterior probability density function (PDF) is assumed to be or approximated as a Gaussian PDF. Under this Gaussian approximation, the marginalizations and expectations of the nonlinear functions, which are required to update the posterior mean and variance, can be efficiently computed using the Gauss quadrature rules. However, in exchange for efficiency, the Gauss quadrature rules cannot provide exact values of the integrals, which implies that the nonlinearity of the system dynamics can only be considered in the approximate sense. Thus, there is still room to extend the boundary of the trade-off between the exact consideration of nonlinearity and reduction of computational cost.

In nonlinear cases, the integrals in the posterior PDF are ones of general nonlinear functions that depend on parameters such as the observed output, which are both analytically and numerically intractable to evaluate. In our previous work [14], a symbolic-numeric method called the holonomic gradient method (HGM) [15] was used to evaluate the integrals efficiently. By utilizing the symbolic computation in terms of differential operators, we can perform integrations offline while considering the nonlinearity of the system exactly. The state estimation is then reduced to solving a set of initial-value problems (IVPs) of nonlinear ordinary differential equations (ODEs), which can be efficiently accomplished online with numerical integration methods such as the fourth order Runge-Kutta (RK4) or Adams-Bashforth-Moulton predictor-corrector (ABM4) methods. However, our previous method requires solving an IVP for each component of the posterior mean and variance in the one-step estimation. This incurs a high computational cost for the online computation and is unacceptable for short sampling intervals.

To overcome the computational limitations of our previous method, we use an integral transform. Integral transforms are useful tools to compute the moments of a probability distribution efficiently. For example, in [16], Idan and Speyer used an integral transform called the characteristic function of the multivariate Cauchy distribution. For linear systems under additive Cauchy noise, the characteristic function can be analytically propagated in time and can be used to derive the explicit forms of the posterior mean and variance. In this paper, we introduce an integral transform, which is different from, but similar to, the moment generating function. This similarity enables us to compute all the components of the posterior mean and variance efficiently from the integral transform and its derivatives. Moreover, the integral transform and its derivatives can be simultaneously evaluated by using the HGM only once. Hence, in contrast to our previous method, only one IVP must be solved, which leads to the reduction of the online computational cost.

Notations

For the field of real numbers 𝐑{\boldsymbol{\mathrm{R}}} and a vector of indeterminates X=[X1Xn]X=[X_{1}\ \cdots\ X_{n}]^{\top}, 𝐑(X){\boldsymbol{\mathrm{R}}}(X) denotes the field of rational functions in the components of XX over 𝐑{\boldsymbol{\mathrm{R}}}. X[X1Xn]\partial_{X}\coloneqq[\partial_{X_{1}}\ \cdots\ \partial_{X_{n}}]^{\top} denotes a vector of differential operators, where Xi=/Xi\partial_{X_{i}}={\partial{}/\partial{X_{i}}}. We abbreviate Xi\partial_{X_{i}} by i\partial_{i} if XX is clearly specified according to the context. For a multi-index vector d=[d1dn]𝐙0nd=[d_{1}\ \cdots\ d_{n}]^{\top}\in{\boldsymbol{\mathrm{Z}}}^{n}_{\geq 0}, XdX^{d} and d\partial^{d} denote X1d1XndnX_{1}^{d_{1}}\cdots X_{n}^{d_{n}} and 1d1ndn\partial_{1}^{d_{1}}\cdots\partial_{n}^{d_{n}}, respectively. The symbol n𝐑(X){\mathcal{R}}_{n}\coloneqq{\boldsymbol{\mathrm{R}}}(X)\langle\partial\rangle denotes the noncommutative ring of differential operators with coefficients in 𝐑(X){\boldsymbol{\mathrm{R}}}(X). The subscript nn of n{\mathcal{R}}_{n} is omitted if it is clear from the context. We denote the action of an element lnl\in{\mathcal{R}}_{n} on a sufficiently smooth function α(X)=α(X1,,Xn)\alpha(X)=\alpha(X_{1},\dots,X_{n}) as lα(X)l\bullet\alpha(X); for instance, iα(X)=α/Xi(X)\partial_{i}\bullet\alpha(X)={\partial{\alpha}/\partial{X_{i}}}(X). We say that a differential operator ll\in{\mathcal{R}} annihilates a function α\alpha if lα=0l\bullet\alpha=0. We denote the set of all positive definite n×nn\times n matrices by PD(n)\mathrm{PD}(n). The PDF of a stochastic variable XX is denoted by p(X;Y)p(X;Y) if it depends on a parameter YY.

II Problem Setting

In the Bayesian filtering approach for discrete-time nonlinear systems, the posterior PDF of the state xk𝐑nx_{k}\in{\boldsymbol{\mathrm{R}}}^{n} at time step kk conditional on the output history y[0:k]𝐑ry_{[0:k]}\subset{\boldsymbol{\mathrm{R}}}^{r} from time steps 0 to kk is recursively updated via the following equations [17].

p(xky[0:k])=p(xk,yky[0:k1])p(yky[0:k1]),\displaystyle p(x_{k}\mid y_{[0:k]})=\frac{p(x_{k},y_{k}\mid y_{[0:k-1]})}{p(y_{k}\mid y_{[0:k-1]})}, (1)
pxy(xk,yky[0:k1])=p(ykxk)\displaystyle p_{xy}(x_{k},y_{k}\mid y_{[0:k-1]})=p(y_{k}\mid x_{k})
×𝐑np(xkxk1)p(xk1y[0:k1])dxk1,\displaystyle\qquad\quad\times\int_{{\boldsymbol{\mathrm{R}}}^{n}}p(x_{k}\mid x_{k-1})p(x_{k-1}\mid y_{[0:k-1]})dx_{k-1},
p(yky[0:k1])=𝐑np(xk,yky[0:k1])𝑑xk.\displaystyle p(y_{k}\mid y_{[0:k-1]})=\int_{{\boldsymbol{\mathrm{R}}}^{n}}p(x_{k},y_{k}\mid y_{[0:k-1]})dx_{k}.

We assume that the system dynamics and observation process are defined by the following state and observation equations.

xk=f(xk1,uk)+wk,\displaystyle x_{k}=f(x_{k-1},u_{k})+w_{k}, (2)
yk=h(xk)+vk,\displaystyle y_{k}=h(x_{k})+v_{k}, (3)

where wk𝐑nw_{k}\in{\boldsymbol{\mathrm{R}}}^{n} and vk𝐑rv_{k}\in{\boldsymbol{\mathrm{R}}}^{r} denote the system and observation noises that are assumed to be independent and identically distributed with the PDFs pw(wk)p_{w}(w_{k}) and pv(vk)p_{v}(v_{k}), respectively. In addition, uk𝐑mu_{k}\in{\boldsymbol{\mathrm{R}}}^{m} denotes a given input of the system. Using the rule of transformation of PDFs [10], the conditional PDFs p(xkxk1)p(x_{k}\mid x_{k-1}) and p(ykxk)p(y_{k}\mid x_{k}) in (1) can be rewritten using functions ff, hh, pwp_{w}, and pvp_{v} as follows:

p(xkxk1;uk)\displaystyle p(x_{k}\mid x_{k-1};u_{k}) =pw(xkf(xk1,uk)),\displaystyle=p_{w}(x_{k}-f(x_{k-1},u_{k})), (4)
p(ykxk)\displaystyle p(y_{k}\mid x_{k}) =pv(ykh(xk)),\displaystyle=p_{v}(y_{k}-h(x_{k})), (5)

where the former PDF is conditional on xk1x_{k-1} and parametrized by uku_{k} owing to (2). By substituting (4) and (5) into (1), we can describe the update law of Bayesian filtering using ff, hh, pwp_{w}, and pvp_{v}.

Since Bayesian filtering is a recursive algorithm, we can focus on the one-step update of the posterior PDF and omit the subscript kk. Moreover, the past history of outputs y[0:k1]y_{[0:k-1]} has nothing to do with the update at time step kk; it only appears in the previous posterior PDF p(xk1y[0:k1])p(x_{k-1}\mid y_{[0:k-1]}), so we also omit the past history y[0:k1]y_{[0:k-1]}. Thus, (1) can be summarized as

p(xy;u)=pxy(x,y;u)𝐑npxy(x,y;u)𝑑x,p(x\mid y;u)=\frac{p_{xy}(x,y;u)}{\int_{{\boldsymbol{\mathrm{R}}}^{n}}p_{xy}(x,y;u)dx}, (6)

where xx, yy, and uu denote the current state, output, and input, respectively; xx^{-} denotes the previous state, and pxy(x,y;u)p_{xy}(x,y;u) denotes the joint PDF of xx and yy defined as the second line in (1).

The update law (6) can be viewed as a functional that maps the previous posterior PDF p(x)p(x^{-}) to the current one p(xy;u){p(x\mid y;u)}. In the linear case with Gaussian noise, this functional can be reduced to simple arithmetic operations of the mean and variance of the previous posterior PDF to yield the prediction and update steps of the Kalman filter. However, in the nonlinear cases, the Gaussian property of the posterior PDF can no longer be guaranteed, which makes it extremely difficult to compute the functional. Therefore, as the first step to overcome this difficulty, we approximate the posterior PDF using a Gaussian.

For the following discussion, p~\tilde{p} denotes an approximation of the PDF pp with the same stochastic variables. Suppose we have a Gaussian p~(x;μ,Σ)=𝒩(x;μ,Σ)\tilde{p}(x^{-};\mu^{-},\Sigma^{-})=\mathcal{N}(x^{-};\mu^{-},\Sigma^{-}) approximating p(x)p(x^{-}), where 𝒩\mathcal{N} is defined as

𝒩(ξ;m,V)1(2π)n|V|exp(12(ξm)V1(ξm)).\mathcal{N}(\xi;m,V)\coloneqq\frac{1}{\sqrt{(2\pi)^{n}|V|}}\exp\left(-\frac{1}{2}(\xi-m)^{\top}V^{-1}(\xi-m)\right).

By replacing p(x)p(x^{-}) with p~(x;μ,Σ)\tilde{p}(x^{-};\mu^{-},\Sigma^{-}) in the definition of pxy(x,y;u)p_{xy}(x,y;u), it can be approximated as

pxy(x,y;u)p~xy(x,y;u,μ,Σ)p(yx)𝐑np(xx;u)p~(x;μ,Σ)𝑑x,p_{xy}(x,y;u)\approx\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})\\ \coloneqq p(y\mid x)\int_{{\boldsymbol{\mathrm{R}}}^{n}}p(x\mid x^{-};u)\tilde{p}(x^{-};\mu^{-},\Sigma^{-})dx^{-}, (7)

and p(xy;u)p(x\mid y;u) can be approximated as

p~(xy;u,μ,Σ)=p~xy(x,y;u,μ,Σ)𝐑np~xy(x,y;u,μ,Σ)𝑑x.\tilde{p}(x\mid y;u,\mu^{-},\Sigma^{-})=\frac{\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})}{\int_{{\boldsymbol{\mathrm{R}}}^{n}}\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})dx}. (8)

Note that the approximating PDF on the left-hand side is not Gaussian since we consider the nonlinearities of functions ff and hh without any approximations. Hence, we further approximate p~(xy;u,μ,Σ)\tilde{p}(x\mid y;u,\mu^{-},\Sigma^{-}) with a Gaussian 𝒩(x;μ,Σ)\mathcal{N}(x;\mu,\Sigma) of the same mean μ\mu and variance Σ\Sigma as those of p~(xy;u,μ,Σ)\tilde{p}(x\mid y;u,\mu^{-},\Sigma^{-}), that is, p~(xy;u,μ,Σ)𝒩(x;μ,Σ)\tilde{p}(x\mid y;u,\mu^{-},\Sigma^{-})\approx\mathcal{N}(x;\mu,\Sigma) where

μ\displaystyle\mu\coloneqq Ep~(xy;u,μ,Σ)[x]\displaystyle E_{\tilde{p}(x\mid y;u,\mu^{-},\Sigma^{-})}[x]
=\displaystyle= 𝐑nxp~xy(x,y;u,μ,Σ)𝑑x𝐑np~xy(x,y;u,μ,Σ)𝑑x\displaystyle\frac{\int_{{\boldsymbol{\mathrm{R}}}^{n}}x\cdot\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})dx}{\int_{{\boldsymbol{\mathrm{R}}}^{n}}\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})dx} (9)

and

Σ\displaystyle\Sigma\coloneqq Ep~(xy;u,μ,Σ)[xx]μμ\displaystyle E_{\tilde{p}(x\mid y;u,\mu^{-},\Sigma^{-})}[xx^{\top}]-\mu\mu^{\top}
=\displaystyle= 𝐑nxxp~xy(x,y;u,μ,Σ)𝑑x𝐑np~xy(x,y;u,μ,Σ)𝑑xμμ.\displaystyle\frac{\int_{{\boldsymbol{\mathrm{R}}}^{n}}xx^{\top}\cdot\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})dx}{\int_{{\boldsymbol{\mathrm{R}}}^{n}}\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})dx}-\mu\mu^{\top}. (10)

Definitions (9) and (10) provide a finite number of functions that map the parameters yy, uu, μ\mu^{-}, and Σ\Sigma^{-} to the current estimates μ\mu and Σ\Sigma. Hence, by evaluating the right-hand sides of (9) and (10) recursively, we can perform Bayesian filtering while exactly considering the nonlinearities of ff and hh. For simplicity, we define Φ[h(x)](y,u,μ,Σ)\Phi\left[h(x)\right](y,u,\mu^{-},\Sigma^{-}) for a scalar-, vector-, or matrix-valued function h(x)h(x) as

Φ[h(x)](y,u,μ,Σ)𝐑nh(x)p~xy(x,y;u,μ,Σ)𝑑x,\Phi\left[h(x)\right](y,u,\mu^{-},\Sigma^{-})\coloneqq\int_{{\boldsymbol{\mathrm{R}}}^{n}}h(x)\cdot\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-})dx,

which is an integral over xx depending on the parameters yy, uu, μ\mu^{-}, and Σ\Sigma^{-}. The computations of (9) and (10) are then reduced to evaluations of three integrals, namely Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}], because μ\mu and Σ\Sigma can be written as μ=Φ[x]/Φ[1]\mu=\Phi[x]/\Phi[1] and Σ=Φ[xx]/Φ[1]μμ\Sigma=\Phi[xx^{\top}]/\Phi[1]-\mu\mu^{\top}, respectively.

For nonlinear KFs, the posterior PDF is usually approximated as Gaussian [12, 18]. The integrals Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}] are also approximated without considering the effects of higher order moments induced by the nonlinearity of ff and hh. More specifically, the nonlinear KFs first approximate the integral in (7) by computing the mean and variance of the nonlinear function f(x,u)f(x^{-},u) and further approximate the integral in the denominator of (8) in the same way using h(x)h(x). By contrast, the proposed method considers the exact integrals using the HGM.

On the other hand, the PF approximates the integrals appearing in (1) using the Monte-Carlo scheme instead of the Gaussian approximation. Therefore, for a sufficiently large number of particles, the PF must exhibit a better accuracy than the proposed method. However, in the numerical example provided in Section V, we observe that the computational cost for the PF is more than that of the proposed method for the same level of accuracy.

Remark 1

Although, as mentioned above, the Gaussian approximation of the posterior distribution is a common approach in nonlinear filtering theory, this assumption is not suited to the cases of noise with heavy-tailed distributions, such as the Cauchy noise. In such cases, however, the proposed method may still be applicable by approximating the posterior distribution using a heavy-tailed distribution with some parameters. Provided that the parameters of the approximating distribution can be expressed as functions of the previous estimates, observed output, and known input (which corresponds to (9) and (10) for the Gaussian approximation), the proposed method can be performed in almost the same manner, which could be considered in future work.

III Holonomic Gradient Method

As described in the previous section, the update law of the mean and variance can be formulated as the evaluation of three integrals Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}]. For linear systems with Gaussian noise, these integrals can be computed analytically using the Gaussian integral. However, for nonlinear systems, we need a numerical method because it is an extremely complex process to compute the integrals analytically. To achieve the trade-off between the exact consideration of nonlinearity and reduction of computational cost, we use a symbolic-numeric method for evaluating the complex integrals, called the HGM. We only provide the outline of the HGM in a specific situation due to limited space.

Let us consider the evaluation of the following integral.

α(X)𝐑mβ(X,Y)γ(X,Y)𝑑Y,\alpha(X)\coloneqq\int_{{\boldsymbol{\mathrm{R}}}^{m}}\beta(X,Y)\cdot\gamma(X,Y)dY, (11)

where β\beta and γ\gamma are smooth scalar-valued functions in X𝐑nX\in{\boldsymbol{\mathrm{R}}}^{n} and Y𝐑mY\in{\boldsymbol{\mathrm{R}}}^{m}. For example, Φ[h(x)](y,u,μ,Σ)\Phi[h(x)](y,u,\mu^{-},\Sigma^{-}) can be regarded as integral (11) with β=h\beta=h, γ=p~xy\gamma=\tilde{p}_{xy}, Y=xY=x, and XX corresponds to (y,u,μ,Σ)(y,u,\mu^{-},\Sigma^{-}).

Definition 1

The function α(X)\alpha(X) is called holonomic if there exists a set of differential operators ={d1,,dq1}(q𝐙0,di𝐙0n)\mathcal{M}=\{\partial^{d_{1}},\dots,\allowbreak\partial^{d_{q-1}}\}\ (q\in{\boldsymbol{\mathrm{Z}}}_{\geq 0},\ d_{i}\in{\boldsymbol{\mathrm{Z}}}_{\geq 0}^{n}) such that the vector-valued function Q(X)Q(X) defined as

Q=[αd1αdq1α]Q=[\alpha\ \partial^{d_{1}}\alpha\ \cdots\ \partial^{d_{q-1}}\alpha]^{\top} (12)

satisfies a set of the following PDEs.

XiQ(X)=AXi(X)Q(X)(i=1,,n),\partial_{X_{i}}Q(X)=A_{X_{i}}(X)Q(X)\qquad(i=1,\dots,n), (13)

where AXi(X)𝐑(X)n×nA_{X_{i}}(X)\in{\boldsymbol{\mathrm{R}}}(X)^{n\times n}. The set of PDEs (13) is called the Pfaffian system of Q(X)Q(X).

We assume that we know \mathcal{M} and AXiA_{X_{i}} in Definition 1 for the holonomic function α(X)\alpha(X). Moreover, we assume that the vector Q(Xinit)Q(X_{\mathrm{init}}) at a certain point XinitX_{\mathrm{init}} is known. For example, the vector Q(Xinit)Q(X_{\mathrm{init}}) can be directly computed from (11) if we use sufficient computational resources. Once Q(Xinit)Q(X_{\mathrm{init}}) is obtained, by virtue of (13), we can efficiently evaluate α(X^)\alpha(\hat{X}) at a point X^Xinit\hat{X}\neq X_{\mathrm{init}} without directly evaluating the multiple integral (11). The target point X^\hat{X} can be arbitrarily selected considering that the following process is successfully performed.

Let X(s)X(s) be a smooth curve [0,1]sX(s)𝐑n[0,1]\ni s\mapsto X(s)\in{\boldsymbol{\mathrm{R}}}^{n} with X(0)=XinitX(0)=X_{\mathrm{init}} and X(1)=X^X(1)=\hat{X}. Considering the Pfaffian system (13) on this curve, we obtain an IVP

dQds=i=1nXiQdXids=i=1nAXi(X(s))Q(X(s))dXids,Q(X(0))=Q(Xinit).\frac{dQ}{ds}=\sum_{i=1}^{n}\partial_{X_{i}}Q\frac{dXi}{ds}=\sum_{i=1}^{n}A_{X_{i}}(X(s))Q(X(s))\frac{dX_{i}}{ds},\\ Q(X(0))=Q(X_{\mathrm{init}}). (14)

This IVP can be solved to obtain Q(X^)Q(\hat{X}) using numerical solution methods such as the RK4 method if none of the denominators in AXiA_{X_{i}} vanish at any s[0,1]s\in[0,1]. Consequently, we can evaluate α(X^)\alpha(\hat{X}) as the first component of Q(X^)Q(\hat{X}).

The finite set \mathcal{M} and matrices AXiA_{X_{i}}, which are supposed to be known, can be computed from a finite set of differential operators called a basis of a zero-dimensional ideal in n\mathcal{R}_{n}. The following theorem provides a way to find such a basis for a given function and to determine whether the function is holonomic or not (We have skipped the details for simplicity).

Theorem 1

[19] A smooth function α(X)\alpha(X) is holonomic if and only if there exist differential operators, described as finite sums of the following form.

lidad(X)Xid(ad𝐑(X),i=1,,n),l_{i}\coloneqq\sum_{d}a_{d}(X)\partial_{X_{i}}^{d}\quad(a_{d}\in{\boldsymbol{\mathrm{R}}}(X),\ i=1,\dots,n), (15)

which annihilate α(X)\alpha(X), i.e., liα(X)=0l_{i}\bullet\alpha(X)=0 for (i=1,,n)(i=1,\dots,n). Moreover, the set of differential operators {l1,,ln}\{l_{1},\dots,l_{n}\} is a basis of a zero-dimensional ideal in n\mathcal{R}_{n}.

Example 1

A function α(X1,X2)=cos(X1X2)\alpha(X_{1},X_{2})=\cos(X_{1}X_{2}) is holonomic because there exist two differential operators of the form (15):

l1X12+X22,l2X22+X12,l_{1}\coloneqq\partial_{X_{1}}^{2}+X_{2}^{2},\qquad l_{2}\coloneqq\partial_{X_{2}}^{2}+X_{1}^{2},

which annihilate α(X1,X2)\alpha(X_{1},X_{2}). Indeed, by using, for example, Risa/Asir [21] and its library yang, the Pfaffian system (13) can be computed as

X1Q=[1/X1X1X2X2/X10]Q,X2Q=[0X1210]Q,\partial_{X_{1}}Q=\left[\begin{array}[]{cc}1/X_{1}&X_{1}X_{2}\\ X_{2}/X_{1}&0\end{array}\right]Q,\quad\partial_{X_{2}}Q=\left[\begin{array}[]{cc}0&X_{1}^{2}\\ 1&0\end{array}\right]Q,

where Q(X1,X2)=[αX2α]Q(X_{1},X_{2})=[\alpha\ \partial_{X_{2}}\alpha]^{\top}, that is, \mathcal{M} in Definition 1 is a singleton {X2}\{\partial_{X_{2}}\}.

It may be difficult to find a basis of a zero-dimensional ideal for α\alpha directly from the expression (11). Using symbolic computation, we can compute a basis of a zero-dimensional ideal for α\alpha from those for β\beta and γ\gamma [20], which implies the following corollary [20, 19].

Corollary 1

If β\beta and γ\gamma in (11) are holonomic, the product βγ\beta\cdot\gamma is also holonomic. In addition, if the product βγ\beta\cdot\gamma is rapidly decreasing with respect to each Yi(i=1,,m)Y_{i}\ (i=1,\dots,m), then α\alpha, defined as (11), is also holonomic.

Example 2

For X𝐑X\in{\boldsymbol{\mathrm{R}}} and Y𝐑Y\in{\boldsymbol{\mathrm{R}}}, let β(X,Y)=cos(XY)\beta(X,Y)=\cos(XY) and γ(X,Y)=exp(X2/2Y2/2)\gamma(X,Y)=\exp(-X^{2}/2-Y^{2}/2). Evidently, β\beta is annihilated by X2+Y2\partial_{X}^{2}+Y^{2} and Y2+X2\partial_{Y}^{2}+X^{2}, and γ\gamma is annihilated by X+X\partial_{X}+X and Y+Y\partial_{Y}+Y. That is, both functions are holonomic. Moreover, γ\gamma is rapidly decreasing with respect to YY and so is the product βγ\beta\cdot\gamma, which indicates that α(X)\alpha(X) defined by (11) is holonomic (Corollary 1). Using Risa/Asir and its library nk_restriction, we can compute X+2X\partial_{X}+2X as a basis of a zero-dimensional ideal for α(X)\alpha(X) from the differential operators annihilating β\beta and γ\gamma. Indeed, it is easy to confirm that in this example α(X)exp(X2)\alpha(X)\propto\exp(-X^{2}), which is annihilated by the basis.

Corollary 1, combined with the following assumption, ensures that p~xy(x,y;u,μ,Σ)\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-}) is holonomic.

Assumption 1

The conditional PDFs p(xx;u)p(x\mid x^{-};u) and p(yx)p(y\mid x) in (7) are holonomic functions.

IV Evaluation of Mean and Variance via Integral Transform

Under Assumption 1, all the components of the integrals Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}] are holonomic functions of uu, yy, μ\mu^{-}, and Σ\Sigma^{-}. Therefore, they can be efficiently evaluated using the HGM if the corresponding Pfaffian systems (13) and initial vectors (12) are computed in advance. The authors propose a method to compute (9) and (10) by directly using the HGM [14]. Although the previous method showed better performance than other existing methods for a numerical example, the IVP (14) has to be solved for each component of the three integrals. If we can decrease the number of IVPs, then it reduces the online computational cost. Hereafter, z𝐑Nz\in{\boldsymbol{\mathrm{R}}}^{N} denotes a vector consisting of all the independent components of y𝐑ry\in{\boldsymbol{\mathrm{R}}}^{r}, u𝐑mu\in{\boldsymbol{\mathrm{R}}}^{m}, μ𝐑n\mu^{-}\in{\boldsymbol{\mathrm{R}}}^{n}, and ΣPD(n)\Sigma^{-}\in\mathrm{PD}(n), where N=r+m+n+n(n+1)/2N=r+m+n+n(n+1)/2. With this notation, for example, p~xy(x,z)\tilde{p}_{xy}(x,z) denotes p~xy(x,y;u,μ,Σ)\tilde{p}_{xy}(x,y;u,\mu^{-},\Sigma^{-}) as a function of xx, yy, uu, μ\mu^{-}, and Σ\Sigma^{-}.

In this work, we use an integral transform of p~xy\tilde{p}_{xy} to reduce the number of IVPs solved in the one-step estimation. Consider an integral transform 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] defined as follows.

𝒯[p~xy](ξ,z)𝐑nexp(ξx)p~xy(x,z)𝑑x.\mathcal{T}[\tilde{p}_{xy}](\xi,z)\coloneqq\int_{{\boldsymbol{\mathrm{R}}}^{n}}\exp(\xi^{\top}x)\cdot\tilde{p}_{xy}(x,z)dx. (16)

Note that although this definition is similar to that of the moment generating function, it is different because p~xy\tilde{p}_{xy} is the PDF of the xx and yy pair rather than xx. We assume that there exists a compact subset of 𝐑n{\boldsymbol{\mathrm{R}}}^{n} including the origin such that for all ξ\xi in the subset, (16) can be defined and smooth.

The integrals Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}] are then obtained from the derivatives of 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] at ξ=0\xi=0 as follows:

𝒯[p~xy](0,z)=𝐑np~xy(x,z)𝑑x=Φ[1](z),\mathcal{T}[\tilde{p}_{xy}](0,z)=\int_{{\boldsymbol{\mathrm{R}}}^{n}}\tilde{p}_{xy}(x,z)dx=\Phi[1](z),
ξi𝒯[p~xy](0,z)\displaystyle\partial_{\xi_{i}}\mathcal{T}[\tilde{p}_{xy}](0,z) =𝐑n[ξiexp(ξx)p~xy(x,z)]|ξ=0dx\displaystyle=\int_{{\boldsymbol{\mathrm{R}}}^{n}}\left.\left[\partial_{\xi_{i}}\exp(\xi^{\top}x)\tilde{p}_{xy}(x,z)\right]\right|_{\xi=0}dx
=Φ[xi](z),\displaystyle=\Phi\left[x_{i}\right](z), (17)
ξiξj𝒯[p~xy](0,z)\displaystyle\partial_{\xi_{i}}\partial_{\xi_{j}}\mathcal{T}[\tilde{p}_{xy}](0,z) =𝐑n[ξiξjexp(ξx)p~xy(x,z)]|ξ=0dx\displaystyle=\int_{{\boldsymbol{\mathrm{R}}}^{n}}\left.\left[\partial_{\xi_{i}}\partial_{\xi_{j}}\exp(\xi^{\top}x)\tilde{p}_{xy}(x,z)\right]\right|_{\xi=0}dx
=Φ[xixj](z).\displaystyle=\Phi\left[x_{i}x_{j}\right](z). (18)

Here, the approximating PDF p~xy\tilde{p}_{xy} is a holonomic function under Assumption 1, and the kernel exp(ξx)\exp(\xi^{\top}x) is also holonomic. This implies that 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] is also a holonomic function by Corollary 1. Therefore, \mathcal{M} exists such that the vector-valued function Q(X)Q(X) defined by \mathcal{M} and 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] satisfies the Pfaffian system (13). Using the Pfaffian system, we can explicitly express 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] and its derivatives (17) as well as (18) as a linear combination of the components of QQ with coefficients in 𝐑(ξ,z){\boldsymbol{\mathrm{R}}}(\xi,z).

First, 𝒯[p~xy](ξ,z)\mathcal{T}[\tilde{p}_{xy}](\xi,z) itself is the first component of Q(ξ,z)Q(\xi,z), thus satisfying the following:

Φ[1]=𝒯[p~xy]=C(0)Q,\Phi[1]=\mathcal{T}[\tilde{p}_{xy}]=C^{(0)}Q, (19)

where C(0)C^{(0)} is a constant coefficient vector [1 0 0]𝐑q[1\ 0\ \cdots\ 0]\in{\boldsymbol{\mathrm{R}}}^{q}. Next, the first derivatives ξi𝒯[p~xy](i=1,,n)\partial_{\xi_{i}}\mathcal{T}[\tilde{p}_{xy}]\ (i=1,\dots,n) are obtained using the first components of the left-hand sides of (13). Hence, the following identities hold.

Φ[xi]=ξi𝒯[p~xy]=Ci(1)Q(i=1,,n),\Phi[x_{i}]=\partial_{\xi_{i}}\mathcal{T}[\tilde{p}_{xy}]=C^{(1)}_{i}Q\quad(i=1,\dots,n), (20)

where the coefficient vector Ci(1)(ξ,z)C^{(1)}_{i}(\xi,z) is the first row vector of Aξi(ξ,z)𝐑(ξ,z)q×qA_{\xi_{i}}(\xi,z)\in{\boldsymbol{\mathrm{R}}}(\xi,z)^{q\times q} in (13). Finally, the second derivatives ξiξj𝒯[p~xy](i,j=1,,n)\partial_{\xi_{i}}\partial_{\xi_{j}}\mathcal{T}[\tilde{p}_{xy}]\ (i,j=1,\dots,n) are obtained by differentiating both sides of (13); ξiξjQ=ξi(AξjQ)=ξiAξjQ+AξjξiQ=(ξiAξj+AξjAξi)Q\partial_{\xi_{i}}\bullet\partial_{\xi_{j}}Q=\partial_{\xi_{i}}\bullet(A_{\xi_{j}}Q)=\partial_{\xi_{i}}A_{\xi_{j}}Q+A_{\xi_{j}}\partial_{\xi_{i}}Q=(\partial_{\xi_{i}}A_{\xi_{j}}+A_{\xi_{j}}A_{\xi_{i}})Q, and hence

Φ[xixj]=ξiξj𝒯[p~xy]=Cij(2)Q(i,j{1,,n}),\Phi[x_{i}x_{j}]=\partial_{\xi_{i}}\partial_{\xi_{j}}\mathcal{T}[\tilde{p}_{xy}]=C^{(2)}_{ij}Q\quad(i,j\in\{1,\dots,n\}), (21)

where the coefficient vector Cij(2)(ξ,z)C^{(2)}_{ij}(\xi,z) is the first row vector of ξjAξi+AξiAξj\partial_{\xi_{j}}A_{\xi_{i}}+A_{\xi_{i}}A_{\xi_{j}}.

We assume the explicit expressions of Ci(1)(ξ,z)C^{(1)}_{i}(\xi,z) and Cij(2)(ξ,z)C^{(2)}_{ij}(\xi,z) are obtained. Using the expressions, we can compute the values of Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}] at z=z^z=\hat{z} from (19)–(21) if we have the vector Q(0,z^)Q(0,\hat{z}) for a given z^\hat{z}. The computation of Q(0,z^)Q(0,\hat{z}) can be accomplished by numerically integrating (14) using numerical integration methods, such as the RK4 method. Therefore, the current mean μ\mu and variance Σ\Sigma for a given data z^\hat{z} can be computed from (9) and (10) as

μ=Φ[x](z^)Φ[1](z^),Σ=Φ[xx](z^)Φ[1](z^)μμ.\mu=\frac{\Phi[x](\hat{z})}{\Phi[1](\hat{z})},\quad\Sigma=\frac{\Phi[xx^{\top}](\hat{z})}{\Phi[1](\hat{z})}-\mu\mu^{\top}. (22)

In the previous method [14], IVP (14) has to be solved for each of the three integrals Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}] even for the one-dimensional case, that is, three IVPs have to be solved in total. By contrast, only a single IVP for 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] needs to be solved in the proposed method. This decrease in the number of IVPs leads to a reduction in the computational cost, which can be seen in Section V.

Algorithm 1 Computation of Pfaffian System for 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] (Offline part)
1:Bases of zero-dimensional ideals annihilating p~(x;μ,Σ)\tilde{p}(x^{-};\mu^{-},\Sigma^{-}), p(xx;u)p(x\mid x^{-};u), p(yx)p(y\mid x), and exp(ξx)\exp(\xi^{\top}x)
2:Explicit expressions of Ci(1)(ξ,z)(i=1,,n)C^{(1)}_{i}(\xi,z)\ (i=1,\dots,n), Cij(2)(ξ,z)(i,j{1,,n})C^{(2)}_{ij}(\xi,z)\ (i,j\in\{1,\dots,n\}), Aλ(ξ,z)(λ{ξ1,,ξn,z1,,zN})A_{\lambda}(\xi,z)\ (\lambda\in\{\xi_{1},\dots,\xi_{n},z_{1},\dots,z_{N}\})
3:Compute basis \mathcal{B} of zero-dimensional ideal for 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}] from inputs
4:Find ={d1,,dq1}\mathcal{M}=\{\partial^{d_{1}},\dots,\partial^{d_{q-1}}\} in Definition 1 from \mathcal{B}
5:Compute coefficient matrices AλA_{\lambda} from \mathcal{B} and \mathcal{M}
6:Compute Ci(1)(ξ,z)C^{(1)}_{i}(\xi,z) and Cij(2)(ξ,z)C^{(2)}_{ij}(\xi,z) from AξiA_{\xi_{i}}
Algorithm 2 One-step Estimation via HGM using (16) (Online part)
1:Explicit expressions of C(0)C^{(0)}, Ci(1)(ξ,z)(i=1,,n)C^{(1)}_{i}(\xi,z)\ (i=1,\dots,n), Cij(2)(ξ,z)(i,j{1,,n})C^{(2)}_{ij}(\xi,z)\ (i,j\in\{1,\dots,n\}), Aλ(ξ,z)(λ{ξ1,,ξn,z1,,zN})A_{\lambda}(\xi,z)\ (\lambda\in\{\xi_{1},\dots,\xi_{n},z_{1},\dots,z_{N}\}), initial point zinitz_{\mathrm{init}} and vector Q(0,zinit)Q(0,z_{\mathrm{init}}), as well as given data z^\hat{z}
2:Failure or estimates μ\mu and Σ\Sigma
3:Define z(s)z(s) such that z(0)=zinitz(0)=z_{\mathrm{init}} and z(1)=z^z(1)=\hat{z}
4:if s[0,1]\exists s\in[0,1] at which denominator in AλA_{\lambda} vanishes, then
5:     return algorithm has failed
6:Solve IVP (14) numerically using initial vector Q(0,zinit)Q(0,z_{\mathrm{init}}) and obtain Q(0,z^)Q(0,\hat{z})
7:return (22) obtained from (19)–(21)

Now, we summarize the offline and online processes of the proposed method as Algorithms 1 and 2. The details of each step in Algorithm 1 can be found in [20, 14].

V Numerical Example

This section presents a numerical example to show the efficiency of the proposed method. In the following demonstration, we use Risa/Asir [21] for the symbolic computation, Maple for the offline numerical computation, and Python for the online numerical computation on a PC (Intel(R) Core(TM) i7-1065G7 CPU @ 1.30 GHz; RAM: 16 GB).

V-A Problem setting and results of offline computation

Consider the following nonlinear stochastic one-dimensional system.

x=45x+u+w,y=2x1+x2+v,x=\frac{4}{5}x^{-}+u+w,\quad y=\frac{2x}{1+x^{2}}+v, (23)

where ww and vv are the system and observation noises with the standard Gaussian PDF, respectively, and uu is given as the function cos(0.6k)\cos(0.6k) of the discrete-time step kk.

First, we symbolically compute the inputs of Algorithm 1, namely, bases of zero-dimensional ideals annihilating p~(x;μ,Σ)=𝒩(x;μ,Σ){\tilde{p}(x^{-};\mu^{-},\Sigma^{-})}=\mathcal{N}(x^{-};\mu^{-},\Sigma^{-}), (4), (5), and exp(ξx)\exp{(\xi x)}. By differentiating p~(x;μ,Σ)\tilde{p}(x^{-};\mu^{-},\Sigma^{-}) with respect to each variable, we can obtain the differential operators annihilating p~(x;μ,Σ)\tilde{p}(x^{-};\mu^{-},\Sigma^{-}) as

Σx+xμ,Σμx+μ,\displaystyle\Sigma^{-}\partial_{x^{-}}+x^{-}-\mu^{-},\quad\Sigma^{-}\partial_{\mu^{-}}-x^{-}+\mu^{-}, (24)
2(Σ)2Σ+Σ(xμ)2.\displaystyle 2(\Sigma^{-})^{2}\partial_{\Sigma^{-}}+\Sigma^{-}-(x^{-}-\mu^{-})^{2}.

The conditional PDF (4) is obtained by substituting the state equation into the PDF of 𝒩(w;0,1)\mathcal{N}(w;0,1) and is annihilated by the following three differential operators:

x+x(4/5)xu,ux+(4/5)x+u,\displaystyle\partial_{x}+x-(4/5)x^{-}-u,\quad\partial_{u}-x+(4/5)x^{-}+u, (25)
(5/4)x+x(4/5)xu.\displaystyle-(5/4)\partial_{x^{-}}+x-(4/5)x^{-}-u.

The other conditional PDF (5) is also obtained by substituting the observation equation into the PDF of 𝒩(v;0,1)\mathcal{N}(v;0,1) and is annihilated by

y+y2x1+x2,(1+x2)22(1x2)x+y2x1+x2.\partial_{y}+y-\frac{2x}{1+x^{2}},\quad-\frac{(1+x^{2})^{2}}{2(1-x^{2})}\partial_{x}+y-\frac{2x}{1+x^{2}}. (26)

Finally, exp(ξx)\exp(\xi x) is annihilated by

ξx,xξ.\partial_{\xi}-x,\quad\partial_{x}-\xi. (27)

By Theorem 1, the sets of differential operators (24)–(27) are bases of zero-dimensional ideals annihilating p~(x;μ,Σ)\tilde{p}(x^{-};\mu^{-},\Sigma^{-}), p(xx;u)p(x\mid x^{-};u), p(yx)p(y\mid x), and exp(ξx)\exp(\xi x), respectively.

Using Algorithm 1, we can derive a basis \mathcal{B} of a zero-dimensional ideal annihilating 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}], a finite set of differential operators \mathcal{M}, and the coefficient vectors Ci(1)(ξ,z),Cij(2)(ξ,z)C^{(1)}_{i}(\xi,z),C^{(2)}_{ij}(\xi,z) and matrices Aλ(λ{ξ,z1,z2,z3,z4})A_{\lambda}\ (\lambda\in\{\xi,z_{1},z_{2},z_{3},z_{4}\}), where z=[z1z4]=[yuμΣ]z=[z_{1}\ \cdots\ z_{4}]^{\top}=[y\ u\ \mu^{-}\ \Sigma^{-}]^{\top}. In this example, \mathcal{B} consists of seven differential operators, which include

2ξz12ξ2z2+2yξz1+ξξ22z1z2+4ξ+ξ.2\partial_{\xi}\partial_{z_{1}}^{2}-\partial_{\xi}^{2}\partial_{z_{2}}+2y\partial_{\xi}\partial_{z_{1}}+\xi\partial_{\xi}^{2}-2\partial_{z_{1}}-\partial_{z_{2}}+4\partial_{\xi}+\xi.

The other operators are omitted because of space limitations. The finite set \mathcal{M} is obtained as {z1,z2,z4,z12,z2z1,z4z2}\left\{\partial_{z_{1}},\partial_{z_{2}},\partial_{z_{4}},\partial_{z_{1}}^{2},\partial_{z_{2}}\partial_{z_{1}},\partial_{z_{4}}\partial_{z_{2}}\right\}. Hence, QQ becomes a seven-dimensional vector-valued function:

Q=[1z1z2z4z12z2z1z4z2]𝒯[p~xy],Q=\left[1\ \partial_{z_{1}}\ \partial_{z_{2}}\ \partial_{z_{4}}\ \partial_{z_{1}}^{2}\ \partial_{z_{2}}\partial_{z_{1}}\ \partial_{z_{4}}\partial_{z_{2}}\right]^{\top}\bullet\mathcal{T}[\tilde{p}_{xy}], (28)

where \bullet represents the actions of all the components on 𝒯[p~xy]\mathcal{T}[\tilde{p}_{xy}]. All entries of the coefficient matrices Aλ𝐑(ξ,z)7×7(λ{ξ,z1,z2,z3,z4})A_{\lambda}\in{\boldsymbol{\mathrm{R}}}(\xi,z)^{7\times 7}\ (\lambda\in\{\xi,z_{1},z_{2},z_{3},z_{4}\}) as well as the coefficient vectors in (20) and (21) are explicitly computed from \mathcal{B}. For example, an overview of AξA_{\xi} is presented bellow.

[z2+45z301625z4+100000z2+45z3000010z2+45z300000162500(z245z3)z1+1000],\left[\begin{array}[]{ccccccc}z_{2}+\frac{4}{5}z_{3}&0&\frac{16}{25}z_{4}+1&0&0&0&0\\ 0&z_{2}+\frac{4}{5}z_{3}&0&0&0&*&0\\ 1&0&z_{2}+\frac{4}{5}z_{3}&*&0&0&0\\ 0&0&\frac{16}{25}&*&0&0&*\\ &(-z_{2}-\frac{4}{5}z_{3})z_{1}+1&*&*&0&*&*\\ &*&*&*&0&*&0\\ &*&*&*&*&*&*\end{array}\right],

where the entries * denote rational functions of ξ,z1,,z4\xi,z_{1},\dots,z_{4} and are omitted because of the space limitations.

Finally, we have to choose the initial points zinitz_{\mathrm{init}} and compute corresponding initial vectors Q(0,zinit)Q(0,z_{\mathrm{init}}). The choice of the initial points depends on the denominators in Aλ(λ{ξ,z1,z2,z3,z4})A_{\lambda}\ (\lambda\in\{\xi,z_{1},z_{2},z_{3},z_{4}\}) because they should not vanish on the integration path; otherwise, the right-hand side of the ODE in (14) diverges and therefore Algorithm 2 fails (line: 5). In this example, however, the least common multiple of all the denominators in AλA_{\lambda} is 16z4+2516z_{4}+25, which must not be zero because z4=Σ>0z_{4}=\Sigma^{-}>0. Therefore, any component of AλA_{\lambda} must not diverge. For this example, we choose the following eight initial points: Zinit{[z1z2z3z4]z1=±1,z2=±1,z3=±1,z4=1}Z_{\mathrm{init}}\coloneqq\left\{[z_{1}\ z_{2}\ z_{3}\ z_{4}]^{\top}\mid z_{1}=\pm 1,z_{2}=\pm 1,z_{3}=\pm 1,z_{4}=1\right\}. The corresponding initial vectors can be computed numerically from the definition (28).

V-B Settings for online computation and estimation results

In the online computations, the data zz consisting of the input uu, output yy, and previous estimates μ\mu^{-} and Σ\Sigma^{-} are given at each time step. Let z^\hat{z} denote the given data at a time step. For this example, we fix the integration path in line 3 of Algorithm 2 to a line segment from an initial point zinitz_{\mathrm{init}} to a given data z^\hat{z}, that is, z(s)sz^+(1s)zinitz(s)\coloneqq s\hat{z}+(1-s)z_{\mathrm{init}}. The initial point zinitz_{\mathrm{init}} is therefore selected for each given data z^\hat{z} such that the length of the line segment is the minimum among the prescribed set ZinitZ_{\mathrm{init}}. The numerical integration along the defined path is performed using the ABM4 method, with the first three steps initialized using the RK4 method. Under these settings, Algorithm 2 is performed at each time step to compute the next estimates μ\mu and Σ\Sigma from the previous ones μ(=z3)\mu^{-}(=z_{3}) and Σ(=z4)\Sigma^{-}(=z_{4}) using the measurement y(=z1)y(=z_{1}) and the known input u(=z2)u(=z_{2}). Figure 1 shows a realization of the system (23) and the estimation results obtained from the proposed method.

Refer to caption
(a) Realization of output sequence
Refer to caption
(b) Trajectories of realized state (solid) and estimate (dashed). Filled area shows intervals between μ±Σ\mu\pm\sqrt{\Sigma}
Figure 1: Realization of system (23) and corresponding estimation result

For comparison, we implemented the EKF, UKF, PF, and our previous method. The number of particles in the PF was set to 100 so that its accuracy would be almost the same as that of the proposed method. Three hundred realizations are generated from the random initial states and sequences of system and observation noises for this comparison.

To compare the performances of all the methods, the negative log-likelihood (NLL) [22] was computed. Figure 2 shows the NLL of each method averaged over all realizations. As can be seen, only the PF of 100 particles shows comparable performance to the proposed and previous methods. The averaged NLL of the previous method is identical to that of the proposed method because both approaches compute the same estimates (22). The computational times for the one-step estimations of all the methods are summarized as boxplots in Fig. 3. It can be seen that the proposed method is faster than the previous method as well as the PF with 100 particles, which show comparable performances for the NLL values. These results indicate that the proposed method outperforms the PF and our previous method.

Refer to caption
Figure 2: Comparison of NLLs for proposed method (solid, thick), previous method (solid, thin), PFs of 100 particles (dashed), UKF (dotted), and EKF (dash-dotted, almost greater than 1.2 after k=4k=4)
Refer to caption
Figure 3: Boxplots of computational times for one-step estimations of all methods

As shown in this example, there exists at least a single problem in which the proposed method outperforms the existing methods, exhibiting the superiority of the proposed method. Currently, it is difficult to demonstrate the superiority for general cases because of limitations such as the instability of ODE (14). Therefore, the theoretical or experimental analysis demonstrating the superiority will be a part of future research.

Despite the complex procedure, the principle behind the proposed method is simple; it is basically the evaluation of the integrals Φ[1]\Phi[1], Φ[x]\Phi[x], and Φ[xx]\Phi[xx^{\top}], which can be naturally obtained by the Gaussian approximation of the posterior PDF in Bayesian filtering. Thus, the proposed method is simple to understand except for the precise evaluation of integrals, which is complex. Moreover, if it is formulated as a set of Algorithms 1 and 2, the implementation of the proposed method can be automated using symbolic computation so that the users do not have to be concerned about its complex procedure. Therefore, we believe that it is worth studying the proposed method, and the improvement of its usability should be a part of the future research.

VI Conclusion

A symbolic-numeric method is proposed herein to perform Bayesian filtering for a class of nonlinear stochastic systems. The posterior PDF of the state is first approximated using a Gaussian PDF, and the mean and variance are then updated while considering the nonlinearity of the system exactly. This update process is formulated as evaluations of several integrals using the HGM. We also propose an integral transform to compute the mean and variance efficiently from the vector-valued function obtained using the HGM.

It is known that the numerical integration in the third step of the HGM may diverge owing to errors in the initial vectors. Therefore, in future work, the selection of integration paths or initial points will be studied in depth to enable more stable numerical integrations, for example, by considering the special structure of the problems considered in this work. Another direction for future work is approximating the posterior PDF more accurately using a PDF with more parameters than the Gaussian, such as the skew Gaussian.

References

  • [1] R. E. Kalman, “A new approach to linear filtering and prediction problems,” ASME Journal of Basic Engineering, vol. 82, pp. 35–45, 1960.
  • [2] R. E. Kalman and R. S. Bucy, “New results in linear filtering and prediction theory,” ASME Journal of Basic Engineering, vol. 83, no. 1, pp. 95–108, 1961.
  • [3] I. Rusnak, “Maximum likelihood optimal estimator of non-autonomous nonlinear dynamic systems,” in Proceedings of European Control Conference, pp. 909–914, 2015.
  • [4] X. Luo, Y. Jiao, W. L. Chiou, and S. S. Yau, “A novel suboptimal method for solving polynomial filtering problems,” Automatica, vol. 62, pp. 26–31, 2015.
  • [5] B. Chen and G. Hu, “Nonlinear state estimation under bounded noises,” Automatica, vol. 98, pp. 159–168, 2018.
  • [6] N. Duong, J. L. Speyer, J. Yoneyama, and M. Idan, “Laplace estimator for linear scalar systems,” in Proceedings of the IEEE Conference on Decision and Control, pp. 2283–2290, 2018.
  • [7] K. Li, F. Pfaff, and U. D. Hanebeck, “Unscented dual quaternion particle filter for SE(3) estimation,” IEEE Control Systems Letters, vol. 5, no. 2, pp. 647–652, 2021.
  • [8] G. Kitagawa, “Monte Carlo filtering and smoothing method for non-Gaussian nonlinear state space model,” Institute of Statistical Mathematics Research Memorandum, vol. 462, 1993.
  • [9] N. J. Gordon, D. J. Salmond, and A. F. M. Smith, “Novel approach to nonlinear/non-Gaussian Bayesian state estimation,” in IEE Proceedings F - Radar and Signal Processing, vol. 140, pp. 107–113, 1993.
  • [10] A. H. Jazwinski Stochastic Processes and Filtering Theory, Elsevier Science, 1970.
  • [11] S. J. Julier and J. K. Uhlmann, “New extension of the Kalman filter to nonlinear systems,” in SPIE, vol. 3068, 1997.
  • [12] I. Arasaratnam and S. Haykin, “Cubature Kalman filters,” IEEE Transactions on Automatic Control, vol. 54, no. 6, pp. 1254–1269, 2009.
  • [13] K. Ito and K. Xiong, “Gaussian filters for nonlinear filtering problems,” IEEE Transactions on Automatic Control, vol. 45, no. 5, pp. 910–927, 2000.
  • [14] T. Iori and T. Ohtsuka, “Symbolic-numeric computation of posterior mean and variance for a class of discrete-time nonlinear stochastic systems,” in Proceedings of the IEEE Conference on Decision and Control, pp. 4814–4821, 2020.
  • [15] H. Nakayama, K. Nishiyama, M. Noro, K. Ohara, T. Sei, N. Takayama, and A. Takemura, “Holonomic gradient descent and its application to the Fisher-Bingham integral,” Advances in Applied Mathematics, vol. 47, no. 3, pp. 639–658, 2011.
  • [16] M. Idan and J. L. Speyer, “Multivariate Cauchy estimator with scalar measurement and process noises,” SIAM Journal on Control and Optimization, vol. 52, no. 2, pp. 1108–1141, 2014.
  • [17] J. A. E. Bryson and Y.-C. Ho Applied Optimal Control, John Wiley & Sons, 1st edition, 1975.
  • [18] H. Fang, N. Tian, Y. Wang, M. Zhou, and M. A. Haile, “Nonlinear Bayesian estimation: from Kalman filter to a broader horizon,” IEEE/CAA Journal of Automatica Sinica, vol. 5, no. 2, pp. 401–417, 2018.
  • [19] T. Hibi ed. Gröbner Bases: Statistics and Software Systems, Springer Japan, 1st edition, 2013.
  • [20] T. Oaku, Y. Shiraki, and N. Takayama, “Algebraic algorithms for D-modules and numerical analysis,” Computer Mathematics (Proceedings of ASCM 2003), vol. 10, pp. 23–39, 2003.
  • [21] M. Noro, N. Takayama, H. Nakayama, K. Nishiyama, and K. Ohara, “Risa/Asir: A computer algebra system,” http://www.math.kobe-u.ac.jp/Asir/asir.html, 2020.
  • [22] M. P. Deisenroth Efficient Reinforcement Learning Using Gaussian Processes, KIT Scientific Publishing, 2010.