This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Generic Stability Implication from Full Information Estimation to Moving-Horizon Estimation

Wuhua Hu W. Hu is with Towngas Energy, Shenzhen, China. The research was conducted in his free time when he was working with Envision Digital in Singapore. E-mail: [email protected]
Abstract

Optimization-based state estimation is useful for handling of constrained linear or nonlinear dynamical systems. It has an ideal form, known as full information estimation (FIE) which uses all past measurements to perform state estimation, and also a practical counterpart, known as moving-horizon estimation (MHE) which uses most recent measurements of a limited length to perform the estimation. Due to the theoretical ideal, conditions for robust stability of FIE are relatively easier to establish than those for MHE, and various sufficient conditions have been developed in literature. This work reveals a generic link from robust stability of FIE to that of MHE, showing that the former implies at least a weaker robust stability of MHE which implements a long enough horizon. The implication strengthens to strict robust stability of MHE if the corresponding FIE satisfies a mild Lipschitz continuity condition. The revealed implications are then applied to derive new sufficient conditions for robust stability of MHE, which further reveal an intrinsic relation between the existence of a robustly stable FIE/MHE and the system being incrementally input/output-to-state stable.

Index Terms:
Nonlinear systems; moving-horizon estimation; full information estimation; state estimation; disturbances; robust stability; incremental input/output-to-state stability

I Introduction

Optimization-based state estimation is an estimation approach which performs state estimation by solving an optimization problem. Compared to conventional approaches like Kalman filtering (KF) which deals with linear dynamical systems, and its extensions like extended KF and unscented KF which can deal with nonlinear dynamical systems based on linearization techniques, optimization-based approach has the advantage of handling linear and nonlinear dynamical systems directly and also including various physical or operational constraints [1]. The optimization formulation also admits flexible definition of the objective function to be optimized, which in some cases is necessary to accurately recover the state [2].

Optimization-based state estimation generally takes one of the two forms: full information estimation (FIE) which uses all past measurements to perform the estimation, and moving-horizon estimation (MHE) which uses most recent measurements of a limited length to perform the estimation. MHE is a practical approximate of FIE which is ideal and computationally intractable. The interest in FIE lies in two aspects: serving as a benchmark to MHE and providing useful insights for the stability analysis of MHE. Recent studies on FIE and MHE are concentrated on analyses of their stability and robustness, as are critical to guide the applications of MHE. While earlier literature assumes restrictive and idealistic conditions such as observability and/or zero or a priori known convergent disturbances [3, 4, 5, 6, 7, 8, 9], recent literature considers more practical conditions like detectability and/or in the presence of bounded disturbances [10, 11, 1].

An important progress was made in [11], which introduced the incremental input/output-to-state stability (i-IOSS) concept on detectability of nonlinear systems (as developed in [12]) to study robust stability of FIE and MHE. Given an i-IOSS system, it was shown that under mild conditions FIE is robustly stable and convergent for convergent disturbances. Whereas, conditions for FIE and MHE to be robustly stable under bounded disturbances were posted as an open research challenge. Reference [13] provided a prompt response to this challenge,111The main results were obtained in the late of 2012 though the paper was not able to be published until 2015. identifying a general set of conditions for FIE to be robustly stable for i-IOSS systems. Onwards, a series of researches have been inspired to close the challenge.

A particular type of cost functions with a max-term was investigated for a class of i-IOSS systems in [14], establishing robust stability of the FIE. The conditions were enhanced in [15], enabling robust stability also of MHE if a sufficiently long horizon is applied. The conclusion was extended to MHE without a max-term in the cost function [16]. Meanwhile, it was shown that MHE is convergent for convergent disturbances, with or without a max-term. On the other hand, reference [2] revealed an implication link from robust stability of FIE to that of MHE, and consequently identified rather general conditions for MHE to be robustly stable by inheriting conditions which ensure robust stability of the corresponding FIE. By making use of a Lipschitz continuity condition introduced in [2], reference [17] streamlined and generalized the analysis and results of [16], showing that the key is essentially to assume that the system satisfies a form of exponential detectability. When global exponential detectability is assumed, the MHE can further be shown to be robustly globally exponentially stable by implementing properly time-discounted stage costs [18].

The reviewed robust stabilities of FIE and MHE were concluded based on the concept defined in [11], which is however found to be flawed in that such defined robustly stable estimator does not imply convergence for convergent disturbances [19, 20]. This motivates a necessary modification of the stability concept in [19], which redefines the estimate error bound with a worst-case time-discounted instead of uniformly-weighted impact of the disturbances. The new concept is an enhancement of the old one, and was shown to be compatible with the original i-IOSS detectability of a system, which is necessary for establishing robust stability of both FIE and MHE. With this new concept, it becomes straightforward to understand the earlier robust local/global exponential stability results reported of FIE and MHE [16, 18]. It also motivates some new results as developed in [1, Ch. 4], which introduced a kind of stabilizability condition to establish robust stability of FIE and MHE. Despite conceptual elegance, the new condition involves an inequality which is uneasy to verify in general.

Motivated by the new and stronger concept of robust stability, this work aims to motivate general sufficient conditions for robust stability of MHE by firstly establishing a generic implication link from robust stability of FIE to that of MHE, and then transforming the challenge into identifying sufficient conditions for ensuring robust stability of FIE. While the reasoning approach is inspired by the ideas introduced in [2], the contributions of this work are three-fold:

  • A generic implication is established from robust stability of FIE to practical robust stability of the corresponding MHE, and the implication becomes stronger to robust stability of the MHE if the FIE admits a certain Lipschitz continuity property;

  • Explicit computations of sufficient MHE horizons are provided for the aforementioned implications to be effective in a local or global sense of robust stability. Analyses on two exemplary cases further illustrate that the upper bound of the MHE estimate error can eventually decrease with respect to (w.r.t.) the horizon size once it is large enough and satisfies certain conditions.

  • Given the revealed implications, new sufficient conditions are derived for robust stability of MHE by firstly developing those for the corresponding FIE. An interesting finding is that a system being i-IOSS is necessary but also sufficient for existence of a robustly stable FIE. Consequently is a similar but weaker conclusion applicable to MHE.

The remaining of this work is organized as follows. Sec. II introduces notation, setup and necessary preliminaries. Sec. III defines general forms of FIE and MHE, and introduces robust stability concepts. Sec. IV reveals the implication from global (or local) robust stability of FIE to that of the corresponding MHE. Sec. V applies the implication to establish robust stability of MHE, by firstly developing conditions for ensuring robust stability of the corresponding FIE. Finally, Sec. VI concludes the work.

II Notation, Setup and Preliminaries

The notation mostly follows the convention in [2, 1]. The symbols \mathbb{R}, 0\mathbb{R}_{\geq 0} and 𝕀0\mathbb{I}_{\geq 0} denote the sets of real numbers, nonnegative real numbers and nonnegative integers, respectively, and 𝕀a:b\mathbb{I}_{a:b} denotes the set of integers from aa to bb. The constraints t0t\geq 0 and t𝕀0t\in\mathbb{I}_{\geq 0} are used interchangeably to refer to the set of discrete times. The symbol ||\left|\cdot\right| denotes the Euclidean norm of a vector. The bold symbol 𝒙a:b\boldsymbol{x}_{a:b}, denotes a sequence of vector-valued variables (xa,xa+1,,xb)(x_{a},\,x_{a+1},\,...,\thinspace x_{b}), and with a function ff acting on a vector xx, f(𝒙a:b)f(\boldsymbol{x}_{a:b}) stands for the sequence of function values (f(xa),f(xa+1),,f(xb))(f(x_{a}),\,f(x_{a+1}),\,...,\thinspace f(x_{b})). Given a scalar function f:×f:\mathbb{R}\times\mathbb{R}\rightarrow\mathbb{R}, any tt\in\mathbb{R} and any scalar function gg, the notation fn(g(),t)f^{\circ n}(g(\cdot),t) refers to the nn fold composition of function gg by function ff subject to the given second argument tt, and f0f^{\circ 0} equals the identity function.

Throughout the paper, tt refers to a discrete time, and as a subscript it indicates dependence on time tt. Whereas, the subscripts or superscripts xx, ww and vv are used exclusively to indicate a function or variable that is associated with the state (xx), process disturbance (ww) or measurement noise (vv). The symbols x\left\lfloor x\right\rfloor and x\left\lceil x\right\rceil refer to the integers that are closest to xx from below and above, respectively. Given two scalars aa and bb, let ab:=max{a,b}a\oplus b:=\max\{a,\,b\}. The operator \oplus is both associative and commutative, i.e., (ab)c=a(bc)(a\oplus b)\oplus c=a\oplus(b\oplus c) and ab=baa\oplus b=b\oplus a, and furthermore is distributive with respect to (w.r.t.) increasing functions. That is, α(abc)=α(a)α(b)α(c)\alpha(a\oplus b\oplus c)=\alpha(a)\oplus\alpha(b)\oplus\alpha(c) if the function α\alpha is increasing in the argument. The frequently used 𝒦\mathcal{K}, \mathcal{L} and 𝒦\mathcal{KL} functions are defined as follows.

Definition 1.

(𝒦\mathcal{K}, \mathcal{L} and 𝒦\mathcal{KL} functions) A function α:00\alpha:\mathbb{R}_{\geq 0}\to\mathbb{R}_{\geq 0} is a 𝒦\mathcal{K} function if it is continuous, zero at zero, and strictly increasing. A function φ:00\varphi:\mathbb{R}_{\geq 0}\to\mathbb{R}_{\geq 0} is a \mathcal{L} function if it is continuous, nonincreasing and satisfies φ(t)0\varphi(t)\to 0 as tt\to\infty. A function β:0×00\beta:\mathbb{R}_{\geq 0}\times\mathbb{R}_{\geq 0}\to\mathbb{R}_{\geq 0} is a 𝒦\mathcal{KL} function if, for each t0t\geq 0, β(,t)\beta(\cdot,t) is a 𝒦\mathcal{K} function and for each s0s\geq 0, β(s,)\beta(s,\cdot) is a \mathcal{L} function.

Consider a discrete-time system described by

xt+1=f(xt,wt),yt=h(xt)+vt,x_{t+1}=f(x_{t},w_{t}),\,\,y_{t}=h(x_{t})+v_{t}, (1)

where xt𝕏nx_{t}\in\mathbb{X}\subseteq\mathbb{R}^{n} is the system state, wt𝕎gw_{t}\in\mathbb{W}\subseteq\mathbb{R}^{g} the process disturbance, yt𝕐py_{t}\in\mathbb{Y}\subseteq\mathbb{R}^{p} the measurement, vt𝕍pv_{t}\in\mathbb{V}\subseteq\mathbb{R}^{p} the measurement disturbance, all at time tt. Here we study state estimation as an independent subject, and so control inputs (if there were any) known up to the estimation time are treated as given constants, which do not cause difficulty to later defined optimization and related analyses and hence are neglected in the problem formulation for brevity [11, 1]. The functions ff and hh are assumed to be continuous and known. The initial state x0x_{0} and the disturbances (wt,vt)(w_{t},v_{t}) are assumed to be unknown but bounded.

Consider two state trajectories. Let

π0:=x0(1)x0(2),πτ+1:=wτ(1)wτ(2),πτ+t+1:=h(xτ(2))h(xτ(1)),\displaystyle\begin{gathered}\pi_{0}:=x_{0}^{(1)}-x_{0}^{(2)},\\ \pi_{\tau+1}:=w_{\tau}^{(1)}-w_{\tau}^{(2)},\,\,\pi_{\tau+t+1}:=h(x_{\tau}^{(2)})-h(x_{\tau}^{(1)}),\end{gathered} (4)

for all τ𝕀0:t1\tau\in\mathbb{I}_{0:t-1}, where πτ+t+1\pi_{\tau+t+1} corresponds to vτ(1)vτ(2)v_{\tau}^{(1)}-v_{\tau}^{(2)} if additive measurement noises are present while identical measurements are assumed. Hence 𝝅0:2t\boldsymbol{\pi}_{0:2t} collects a sequence of deviation vectors, and its domain is denoted as Π\Pi. Let ι(π)\iota(\pi_{\cdot}) extract the time index (i.e., the original index τ\tau above) of π\pi_{\cdot}. Specifically, we have

ι(πτ+1)=ι(πτ+t+1)=τ,τ𝕀1:t1.\iota(\pi_{\tau+1})=\iota(\pi_{\tau+t+1})=\tau,\,\,\forall\tau\in\mathbb{I}_{-1:t-1}. (5)

Note that the time index of π0\pi_{0} is defined as 1-1 which refers to the time associated with a priori information.

With the above notation, we can have a concise statement of the i-IOSS definition given in [1, 19].

Definition 2.

(Concise definition of i-IOSS) The system xt+1=f(xt,wt)x_{t+1}=f(x_{t},w_{t}), yt=h(xt)y_{t}=h(x_{t}) is i-IOSS if there exists α𝒦\alpha\in\mathcal{KL} such that for every two initial states x0(1),x0(2)x_{0}^{(1)},x_{0}^{(2)}, and two sequences of disturbances 𝒘0:t1(1),𝒘0:t1(2)\boldsymbol{w}_{0:t-1}^{(1)},\boldsymbol{w}_{0:t-1}^{(2)}, the following inequality holds for all t𝕀0t\in\mathbb{I}_{\geq 0}:

|xt(1)xt(2)|maxi𝕀0:2tα(|π|i,tι(πi)1),\left|x_{t}^{(1)}-x_{t}^{(2)}\right|\leq\max_{i\in\mathbb{I}_{0:2t}}\alpha\left(\left|\pi{}_{i}\right|,\,t-\iota(\pi_{i})-1\right), (6)

where xt(i)x_{t}^{(i)} is a shorthand of xt(x0(i),𝒘0:t1(i))x_{t}(x_{0}^{(i)},\boldsymbol{w}_{0:t-1}^{(i)}) for i{1, 2}i\in\{1,\,2\}. Furthermore, the system is exp-i-IOSS if in the above inequality α\alpha admits an exponential form as α(s,τ):=csλτ\alpha(s,\tau):=cs\lambda^{\tau} with certain λ(0,1)\lambda\in(0,1), c>0c>0 and for all s,τ0s,\tau\geq 0.

As proved in [19], this definition of i-IOSS is equivalent to the original one introduced in [12] which applies 𝒦\mathcal{K} instead of 𝒦\mathcal{KL} function to each uncertainty |πi||\pi_{i}| for all i1i\geq 1.

The following defines Lipschitz continuity of a function at a given point, and presents a related proposition.

Definition 3.

(Lipschitz continuity at a point) Given a subset 𝕊n\mathbb{S}\subseteq\mathbb{R}^{n}, a function f:nmf:\mathbb{R}^{n}\to\mathbb{R}^{m} is said to be Lipschitz continuous at a point x𝕊x^{*}\in\mathbb{S} over the subset 𝕊\mathbb{S} if there is a constant cc such that |f(x)f(x)|c|xx||f(x)-f(x^{*})|\leq c|x-x^{*}| for all x𝕊x\in\mathbb{S}.

Proposition 1.

(Properties related to Lipschitz continuity) Given any β𝒦\beta\in\mathcal{KL}, η(0,1)\eta\in(0,1), s¯>0\bar{s}>0 and s¯(0,s¯]\underline{s}\in(0,\,\bar{s}], define τmin=min{τ:β(s,τ)ηs,s[s¯,s¯],τ𝕀0}\tau_{\min}=\min\{\tau:\beta(s,\tau)\leq\eta s,\,s\in[\underline{s},\bar{s}],\,\tau\in\mathbb{I}_{\geq 0}\}. The following conclusions hold: a) the minimum τmin\tau_{\min} is well defined, i.e., it exists; b) the inequality β(s,τmin)η(ss¯)\beta(s,\tau_{\min})\leq\eta\cdot(s\oplus\underline{s}) holds for all s[0,s¯]s\in[0,\bar{s}]; and c) if β(,0)\beta(\cdot,0) is Lipschitz continuous at the origin over the range [0,s¯][0,\,\bar{s}] , then a) and b) hold also with s¯=0\underline{s}=0.

Proof:

a) With s¯>0\underline{s}>0, if we take τ¯:=min{τ:β(s¯,τ)ηs¯}\bar{\tau}:=\min\{\tau:\beta(\bar{s},\tau)\leq\eta\underline{s}\}, which certainly exists and has a finite value by the definition of a 𝒦\mathcal{KL} function, then we must have β(s,τ¯)β(s¯,τ¯)ηs¯ηs\beta(s,\bar{\tau})\leq\beta(\bar{s},\bar{\tau})\leq\eta\underline{s}\leq\eta s for all s[s¯,s¯]s\in[\underline{s},\bar{s}]. That is, τ¯\bar{\tau} is a feasible solution of τ\tau and hence also an upper bound for the considered minimization problem. Consequently, it is sufficient to consider the decision variable τ\tau in the finite set 𝕀0:τ¯\mathbb{I}_{0:\bar{\tau}}. By enumerating τ\tau in the set 𝕀0:τ¯\mathbb{I}_{0:\bar{\tau}}, the inequality β(s,τ)ηs\beta(s,\tau)\leq\eta s can be checked for all s[s¯,s¯]s\in[\underline{s},\bar{s}] subject to each given τ\tau. All valid values of τ\tau can then be collected in a set 𝒯\mathcal{T}. Since 𝒯\mathcal{T} is nonempty and finite, there must exist a minimum value, that is, τmin:=min{τ:τ𝒯}\tau_{\min}:=\min\{\tau:\tau\in\mathcal{T}\} is well defined. This completes the proof of a).

b) Since the inequality β(s,τmin)ηs=η(ss¯)\beta(s,\tau_{\min})\leq\eta s=\eta\cdot(s\oplus\underline{s}) holds for all s[s¯,s¯]s\in[\underline{s},\bar{s}] by the definition of τmin\tau_{\min}, it is remaining to show that it holds also for all s[0,s¯)s\in[0,\underline{s}). In the latter case, we have β(s,τmin)<β(s¯,τmin)ηs¯=η(ss¯)\beta(s,\tau_{\min})<\beta(\underline{s},\tau_{\min})\leq\eta\underline{s}=\eta\cdot(s\oplus\underline{s}), where the first inequality owes to the property of a 𝒦\mathcal{KL} function and the second results from a). This completes the proof of b).

c) Since β\beta is a 𝒦\mathcal{KL} function, β(,0)\beta(\cdot,0) being Lipschitz continuous at the origin over [0,s¯][0,\bar{s}] implies that β(s,τ)a(τ)s\beta(s,\tau)\leq a(\tau)s for all s[0,s¯]s\in[0,\bar{s}] and any τ𝕀0\tau\in\mathbb{I}_{\geq 0}, where a(τ)a(\tau) is the Lipschitz constant which is nonincreasing in τ\tau and furthermore goes to zero as τ\tau\rightarrow\infty, i.e., aa\in\mathcal{L}. Consequently, τmin=min{τ:a(τ)η,τ𝕀0}\tau_{\min}=\min\{\tau:a(\tau)\leq\eta,\,\tau\in\mathbb{I}_{\geq 0}\} is well defined, which implies that β(s,τmin)a(τmin)sηs\beta(s,\tau_{\min})\leq a(\tau_{\min})s\leq\eta s for all s[0,s¯]s\in[0,\bar{s}]. Therefore, in this case both a) and b) hold true for all s¯0\underline{s}\geq 0, which includes the origin. ∎

III Optimization-based State Estimation

Consider the system described in (1). Given a present time tt, the state estimation problem is to find an optimal estimate of state xtx_{t} based on historical measurements {yτ}\{y_{\tau}\} for τ\tau in a time set. Ideally, all measurements up to time tt are used, leading to the so-called FIE; and practically, only measurements are used within a limited distance backwards from time tt, yielding the so-called MHE. Both FIE and MHE can be cast as optimization problems.222Readers are referred to [11] for a brief introduction of their connection to control problems and difference from probabilistic formulations. To be concise, we will first define MHE and then treat FIE as a variant.

Let MHE implement a moving horizon of size TT. The decision variables are denoted as (𝝌tT:t,𝝎tT:t1,𝝂tT:t1)(\boldsymbol{\chi}_{t-T:t},\boldsymbol{\omega}_{t-T:t-1},\boldsymbol{\nu}_{t-T:t-1}), which correspond to the system variables (𝒙tT:t,𝒘tT:t1,𝒗tT:t1)(\boldsymbol{x}_{t-T:t},\boldsymbol{w}_{t-T:t-1},\boldsymbol{v}_{t-T:t-1}).333As in [1], the last measurement is not considered for ease of presentation, though the inclusion does not change the conclusions. And let the optimal decision variables be (𝒙^tT:t,𝒘^tT:t1,𝒗^tT:t1)(\hat{\boldsymbol{x}}_{t-T:t},\hat{\boldsymbol{w}}_{t-T:t-1},\hat{\boldsymbol{v}}_{t-T:t-1}). Since 𝒙^tT+1:t\hat{\boldsymbol{x}}_{t-T+1:t} is uniquely determined from x^tT\hat{x}_{t-T} and 𝒘^tT:t1\hat{\boldsymbol{w}}_{t-T:t-1}, the decision variables essentially reduce to (χtT,𝝎tT:t1,𝝂tT:t1)(\chi_{t-T},\boldsymbol{\omega}_{t-T:t-1},\boldsymbol{\nu}_{t-T:t-1}).

In addition, let x¯tT\bar{x}_{t-T} be a priori estimate of xtTx_{t-T}, and in particular x¯0\bar{x}_{0} be bounded. Without loss of generality, the prior estimates of the disturbances are assumed to be zero. Denote the cost function as VT(χtTx¯tT,𝝎tT:t1,𝝂tT:t1)V_{T}(\chi_{t-T}-\bar{x}_{t-T},\boldsymbol{\omega}_{t-T:t-1},\boldsymbol{\nu}_{t-T:t-1}), which penalizes uncertainties in the initial state, the process and the measurements. Then, the MHE instance at time tt is defined by the following optimization problem:

MHE (or FIEif Tt):\displaystyle\begin{array}[]{c}\text{MHE (or FIE}\\ \text{if }T\leftarrow t):\end{array} minVT(χtTx¯tT,𝝎tT:t1,𝝂tT:t1)\displaystyle\min V_{T}(\chi_{t-T}-\bar{x}_{t-T},\boldsymbol{\omega}_{t-T:t-1},\boldsymbol{\nu}_{t-T:t-1}) (7)
s.t., χτ+1=f(χτ,ωτ),τ𝕀tT:t1,\displaystyle\chi_{\tau+1}=f(\chi_{\tau},\omega_{\tau}),\,\,\forall\tau\in\mathbb{I}_{t-T:t-1},
yτ=h(χτ)+ντ,τ𝕀tT:t1,\displaystyle y_{\tau}=h(\chi_{\tau})+\nu_{\tau},\,\,\forall\tau\in\mathbb{I}_{t-T:t-1},
χtT𝕏,𝝎tT:t1𝕎T,𝝂tT:t1𝕍T.\displaystyle\chi_{t-T}\in\mathbb{X},\,\boldsymbol{\omega}_{t-T:t-1}\in\mathbb{W}^{T},\,\boldsymbol{\nu}_{t-T:t-1}\in\mathbb{V}^{T}.

As 𝝂tT:t1\boldsymbol{\nu}_{t-T:t-1} is uniquely determined by χtT\chi_{t-T} and 𝝎tT:t1\boldsymbol{\omega}_{t-T:t-1}, it is kept mainly for the convenience of expressing the disturbance set and the objective function. Since the global optimal solution x^τ\hat{x}_{\tau}, for any τt\tau\leq t, is dependent on time tt when the MHE instance is defined, to be unambiguous we use x^t\hat{x}_{t}^{\star} to represent x^t\hat{x}_{t} that is solved from the instance defined at time tt. This keeps x^t\hat{x}_{t}^{\star} unchanged, while the realization x^t\hat{x}_{t} varies as MHE renews itself in time.

To define FIE, it suffices to adapt the horizon size TT in the MHE formulation to taking the time-varying value tt. The yielded FIE has the form of an MHE but with complete data originating from the zero initial time. To link them easily, an FIE is called the corresponding FIE of an MHE based on which the FIE is derived, and conversely is the MHE called a corresponding MHE of the FIE. Since it becomes computationally intractable as time elapses, FIE is studied mainly for its theoretical interest: its performance is viewed as a limit or benchmark that MHE attempts to approach, and its stability can be a good start point for analysis of MHE.

An important issue in designing FIE or MHE is to identify conditions under which the associated optimization admits optimal estimates that satisfy a robust stability property defined below. Let 𝒙0:t(x0,𝒘0:t1)\boldsymbol{x}_{0:t}(x_{0},\boldsymbol{w}_{0:t-1}) denote a state sequence generated from an initial state x0x_{0}, and a disturbance sequence 𝒘0:t1\boldsymbol{w}_{0:t-1}. In addition, define a bounded set for any given δ0>0\delta_{0}>0:

𝕏δ0:={(x1,x2):|x1x2|δ0,x1,x2𝕏}.\mathbb{X}_{\delta_{0}}:=\{(x_{1},\,x_{2}):|x_{1}-x_{2}|\leq\delta_{0},\,x_{1},x_{2}\in\mathbb{X}\}. (8)
Definition 4.

(Robust stable estimation) The estimate x^t\hat{x}_{t} of state xtx_{t} is based on partial or full sequence of the noisy measurements, 𝒚0:t=h(𝒙0:t(x0,𝒘0:t1))+𝒗0:t\boldsymbol{y}_{0:t}=h(\boldsymbol{x}_{0:t}(x_{0},\boldsymbol{w}_{0:t-1}))+\boldsymbol{v}_{0:t}. The estimate is robustly asymptotically stable (RAS) if given any δ0>0\delta_{0}>0, there exist functions βx,βw,βv𝒦\beta_{x},\,\beta_{w},\,\beta_{v}\in\mathcal{KL} such that the following inequality holds for all (x0,x¯0)𝕏δ0(x_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}}, 𝝎0:t1𝕎t\boldsymbol{\omega}_{0:t-1}\in\mathbb{W}^{t}, 𝝂0:t1𝕍t\boldsymbol{\nu}_{0:t-1}\in\mathbb{V}^{t} and t𝕀0t\in\mathbb{I}_{\geq 0}:

|xtx^t|\displaystyle\left|x_{t}-\hat{x}_{t}\right| βx(|x0x¯0|,t)\displaystyle\leq\beta_{x}(\left|x_{0}-\bar{x}_{0}\right|,t)
max{w,v}maxτ𝕀0:t1β(|τ|,tτ1).\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\beta_{*}(\left|*_{\tau}\right|,\,t-\tau-1). (9)

The estimate is further said to be robustly globally asymptotically stable (RGAS) if the inequality is satisfied for all x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}. Moreover, if the estimate is RAS (or RGAS) and the 𝒦\mathcal{KL} function admits an exponential form as β(s,τ):=csλt\beta_{*}(s,\tau):=c_{*}s\lambda^{t} with certain λ(0,1)\lambda\in(0,1), c>0c_{*}>0 and for all {x,w,v}*\in\{x,w,v\}, then the estimate is said to be robustly exponentially stable (RES) (or robustly globally exponentially stable (RGES)).

The last measurement yty_{t} and hence the corresponding noise vtv_{t} is not considered in the above inequality, to keep the definition consistent with the formulations of FIE and MHE. Here the definition of RGAS strengthens the one introduced in [11] which applies 𝒦\mathcal{K} instead of 𝒦\mathcal{KL} functions to the disturbances. This change is necessary to enable a desirable feature that a state estimator which is RGAS must be convergent under convergent disturbances [1, 19]. The next definition presents a weaker alternative of the above robust stability, which will also be needed in later analysis.

Definition 5.

(Practical robust stable estimations) The estimate defined in Definition 4 is practically RAS (pRAS) if given any ϵ,δ0>0\epsilon,\delta_{0}>0, there exist functions βx,βw,βv𝒦\beta_{x},\,\beta_{w},\,\beta_{v}\in\mathcal{KL} such that the following inequality holds for all (x0,x¯0)𝕏δ0(x_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}}, 𝝎0:t1𝕎t\boldsymbol{\omega}_{0:t-1}\in\mathbb{W}^{t}, 𝝂0:t1𝕍t\boldsymbol{\nu}_{0:t-1}\in\mathbb{V}^{t} and t𝕀0t\in\mathbb{I}_{\geq 0}:

|xtx^t|\displaystyle\left|x_{t}-\hat{x}_{t}\right| ϵβx(|x0x¯0|,t)\displaystyle\leq\epsilon\oplus\beta_{x}(\left|x_{0}-\bar{x}_{0}\right|,t)
max{w,v}maxτ𝕀0:t1β(|τ|,tτ1).\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\beta_{*}(\left|*_{\tau}\right|,\,t-\tau-1). (10)

The estimate is further said to be practically RGAS (pRGAS) if the inequality is satisfied for all x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}.

The adverb “practically” before RGAS is employed to keep it in line with the practical stability concept developed in control literature (e.g., [21]). Compared to an RGAS estimate, the pRGAS estimate admits a looser bound with a non-vanishing constant term ϵ\epsilon. As will be shown later, this term can be made arbitrarily small if MHE implements a long enough horizon. In addition, it is worthwhile to remark that there is no need to define such stability variant for the case of RES or RGES because in either case the exponential property ensures that the ϵ\epsilon term will be fully tempered and removed.

By applying the notation used in the concise definition of i-IOSS in (6), conciser forms of the above stability definitions can also be obtained, which will be useful to convey stability conditions for both FIE and MHE later on. Towards that, let x0(1):=x0x_{0}^{(1)}:=x_{0}, wτ(1):=wτw_{\tau}^{(1)}:=w_{\tau}, h(xτ(1)):=yτvτh(x_{\tau}^{(1)}):=y_{\tau}-v_{\tau}, x0(2):=x¯0x_{0}^{(2)}:=\bar{x}_{0}, wτ(2):=0w_{\tau}^{(2)}:=0, and h(xτ(2)):=yτh(x_{\tau}^{(2)}):=y_{\tau} for all τ𝕀0:t1\tau\in\mathbb{I}_{0:t-1}. Consequently, the notation 𝝅0:2t\boldsymbol{\pi}_{0:2t} defined in (4) has a new realization, and its corresponding domain is denoted as Π\Pi if χ0,x¯0𝕏\chi_{0},\,\bar{x}_{0}\in\mathbb{X} or Πδ0\Pi_{\delta_{0}} if (χ0,x¯0)𝕏δ0(\chi_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}}.

With the new notation, it is straightforward to prove the following equivalent definition of robust stabilities.

Lemma 1.

(Concise equivalent definition of robust stable estimations) The estimate defined in Definition 4 is RAS if and only if given any δ0>0\delta_{0}>0, there exists a function β𝒦\beta\in\mathcal{KL} such that the following inequality holds for all 𝛑0:2tΠδ0\boldsymbol{\pi}_{0:2t}\in\Pi_{\delta_{0}} and t𝕀0t\in\mathbb{I}_{\geq 0}:

|xtx^t|maxi𝕀0:2tβ(|πi|,tι(πi)1).\left|x_{t}-\hat{x}_{t}\right|\leq\max_{i\in\mathbb{I}_{0:2t}}\beta(\left|\pi_{i}\right|,t-\iota(\pi_{i})-1). (11)

And it is RGAS if and only if the above inequality holds for all 𝛑0:2tΠ\boldsymbol{\pi}_{0:2t}\in\Pi. The estimate being RAS (or RGAS) is further said to be RES (or RGES), if and only if the 𝒦\mathcal{KL} function β\beta admits the form as β(s,τ):=csλτ\beta(s,\tau):=cs\lambda^{\tau} with certain c>0c>0, λ(0,1)\lambda\in(0,1) and for all τ0\tau\geq 0 and ss in the corresponding domain.

Similar concise definitions of pRAS and pRGAS can both be obtained by adding a constant ϵ>0\epsilon>0 to the right hand side of (11), in which the 𝒦\mathcal{KL} function β\beta will then be dependent on ϵ\epsilon. For brevity, we do not state them with another lemma.

IV Stability Implication from FIE to MHE

At any discrete time, an MHE instance can be interpreted as the corresponding FIE initiating from the start of the horizon over which the MHE instance is defined. Thus, the corresponding FIE being robustly stable implies that each MHE instance is robustly stable within the time horizon over which the instance is defined. If we interpret this as MHE being instance-wise robustly stable, then the challenge reduces to identifying conditions under which instance-wise robust stability implies robust stability of MHE. This observation was made in [2], and the challenge was solved there for a weaker definition of RGAS. This section resolves the challenge subject to the stronger stability concept given by Definition 4.

To that end, as in [2], we apply an ordinary assumption on the prior estimate x¯tT\bar{x}_{t-T} of the initial state xtTx_{t-T} of an MHE instance.

Assumption 1.

Given any time tT+1t\geq T+1, the prior estimate x¯tT\bar{x}_{t-T} of xtTx_{t-T} is given such that

|xtTx¯tT||xtTx^tT|.|x_{t-T}-\bar{x}_{t-T}|\leq|x_{t-T}-\hat{x}_{t-T}^{\star}|.

The assumption is obviously true if x¯tT\bar{x}_{t-T} is set to x^tT\hat{x}_{t-T}^{\star}, which is the MHE estimate obtained at time tTt-T. Alternatively, a better x¯tT\bar{x}_{t-T} might be obtained with smoothing techniques which use measurements both before and after time tTt-T [22, 1].

Next, we present an important lemma which links global robust stability of MHE with that of its corresponding FIE.

Lemma 2.

(Global stability implication from FIE to MHE) Consider MHE under Assumption 1. The following two conclusions hold:

a) (RGAS–>pRGAS/RGAS/RGES) If FIE is RGAS as per (9), then there exists T¯𝕀0\underline{T}\in\mathbb{I}_{\geq 0} such that the corresponding MHE under Assumption 1 is pRGAS for all TT¯T\geq\text{$\underline{T}$}. If further the 𝒦\mathcal{K} function βx(,0)\beta_{x}(\cdot,0) in (9) is globally Lipschitz continuous at the origin, then the implication strengthens to the existence of T¯𝕀0\underline{T}^{\prime}\in\mathbb{I}_{\geq 0} such that the MHE is RGAS for all TT¯T\geq\text{$\underline{T}$}^{\prime}; and if furthermore the 𝒦\mathcal{K} functions βw(,0)\beta_{w}(\cdot,0) and βv(,0)\beta_{v}(\cdot,0) are also globally Lipschitz continuous at the origin, then the MHE is RGES for all TT¯T\geq\text{$\underline{T}$}^{\prime}.

b) (RGES–>RGES) If FIE is RGES as per (9), then there exists T¯𝕀0\underline{T}\in\mathbb{I}_{\geq 0} such that the corresponding MHE is RGES for all TT¯T\geq\text{$\underline{T}$}. In particular, with the 𝒦\mathcal{KL} function βx(s,t):=cxsλt\beta_{x}(s,t):=c_{x}s\lambda^{t} in (9) for certain cx>0c_{x}>0 and λ(0,1)\lambda\in(0,1), it is feasible to set T¯=1(logλ1cx+1)\underline{T}=1\oplus(\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor+1).

Proof:

a) RGAS–>pRGAS. Let n:=tTn:=\left\lfloor\frac{t}{T}\right\rfloor, and so 0tnTT10\leq t-nT\leq T-1. For all τ𝕀0:T1\tau\in\mathbb{I}_{0:T-1}, the MHE and the corresponding FIE estimates are the same, both denoted as x^τ\hat{x}_{\tau}^{\star}. So, given any τ𝕀0:tnT\tau\in\mathbb{I}_{0:t-nT}, the absolute estimation error |xτx^τ|\left|x_{\tau}-\hat{x}_{\tau}^{\star}\right| satisfies the RGAS inequality given by (9). That is, we have

|xτx^τ|βx(|x0x¯0|,τ)\displaystyle\left|x_{\tau}-\hat{x}_{\tau}^{\star}\right|\leq\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace\tau)
max{w,v}maxτ𝕀0:τ1β(|τ|,ττ1),\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau^{\prime}\in\mathbb{I}_{0:\tau-1}}\beta_{*}(\left|*_{\tau^{\prime}}\right|,\,\tau-\tau^{\prime}-1), (12)

for all τ𝕀0:tnT\tau\in\mathbb{I}_{0:t-nT}. Next, we proceed to show that the RGAS property is maintained for all τ𝕀tnT+1:t\tau\in\mathbb{I}_{t-nT+1:t}.

Let t1:=0t_{-1}:=0 and ti:=t(ni)Tt_{i}:=t-(n-i)T for all i𝕀0:ni\in\mathbb{I}_{0:n}. The MHE instance at time t1t_{1} can be viewed as the corresponding FIE confined to time interval [t0,t1][t_{0},\thinspace t_{1}]. Thus, the MHE satisfies the RGAS property within this interval. That is, by (9) we have:

|xt1x^t1|βx(|xt0x^t0|,T)\displaystyle\left|x_{t_{1}}-\hat{x}_{t_{1}}^{\star}\right|\leq\beta_{x}(\left|x_{t_{0}}-\hat{x}_{t_{0}}^{\star}\right|,\thinspace T)
max{w,v}maxτ𝕀t0:t11β(|τ|,t1τ1),\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{t_{0}:t_{1}-1}}\beta_{*}(\left|*_{\tau}\right|,\,t_{1}-\tau-1),

where Assumption 1 has been applied to produce the first term of the right hand side of the inequality. Repeat this reasoning for the MHE instance defined at time t2t_{2} and then apply the above inequality, yielding

|xt2x^t2|βx(|xt1x^t1|,T)\displaystyle\left|x_{t_{2}}-\hat{x}_{t_{2}}^{\star}\right|\leq\beta_{x}(\left|x_{t_{1}}-\hat{x}_{t_{1}}^{\star}\right|,\thinspace T)
max{w,v}maxτ𝕀t1:t21β(|τ|,t2τ1)\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{t_{1}:t_{2}-1}}\beta_{*}(\left|*_{\tau}\right|,\,t_{2}-\tau-1)
βx2(|xt0x^t0|,T)\displaystyle\leq\beta_{x}^{\circ 2}(\left|x_{t_{0}}-\hat{x}_{t_{0}}^{\star}\right|,\thinspace T)
max{w,v}maxτ𝕀t0:t11βx(β(|τ|,t1τ1),T)\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{t_{0}:t_{1}-1}}\beta_{x}\left(\beta_{*}(\left|*_{\tau}\right|,\,t_{1}-\tau-1),\,T\right)
max{w,v}maxτ𝕀t1:t21β(|τ|,t2τ1).\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{t_{1}:t_{2}-1}}\beta_{*}(\left|*_{\tau}\right|,\,t_{2}-\tau-1).

By induction, we obtain

|xtx^t|=\displaystyle\left|x_{t}-\hat{x}_{t}^{\star}\right|= |xtnx^tn|βxn(|xt0x^t0|,T)\displaystyle\left|x_{t_{n}}-\hat{x}_{t_{n}}^{\star}\right|\leq\beta_{x}^{\circ n}(\left|x_{t_{0}}-\hat{x}_{t_{0}}^{\star}\right|,\thinspace T)
max{w,v}maxi𝕀0:n1maxτ𝕀tni1:tni1\displaystyle\oplus\max_{*\in\{w,v\}}\max_{i\in\mathbb{I}_{0:n-1}}\max_{\tau\in\mathbb{I}_{t_{n-i-1}:t_{n-i}-1}}
βxi(β(|τ|,tniτ1),T).\displaystyle\quad\beta_{x}^{\circ i}\left(\beta_{*}(\left|*_{\tau}\right|,\,t_{n-i}-\tau-1),\,T\right).

Since |xt0x^t0|\left|x_{t_{0}}-\hat{x}_{t_{0}}^{\star}\right| satisfies inequality (12), this implies that

|xtx^t|\displaystyle\left|x_{t}-\hat{x}_{t}^{\star}\right|\leq βxn(βx(|x0x¯0|,t0),T)\displaystyle\beta_{x}^{\circ n}\left(\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t_{0}),\,T\right)
max{w,v}maxi𝕀0:nmaxτ𝕀tni1:tni1\displaystyle\oplus\max_{*\in\{w,v\}}\max_{i\in\mathbb{I}_{0:n}}\max_{\tau\in\mathbb{I}_{t_{n-i-1}:t_{n-i}-1}}
βxi(β(|τ|,tniτ1),T)\displaystyle\quad\beta_{x}^{\circ i}\left(\beta_{*}(\left|*_{\tau}\right|,\,t_{n-i}-\tau-1),\,T\right)
=\displaystyle= βxtT(βx(|x0x¯0|,ttTT),T)\displaystyle\beta_{x}^{\circ\left\lfloor\frac{t}{T}\right\rfloor}\left(\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t-\left\lfloor\frac{t}{T}\right\rfloor T),\,T\right)
max{w,v}maxτ𝕀0:t1\displaystyle\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}
βxtτT\displaystyle\beta_{x}^{\circ\left\lfloor\frac{t-\tau}{T}\right\rfloor} (β(|τ|,ttτTTτ1),T)\displaystyle\left(\beta_{*}(\left|*_{\tau}\right|,\,t-\left\lfloor\frac{t-\tau}{T}\right\rfloor T-\tau-1),\,T\right) (13)
=:\displaystyle=: βx,T(|x0x¯0|,t)\displaystyle\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t)
max{w,v}maxτ𝕀0:t1β,T(|τ|,tτ1),\displaystyle\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\beta_{*,T}(\left|*_{\tau}\right|,\,t-\tau-1),

where the relation that tτT=i\left\lfloor\frac{t-\tau}{T}\right\rfloor=i for all τ𝕀tni1:tni\tau\in\mathbb{I}_{t_{n-i-1}:t_{n-i}} and i𝕀0:ni\in\mathbb{I}_{0:n} has been used to establish the first equality, and the three functions βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are defined by one-to-one correspondence to the three preceding terms, each of which is dependent on the horizon size TT. By definition, it is obvious that βx,T(,τ)\beta_{x,T}(\cdot,\tau), βw,T(,τ)\beta_{w,T}(\cdot,\tau) and βv,T(,τ)\beta_{v,T}(\cdot,\tau) are 𝒦\mathcal{K} functions for any given τ0\tau\geq 0. However, it is not necessary that βx,T(s,)\beta_{x,T}(s,\cdot), βw,T(s,)\beta_{w,T}(s,\cdot) and βv,T(s,)\beta_{v,T}(s,\cdot) are \mathcal{L} functions for all s,T0s,T\geq 0, and so βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are not necessarily 𝒦\mathcal{KL} functions. The remaining proof is to show that there exists T¯\underline{T} such that βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are bounded by 𝒦\mathcal{KL} functions plus a positive constant for all TT¯T\geq\underline{T}.

As TT increases from zero to infinity, the function instances form a sequence 𝜷,0:\boldsymbol{\beta}_{*,0:\infty}, for each {x,w,v}*\in\{x,\,w,\,v\}. By the definition of β,T\beta_{*,T}, given any ss and τ\tau, we have limTβ,T(s,τ)=β(s,τ)\lim_{T\rightarrow\infty}\beta_{*,T}(s,\tau)=\beta_{*}(s,\tau) for each {x,w,v}*\in\{x,\,w,\,v\}. That is, β,T\beta_{*,T} point-wisely converges to β\beta_{*} as TT\rightarrow\infty. Next, we prove that the convergence is uniform w.r.t. the arguments ss and τ\tau. Since |x0x¯0||x_{0}-\bar{x}_{0}| is assumed to be bounded, there exists M>0M>0 such that |x0x¯0|M|x_{0}-\bar{x}_{0}|\leq M. Given any 0<ζ<2βx(M,0)0<\zeta<2\beta_{x}(M,0), there must exist TT^{\prime} satisfying βx(Mβx(M, 0),T)<ζ2\beta_{x}(M\oplus\beta_{x}(M,\thinspace 0),\thinspace T^{\prime})<\frac{\zeta}{2}, such that for all TTT\geq T^{\prime} we will have for any t<Tt<T,

supx0,x¯0,t|βx,T(|x0x¯0|,t)βx(|x0x¯0|,t)|=0<ζ,\sup_{x_{0},\bar{x}_{0},t}\left|\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t)-\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t)\right|=0<\zeta,

and for any tTt\geq T,

supx0,x¯0,t|βx,T(|x0x¯0|,t)βx(|x0x¯0|,t)|\displaystyle\sup_{x_{0},\bar{x}_{0},t}\left|\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t)-\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t)\right|
βx,T(M,t)+βx(M,t)<βxtT(βx(M, 0),T)+ζ2\displaystyle\leq\beta_{x,T}(M,\thinspace t)+\beta_{x}(M,\thinspace t)<\beta_{x}^{\circ\left\lfloor\frac{t}{T}\right\rfloor}\left(\beta_{x}(M,\thinspace 0),\,T\right)+\frac{\zeta}{2}
<βxmax{0,tT1}(ζ2,T)+ζ2ζ.\displaystyle<\beta_{x}^{\circ\max\{0,\,\left\lfloor\frac{t}{T}\right\rfloor-1\}}\left(\frac{\zeta}{2},\,T\right)+\frac{\zeta}{2}\leq\zeta.

That is, supx0,x¯0,t|βx,T(|x0x¯0|,t)βx(|x0x¯0|,t)|<ζ\sup_{x_{0},\bar{x}_{0},t}\left|\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t)-\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t)\right|<\zeta holds true for all TTT\geq T^{\prime}. This implies that βx,T\beta_{x,T} uniformly converges to βx\beta_{x}. So can βw,T\beta_{w,T} and βv,T\beta_{v,T} be shown in the same way to uniformly converge to βw\beta_{w} and βv\beta_{v}, respectively.444If any of the uncertainties does not have a certain bound, then the valid TT^{\prime} will depend on the actual uncertainty magnitude and consequently only semi-uniformly convergence will be proved.

Consequently, given any ϵ>0\epsilon>0, there exists T¯>0\underline{T}>0 such that βx,T(|x0x¯0|,t)<βx(|x0x¯0|,t)+ϵ/2\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t)<\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t)+\epsilon/2 and β,T(|τ|,tτ1)<β(|τ|,tτ1)+ϵ/2\beta_{*,T}(\left|*_{\tau}\right|,\,t-\tau-1)<\beta_{*}(\left|*_{\tau}\right|,\,t-\tau-1)+\epsilon/2 for all {w,v}*\in\{w,\,v\}, TT¯T\geq\underline{T}, x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}, wτ𝕎w_{\tau}\in\mathbb{W}, vτ𝕍v_{\tau}\in\mathbb{V}, τ𝕀0:t1\tau\in\mathbb{I}_{0:t-1} and t𝕀0t\in\mathbb{I}_{\geq 0}. Given each {x,w,v}*\in\{x,w,v\}, let

β(s,τ):={β(s,τ)+ϵ/2if β(s,τ)>ϵ/2,β(s,τ)if β(s,τ)ϵ/2,\beta_{*}^{\prime}(s,\tau):=\begin{cases}\beta_{*}(s,\tau)+\epsilon/2&\text{if }\beta_{*}(s,\tau)>\epsilon/2,\\ \beta_{*}(s,\tau)&\text{if }\beta_{*}(s,\tau)\leq\epsilon/2,\end{cases}

for all s,τ0s,\tau\geq 0. Given any {x,w,v}*\in\{x,w,v\}, it is easy to verify that β\beta_{*}^{\prime} is a 𝒦\mathcal{KL} function as β\beta_{*} is, and so for any TT¯T\geq\underline{T} we will have β,T(s,τ)β(s,τ)+ϵ/2ϵβ(s,τ)\beta_{*,T}(s,\tau)\leq\beta_{*}(s,\tau)+\epsilon/2\leq\epsilon\oplus\beta_{*}^{\prime}(s,\tau) for all s,τs,\tau in aforementioned domains. By applying these inequalities to the last inequality of |xtx^t|\left|x_{t}-\hat{x}_{t}^{\star}\right| above, we conclude that given any ϵ>0\epsilon>0, there exists T¯>0\underline{T}>0 such that

|xtx^t|\displaystyle\left|x_{t}-\hat{x}_{t}^{\star}\right| ϵβx(|x0x¯0|,t)\displaystyle\leq\epsilon\oplus\beta_{x}^{\prime}(|x_{0}-\bar{x}_{0}|,\thinspace t)
max{w,v}maxτ𝕀0:t1β(|τ|,tτ1)\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\begin{array}[]{c}\beta_{*}^{\prime}(\left|*_{\tau}\right|,\,t-\tau-1)\end{array}

for all TT¯T\geq\underline{T}, where β𝒦\beta_{*}^{\prime}\in\mathcal{KL} for all {x,w,v}*\in\{x,w,v\}. This implies that the MHE is pRGAS by definition, and hence completes proof of the first conclusion of part a).

RGAS–>RGAS. If further βx(,0)\beta_{x}(\cdot,0) from (9) is globally Lipschitz continuous at the origin, then for any τ0\tau\geq 0 there must exist a \mathcal{L} function μx\mu_{x} such that βx(s,τ)μx(τ)s\beta_{x}(s,\tau)\leq\mu_{x}(\tau)s for all s𝕏s\in\mathbb{X} because βx\beta_{x}\in\mathcal{L}. By applying this property to (13), for all TT¯T\geq\text{$\underline{T}$}^{\prime} satisfying μx(T¯)η\mu_{x}(\text{$\underline{T}$}^{\prime})\leq\eta with certain η(0,1)\eta\in(0,1), the inequality there proceeds as

|xtx^t|\displaystyle\left|x_{t}-\hat{x}_{t}^{\star}\right| μxtT(T)μ(0)|x0x¯0|max{w,v}maxτ𝕀0:t1\displaystyle\leq\mu_{x}^{\left\lfloor\frac{t}{T}\right\rfloor}(T)\mu(0)|x_{0}-\bar{x}_{0}|\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}
μxtτT(T)β(|τ|,ttτTTτ1)\displaystyle\quad\mu_{x}^{\left\lfloor\frac{t-\tau}{T}\right\rfloor}(T)\beta_{*}(\left|*_{\tau}\right|,\,t-\left\lfloor\frac{t-\tau}{T}\right\rfloor T-\tau-1)
ηtT1μx(0)|x0x¯0|\displaystyle\leq\eta^{\frac{t}{T}-1}\mu_{x}(0)|x_{0}-\bar{x}_{0}|
max{w,v}maxτ𝕀0:t1ηtτT1β(|τ|, 0).\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\eta^{\frac{t-\tau}{T}-1}\beta_{*}(\left|*_{\tau}\right|,\,0). (14)

Consequently, the MHE is RGAS by definition, which completes the proof of the second conclusion of part a).

RGAS–>RGES. If in addition to βx(,0)\beta_{x}(\cdot,0), the 𝒦\mathcal{K} functions βw(,0)\beta_{w}(\cdot,0) and βv(,0)\beta_{v}(\cdot,0) from (9) are globally Lipschitz continuous at the origin, then inequality (14) proceeds as

|xtx^t|\displaystyle\left|x_{t}-\hat{x}_{t}^{\star}\right| ηtT1μx(0)|x0x¯0|\displaystyle\leq\eta^{\frac{t}{T}-1}\mu_{x}(0)|x_{0}-\bar{x}_{0}|
max{w,v}maxτ𝕀0:t1ηtτT1μ,0|τ|,\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\eta^{\frac{t-\tau}{T}-1}\mu_{*,0}\left|*_{\tau}\right|,

where μ,0\mu_{*,0} is the Lipschitz constant of β(,0)\beta_{*}(\cdot,0) at the origin for all {w,v}*\in\{w,v\}. Consequently, the MHE is RGES by definition, which completes the third conclusion of part a).

b) RGES–>RGES. In this case, we have β(s,τ)=csλτ\beta_{*}(s,\tau)=c_{*}s\lambda^{\tau} with certain c>0c_{*}>0 and λ(0,1)\lambda\in(0,1) for each {x,w,v}*\in\{x,w,v\}. Since RGES implies RGAS, the deduction in the proof of part a) is applicable. In particular, the three functions βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are derived explicitly as:

βx,T(|x0x¯0|,t)\displaystyle\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t) =cxtT+1|x0x¯0|λt,\displaystyle=c_{x}^{\left\lfloor\frac{t}{T}\right\rfloor+1}|x_{0}-\bar{x}_{0}|\lambda^{t}, (15)
βw,T(|wτ|,tτ1)\displaystyle\beta_{w,T}(\left|w_{\tau}\right|,\,t-\tau-1) =cxtτTcw|wτ|λtτ1,\displaystyle=c_{x}^{\left\lfloor\frac{t-\tau}{T}\right\rfloor}c_{w}\left|w_{\tau}\right|\lambda^{t-\tau-1},
βv,T(|vτ|,tτ1)\displaystyle\beta_{v,T}(\left|v_{\tau}\right|,\,t-\tau-1) =cxtτTcv|vτ|λtτ1,\displaystyle=c_{x}^{\left\lfloor\frac{t-\tau}{T}\right\rfloor}c_{v}\left|v_{\tau}\right|\lambda^{t-\tau-1},

for all x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}, wτ𝕎w_{\tau}\in\mathbb{W}, vτ𝕍v_{\tau}\in\mathbb{V}, τ𝕀0:t1\tau\in\mathbb{I}_{0:t-1} and t𝕀0t\in\mathbb{I}_{\geq 0}. If cx1c_{x}\leq 1, it is obvious that βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are all upper bounded by 𝒦\mathcal{KL} functions in exponential forms, which implies RGES of the MHE for all TT¯:=11+logλ1cxT\geq\underline{T}:=1\geq 1+\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor. And if cx>1c_{x}>1, by defining T¯=1+logλ1cx\underline{T}=1+\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor (1\geq 1) we have cxtTλtcxtTλt(cx1T¯λ)t((cx)1logλ1cxλ)t=1c_{x}^{\left\lfloor\frac{t}{T}\right\rfloor}\lambda^{t}\leq c_{x}^{\frac{t}{T}}\lambda^{t}\leq\left(c_{x}^{\frac{1}{\underline{T}}}\lambda\right)^{t}\leq\left(\left(c_{x}\right)^{\frac{1}{\log_{\lambda}\frac{1}{c_{x}}}}\lambda\right)^{t}=1 for all TT¯T\geq\underline{T}, where the penultimate equality holds only for t=0t=0. Consequently βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are again all upper bounded by 𝒦\mathcal{KL} functions in exponential forms which imply RGES of the MHE. Combining the two cases, we conclude that the MHE is RGES for all TT¯:=1(logλ1cx+1)T\geq\underline{T}:=1\oplus(\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor+1), which completes the proof of b). ∎

Lemma 2.a) indicates that RGAS of FIE implies pRGAS of the corresponding MHE which implements a long enough horizon, and that the implication strengthens to RGAS or RGES of the MHE if the FIE additionally satisfies certain global Lipschitz continuity conditions. In the latter case, it is not certain that the MHE will be RGES unless the three bound functions {βx\beta_{x}, βw\beta_{w}, βv\beta_{v}} of the FIE all satisfy the global Lipschitz conditions. Also note that the main conclusion of Lemma 2.b) is implied by the third conclusion of Lemma 2.a) and that its proof above is presented mainly for derivation of an explicit bound of the sufficient horizon size.

Remark 1.

In the second case of Lemma 2, an FIE being RGAS and its bound function βx(,0)\beta_{x}(\cdot,0) being globally Lipschitz at the origin imply that the system is exp-i-IOSS. This can be proved by extending the proof for the local case in [17, Proposition 2] to the global case. However, the two conditions do not necessarily imply that the FIE will be RGES (or at least the proof is unclear yet), though practical RGES is guaranteed by Lemma 2 if βw(,0)\beta_{w}(\cdot,0) and βv(,0)\beta_{v}(\cdot,0) are also globally Lipschitz at the origin. On the other hand, if an FIE is RGES, then the Lipschitz conditions in the third case are valid and hence the corresponding MHE will be RGES by Lemma 2, as is in line with the existing results reported in [18, 23].

Remark 2.

Reference [24] proved existence of a finite horizon for MHE to be RGAS if the corresponding FIE induces a global contraction for the estimation error which essentially requires the dynamical system to be exp-i-IOSS. As referred to Remark 2, this conveys the same necessary condition as in the second case of Lemma 2. Overall, Lemma 2 covers more general conclusions, which links RGAS of FIE to pRGAS of the corresponding MHE but also RGAS and RGES of MHE as two naturally induced cases subject to extra mild conditions.

Remark 3.

Reference [25] proved existence of a finite horizon for a model predictive control scheme to exponentially stabilize an unperturbed nonlinear system via state feedback. Though the result could be relevant to the strongest conclusion of Lemma 2 when the FIE is RGES, a strict and meaningful connection is yet to be explored for establishing robustly stable estimators here when both disturbances and i-IOSS stability are in place which are yet absent in the system setup of [25].

Next, we present additional results when the conditions of Lemma 2 are relaxed and confined to local regions of the estimation problem. In this case, as a byproduct, an explicit bound for the sufficient MHE horizon size is derived when the corresponding FIE is known to be RAS or RGAS.

Lemma 3.

(Local/global stability implication from FIE to MHE) Consider MHE under Assumption 1. Let the disturbances be bounded as |wt|δw|w_{t}|\leq\delta_{w} and |vt|δv|v_{t}|\leq\delta_{v} for all t𝕀0t\in\mathbb{I}_{\geq 0}. Given any η(0, 1)\eta\in(0,\thinspace 1), the following four conclusions hold:

a) (RAS–>pRAS/RAS) Given any δ0>0\delta_{0}>0 and ϵ(0,ηs¯]\epsilon\in(0,\eta\bar{s}], where s¯:=βx(δ0, 0)βw(δw, 0)βv(δv, 0)\bar{s}:=\beta_{x}(\delta_{0},\thinspace 0)\oplus\beta_{w}(\delta_{w},\,0)\oplus\beta_{v}(\delta_{v},\,0), if FIE is RAS as per (9), then the corresponding MHE is pRAS for all (x0,x¯0)𝕏δ0(x_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}} and TT¯(ϵ):=min{τ:βx(s,τ)ηs,s[ϵη,s¯]}T\geq\underline{T}(\epsilon):=\min\{\tau:\beta_{x}(s,\tau)\leq\eta s,\,s\in[\frac{\epsilon}{\eta},\bar{s}]\}. If in addition βx(,τ)\beta_{x}(\cdot,\tau) is Lipschitz continuous at the origin over [0,s¯][0,\bar{s}] for all τ0\tau\geq 0, then the corresponding MHE is RAS for all (x0,x¯0)𝕏δ0(x_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}} and TT¯(0)T\geq\underline{T}(0).

b) (RGAS–>pRGAS/RGAS) If the conditions in a) are satisfied for all x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X} (which replaces (x0,x¯0)𝕏δ0(x_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}}), then the two conclusions in a) hold globally.

c) (RES–>RES) Given s¯\bar{s} defined in a) and any δ0>0\delta_{0}>0, if FIE is RES as per (9), in which for each {x,w,v}*\in\{x,w,v\} the 𝒦\mathcal{KL} function β\beta_{*} has an exponential form as β(s,τ):=csλτ\beta_{*}(s,\tau):=c_{*}s\lambda^{\tau} for certain c>0c_{*}>0, λ(0,1)\lambda\in(0,1) and all s[0,s¯]s\in[0,\bar{s}] and τ0\tau\geq 0, then the corresponding MHE is RES for all (x0,x¯0)𝕏δ0(x_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}} and TT¯:=1(logλ1cx+1)T\geq\underline{T}:=1\oplus(\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor+1).

d) (RGES–>RGES) If the conditions in c) are satisfied for all x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}, then the conclusion in c) holds globally.

Proof:

a) RAS–>pRAS. Given any η(0, 1)\eta\in(0,\,1) and ϵ(0,ηs¯]\epsilon\in(0,\eta\bar{s}], define β~x(s)=η(sϵη)\tilde{\beta}_{x}(s)=\eta\cdot(s\oplus\frac{\epsilon}{\eta}) for all s[0,s¯]s\in[0,\bar{s}]. The proof follows the same procedure of the proof for Lemma 2.a), but further bounds the term after the function composition with βx\beta_{x}\circ by β~x(s)\tilde{\beta}_{x}(s) defined here in each induction step. That the value of β~x(s)\tilde{\beta}_{x}(s) lies in [0,s¯][0,\bar{s}] makes the same compose-and-then-bound approach applicable throughout the induction steps.

Let T¯(ϵ):=min{τ:βx(s,τ)ηs,s[ϵη,s¯]}\underline{T}(\epsilon):=\min\{\tau:\beta_{x}(s,\tau)\leq\eta s,\,s\in[\frac{\epsilon}{\eta},\bar{s}]\}, which is well defined by Proposition 1.a). For any TT¯(ϵ)T\geq\underline{T}(\epsilon), Proposition 1.b) is applicable, and so the aforementioned induction leads to the following bound of the MHE estimate error:

|xtx^t|ϵηtTβx(|x0x¯0|,ttTT)\displaystyle\left|x_{t}-\hat{x}_{t}^{\star}\right|\leq\epsilon\oplus\eta^{\left\lfloor\frac{t}{T}\right\rfloor}\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace t-\left\lfloor\frac{t}{T}\right\rfloor T)
max{w,v}maxτ𝕀0:t1ηtτTβ(|τ|,ttτTTτ1)\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\eta^{\left\lfloor\frac{t-\tau}{T}\right\rfloor}\beta_{*}(\left|*_{\tau}\right|,\,t-\left\lfloor\frac{t-\tau}{T}\right\rfloor T-\tau-1)
ϵ1ηβx(|x0x¯0|, 0)ηtT\displaystyle\leq\epsilon\oplus\frac{1}{\eta}\beta_{x}(|x_{0}-\bar{x}_{0}|,\thinspace 0)\eta^{\frac{t}{T}}
max{w,v}maxτ𝕀0:t11ηβ(|τ|, 0)ηtτT\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\frac{1}{\eta}\beta_{*}(\left|*_{\tau}\right|,\,0)\eta^{\frac{t-\tau}{T}}
=:ϵ1ηβx,T(|x0x¯0|,t)\displaystyle=:\epsilon\oplus\frac{1}{\eta}\beta_{x,T}(|x_{0}-\bar{x}_{0}|,\thinspace t)
max{w,v}maxτ𝕀0:t11ηβ,T(|τ|,tτ1),\displaystyle\quad\oplus\max_{*\in\{w,v\}}\max_{\tau\in\mathbb{I}_{0:t-1}}\frac{1}{\eta}\beta_{*,T}(\left|*_{\tau}\right|,\,t-\tau-1), (16)

where the three functions βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are defined by one-to-one correspondence to the three preceding terms, each of which is dependent on the horizon size TT. With η(0,1)\eta\in(0,1), it is obvious that βx,T\beta_{x,T}, βw,T\beta_{w,T} and βv,T\beta_{v,T} are all 𝒦\mathcal{KL} functions. Consequently, the above estimate error bound implies that the MHE is pRAS by definition.

RAS–>RAS. When βx(,τ)\beta_{x}(\cdot,\tau) is Lipschitz continuous at the origin over the domain [0,s¯][0,\bar{s}] for all τ0\tau\geq 0, by Proposition 1.c) the value T¯(ϵ)\underline{T}(\epsilon) is well defined also at ϵ=0\epsilon=0. As a consequence, the same proof above will show that the implied pRAS collapses to RAS, as the ambiguity caused by ϵ\epsilon disappears.

b) RGAS–>pRGAS/RGAS. The proof is straightforward by extending the reasoning in the proof of a) to the entire applicable domain with x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}.

c) RES–>RES. In this case, we have βx(s,τ)=cxsλτ\beta_{x}(s,\tau)=c_{x}s\lambda^{\tau} with certain cx>0c_{x}>0, λ(0,1)\lambda\in(0,1) and for all s[0,s¯]s\in[0,\bar{s}] and τ0\tau\geq 0. If cx1c_{x}\leq 1, let T¯:=1\underline{T}:=1 and hence βx(s,T)βx(s,T¯)=cxsλ<s\beta_{x}(s,T)\leq\beta_{x}(s,\text{$\underline{T}$})=c_{x}s\lambda<s for all TT¯T\geq\underline{T} and s[0,s¯]s\in[0,\bar{s}]. And if cx>1c_{x}>1, let T¯:=logλ1cx+1\text{$\underline{T}$}:=\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor+1 and hence βx(s,T)βx(s,T¯)=cxsλlogλ1cx+1<cxsλlogλ1cx=s\beta_{x}(s,T)\leq\beta_{x}(s,\text{$\underline{T}$})=c_{x}s\lambda^{\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor+1}<c_{x}s\lambda^{\log_{\lambda}\frac{1}{c_{x}}}=s for all TT¯T\geq\underline{T} and s[0,s¯]s\in[0,\bar{s}]. In either case, the induction steps in the proof of Lemma 2.b) are applicable as the domain of the function composition via βx\beta_{x}\circ is kept within [0,s¯][0,\bar{s}] due to the contraction induced by the inequality βx(s,T)<s\beta_{x}(s,T)<s. As a consequence, the state estimate error will again be bounded as per (15). By applying the same argument right after (15) subject to T¯=1(logλ1cx+1)\underline{T}=1\oplus(\left\lfloor\log_{\lambda}\frac{1}{c_{x}}\right\rfloor+1) (which consolidates the two cases), we conclude that the MHE corresponding to the FIE is RES. This completes the proof of c).

d) RGES–>RGES. The proof is straightforward by extending the reasoning in the proof of c) to the entire problem domain with x0,x¯0𝕏x_{0},\bar{x}_{0}\in\mathbb{X}. ∎

Remark 4.

Two remarks follow on the proof of Lemma 3.a). a) The proof shows that the upper bound of the estimate error is scaled by the factor ηtT\eta^{\left\lfloor\frac{t}{T}\right\rfloor}, where η\eta is a value that bonds with the horizon bound T¯(ϵ)\underline{T}(\epsilon). Given any TT¯(ϵ)T\geq\underline{T}(\epsilon), the error bound can in fact be tightened by replacing the factor ηtT\eta^{\left\lfloor\frac{t}{T}\right\rfloor} with (ηϕ(TT¯))tT\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor}, where ϕ\phi is a certain \mathcal{L} function (and T¯\underline{T} is short for T¯(ϵ)\underline{T}(\epsilon)). As shown in Appendix A, as time tt is large enough, the derivative of this tighter factor will be negative and hence the error bound will be decreasing w.r.t. TT if the horizon size TT and the associated bound T¯(ϵ)\underline{T}(\epsilon) satisfy certain conditions that depend on the 𝒦\mathcal{KL} function βx\beta_{x}. b) The proof indicates that it is valid for MHE to change its horizon size over time according to the amplitude of uncertainty in an intermediate initial state, as long as sufficient contraction of the uncertainty is maintained. This leaves space for adapting the horizon size of MHE such that a flexible and time-varying tradeoff can be made between estimation quality and computational complexity.

Lemma 3 indicates that there always exists a large enough horizon size TT such that RAS/RGAS of its corresponding FIE implies pRAS/pRGAS of MHE, and further RAS/RGAS of the MHE if the FIE admits a certain Lipschitz continuity property. When the corresponding FIE is RES/RGES, the implication directly strengthens to RES/RGES of the MHE.

Note that the implication for pRGAS of MHE in Lemma 3.b) is equivalent to the one concluded in Lemma 2.a), but here it is derived from a different proof. The advantage is to have an explicit way of computing a sufficient horizon size. The implication for RGES of MHE in Lemma 3.d) is again equivalent to that presented in Lemma 2.b), while here it is extended naturally from a local implication.

The bound T¯\underline{T} for a sufficient horizon size as given in Lemma 3.a) can be derived explicitly if the 𝒦\mathcal{KL} function βx\beta_{x} has particular forms. Given a1,a21,b1(0,1),b2>0a_{1},a_{2}\geq 1,\,b_{1}\in(0,1),\,b_{2}>0 and c1,c2>0c_{1},c_{2}>0, two expressions of T¯\underline{T} are derived as follows:

T¯={logb1ηc1s¯a11,if βx(s,τ):=c1sa1b1τ,c2s¯a21ηb21,if βx(s,τ):=c2sa2(τ+1)b2.\underline{T}=\begin{cases}\left\lceil\log_{b_{1}}\frac{\eta}{c_{1}\bar{s}^{a_{1}-1}}\right\rceil,&\text{if }\beta_{x}(s,\tau):=c_{1}s^{a_{1}}b_{1}^{\tau},\\ \left\lceil\sqrt[b_{2}]{\frac{c_{2}\bar{s}^{a_{2}-1}}{\eta}}-1\right\rceil,&\text{if }\beta_{x}(s,\tau):=c_{2}s^{a_{2}}(\tau+1)^{-b_{2}}.\end{cases}

The given constants {ai,bi,ci}\{a_{i},b_{i},c_{i}\} for i{1,2}i\in\{1,2\} are such that the two βx(,τ)\beta_{x}(\cdot,\tau)’s above are Lipschitz continuous at the origin over [0,s¯][0,\,\bar{s}] for all τ𝕀0\tau\in\mathbb{I}_{\geq 0}.

V Robust Stability of FIE and MHE

This section presents new sufficient conditions for robust stability of FIE, and also for that of the corresponding MHE by applying the stability implication revealed in last section.

To make the conditions easy to interpret and the proof concise to present, the following notations are introduced in the spirit of (4):

π0\displaystyle\pi_{0} :=x0x¯0,πτ+1:=wτ,πτ+t+1:=vτ,\displaystyle:=x_{0}-\bar{x}_{0},\,\pi_{\tau+1}:=w_{\tau},\,\pi_{\tau+t+1}:=v_{\tau}, (17)
π^0\displaystyle\hat{\pi}_{0} :=x0x^0,π^τ+1:=wτw^τ,π^τ+t+1:=vτv^τ\displaystyle:=x_{0}-\hat{x}_{0},\,\hat{\pi}_{\tau+1}:=w_{\tau}-\hat{w}_{\tau},\,\hat{\pi}_{\tau+t+1}:=v_{\tau}-\hat{v}_{\tau}
π~0\displaystyle\tilde{\pi}_{0} :=χ0x¯0,π~τ+1:=ωτ,π~τ+t+1:=ντ,\displaystyle:=\chi_{0}-\bar{x}_{0},\,\tilde{\pi}_{\tau+1}:=\omega_{\tau},\,\tilde{\pi}_{\tau+t+1}:=\nu_{\tau},
π~^0\displaystyle\hat{\tilde{\pi}}_{0} :=x^0x¯0,π~^τ+1:=w^τ,π~^τ+t+1:=v^τ,\displaystyle:=\hat{x}_{0}-\bar{x}_{0},\,\hat{\tilde{\pi}}_{\tau+1}:=\hat{w}_{\tau},\,\hat{\tilde{\pi}}_{\tau+t+1}:=\hat{v}_{\tau},

for all τ𝕀0:t1\tau\in\mathbb{I}_{0:t-1}, where π^\hat{\pi}_{\cdot} and π~^\hat{\tilde{\pi}}_{\cdot} refer to the optimal estimates of π\pi_{\cdot} and π~\tilde{\pi}_{\cdot}, respectively. Given any {π,π^,π~,π~^}*\in\{\pi,\,\hat{\pi},\,\tilde{\pi},\,\hat{\tilde{\pi}}\}, notation 0:2t\boldsymbol{*}_{0:2t} collects the sequence of vector variables ()0:2t(*_{\cdot})_{0:2t} and the corresponding domain is denoted by Π\Pi if χ0,x¯0𝕏\chi_{0},\,\bar{x}_{0}\in\mathbb{X} and as Πδ0\Pi_{\delta_{0}} if (χ0,x¯0)𝕏δ0(\chi_{0},\,\bar{x}_{0})\in\mathbb{X}_{\delta_{0}}, where 𝕏δ0\mathbb{X}_{\delta_{0}} is defined in (8). Function ι()\iota(*_{\cdot}) extracts the time index (i.e., the original index τ\tau) of *_{\cdot} as per (5). Given any i𝕀0:2ti\in\mathbb{I}_{0:2t} , it is easy to verify that ι(πi)=ι(π^i)=ι(π~i)=ι(π~^i)\iota(\pi_{i})=\iota(\hat{\pi}_{i})=\iota(\tilde{\pi}_{i})=\iota(\hat{\tilde{\pi}}_{i}) and that πi=π^i+π~^i\pi_{i}=\hat{\pi}_{i}+\hat{\tilde{\pi}}_{i}.

Given any δ0>0\delta_{0}>0, the next two assumptions are introduced to establish robust stability of FIE.

Assumption 2.

There exists ρ¯,ρ𝒦\underline{\rho},\rho\in\mathcal{KL} such that the cost function of FIE Vt(𝝅~0:2t)V_{t}(\tilde{\boldsymbol{\pi}}_{0:2t}), which is continuous, satisfies the following inequality for all 𝝅~0:2tΠδ0\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi_{\delta_{0}} and t𝕀0t\in\mathbb{I}_{\geq 0}:

maxi𝕀0:2tρ¯(|π~i|,tι(π~i)1)Vt(𝝅~0:2t)\displaystyle\max_{i\in\mathbb{I}_{0:2t}}\underline{\rho}(\left|\tilde{\pi}_{i}\right|,\,t-\iota(\tilde{\pi}_{i})-1)\leq V_{t}(\tilde{\boldsymbol{\pi}}_{0:2t}) (18)
maxi𝕀0:2tρ(|π~i|,tι(π~i)1).\displaystyle\quad\leq\max_{i\in\mathbb{I}_{0:2t}}\rho(\left|\tilde{\pi}_{i}\right|,\,t-\iota(\tilde{\pi}_{i})-1).
Assumption 3.

𝒦\mathcal{KL} function α\alpha from the i-IOSS property (6) and 𝒦\mathcal{KL} functions ρ¯\underline{\rho} and ρ\rho from Assumption 2 satisfy the following inequality for all 𝝅~0:2tΠδ0\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi_{\delta_{0}} and τ,τ𝕀0:t\tau,\,\tau^{\prime}\in\mathbb{I}_{0:t}:

α(2ρ¯1(ρ(|π~i|,τ),τ),τ)α¯(|π~i|,τ),\alpha\left(2\underline{\rho}^{-1}\left(\rho(|\tilde{\pi}_{i}|,\,\tau),\,\tau^{\prime}\right),\,\tau^{\prime}\right)\leq\bar{\alpha}(|\tilde{\pi}_{i}|,\,\tau), (19)

for certain α¯𝒦\bar{\alpha}\in\mathcal{KL}, in which ρ¯1(,τ)\underline{\rho}^{-1}(\cdot,\,\tau) is the inverse of ρ¯(,τ)\underline{\rho}(\cdot,\,\tau) w.r.t. its first argument given the second argument τ\tau.

Overall, Assumption 2 requires that the FIE has a property that mimics the i-IOSS property of the system, while Assumption 3 ensures that the FIE is more sensitive than the system to the uncertainties so that accurate inference of the state is possible. The interpretation of the relatively more obscure Assumption 3 becomes clear with a concrete realization below.

Lemma 4.

(Concrete realization of Assumption 3) Assumption 3 is true if the 𝒦\mathcal{KL} function ρ¯\underline{\rho} satisfies

ρ¯(|π~i|,τ)α(2|π~i|,τ),\underline{\rho}(|\tilde{\pi}_{i}|,\,\tau)\geq\alpha(2|\tilde{\pi}_{i}|,\,\tau), (20)

for all 𝛑~0:2tΠδ0\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi_{\delta_{0}} and τ𝕀0:t\tau\in\mathbb{I}_{0:t}.

Proof:

It suffices to show that inequality (19) is satisfied. Subject to (20), we have

α(2ρ¯1(ρ(s,τ),τ),τ)ρ¯(ρ¯1(ρ(s,τ),τ),τ)=ρ(s,τ).\displaystyle\alpha\left(2\underline{\rho}^{-1}\left(\rho(s,\tau),\tau^{\prime}\right),\tau^{\prime}\right)\leq\underline{\rho}\left(\underline{\rho}^{-1}\left(\rho(s,\tau),\tau^{\prime}\right),\tau^{\prime}\right)=\rho(s,\tau).

This implies that inequality (19) is satisfied with α¯:=ρ\bar{\alpha}:=\rho, and hence completes the proof. ∎

Robust stability of FIE/MHE can then be established under Assumptions 2 and 3.

Theorem 1.

The following two conclusions hold:

a) (RAS/RGAS/RES/RGES of FIE) The FIE is RAS if the system is i-IOSS and the cost function of FIE satisfies Assumptions 2 and 3, and is RGAS if further the two assumptions are valid for all 𝛑~0:2tΠ\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi. And if, furthermore, the 𝒦\mathcal{KL} functions in Assumption 3 admit specific forms as α(s,τ)=csbτ\alpha(s,\tau)=csb^{\tau} and α¯(s,τ)=c¯sb¯τ\bar{\alpha}(s,\tau)=\bar{c}s\bar{b}^{\tau} with certain c,c¯>0c,\bar{c}>0, b,b¯(0,1)b,\bar{b}\in(0,1) and for all τ0\tau\geq 0 and ss in the applicable domain, then the FIE which is RAS or RGAS will be RES or RGES, respectively.

b) (pRAS/pRGAS/RES/RGES/RGAS of MHE) In the four cases of a), the corresponding MHE under additional Assumption 1 with a sufficiently long horizon is pRAS, pRGAS, RES and RGES, respectively. And the pRGAS of MHE strengthens to RGAS if the 𝒦\mathcal{K} functions α(,0)\alpha(\cdot,0) and α¯(,0)\bar{\alpha}(\cdot,0) in Assumption 3 are globally Lipschitz continuous at the origin.

Proof:

a) RAS/RGAS of FIE. The global optimal solution of 𝝅~0:2t\tilde{\boldsymbol{\pi}}_{0:2t} for the FIE is denoted as 𝝅~^0:2t\hat{\tilde{\boldsymbol{\pi}}}_{0:2t} (cf. Eq. (17)), yielding a minimum cost VtoV_{t}^{o}. It follows that for all t0t\geq 0,

Vto\displaystyle V_{t}^{o} =Vt(𝝅~^0:2t)Vt(𝝅0:2t)\displaystyle=V_{t}(\hat{\tilde{\boldsymbol{\pi}}}_{0:2t})\leq V_{t}(\boldsymbol{\pi}_{0:2t})
Assump. 2maxi𝕀0:2tρ(|πi|,tι(πi)1)=:V¯t\displaystyle\stackrel{{\scriptstyle\text{Assump. \ref{assump: A2}}}}{{\leq}}\max_{i\in\mathbb{I}_{0:2t}}\rho(\left|\pi_{i}\right|,\,t-\iota(\pi_{i})-1)=:\bar{V}_{t}

Consequently, by Assumption 2 we have ρ¯(|π~^i|,tι(π~^i)1)VtoV¯t\underline{\rho}(|\hat{\tilde{\pi}}_{i}|,\,t-\iota(\hat{\tilde{\pi}}_{i})-1)\leq V_{t}^{o}\leq\bar{V}_{t}, and further |π~^i|ρ¯t1(V¯t,tι(π~^i)1)|\hat{\tilde{\pi}}_{i}|\leq\underline{\rho}_{t}^{-1}(\bar{V}_{t},\,t-\iota(\hat{\tilde{\pi}}_{i})-1).

Since π^i=πiπ~^i\hat{\pi}_{i}=\pi_{i}-\hat{\tilde{\pi}}_{i} for each i𝕀0:2ti\in\mathbb{I}_{0:2t}, by applying the triangle inequality this implies that

|π^i|\displaystyle\left|\hat{\pi}_{i}\right| |πi|+|π~^i||πi|+ρ¯t1(V¯t,tι(π~^i)1)\displaystyle\leq\left|\pi_{i}\right|+\left|\hat{\tilde{\pi}}_{i}\right|\leq\left|\pi_{i}\right|+\underline{\rho}_{t}^{-1}(\bar{V}_{t},\,t-\iota(\hat{\tilde{\pi}}_{i})-1)
2|πi|2ρ¯t1(V¯t,tι(π~^i)1),\displaystyle\leq 2\left|\pi_{i}\right|\oplus 2\underline{\rho}_{t}^{-1}(\bar{V}_{t},\,t-\iota(\hat{\tilde{\pi}}_{i})-1), (21)

for all i𝕀0:2ti\in\mathbb{I}_{0:2t}. Substitute (21) into the i-IOSS property of (6), yielding

|x(t;x0,𝒘0:t1)x(t;x^0,𝒘^0:t1)|\displaystyle\left|x(t;\,x_{0},\,\boldsymbol{w}_{0:t-1})-x(t;\,\hat{x}_{0},\,\hat{\boldsymbol{w}}_{0:t-1})\right|
maxi𝕀0:2tα(|π^i|,tι(π^i)1)maxi𝕀0:2tα(2|πi|,tι(π^i)1)\displaystyle\leq\max_{i\in\mathbb{I}_{0:2t}}\alpha(\left|\hat{\pi}_{i}\right|,\,t-\iota(\hat{\pi}_{i})-1)\leq\max_{i\in\mathbb{I}_{0:2t}}\alpha\left(2\left|\pi_{i}\right|,\,t-\iota(\hat{\pi}_{i})-1\right)
maxi𝕀0:2tα(2ρ¯1(V¯t,tι(π~^i)1),tι(π^i)1)\displaystyle\quad\quad\quad\quad\quad\oplus\max_{i\in\mathbb{I}_{0:2t}}\alpha\left(2\underline{\rho}^{-1}(\bar{V}_{t},\,t-\iota(\hat{\tilde{\pi}}_{i})-1),\,t-\iota(\hat{\pi}_{i})-1\right)
=maxi𝕀0:2tα(2|πi|,tι(π^i)1)\displaystyle=\max_{i\in\mathbb{I}_{0:2t}}\alpha\left(2\left|\pi_{i}\right|,\,t-\iota(\hat{\pi}_{i})-1\right)
maxi𝕀0:2tmaxj𝕀0:2tα(2ρ¯1(ρ(|πj|,tι(πj)1),tι(π~^i)1),tι(π^i)1)\displaystyle\quad\oplus\max_{i\in\mathbb{I}_{0:2t}}\max_{j\in\mathbb{I}_{0:2t}}\alpha\left(\begin{array}[]{c}2\underline{\rho}^{-1}\left(\begin{array}[]{c}\rho(\left|\pi_{j}\right|,\,t-\iota(\pi_{j})-1),\\ t-\iota(\hat{\tilde{\pi}}_{i})-1\end{array}\right),\\ t-\iota(\hat{\pi}_{i})-1\end{array}\right)
maxi𝕀0:2t(α(2|πi|,tι(πi)1)α¯(|πi|,tι(πi)1))\displaystyle\leq\max_{i\in\mathbb{I}_{0:2t}}\left(\alpha\left(2\left|\pi_{i}\right|,\,t-\iota(\pi_{i})-1\right)\oplus\bar{\alpha}\left(\left|\pi_{i}\right|,\,t-\iota(\pi_{i})-1\right)\right)
=:maxi𝕀0:2tβ(|πi|,tι(πi)1),\displaystyle=:\max_{i\in\mathbb{I}_{0:2t}}\beta\left(\left|\pi_{i}\right|,\,t-\iota(\pi_{i})-1\right),

where the equality ι(π^)=ι(π)=ι(π~^)\iota(\hat{\pi}_{\cdot})=\iota(\pi_{\cdot})=\iota(\hat{\tilde{\pi}}_{\cdot}) and Assumption 3 have been used to derive the last inequality, and β(|πi|,τ):=α(2|πi|,τ)α¯(|πi|,τ)\beta\left(|\pi_{i}|,\tau\right):=\alpha\left(2|\pi_{i}|,\tau\right)\oplus\bar{\alpha}\left(|\pi_{i}|,\tau\right) for all 𝝅0:2tΠδ0\boldsymbol{\pi}_{0:2t}\in\Pi_{\delta_{0}} and τ0\tau\geq 0. Since β\beta is a 𝒦\mathcal{KL} function, the FIE is RAS by definition. With the same reasoning, the conclusion immediately extends to that the FIE is RGAS when Assumptions 2 and 3 are applicable to the entire problem domain as specified by 𝝅~0:2tΠ\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi.

RES/RGES of FIE. If in addition, α(2s,τ)=2csbτ\alpha\left(2s,\,\tau\right)=2csb^{\tau} and α¯(s,τ)=c¯sb¯τ\bar{\alpha}\left(s,\,\tau\right)=\bar{c}s\bar{b}^{\tau} with certain c,c¯>0c,\bar{c}>0 and b,b¯(0,1)b,\bar{b}\in(0,1) and for all τ0\tau\geq 0 and ss in an applicable domain, then it is immediate that α(2s,τ)α¯(s,τ)c¯sb¯τ=:β(s,τ)\alpha\left(2s,\,\tau\right)\oplus\bar{\alpha}\left(s,\,\tau\right)\leq\bar{c}^{\prime}s\bar{b}^{\prime\tau}=:\beta(s,\,\tau), with c¯:=max{2c,c¯}\bar{c}^{\prime}:=\max\{2c,\,\bar{c}\} and b¯:=max{b,b¯}\bar{b}^{\prime}:=\max\{b,\,\bar{b}\}. Here β\beta is a 𝒦\mathcal{KL} function in an exponential form, so the FIE which is RAS (or RGAS) in this case will be RES (or RGES) by Lemma 1.

b) pRAS/pRGAS/RES/RGES/RGAS of MHE. The conclusions are direct application of RAS/RGAS/RES/RGES of FIE and Lemma 2 and 3, subject to the given conditions. ∎

Next, we present a lemma indicating that Assumptions 2 and 3 do not impose special difficulty as there always exists a cost function satisfying both of them if the system is i-IOSS.

Lemma 5.

(Satisfaction of Assumptions 2 and 3) If the system is i-IOSS as per (6), then, given any t0t\geq 0 and ρ¯𝒦\underline{\rho}\in\mathcal{KL} satisfying ρ¯(|π~i|,τ)α(2|π~i|,τ)\underline{\rho}(|\tilde{\pi}_{i}|,\,\tau)\geq\alpha(2|\tilde{\pi}_{i}|,\,\tau) for all 𝛑~0:2tΠδ0\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi_{\delta_{0}} and τ𝕀0:t\tau\in\mathbb{I}_{0:t}, it is feasible to specify the cost function of FIE as

Vt(𝝅~0:2t):=maxi𝕀0:2tρ¯(|π~i|,tι(π~i)1),V_{t}(\tilde{\boldsymbol{\pi}}_{0:2t}):=\max_{i\in\mathbb{I}_{0:2t}}\underline{\rho}(\left|\tilde{\pi}_{i}\right|,\,t-\iota(\tilde{\pi}_{i})-1), (22)

such that Assumptions 2 and 3 hold true. The same form of cost function remains valid for Assumptions 2 and 3 to hold globally with 𝛑~0:2tΠ\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi. In either the local or the global case, if the system is exp-i-IOSS, then the 𝒦\mathcal{KL} function ρ¯\underline{\rho} admits a form as ρ¯(s,τ)=c|π~i|bτ\underline{\rho}(s,\,\tau)=c|\tilde{\pi}_{i}|b^{\tau} with certain c>0c>0 and b(0,1)b\in(0,1) and for τ0\tau\geq 0 and π~i\tilde{\pi}_{i} in the applicable domain.

Proof:

Given the cost function specified as per (22), Assumption 2 is automatically met. With ρ¯(|π~i|,τ)α(2|π~i|,τ)\underline{\rho}(|\tilde{\pi}_{i}|,\,\tau)\geq\alpha(2|\tilde{\pi}_{i}|,\,\tau) for all 𝝅~0:2tΠδ0\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi_{\delta_{0}} and τ𝕀0:t\tau\in\mathbb{I}_{0:t}, Assumption 3 is also met by Lemma 4. While, existence of such 𝒦\mathcal{KL} function ρ¯\underline{\rho} is guaranteed as it is always feasible to let ρ¯(|π~i|,τ):=α(2|π~i|,τ)\underline{\rho}(|\tilde{\pi}_{i}|,\,\tau):=\alpha(2|\tilde{\pi}_{i}|,\,\tau). Since the reasoning approach does not rely on the domain of |π~i||\tilde{\pi}_{i}|, the conclusion remains valid if Assumptions 2 and 3 are extended to the entire problem domain with 𝝅~0:2tΠ\tilde{\boldsymbol{\pi}}_{0:2t}\in\Pi.

If the system is exp-i-IOSS with α(|π~i|,τ):=c|π~i|bτ\alpha(|\tilde{\pi}_{i}|,\,\tau):=c^{\prime}|\tilde{\pi}_{i}|b^{\tau} for certain c>0c^{\prime}>0 and b(0,1)b\in(0,1), then it is valid to let ρ¯(|π~i|,τ):=α(2|π~i|,τ)=c|π~i|bτ\underline{\rho}(|\tilde{\pi}_{i}|,\,\tau):=\alpha(2|\tilde{\pi}_{i}|,\,\tau)=c|\tilde{\pi}_{i}|b^{\tau} with c:=2cc:=2c^{\prime} for all τ0\tau\geq 0 and π~i\tilde{\pi}_{i} in the applicable local/global domain. ∎

Lemma 5 indicates that a valid cost function can always be designed from the i-IOSS bound function α\alpha, and hence implies an important result below.

Corollary 1.

There exists a cost function for FIE to be RGAS (or RGES) if and only if the system is i-IOSS (or exp-i-IOSS).

Proof:

Sufficiency. By Lemma 5, the system being i-IOSS implies that the FIE admits a cost function such that Assumptions 2 and 3 hold true, which consequently implies that the FIE is RGAS by Theorem 1. When the system is exp-i-IOSS, the conclusion trivially strengthens to that the FIE is RGES by following the same approach of reasoning.

Necessity. The necessity in the RGAS case has been proved in existing literature, e.g., [1, Proposition 4.6] and [23, Proposition 2.4], while the proof in the RGES case is implied by the same proof there. ∎

By Lemma 2, it follows immediately from Corollary 1 that there exists a cost function for the MHE to be pRGAS (or RGES) if the system is i-IOSS (or exp-i-IOSS).

Remark 5.

The conclusion of Corollary 1 coincides with a key finding reported in a latest paper [24], which had been submitted for review. The derivation approaches are, however, quite different. Here, we apply the reasoning approach of [2], focusing on developing most general conditions for robust stability of FIE, and the aforementioned conclusion appears as a corollary for an endeavour to understand the developed conditions. In contrast, reference [24] reaches the conclusion by starting with a particular cost function which is constructed directly from the i-IOSS property of the system and is not necessarily the only form admitted by our derived conditions.

VI Conclusion

This work proved that robust global (or local) asymptotic stability of full information estimation (FIE) implies practical robust global (or local) asymptotic stability of moving horizon estimation (MHE) which implements a sufficiently long horizon. The “practical” becomes exact if the FIE admits a certain Lipschitz continuity. In both exact and inexact cases, explicit ways were also provided of computing a sufficient horizon size for a robustly stable MHE. With the revealed implication, sufficient conditions for the MHE to be robustly stable were derived by firstly developing those for ensuring robust stability of the corresponding FIE. A particular realization of these conditions indicates that the system being i-IOSS is not only necessary but also sufficient to ensure the existence of a robustly globally asymptotically stable of FIE. With the revealed implication, the sufficiency remains valid for existence of a practically robustly stable MHE.

Since it is generic and relies only on robust stability of the corresponding FIE, the revealed stability link implies that existing conditions which ensure robust stability of FIE can all be inherited to establish that of the corresponding MHE, but also paves the way for derivation of new sufficient conditions via deeper analysis of FIE. This may also contribute to developing robustly stable MHE which exploits sub-optimal solutions for resource-constrained or faster estimations. Readers are referred to Sec. VII of [2] for related discussions, and [26, 27, 28, 29] for some recent developments in this line of research.

Appendix A. Derivative of the Error Bound Factor w.r.t. the Moving Horizon Size TT

The symbols mostly come from Lemma 3.a) and its proof. Given TT¯(ϵ)T\geq\underline{T}(\epsilon), it is easy to show that the estimate error bound in (16) can be tightened by replacing ηtT\eta^{\left\lfloor\frac{t}{T}\right\rfloor} with (ηβx(s,T)βx(s,T¯))tT\left(\frac{\eta\beta_{x}(s,T)}{\beta_{x}(s,\underline{T})}\right)^{\left\lfloor\frac{t}{T}\right\rfloor}. This appendix analyzes the derivative of this new bound factor w.r.t. TT, in order to understand monotonicity of the bound w.r.t. the horizon size TT.

Firstly, we show that there exists ϕ\phi\in\mathcal{L} such that ϕ(TT¯)=βx(s,T)βx(s,T¯)\phi(T-\underline{T})=\frac{\beta_{x}(s,T)}{\beta_{x}(s,\underline{T})} for all TT¯T\geq\underline{T} . By Lemma 8 in [30], given βx𝒦\beta_{x}\in\mathcal{KL}, there exists α𝒦\alpha\in\mathcal{K} and φ\varphi\in\mathcal{L} such that βx(s,t)α(s)φ(t)\beta_{x}(s,t)\leq\alpha(s)\varphi(t), for all s,t0s,t\geq 0. This implies that the RGAS property of FIE can equivalently be expressed with the product of 𝒦\mathcal{K} and \mathcal{L} functions. Therefore, it does not lose generality by assuming βx(s,t):=α(s)φ(t)\beta_{x}(s,t):=\alpha(s)\varphi(t). Consequently, βx(s,T)βx(s,T¯)=φ(T)φ(T¯)\frac{\beta_{x}(s,T)}{\beta_{x}(s,\underline{T})}=\frac{\varphi(T)}{\varphi(\underline{T})}. Since φ\varphi is a \mathcal{L} function, given T¯0\underline{T}\geq 0, it is feasible to define ϕ(TT¯)=φ(T)φ(T¯)\phi(T-\underline{T})=\frac{\varphi(T)}{\varphi(\underline{T})} and ϕ\phi will be a \mathcal{L} function for all TT¯T\geq\underline{T}.

Next, we derive the derivative of (ηϕ(TT¯))tT\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor} w.r.t. TT. By definition we have 0<ϕ(TT¯)10<\phi(T-\underline{T})\leq 1 for all TT¯T\geq\underline{T}. Let z:=ln(ηϕ(TT¯))tT=tTln(ηϕ(TT¯))z:=\ln\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor}=\left\lfloor\frac{t}{T}\right\rfloor\ln(\eta\phi(T-\underline{T})). Hence (ηϕ(TT¯))tT=ez\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor}=e^{z}. The chain rule of derivatives implies

d(ηϕ(TT¯))tTdT=ezdzdT\displaystyle\frac{d\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor}}{dT}=e^{z}\frac{dz}{dT}
=(ηϕ(TT¯))tT(dtTdTlnηϕ(TT¯)+tTϕ(TT¯)dϕ(TT¯)dT)\displaystyle=\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor}\cdot\left(\begin{array}[]{c}\frac{d\left\lfloor\frac{t}{T}\right\rfloor}{dT}\ln\eta\phi(T-\underline{T})\\ +\frac{\left\lfloor\frac{t}{T}\right\rfloor}{\phi(T-\underline{T})}\frac{d\phi(T-\underline{T})}{dT}\end{array}\right)
=tT(ηϕ(TT¯))tTϕ(TT¯)(ϕ(TT¯)tTdtTdTlnηϕ(TT¯)+dϕ(TT¯)dT).\displaystyle=\frac{\left\lfloor\frac{t}{T}\right\rfloor\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor}}{\phi(T-\underline{T})}\cdot\left(\begin{array}[]{c}\frac{\phi(T-\underline{T})}{\left\lfloor\frac{t}{T}\right\rfloor}\frac{d\left\lfloor\frac{t}{T}\right\rfloor}{dT}\ln\eta\phi(T-\underline{T})\\ +\frac{d\phi(T-\underline{T})}{dT}\end{array}\right).

The first factor of the last expression is positive, so the sign of the derivative is uniquely determined by the second factor. When tt is much larger than TT, we have tTtT\left\lfloor\frac{t}{T}\right\rfloor\rightarrow\frac{t}{T}. Then, the second factor approximates to

dϕ(TT¯)dT1Tϕ(TT¯)lnηϕ(TT¯)\displaystyle\frac{d\phi(T-\underline{T})}{dT}-\frac{1}{T}\phi(T-\underline{T})\ln\eta\phi(T-\underline{T})
=1φ(T¯)(dφ(T)dTφ(T)Tlnηφ(T)φ(T¯))\displaystyle=\frac{1}{\varphi(\underline{T})}\left(\frac{d\varphi(T)}{dT}-\frac{\varphi(T)}{T}\ln\frac{\eta\varphi(T)}{\varphi(\underline{T})}\right)

As 1φ(T¯)\frac{1}{\varphi(\underline{T})} is positive, it is sufficient to analyze the sign of

κ(T):=dφ(T)dTφ(T)Tlnηφ(T)φ(T¯).\kappa(T):=\frac{d\varphi(T)}{dT}-\frac{\varphi(T)}{T}\ln\frac{\eta\varphi(T)}{\varphi(\underline{T})}.

As examples, two particular forms of φ\varphi are considered. Firstly, let us consider the exponential form: φ(T):=b1T\varphi(T):=b_{1}^{T} with b1(0, 1)b_{1}\in(0,\,1). Then, κ(T)=b1TTlnb1T¯η\kappa(T)=\frac{b_{1}^{T}}{T}\ln\frac{b_{1}^{\underline{T}}}{\eta}. So, κ(T)<0\kappa(T)<0 and hence (ηϕ(TT¯))tT\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor} will be strictly decreasing w.r.t. TT if TT¯1+logb1ηT\geq\underline{T}\geq 1+\left\lceil\log_{b_{1}}\eta\right\rceil. Secondly, let us consider the fractional form: φ(T):=(T+1)b2\varphi(T):=(T+1)^{-b_{2}} with b2>0b_{2}>0. Then, κ(T)=b2(T+1)b2(1T+11Tlnηb2(T+1)T¯+1).\kappa(T)=\frac{-b_{2}}{(T+1)^{b_{2}}}\left(\frac{1}{T+1}-\frac{1}{T}\ln\frac{\sqrt[b_{2}]{\eta}(T+1)}{\underline{T}+1}\right). So, κ(T)<0\kappa(T)<0 and hence (ηϕ(TT¯))tT\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor} will again be strictly decreasing w.r.t. TT if TT¯>ηb2(T+1)eTT+11T\geq\underline{T}>\sqrt[b_{2}]{\eta}(T+1)e^{-\frac{T}{T+1}}-1. Since the condition is more restrictive in this case, the monotonic decrement property seems harder to establish than that in the previous case.

To conclude, the analysis shows that if the moving horizon size TT satisfies certain conditions (being large enough in general), the bound of the estimate error as controlled by the factor (ηϕ(TT¯))tT\left(\eta\phi(T-\underline{T})\right)^{\left\lfloor\frac{t}{T}\right\rfloor} will eventually decrease with TT.

References

  • [1] J. B. Rawlings, D. Q. Mayne, and M. Diehl, Model Predictive Control: Theory, Computation, and Design, 2nd ed.   Nob Hill Publishing Madison, WI, 2020.
  • [2] W. Hu, “Robust stability of optimization-based state estimation,” https://arxiv.org/abs/1702.01903, 2017.
  • [3] H. Michalska and D. Mayne, “Moving horizon observers,” in Nonlinear Control Systems Design 1992.   Elsevier, 1993, pp. 185–190.
  • [4] K. R. Muske, J. B. Rawlings, and J. H. Lee, “Receding horizon recursive state estimation,” in American Control Conference (ACC), San Francisco, CA, USA, Jun 1993, pp. 900–904.
  • [5] H. Michalska and D. Q. Mayne, “Moving horizon observers and observer-based control,” IEEE Trans. on Autom. Control, vol. 40, no. 6, pp. 995–1006, 1995.
  • [6] C. Rao, J. Rawlings, and D. Mayne, “Constrained state estimation for nonlinear discrete-time systems: Stability and moving horizon approximations,” IEEE Trans. on Autom. Control, vol. 48, no. 2, pp. 246–258, 2003.
  • [7] A. Alessandri, M. Baglietto, and G. Battistelli, “Moving-horizon state estimation for nonlinear discrete-time systems: New stability results and approximation schemes,” Automatica, vol. 44, no. 7, pp. 1753–65, 2008.
  • [8] J. B. Rawlings and D. Q. Mayne, Model Predictive Control: Theory and Design.   Madison, WI: Nob Hill Pub., 2009.
  • [9] A. Alessandri, M. Baglietto, G. Battistelli, and V. Zavala, “Advances in moving horizon estimation for nonlinear systems,” in The 49th IEEE Conference on Decision and Control (CDC), Atlanta, Georgia, USA, Dec. 2010, pp. 5681–5688.
  • [10] J. Liu, “Moving horizon state estimation for nonlinear systems with bounded uncertainties,” Chem. Eng. Sci., vol. 93, pp. 376–386, 2013.
  • [11] J. Rawlings and L. Ji, “Optimization-based state estimation: Current status and some new results,” Journal of Process Control, vol. 22, pp. 1439–1444, 2012.
  • [12] E. Sontag and Y. Wang, “Output-to-state stability and detectability of nonlinear systems,” Systems & Control Letters, vol. 29, no. 5, pp. 279–290, 1997.
  • [13] W. Hu, L. Xie, and K. You, “Optimization-based state estimation under bounded disturbances,” in The 54th IEEE Conference on Decision and Control (CDC), Osaka, Japan, Dec 2015.
  • [14] L. Ji, J. B. Rawlings, W. Hu, A. Wynn, and M. Diehl, “Robust stability of moving horizon estimation under bounded disturbances,” IEEE Trans. on Autom. Control, vol. 61, no. 11, pp. 3509 – 3514, 2016.
  • [15] M. A. Müller, “Nonlinear moving horizon estimation for systems with bounded disturbances,” in American Control Conference (ACC), Boston, USA, Jul 2016, pp. 883–888.
  • [16] ——, “Nonlinear moving horizon estimation in the presence of bounded disturbances,” Automatica, vol. 79, pp. 306–314, 2017.
  • [17] D. A. Allan and J. B. Rawlings, Handbook of Model Predictive Conttrol.   Springer Nature Switzerland AG, 2019, ch. Moving Horizon Estimation, pp. 99–124.
  • [18] S. Knüfer and M. A. Müller, “Robust global exponential stability for moving horizon estimation,” in The 57th IEEE Conference on Decision and Control, Miami Beach, FL, USA, Dec. 2018, pp. 3477–3482.
  • [19] D. A. Allan, J. Rawlings, and A. R. Teel, “Nonlinear detectability and incremental input/output-to-state stability,” SIAM Journal on Control and Optimization, vol. 59, no. 4, pp. 3017–3039, 2021.
  • [20] S. Knüfer and M. A. Müller, “Time-discounted incremental input/output-to-state stability,” in The 59th IEEE Conference on Decision and Control (CDC), Jeju Island, Korea, Dec. 2020, pp. 5394–5400.
  • [21] L. Grüne and J. Pannek, “Nonlinear model predictive control: Theory and algorithms,” in Nonlinear Model Predictive Control, 2nd ed.   Springer, 2017, pp. 45–69.
  • [22] A. Aravkin, J. V. Burke, L. Ljung, A. Lozano, and G. Pillonetto, “Generalized kalman smoothing: Modeling and algorithms,” arXiv preprint arXiv:1609.06369, 2016.
  • [23] D. A. Allan, “A lyapunov-like function for analysis of model predictive control and moving horizon estimation,” Ph.D. dissertation, 2020.
  • [24] S. Knuefer and M. A. Mueller, “Nonlinear full information and moving horizon estimation: Robust global asymptotic stability,” arXiv preprint arXiv:2105.02764, 2021.
  • [25] A. Jadbabaie and J. Hauser, “On the stability of receding horizon control with a general terminal cost,” IEEE Trans. on Autom. Control, vol. 50, no. 5, pp. 674–678, 2005.
  • [26] A. Wynn, M. Vukov, and M. Diehl, “Convergence guarantees for moving horizon estimation based on the real-time iteration scheme,” IEEE Trans. on Autom. Control, vol. 59, no. 8, pp. 2215–2221, 2014.
  • [27] A. Alessandri and M. Gaggero, “Fast moving horizon state estimation for discrete-time systems using single and multi iteration descent methods,” IEEE Trans. on Autom. Control, vol. 62, no. 9, pp. 4499–4511, 2017.
  • [28] M. Gharbi, B. Gharesifard, and C. Ebenbauer, “Anytime proximity moving horizon estimation: Stability and regret,” arXiv preprint arXiv:2006.14303, 2020.
  • [29] J. D. Schiller, S. Knüfer, and M. A. Müller, “Robust stability of suboptimal moving horizon estimation using an observer-based candidate solution,” arXiv preprint arXiv:2011.08723, 2020.
  • [30] E. D. Sontag, “Comments on integral variants of ISS,” Systems & Control Letters, vol. 34, no. 1, pp. 93–100, 1998.