This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Enhancing interval observers for state estimation using constraints thanks: This work was funded by Novartis Pharmaceuticals as part of the Novartis-MIT Center for Continuous Manufacturing.

Stuart M. Harwood    Paul I. Barton Process Systems Engineering Laboratory, Massachusetts Institute of Technology, Cambridge, MA 02139 ([email protected])
Abstract

This work considers the problem of calculating an interval-valued state estimate for a nonlinear system subject to bounded inputs and measurement errors. Such state estimators are often called interval observers. Interval observers can be constructed using methods from reachability theory. Recent advances in the construction of interval enclosures of reachable sets for nonlinear systems inspire the present work. These advances can incorporate constraints on the states to produce tighter interval enclosures. When applied to the state estimation problem, bounded-error measurements may be used as state constraints in these new theories. The result is a method that is easily implementable and which generally produces better, tighter interval state estimates. Furthermore, a novel linear programming-based method is proposed for calculating the observer gain, which must be tuned in practice. In contrast with previous approaches, this method does not rely on special system structure. The new approaches are demonstrated with numerical examples.

1 Introduction

This work considers the problem of bounded-error state estimation for nonlinear dynamic systems with continuous-time measurements. The goal of state estimation is to determine or estimate the internal state of a real system. While only certain states or functions of the states can be measured directly, a mathematical model of the system is available. When the system is a dynamic one, a history of measurements is typically available. Estimating the state using a dynamic model and measurements goes back to the Kalman filter [12], where a more statistical estimate of the state is obtained. More recently, work has focused on estimates in the form of guaranteed bounds on the state in the presence of bounded uncertainties and measurement errors [5, 11, 19, 21, 26]. This work continues in this direction.

Interval observers are a type of state estimator most closely related to the present work. The approach to constructing interval observers taken in [19, 18, 20, 21] is similar to the theory that we will use in the proposed approach. Underpinning these methods are theorems for general estimation of reachable sets. These theories are in the vein of comparison theorems involving differential inequalities. Such theorems have a long history, going back to “Müller’s theorem” [27], which subsequently was generalized to control systems in [6]. Recent work from [9, 7, 8, 24] develops these kinds of comparison theorems further. In particular, they incorporate state constraints into the bounding theory in a fundamental way; this information typically produces a tighter estimate of the reachable set without significant extra computational effort.

While the work in [19, 18, 20, 21] relies on the classic Müller’s theorem or the theory of cooperative systems, the present work depends on the more recent method for reachability analysis in [8]. When this theory is applied to state estimation, bounded error measurements take the role of state constraints, and tighter state estimates result.

Related literature includes work on general reachability analysis with constraints, which has been addressed in [16]. In that work, the focus is on linear systems and ellipsoidal enclosures of the reachable sets. Related work deals with control problems with state constraints [1, 15], in which the theoretical basis of the approaches is on the Hamilton-Jacobi-Bellman partial differential equation (PDE); the authors of [15] note that the solution of such a PDE is in general complicated for nonlinear systems.

The contributions and structure of this work are as follows. Section 2 discusses notation and formally states the problem of interest. Section 3 briefly discusses some of the ideas behind interval state estimates, and then proceeds to detail the proposed method. Again, the novelty of the proposed state estimation method is the incorporation of measurements as constraints and applying recent advances in reachability analysis. Further, a new procedure for calculating the “gain” of the observer system is discussed. In contrast with previous work, this method does not rely on any special system structure such as cooperativity. Section 4 demonstrates a numerical implementation of the new method for state estimation. Comparing it with related methods, it is shown that better, tighter state estimates can be obtained with the new method. Section 5 concludes with some final remarks.

2 Problem statement and preliminaries

2.1 Notation

First, notation that is used in this work includes lowercase bold letters for vectors and uppercase bold letters for matrices. The transposes of a vector 𝐯\mathbf{v} and matrix 𝐌\mathbf{M} are denoted 𝐯T\mathbf{v}^{\textnormal{T}} and 𝐌T\mathbf{M}^{\textnormal{T}}, respectively. The exception is that 𝟎\mathbf{0} may denote either a matrix or vector of zeros, but it should be clear from context what the appropriate dimensions are. Similarly, 𝟏\mathbf{1} denotes a vector of ones whose dimension should be clear from context. The jthj^{th} component of a vector 𝐯\mathbf{v} will be denoted vj{v}_{j}. For a matrix 𝐌p×n\mathbf{M}\in\mathbb{R}^{p\times n}, the notation 𝐌=[𝐦iT]\mathbf{M}=[\mathbf{m}_{i}^{\textnormal{T}}] may be used to emphasize that the ithi^{th} row of 𝐌\mathbf{M} is 𝐦i\mathbf{m}_{i}, for i{1,,p}i\in\{1,\dots,p\}. Similarly, 𝐌=[mi,j]\mathbf{M}=[m_{i,j}] emphasizes that the element in the ithi^{th} row and jthj^{th} column is mi,jm_{i,j}. Inequalities between vectors hold componentwise. For 𝐯\mathbf{v}, 𝐰n\mathbf{w}\in\mathbb{R}^{n} such that 𝐯𝐰\mathbf{v}\leq\mathbf{w}, [𝐯,𝐰][\mathbf{v},\mathbf{w}] denotes a nonempty interval in n\mathbb{R}^{n}. For a set YnY\subset\mathbb{R}^{n}, denote the set of nonempty interval subsets of YY by 𝕀Y\mathbb{I}Y. The equivalence of norms on n\mathbb{R}^{n} is used often; when a statement or result holds for any choice of norm, it is denoted \left\|\cdot\right\|. In some cases, it is useful to reference a specific norm, in which case it is subscripted; for instance, 1\left\|\cdot\right\|_{1} denotes the 1-norm. The dual norm of a norm \left\|\cdot\right\| is denoted \left\|\cdot\right\|_{*}. In a metric space, a neighborhood of a point xx is denoted N(x)N(x) and refers to an open ball centered at xx with some nonzero radius.

2.2 Problem statement

The problem of interest is one in estimating the state of a continuous-time dynamic system; in particular, we seek a rigorous enclosure of the states or a bounded error estimate. The dynamic model of the system is defined by: for positive integers nx,ny,nun_{x},n_{y},n_{u}, nonempty interval T=[t0,tf]T=[t_{0},t_{f}]\subset\mathbb{R}, DxnxD_{x}\subset\mathbb{R}^{n_{x}}, DunuD_{u}\subset\mathbb{R}^{n_{u}}, 𝐟:T×Du×Dxnx\mathbf{f}:T\times D_{u}\times D_{x}\to\mathbb{R}^{n_{x}}, and 𝐂ny×nx\mathbf{C}\in\mathbb{R}^{n_{y}\times n_{x}}. The system obeys the following equations:

𝐱˙(t)\displaystyle\dot{\mathbf{x}}(t) =𝐟(t,𝐮(t),𝐱(t)),a.e.tT,\displaystyle=\mathbf{f}(t,\mathbf{u}(t),\mathbf{x}(t)),\quad a.e.\;t\in T, (1a)
𝐲(t)\displaystyle\mathbf{y}(t) =𝐂𝐱(t)+𝐯(t),tT,\displaystyle=\mathbf{C}\mathbf{x}(t)+\mathbf{v}(t),\quad\forall t\in T, (1b)

where 𝐱(t)\mathbf{x}(t) is the state of the system, 𝐲(t)\mathbf{y}(t) is the measurement or output of the system, 𝐮(t)\mathbf{u}(t) is a disturbance to the system, and 𝐯(t)\mathbf{v}(t) is the measurement noise.

In this work, we exclusively consider absolutely continuous solutions 𝐱\mathbf{x} of the IVP in ODEs (1a). Further, we assume that it is known that the initial conditions satisfy 𝐱(t0)X0𝕀Dx\mathbf{x}(t_{0})\in X_{0}\in\mathbb{I}D_{x}, 𝐯\mathbf{v} is measurable and 𝐯(t)V(t)𝕀ny\mathbf{v}(t)\in V(t)\in\mathbb{I}\mathbb{R}^{n_{y}} for all tTt\in T, 𝐮\mathbf{u} is measurable and 𝐮(t)U(t)𝕀Du\mathbf{u}(t)\in U(t)\in\mathbb{I}D_{u} for all tTt\in T, where V(t)KvV(t)\subset K_{v} and U(t)KuU(t)\subset K_{u}, for all tTt\in T, for compact KvK_{v} and KuK_{u}. In other words, interval enclosures of the possible initial conditions, measurement noise, and disturbance values are known. A fundamental assumption on the dynamics 𝐟\mathbf{f} is a kind of local Lipschitz condition: For any 𝐳Dx\mathbf{z}\in D_{x}, there exists a neighborhood N(𝐳)N(\mathbf{z}) and αL1(T)\alpha\in L^{1}(T) such that for almost every tTt\in T and every 𝐰U(t)\mathbf{w}\in U(t)

𝐟(t,𝐰,𝐳1)𝐟(t,𝐰,𝐳2)α(t)𝐳1𝐳2,\left\|\mathbf{f}(t,\mathbf{w},\mathbf{z}_{1})-\mathbf{f}(t,\mathbf{w},\mathbf{z}_{2})\right\|\leq\alpha(t)\left\|\mathbf{z}_{1}-\mathbf{z}_{2}\right\|,

for every 𝐳1,𝐳2N(𝐳)Dx\mathbf{z}_{1},\mathbf{z}_{2}\in N(\mathbf{z})\cap D_{x}.

The challenge, then is to construct an interval [𝐱L(tf),𝐱U(tf)][\mathbf{x}^{L}(t_{f}),\mathbf{x}^{U}(t_{f})] such that

𝐱(tf)[𝐱L(tf),𝐱U(tf)]\mathbf{x}(t_{f})\in[\mathbf{x}^{L}(t_{f}),\mathbf{x}^{U}(t_{f})]

for any possible realization of noise 𝐯\mathbf{v}, disturbance 𝐮\mathbf{u}, and initial condition.

For future reference, since VV and UU are interval-valued, we can write V(t)=[𝐯L(t),𝐯U(t)]V(t)=[\mathbf{v}^{L}(t),\mathbf{v}^{U}(t)] and U(t)=[𝐮L(t),𝐮U(t)]U(t)=[\mathbf{u}^{L}(t),\mathbf{u}^{U}(t)] for appropriate functions.

3 State estimation

3.1 Motivation

Previous work on bounded-error state estimation includes interval observers; we refer to [4] for a recent review of such work and some related topics. The essence of this type of method is easily understood for a linear system. Consider

𝐱˙(t)\displaystyle\dot{\mathbf{x}}(t) =𝐀𝐱(t),\displaystyle=\mathbf{A}\mathbf{x}(t), (2)
𝐲(t)\displaystyle\mathbf{y}(t) =𝐂𝐱(t)+𝐯(t).\displaystyle=\mathbf{C}\mathbf{x}(t)+\mathbf{v}(t). (3)

Then for any matrix 𝐋nx×ny\mathbf{L}\in\mathbb{R}^{n_{x}\times n_{y}}, we can write

𝐱˙(t)=(𝐀𝐋𝐂)𝐱(t)+𝐋𝐂𝐱(t),\dot{\mathbf{x}}(t)=(\mathbf{A}-\mathbf{LC})\mathbf{x}(t)+\mathbf{LC}\mathbf{x}(t),

and subsequently

𝐱˙(t)=(𝐀𝐋𝐂)𝐱(t)𝐋𝐯(t)+𝐋𝐲(t).\dot{\mathbf{x}}(t)=(\mathbf{A}-\mathbf{LC})\mathbf{x}(t)-\mathbf{L}\mathbf{v}(t)+\mathbf{L}\mathbf{y}(t).

The theory of interval observers often focuses on the case that the “gain matrix” 𝐋\mathbf{L} is chosen so that 𝐀𝐋𝐂\mathbf{A}-\mathbf{LC} has positive off-diagonal components, and that 𝐰vN\left\|\mathbf{w}\right\|_{\infty}\leq v^{N}, for all 𝐰V(t)\mathbf{w}\in V(t), for all tt (so that 𝐋𝐯(t)𝐋𝐯(t)𝟏𝐋vN𝟏\mathbf{L}\mathbf{v}(t)\leq\left\|\mathbf{L}\mathbf{v}(t)\right\|_{\infty}\mathbf{1}\leq\left\|\mathbf{L}\right\|_{\infty}v^{N}\mathbf{1}, implying 𝐋vN𝟏𝐋𝐯(t)𝐋vN𝟏-\left\|\mathbf{L}\right\|_{\infty}v^{N}\mathbf{1}\leq\mathbf{L}\mathbf{v}(t)\leq\left\|\mathbf{L}\right\|_{\infty}v^{N}\mathbf{1}). In this case one can write

𝐱˙L(t)\displaystyle\dot{\mathbf{x}}^{L}(t) =(𝐀𝐋𝐂)𝐱L(t)𝐋vN𝟏+𝐋𝐲(t),\displaystyle=(\mathbf{A}-\mathbf{LC})\mathbf{x}^{L}(t)-\left\|\mathbf{L}\right\|_{\infty}v^{N}\mathbf{1}+\mathbf{L}\mathbf{y}(t),
𝐱˙U(t)\displaystyle\dot{\mathbf{x}}^{U}(t) =(𝐀𝐋𝐂)𝐱U(t)+𝐋vN𝟏+𝐋𝐲(t).\displaystyle=(\mathbf{A}-\mathbf{LC})\mathbf{x}^{U}(t)+\left\|\mathbf{L}\right\|_{\infty}v^{N}\mathbf{1}+\mathbf{L}\mathbf{y}(t).

Then assuming X0[𝐱L(t0),𝐱U(t0)]X_{0}\subset[\mathbf{x}^{L}(t_{0}),\mathbf{x}^{U}(t_{0})], the theory of monotone dynamic systems ensures that for any initial condition and realization of measurement noise, 𝐱(t)[𝐱L(t),𝐱U(t)]\mathbf{x}(t)\in[\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)], for all tTt\in T.

Our approach is to extend this idea with more general theorems for reachability analysis based on the theory of differential inequalities, and in particular, to leverage the recent developments that take advantage of constraint information to tighten the estimates of the reachable sets. To this end, we interpret the state estimation problem as a reachability problem for an IVP in constrained ODEs. This has been considered most recently in [8]. Similar techniques from [9, 7, 24] also apply; those articles use “a priori enclosures” of the reachable set, which are in effect constraints, although they do not necessarily exclude any solutions. As demonstrated in [8], these techniques can be applied to constrained systems, if one is only interested in solutions that also satisfy the constraints.

Thus, the proposed approach is to apply reachability analysis to the system

𝐱˙(t)\displaystyle\dot{\mathbf{x}}(t) =𝐟(t,𝐮(t),𝐱(t))𝐋𝐂𝐱(t)𝐋𝐯(t)+𝐋𝐲(t),a.e.tT,\displaystyle=\mathbf{f}(t,\mathbf{u}(t),\mathbf{x}(t))-\mathbf{LC}\mathbf{x}(t)-\mathbf{L}\mathbf{v}(t)+\mathbf{L}\mathbf{y}(t),\quad a.e.\;t\in T,
𝐱(t)\displaystyle\mathbf{x}(t) Xc(t)tT,\displaystyle\in X_{c}(t)\quad\forall t\in T,

for some choice of matrix 𝐋\mathbf{L}, and constraint mapping XcX_{c} defined by Xc(t)={𝐳:𝐲(t)𝐯U(t)𝐂𝐳𝐲(t)𝐯L(t)}X_{c}(t)=\{\mathbf{z}:\mathbf{y}(t)-\mathbf{v}^{U}(t)\leq\mathbf{C}\mathbf{z}\leq\mathbf{y}(t)-\mathbf{v}^{L}(t)\}.

Recent developments in the design of interval observers from [5, 26] allow the gain matrix 𝐋\mathbf{L} to vary, depending on time and measurements. There is no significant hurdle to incorporating such a generalization in the method that follows; we choose not to do so in order to keep notation simple and to highlight the essence of our contribution.

3.2 Proposed method

To state the method, we must introduce a few definitions and operations. Central to the method for estimating reachable sets that we will use, from [24], is the ability to overestimate or bound the dynamics. We will do so with an inclusion function constructed using principles from interval arithmetic. See [22] for an introduction to interval arithmetic and inclusion functions. As in the discussion above, the dynamics we care about are for a modified system depending on the gain matrix 𝐋\mathbf{L}.

Definition 1.

For 𝐋nx×ny\mathbf{L}\in\mathbb{R}^{n_{x}\times n_{y}}, define

𝐟^:(t,𝐮,𝐳,𝐯;𝐋)𝐟(t,𝐮,𝐳)𝐋𝐂𝐳𝐋𝐯,\widehat{\mathbf{f}}:(t,\mathbf{u},\mathbf{z},\mathbf{v};\mathbf{L})\mapsto\mathbf{f}(t,\mathbf{u},\mathbf{z})-\mathbf{LC}\mathbf{z}-\mathbf{L}\mathbf{v},

and let 𝐟^L\widehat{\mathbf{f}}^{L} and 𝐟^U\widehat{\mathbf{f}}^{U} be the endpoints of an inclusion monotonic interval extension of 𝐟^(,,,;𝐋)\widehat{\mathbf{f}}(\cdot,\cdot,\cdot,\cdot;\mathbf{L}).

The critical property of an inclusion monotonic interval extension is that it is an inclusion function; that is

𝐟^L([t,t],U,Z,V;𝐋)𝐟^(t,𝐮,𝐳,𝐯;𝐋)𝐟^U([t,t],U,Z,V;𝐋)\widehat{\mathbf{f}}^{L}([t,t],U^{\prime},Z^{\prime},V^{\prime};\mathbf{L})\leq\widehat{\mathbf{f}}(t,\mathbf{u},\mathbf{z},\mathbf{v};\mathbf{L})\leq\widehat{\mathbf{f}}^{U}([t,t],U^{\prime},Z^{\prime},V^{\prime};\mathbf{L})

for any U𝕀DuU^{\prime}\in\mathbb{I}D_{u}, Z𝕀DxZ^{\prime}\in\mathbb{I}D_{x}, V𝕀nyV^{\prime}\in\mathbb{IR}^{n_{y}}, and (t,𝐮,𝐳,𝐯)T×U×Z×V(t,\mathbf{u},\mathbf{z},\mathbf{v})\in T\times U^{\prime}\times Z^{\prime}\times V^{\prime}. Inclusion functions for a broad class of functions can be automatically and cheaply evaluated with a number of numerical libraries implementing interval arithmetic, for instance, MC++ [2], INTLAB [23], or PROFIL [14]. The proposed method relies on these automatically computable inclusion functions, which reduces the amount of analysis that must be performed compared to other work. For instance, the work in [19, 18] requires analysis of the dynamics to derive a hybrid automata which is then used to construct the state estimate.

Since it will be useful later, we mention that the natural interval extension of a function is an inclusion monotonic interval extension. For a linear function g:𝐳𝐚T𝐳g:\mathbf{z}\mapsto\mathbf{a}^{\textnormal{T}}\mathbf{z}, the natural interval extension [gL,gU][g^{L},g^{U}] is computed by applying the rules of interval arithmetic:

[viL,viU]\displaystyle[v_{i}^{L},v_{i}^{U}] ={[aiziL,aiziU]if ai0,[aiziU,aiziL]else,\displaystyle=\begin{cases}[a_{i}z_{i}^{L},a_{i}z_{i}^{U}]&\text{if }a_{i}\geq 0,\\ [a_{i}z_{i}^{U},a_{i}z_{i}^{L}]&\text{else},\end{cases}
[gL,gU]\displaystyle[g^{L},g^{U}] =[iviL,iviU].\displaystyle=[{\textstyle{\sum}}_{i}v_{i}^{L},{\textstyle{\sum}}_{i}v_{i}^{U}].

We will also require the following definition. If 𝐯𝐰\mathbf{v}\leq\mathbf{w}, then BiL(𝐯,𝐰)B_{i}^{L}(\mathbf{v},\mathbf{w}) returns the ithi^{th} lower face of the interval [𝐯,𝐰][\mathbf{v},\mathbf{w}]. If [𝐯,𝐰][\mathbf{v},\mathbf{w}] is empty, then a nonempty interval [𝐯^,𝐰^][\widehat{\mathbf{v}},\widehat{\mathbf{w}}] is constructed and the faces of that interval are returned.

Definition 2.

Define for each i{1,2,,nx}i\in\{1,2,\dots,n_{x}\}

BiL\displaystyle B_{i}^{L} :(𝐯,𝐰){𝐳[𝐯^,𝐰^]:zi=v^i},\displaystyle:(\mathbf{v},\mathbf{w})\mapsto\{\mathbf{z}\in[\widehat{\mathbf{v}},\widehat{\mathbf{w}}]:z_{i}=\widehat{v}_{i}\},
BiU\displaystyle B_{i}^{U} :(𝐯,𝐰){𝐳[𝐯^,𝐰^]:zi=w^i},\displaystyle:(\mathbf{v},\mathbf{w})\mapsto\{\mathbf{z}\in[\widehat{\mathbf{v}},\widehat{\mathbf{w}}]:z_{i}=\widehat{w}_{i}\},

where 𝐯^\widehat{\mathbf{v}}, 𝐰^\widehat{\mathbf{w}} are given componentwise by v^j=min{vj,(vj+wj)/2}\widehat{v}_{j}=\min\{v_{j},(v_{j}+w_{j})/2\} and w^j=max{wj,(vj+wj)/2}\widehat{w}_{j}=\max\{w_{j},(v_{j}+w_{j})/2\}.

The operation defined in Algorithm 1 is required. Originally from Definition 4 in [24], this algorithm defines the operation ItI_{t}, which tightens an interval [𝐯,𝐰][\mathbf{v},\mathbf{w}] by excluding points which cannot satisfy a given set of linear constraints 𝐌𝐳𝐝\mathbf{M}\mathbf{z}\leq\mathbf{d}. Specifically, the discussion in §5.2 of [24] establishes that the tightened interval It([𝐯,𝐰],𝐝;𝐌)I_{t}([\mathbf{v},\mathbf{w}],\mathbf{d};\mathbf{M}) satisfies

{𝐳[𝐯,𝐰]:𝐌𝐳𝐝}It([𝐯,𝐰],𝐝;𝐌)[𝐯,𝐰].\{\mathbf{z}\in[\mathbf{v},\mathbf{w}]:\mathbf{M}\mathbf{z}\leq\mathbf{d}\}\subset I_{t}([\mathbf{v},\mathbf{w}],\mathbf{d};\mathbf{M})\subset[\mathbf{v},\mathbf{w}].
Algorithm 1 Definition of the interval-tightening operator ItI_{t}
0:  positive integers nm,nn_{m},n, 𝐌=[mi,j]nm×n\mathbf{M}=[m_{i,j}]\in\mathbb{R}^{n_{m}\times n}, 𝐝nm\mathbf{d}\in\mathbb{R}^{n_{m}}, (𝐯,𝐰)n×n(\mathbf{v},\mathbf{w})\in\mathbb{R}^{n}\times\mathbb{R}^{n}, 𝐯𝐰\mathbf{v}\leq\mathbf{w}
  (𝐯^,𝐰^)(𝐯,𝐰)(\widehat{\mathbf{v}},\widehat{\mathbf{w}})\leftarrow(\mathbf{v},\mathbf{w})
  for  i{1,,nm}i\in\{1,\dots,n_{m}\}  do
     for  j{1,,n}j\in\{1,\dots,n\}  do
        if  mi,j0m_{i,j}\neq 0  then
            γmedian{v^j,w^j,1/mi,j(di+kjmax{mi,kv^k,mi,kw^k})}\gamma\leftarrow\textnormal{median}{}\left\{\widehat{v}_{j},\widehat{w}_{j},\nicefrac{{1}}{{m_{i,j}}}\big{(}d_{i}+\sum_{k\neq j}\max\{-m_{i,k}\widehat{v}_{k},-m_{i,k}\widehat{w}_{k}\}\big{)}\right\}
           if  mi,j>0m_{i,j}>0  then
               w^jγ\widehat{w}_{j}\leftarrow\gamma
           end if
           if  mi,j<0m_{i,j}<0  then
               v^jγ\widehat{v}_{j}\leftarrow\gamma
           end if
        end if
     end for
  end for
  return  It([𝐯,𝐰],𝐝;𝐌)[𝐯^,𝐰^]I_{t}([\mathbf{v},\mathbf{w}],\mathbf{d};\mathbf{M})\leftarrow[\widehat{\mathbf{v}},\widehat{\mathbf{w}}]

With the definition of the interval tightening operator, we can define an operation specific to our problem which tightens an interval based on the constraints 𝐂𝐳𝐲(t)𝐯L(t)\mathbf{C}\mathbf{z}\leq\mathbf{y}(t)-\mathbf{v}^{L}(t), 𝐂𝐳𝐲(t)𝐯U(t)\mathbf{C}\mathbf{z}\geq\mathbf{y}(t)-\mathbf{v}^{U}(t).

Definition 3.

Let

𝐝:(t,𝐲)[𝐲𝐯L(t)𝐲+𝐯U(t)],𝐌=[𝐂𝐂].\mathbf{d}:(t,\mathbf{y})\mapsto\begin{bmatrix}\mathbf{y}-\mathbf{v}^{L}(t)\\ -\mathbf{y}+\mathbf{v}^{U}(t)\end{bmatrix},\qquad\mathbf{M}=\begin{bmatrix}\mathbf{C}\\ -\mathbf{C}\end{bmatrix}.

Define

Ic:([𝐯,𝐰],t,𝐲)It([𝐯,𝐰],𝐝(t,𝐲);𝐌).I_{c}:([\mathbf{v},\mathbf{w}],t,\mathbf{y})\mapsto I_{t}([\mathbf{v},\mathbf{w}],\mathbf{d}(t,\mathbf{y});\mathbf{M}).

We can now state the specific method for constructing a state estimator. The main difference between this method and a classic comparison theorem for reachability (see [6]) is the application of the IcI_{c} operator. While the classic method would overstimate, for instance, f^i\widehat{f}_{i} on the set {𝐳[𝐱L(t),𝐱U(t)]:zi=xiU(t)}\{\mathbf{z}\in[\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)]:z_{i}=x_{i}^{U}(t)\} (the ithi^{th} upper face of the interval), the proposed method applies IcI_{c} to that set, and overestimates f^i\widehat{f}_{i} on the resulting (no larger) interval.

Theorem 1.

For any 𝐋nx×ny\mathbf{L}\in\mathbb{R}^{n_{x}\times n_{y}}, let (𝐱L,𝐱U)(\mathbf{x}^{L},\mathbf{x}^{U}) be absolutely continuous and satisfy

x˙iL(t)\displaystyle\dot{x}_{i}^{L}(t) =f^iL([t,t],U(t),Ic(BiL(𝐱L(t),𝐱U(t)),t,𝐲(t)),V(t);𝐋)+𝐋𝐲(t),a.e.tT,\displaystyle=\widehat{f}_{i}^{L}([t,t],U(t),I_{c}(B_{i}^{L}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)),t,\mathbf{y}(t)),V(t);\mathbf{L})+\mathbf{L}\mathbf{y}(t),\quad a.e.\;t\in T,
x˙iU(t)\displaystyle\dot{x}_{i}^{U}(t) =f^iU([t,t],U(t),Ic(BiU(𝐱L(t),𝐱U(t)),t,𝐲(t)),V(t);𝐋)+𝐋𝐲(t),a.e.tT,\displaystyle=\widehat{f}_{i}^{U}([t,t],U(t),I_{c}(B_{i}^{U}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)),t,\mathbf{y}(t)),V(t);\mathbf{L})+\mathbf{L}\mathbf{y}(t),\quad a.e.\;t\in T,

for each i{1,2,,nx}i\in\{1,2,\dots,n_{x}\}, and with initial conditions [𝐱L(t0),𝐱U(t0)]=X0[\mathbf{x}^{L}(t_{0}),\mathbf{x}^{U}(t_{0})]=X_{0}. Then 𝐱(t)[𝐱L(t),𝐱U(t)]\mathbf{x}(t)\in[\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)] for all tTt\in T; that is, [𝐱L(t),𝐱U(t)][\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)] is an interval estimate of 𝐱(t)\mathbf{x}(t) for all tt.

Proof.

As mentioned earlier, the claim follows by applying an established method for estimating reachable sets for the system

𝐱˙(t)=𝐟(t,𝐮(t),𝐱(t))𝐋𝐂𝐱(t)𝐋𝐯(t)+𝐋𝐲(t),a.e.tT,\dot{\mathbf{x}}(t)=\mathbf{f}(t,\mathbf{u}(t),\mathbf{x}(t))-\mathbf{LC}\mathbf{x}(t)-\mathbf{L}\mathbf{v}(t)+\mathbf{L}\mathbf{y}(t),\quad a.e.\;t\in T,

where 𝐱(t0)X0\mathbf{x}(t_{0})\in X_{0}, (𝐮(t),𝐯(t))U(t)×V(t)(\mathbf{u}(t),\mathbf{v}(t))\in U(t)\times V(t). The specific method is from [24, §5.2]. With the assumptions on the problem setting in §2.2 and the definitions of the dynamics defining (𝐱L,𝐱U)(\mathbf{x}^{L},\mathbf{x}^{U}), all the assumptions and hypotheses of the method from [24] are satisfied, and the conclusion holds that 𝐱(t)[𝐱L(t),𝐱U(t)]\mathbf{x}(t)\in[\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)] for any solution 𝐱\mathbf{x} satisfying 𝐲(t)𝐯U(t)𝐂𝐱(t)𝐲(t)𝐯L(t)\mathbf{y}(t)-\mathbf{v}^{U}(t)\leq\mathbf{C}\mathbf{x}(t)\leq\mathbf{y}(t)-\mathbf{v}^{L}(t). ∎

We note that Theorem 1 does not guarantee the existence of 𝐱L\mathbf{x}^{L} and 𝐱U\mathbf{x}^{U}; it provides a construction that, if successful, yields an interval estimate. Conditions that we have not stated explicitly, such as continuity of 𝐟^L\widehat{\mathbf{f}}^{L} and 𝐟^U\widehat{\mathbf{f}}^{U} and measurability of 𝐯L\mathbf{v}^{L}, 𝐯U\mathbf{v}^{U}, 𝐮L\mathbf{u}^{L}, 𝐮U\mathbf{u}^{U} (defining the set-valued mappings VV and UU), may be required to apply standard existence results for the solutions of IVPs in ODEs. Perhaps more important are the typically stronger conditions that guarantee the applicability of numerical methods for the solution of IVPs in ODEs; see, for instance, Theorem 1.1 of Section 1.4 of [17]. For example, in the numerical examples that follow, we use measurements 𝐲\mathbf{y} that are a continuous function on TT.

Comparisons

In §4, we will compare the proposed method established above against its variants and others from the literature. The method from Theorem 1 with 𝐋=𝟎\mathbf{L}=\mathbf{0} is called the “No Measurements” method; while it still uses measurement information as constraints, it does not use the measurement values 𝐲(t)\mathbf{y}(t) directly. Meanwhile, the “No Constraints” method is a variant of the proposed method that does not use the constraint information; that is, the endpoints of the interval estimate satisfy the differential equations

x˙iL(t)\displaystyle\dot{x}_{i}^{L}(t) =f^iL([t,t],U(t),BiL(𝐱L(t),𝐱U(t)),V(t);𝐋)+𝐋𝐲(t),i,\displaystyle=\widehat{f}_{i}^{L}([t,t],U(t),B_{i}^{L}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)),V(t);\mathbf{L})+\mathbf{L}\mathbf{y}(t),\quad\forall i,
x˙iU(t)\displaystyle\dot{x}_{i}^{U}(t) =f^iU([t,t],U(t),BiU(𝐱L(t),𝐱U(t)),V(t);𝐋)+𝐋𝐲(t),i.\displaystyle=\widehat{f}_{i}^{U}([t,t],U(t),B_{i}^{U}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)),V(t);\mathbf{L})+\mathbf{L}\mathbf{y}(t),\quad\forall i.

Note that application of the operation IcI_{c} has been omitted.

3.3 Linear stability analysis and automatic gain calculation

Methods from the literature for calculating the gain matrix 𝐋\mathbf{L} often focus on the case that the system is linear, e.g. 𝐱˙=𝐀𝐱\dot{\mathbf{x}}=\mathbf{A}\mathbf{x}. The methods often involve finding a gain matrix 𝐋\mathbf{L} so that 𝐀𝐋𝐂\mathbf{A}-\mathbf{LC} has nonnegative off-diagonal components; a matrix with this last property is often called a Metzler matrix. Consequently, the theory of cooperative or monotone systems can be applied to derive interval estimates.

In the present section, our goal is to state a method for calculating the gain matrix that does not depend on the system being cooperative. We begin with a fundamental result which will motivate the method; we show that the No Constraints method with a gain matrix satisfying certain constraints will produce an asymptotically exact interval estimate for an idealized linear system.

Theorem 2.

Suppose that the system of interest has a homogeneous, linear time invariant form, with exact measurements:

𝐱˙(t)\displaystyle\dot{\mathbf{x}}(t) =𝐀𝐱(t),a.e.t,\displaystyle=\mathbf{A}\mathbf{x}(t),\quad a.e.\;t,
𝐲(t)\displaystyle\mathbf{y}(t) =𝐂𝐱(t),t.\displaystyle=\mathbf{C}\mathbf{x}(t),\quad\forall t.

Consider the No Constraints method for calculating an interval state estimate; assume that the natural interval extension is used to calculate the inclusion monotonic interval extension function required in Definition 1. Let the columns of 𝐂\mathbf{C} be 𝐂i\mathbf{C}_{i}. Then for any gain matrix 𝐋=[iT]\mathbf{L}=[\boldsymbol{\ell}_{i}^{\textnormal{T}}] which satisfies

ai,iiT𝐂i+ji|ai,jiT𝐂j|<0,i,{a}_{i,i}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{i}+\sum_{j\neq i}\left|{a}_{i,j}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{j}\right|<0,\quad\forall i, (4)

the interval state estimate [𝐱L,𝐱U][\mathbf{x}^{L},\mathbf{x}^{U}] resulting from the No Constraints method is asymptotically exact; i.e. 𝐱U(t)𝐱L(t)𝟎\mathbf{x}^{U}(t)-\mathbf{x}^{L}(t)\to\mathbf{0} as t+t\to+\infty.

Proof.

For any 𝐋\mathbf{L}, let 𝐌=𝐀𝐋𝐂\mathbf{M}=\mathbf{A}-\mathbf{LC} (clearly 𝐌\mathbf{M} depends on 𝐋\mathbf{L} but we will suppress this dependence in the notation). Let 𝐌=[𝐦iT]\mathbf{M}=[\mathbf{m}_{i}^{\textnormal{T}}] so that the state estimates (𝐱L,𝐱U)(\mathbf{x}^{L},\mathbf{x}^{U}) resulting from the No Constraints method satisfy the following differential equations for each ii and almost every tt:

x˙iL(t)\displaystyle\dot{x}_{i}^{L}(t) =𝐦iT𝐁iL(𝐱L(t),𝐱U(t))+𝐋𝐲(t),\displaystyle=\mathbf{m}_{i}^{\textnormal{T}}\mathbf{B}_{i}^{L}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t))+\mathbf{L}\mathbf{y}(t), (5)
x˙iU(t)\displaystyle\dot{x}_{i}^{U}(t) =𝐦iT𝐁iU(𝐱L(t),𝐱U(t))+𝐋𝐲(t).\displaystyle=\mathbf{m}_{i}^{\textnormal{T}}\mathbf{B}_{i}^{U}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t))+\mathbf{L}\mathbf{y}(t). (6)

Here, 𝐁iL:2nn\mathbf{B}_{i}^{L}:\mathbb{R}^{2n}\to\mathbb{R}^{n} is a linear mapping whose effect is the following: if 𝐳L=𝐁iL(𝐱L,𝐱U),\mathbf{z}^{L}=\mathbf{B}_{i}^{L}(\mathbf{x}^{L},\mathbf{x}^{U}), then

zjL={xiLif j=i,xjLif ji and mi,j0,xjUelse (ji and mi,j<0).z_{j}^{L}=\begin{cases}x_{i}^{L}&\text{if }j=i,\\ x_{j}^{L}&\text{if }j\neq i\text{ and }m_{i,j}\geq 0,\\ x_{j}^{U}&\text{else }(j\neq i\text{ and }m_{i,j}<0).\end{cases}

Simply, 𝐦iT𝐁iL(𝐱L(t),𝐱U(t))\mathbf{m}_{i}^{\textnormal{T}}\mathbf{B}_{i}^{L}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)) is exactly the value of the lower bound of the natural interval extension of 𝐱𝐦iT𝐱\mathbf{x}\mapsto\mathbf{m}_{i}^{\textnormal{T}}\mathbf{x} on the set BiL(𝐱L(t),𝐱U(t))B_{i}^{L}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)), as we would expect of the No Constraints method. Similarly, if 𝐳U=𝐁iU(𝐱L,𝐱U),\mathbf{z}^{U}=\mathbf{B}_{i}^{U}(\mathbf{x}^{L},\mathbf{x}^{U}), then

zjU={xiUif j=i,xjUif ji and mi,j0,xjLelse (ji and mi,j<0),z_{j}^{U}=\begin{cases}x_{i}^{U}&\text{if }j=i,\\ x_{j}^{U}&\text{if }j\neq i\text{ and }m_{i,j}\geq 0,\\ x_{j}^{L}&\text{else }(j\neq i\text{ and }m_{i,j}<0),\end{cases}

and again, 𝐦iT𝐁iU(𝐱L(t),𝐱U(t))\mathbf{m}_{i}^{\textnormal{T}}\mathbf{B}_{i}^{U}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)) equals the upper bound of the natural interval extension of 𝐱𝐦iT𝐱\mathbf{x}\mapsto\mathbf{m}_{i}^{\textnormal{T}}\mathbf{x} on the set BiU(𝐱L(t),𝐱U(t))B_{i}^{U}(\mathbf{x}^{L}(t),\mathbf{x}^{U}(t)).

Then we have that for each ii

x˙iU(t)x˙iL(t)=mi,i(xiU(t)xiL(t))+ji|mi,j|(xjU(t)xjL(t)),\dot{x}_{i}^{U}(t)-\dot{x}_{i}^{L}(t)=m_{i,i}(x_{i}^{U}(t)-x_{i}^{L}(t))+\sum_{j\neq i}\left|m_{i,j}\right|(x_{j}^{U}(t)-x_{j}^{L}(t)),

or in matrix form,

𝐱˙U(t)𝐱˙L(t)=𝐌~(𝐱U(t)𝐱L(t))\dot{\mathbf{x}}^{U}(t)-\dot{\mathbf{x}}^{L}(t)=\widetilde{\mathbf{M}}(\mathbf{x}^{U}(t)-\mathbf{x}^{L}(t))

where each off-diagonal component of 𝐌~\widetilde{\mathbf{M}} equals the absolute value of the corresponding component of 𝐌\mathbf{M}, and the diagonals of 𝐌~\widetilde{\mathbf{M}} and 𝐌\mathbf{M} are equal.

If every eigenvalue of 𝐌~\widetilde{\mathbf{M}} has negative real part, then 𝐱U(t)𝐱L(t)𝟎\mathbf{x}^{U}(t)-\mathbf{x}^{L}(t)\to\mathbf{0} as t+t\to+\infty (see for instance [13, Thm. 3.5]. Gershgorin’s circle theorem [25, §7.4] provides a way to bound the eigenvalues; for any eigenvalue λ\lambda, there is some ii such that

|λm~i,i|ji|m~i,j|.\left|\lambda-\widetilde{m}_{i,i}\right|\leq\sum_{j\neq i}\left|\widetilde{m}_{i,j}\right|.

Since each mi,im_{i,i} is real, if we require that for all ii

m~i,i+ji|m~i,j|<0,\widetilde{m}_{i,i}+\sum_{j\neq i}\left|\widetilde{m}_{i,j}\right|<0,

then we can be sure that all eigenvalues of 𝐌~\widetilde{\mathbf{M}} have negative real part.

Of course, we realize that these conditions simplify to mi,i+ji|mi,j|<0{m}_{i,i}+\sum_{j\neq i}\left|{m}_{i,j}\right|<0 for all ii. Finally, recall the definition 𝐌=𝐀𝐋𝐂\mathbf{M}=\mathbf{A}-\mathbf{LC}, let the rows of 𝐋\mathbf{L} be iT\boldsymbol{\ell}_{i}^{\textnormal{T}}, and let the columns of 𝐂\mathbf{C} be 𝐂i\mathbf{C}_{i}. Then these conditions become

ai,iiT𝐂i+ji|ai,jiT𝐂j|<0.{a}_{i,i}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{i}+\sum_{j\neq i}\left|{a}_{i,j}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{j}\right|<0.

Related results are in the literature, like [4, Thm. 1]. Stability/asymptotic properties of the interval state estimate are often stated as requiring 𝐀𝐋𝐂\mathbf{A}-\mathbf{LC} to have eigenvalues with negative real part; it is clear, again using Gershgorin’s circle theorem, that conditions (4), if satisfied, imply exactly this. Of course, it is not the stability of the modified system 𝐱˙=(𝐀𝐋𝐂)𝐱+𝐋𝐲\dot{\mathbf{x}}=(\mathbf{A}-\mathbf{LC})\mathbf{x}+\mathbf{L}\mathbf{y} that matters – this is mostly a coincidence; it is the stability of the dynamics defining 𝐱U𝐱L\mathbf{x}^{U}-\mathbf{x}^{L} that determines their asymptotic properties.

To turn the conditions in Theorem 2 into an implementable numerical procedure, the next result states that a gain matrix satisfying the conditions may be found as the solution of a linear program (LP).

Corollary 3.

Let the columns of 𝐂\mathbf{C} be 𝐂i\mathbf{C}_{i}. Consider the LP

min𝐋,𝐁,s\displaystyle\min_{\mathbf{L},\mathbf{B},s}\; s\displaystyle s (7)
s.t. iT𝐂jbi,jai,j,i,j:ji,\displaystyle\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{j}-b_{i,j}\leq a_{i,j},\quad\forall i,\forall j:j\neq i,
\displaystyle- iT𝐂jbi,jai,j,i,j:ji,\displaystyle\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{j}-b_{i,j}\leq-a_{i,j},\quad\forall i,\forall j:j\neq i,
\displaystyle- iT𝐂i+jibi,jsai,i,i,\displaystyle\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{i}+{\textstyle{\sum}}_{j\neq i}b_{i,j}-s\leq-a_{i,i},\quad\forall i,

where the variables are 𝐋=[iT]\mathbf{L}=[\boldsymbol{\ell}_{i}^{\textnormal{T}}], 𝐁=[bi,j]\mathbf{B}=[b_{i,j}], and ss. Any 𝐋\mathbf{L} which is a solution of this LP with optimal objective value s<0s^{*}<0 satisfies conditions (4).

Proof.

Note that the first two sets of constraints imply bi,j|ai,jiT𝐂j|b_{i,j}\geq\left|a_{i,j}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{j}\right|, for all iji\neq j. Then, with the last set of constraints, we get

ai,iiT𝐂i+ji|ai,jiT𝐂j|ai,iiT𝐂i+jibi,js,a_{i,i}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{i}+{\textstyle{\sum}}_{j\neq i}\left|a_{i,j}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{j}\right|\leq a_{i,i}-\boldsymbol{\ell}_{i}^{\textnormal{T}}\mathbf{C}_{i}+{\textstyle{\sum}}_{j\neq i}b_{i,j}\leq s,

for each ii. Thus, if the optimal ss^{*} is negative, we certainly have that a corresponding solution 𝐋\mathbf{L} satisfies conditions (4). ∎

As a practical note, we would add a (negative) lower bound on ss to the above LP in order to prevent the possibility that the LP is unbounded. Further, the variables bi,ib_{i,i} do not appear in the constraints or objective and could be eliminated; a decent numerical LP solver will likely identify this automatically and eliminate them in a presolve step.

Other results for calculating the gain matrix based on the solution of an LP have been proposed; see for instance [3]. Again, these results rely on something like 𝐀𝐋𝐂\mathbf{A}-\mathbf{LC} being Metzler, although this extra structure permits more detailed claims about the input-output gains of the system.

4 Examples

We evaluate the performance of the proposed estimation method on some examples. At the heart of the method is the solution of the IVP in ODEs in Theorem 1. This initial value problem is solved with a C/C++ code employing the CVODE component of the SUNDIALS suite [10]. Specifically, the numerical method uses the implementation of the Backwards Differentiation Formulae (BDF) in CVODE, using a Newton iteration for the corrector, with relative and absolute integration tolerances equal to 10910^{-9}. The implementation of interval arithmetic in MC++ [2] produces the inclusion function in Definition 1.

The proposed method will be referred to as GMAC, indicating that it incorporates Gained Measurement And Constraint information when constructing the state estimate. The No Constraints and No Measurement methods will be compared (recall discussion following Theorem 1).

4.1 Bioreactor

We consider an example involving a bioreactor originally from [20, 19]. The dynamic model describes the evolution in time of the concentrations of biomass and feed substrate. The dynamic equations on the time domain T=[0,20]T=[0,20] (day) are

x˙(t)\displaystyle\dot{x}(t) =(μ0(t)s(t)s(t)+ks+s(t)2/kiαD(t))x(t),\displaystyle=\left(\mu_{0}(t)\frac{s(t)}{s(t)+k_{s}+s(t)^{2}/k_{i}}-\alpha D(t)\right)x(t), x(0)[0,10] (mmol/L),\displaystyle x(0)\in[0,10]\text{ (mmol/L)},
s˙(t)\displaystyle\dot{s}(t) =kμ0(t)x(t)s(t)s(t)+ks+s(t)2/ki+D(t)(sin(t)s(t)),\displaystyle=-k\mu_{0}(t)x(t)\frac{s(t)}{s(t)+k_{s}+s(t)^{2}/k_{i}}+D(t)(s_{in}(t)-s(t)), s(0)[0,100] (mmol/L),\displaystyle s(0)\in[0,100]\text{ (mmol/L)},

where x(t)x(t) and s(t)s(t) are the biomass and substrate concentrations, respectively, at time tt, and the known parameters are

(ks,ki)=(9.28,256) (mmol/L),\displaystyle(k_{s},k_{i})=(9.28,256)\text{ (mmol/L)},
(k,α)=(42.14,0.5),\displaystyle(k,\alpha)=(42.14,0.5),
D(t)={2 (day1),if t[0,5],0.5 (day1),if t(5,10],1.067 (day1),if t(10,20].\displaystyle D(t)=\begin{cases}2\text{ (day}^{-1}),&\text{if }t\in[0,5],\\ 0.5\text{ (day}^{-1}),&\text{if }t\in(5,10],\\ 1.067\text{ (day}^{-1}),&\text{if }t\in(10,20].\end{cases}

Meanwhile, the unknown parameters/disturbances are (μ0,sin)(\mu_{0},s_{in}) which satisfy for all tTt\in T

μ0(t)\displaystyle\mu_{0}(t) [0.703,0.777] (day1),\displaystyle\in[0.703,0.777]\text{ (day}^{-1}),
sin(t)\displaystyle s_{in}(t) [0.95s^in(t),1.05s^in(t)] (mmol/L),\displaystyle\in[0.95\widehat{s}_{in}(t),1.05\widehat{s}_{in}(t)]\text{ (mmol/L)},

where s^in:t50+15cos(t/5)\widehat{s}_{in}:t\mapsto 50+15\cos(t/5) (so take U:t[0.703,0.777]×[0.95s^in(t),1.05s^in(t)]U:t\mapsto[0.703,0.777]\times[0.95\widehat{s}_{in}(t),1.05\widehat{s}_{in}(t)]).

Continuous measurements of x(t)x(t), the biomass concentration, are available (so 𝐂=[1 0]\mathbf{C}=[1\;0]). The error111We note that the error in the initial measurements of the biomass concentration are much larger than the subsequent measurements. While in practice it seems likely that the initial measurements should be known at least as accurately as any subsequent online measurements, for consistency we follow this example as it appears in [20, 19]. in these measurements satifies v(t)[0.25,0.25]v(t)\in[-0.25,0.25]. For simulation purposes, we obtain measurements from a numerical solution with μ0(t)=0.74\mu_{0}(t)=0.74 for all tt, sin(t)=s^in(t)s_{in}(t)=\widehat{s}_{in}(t) for all tt, v(t)=0v(t)=0 for all tt, and initial conditions x(0)=5x(0)=5 and s(0)=40s(0)=40. These measurements are obtained at 500 equally spaced time points from the interval TT, and then made continuous by linearly interpolating between them.

In [20], a “bundle” of interval observers are constructed, essentially by running independent estimates with different gain matrices 𝐋\mathbf{L}. We choose one of these gain matrices 𝐋=[2 0]T\mathbf{L}=[2\;0]^{\textnormal{T}}.

Figure 1 summarizes the results; the upper and lower bounds of the interval estimate for each state is plotted versus time. The proposed GMAC method is strictly tighter than either the No Measurements or No Constraints methods for at least one of the states. The interval estimates at the final time are given in Table 1.

Compared to the bundle of observers from [20], the estimates from GMAC are of comparable quality. As mentioned, however, we achieve our results with only one observer gain matrix. The same example is considered in [19], and again we achieve similar state estimates. The work in [19] is based on a hybrid automata approach, and some extra analysis to determine the switching conditions is required. Meanwhile, we rely on automatic construction of inclusion functions through implementations of interval arithmetic and little extra analysis is required.

Refer to caption
Figure 1: State estimates for the bioreactor system from §4.1: biomass (top) and substrate (bottom). The upper bounds for biomass, xx, from the No Measurements and No Constraints methods are mostly off the bounds of the figure. The estimates from the proposed GMAC and No Measurements methods are nearly indistinguishable for the substrate, ss.
Table 1: State estimates from §4.1 for various methods at final time: (x(20),s(20))[xL,xU]×[sL,sU](x(20),s(20))\in[x^{L},x^{U}]\times[s^{L},s^{U}]
method [xL,xU][x^{L},x^{U}] [sL,sU][s^{L},s^{U}]
GMAC [0.449,1.19][0.449,1.19] [17.4,30.3][17.4,30.3]
No Measurements [0,10400][0,10400] [17.4,30.3][17.4,30.3]
No Constraints [0.398,314000][0.398,314000] [0,32.1][0,32.1]

4.2 Linearized system

This example comes from Example 1 of [4]. We have a three state system with nonlinear dynamics; 𝐱\mathbf{x} satisfies

𝐱˙(t)=[200143134]𝐱(t)+[2u1(t)x1(t)x2(t)β(t)0u2(t)x1(t)x2(t)β(t)],a.e.t,\dot{\mathbf{x}}(t)=\begin{bmatrix}2&0&0\\ 1&-4&\sqrt{3}\\ -1&-\sqrt{3}&-4\end{bmatrix}\mathbf{x}(t)+\begin{bmatrix}-2u_{1}(t)x_{1}(t)x_{2}(t)\beta(t)\\ 0\\ u_{2}(t)x_{1}(t)x_{2}(t)\beta(t)\\ \end{bmatrix},\quad a.e.\;t,

where

β:t1+sin(2t).\beta:t\mapsto 1+\sin(2t).

The time interval of interest is T=[0,5]T=[0,5], with the state known at t=0t=0: 𝐱(0)=(1,1,0)\mathbf{x}(0)=(1,1,0). The disturbances satisfy (u1(t),u2(t))[4.48,6.12]×[3.2,3.6](u_{1}(t),u_{2}(t))\in[4.48,6.12]\times[3.2,3.6]. Measurements of the first state are available, so that 𝐂=[1 0 0]\mathbf{C}=[1\;0\;0]. The error in these measurements satisfies v(t)[0.1,0.1]v(t)\in[-0.1,0.1]. For simulation purposes, we obtain measurements from a numerical solution of the system with u1(t)=5.3u_{1}(t)=5.3 for all tt, u2(t)=3.4u_{2}(t)=3.4 for all tt, and v(t)=0.1sin(10t)v(t)=0.1\sin(10t). These measurements are obtained at 500 equally spaced time points from the interval TT, and then made continuous by linearly interpolating between them.

Although this is a nonlinear system, we attempt to design an observer gain matrix based on the linear part. The original analysis in [4, Example 1] suggests

𝐋=𝐋1[3  0  0]T.\mathbf{L}=\mathbf{L}_{1}\equiv[3\;\;0\;\;0]^{\textnormal{T}}.

However, solving LP (7) yields

𝐋=𝐋2[4.27  11]T,\mathbf{L}=\mathbf{L}_{2}\equiv[4.27\;\;1\;\;{-}1]^{\textnormal{T}},

and a negative optimal objective value is obtained. We will calculate state estimates with both separately. Refer to these as “gain 1” and “gain 2,” respectively. Recall that the No Measurements method is defined by using 𝐋=𝟎\mathbf{L}=\mathbf{0}, and so it remains the same.

Figure 2 summarizes the results; the upper and lower bounds of the interval estimate for each state is plotted versus time. For this example, omitting constraints produces a much more conservative interval state estimate. In fact, the bounds produced by the No Constraints method with gain 1 diverge and the numerical integration procedure prematurely terminates shortly after t=3.5t=3.5 (specifically, the corrector iteration of the BDF in CVODE fails). Using gain 2, the No Constraints method performs much better and can produce state estimates for the entire time interval.

Meanwhile, the proposed GMAC method (with either gain- the results are largely the same) more consistently produces a tight interval estimate for each state. The No Measurements method is comparable. Table 2 lists specific values at the final time. The main difference is that the estimate of the first state from GMAC is much tighter. This is somewhat of a moot point, however, as bounded-error measurements of the first state are directly available; for our simulated measurement values, this implies the interval estimate x1(5)[0.692,0.892]x_{1}(5)\in[0.692,0.892].

Refer to caption
Figure 2: Interval state estimates over time for each state in the example from §4.2: x1x_{1} (top), x2x_{2} (middle), x3x_{3} (bottom). The estimates from the proposed GMAC method and No Measurements method are nearly indistinguishable for x2x_{2} and x3x_{3}. “Gain i” refers to using a gain matrix 𝐋=𝐋i\mathbf{L}=\mathbf{L}_{i} (for ii equal to 1 or 2). GMAC with gain 1 is nearly identical to GMAC with gain 2.
Table 2: State estimates from §4.2 for various methods at final time: 𝐱(5)[𝐱L,𝐱U]\mathbf{x}(5)\in[\mathbf{x}^{L},\mathbf{x}^{U}]
method [x1L,x1U][x_{1}^{L},x_{1}^{U}] [x2L,x2U][x_{2}^{L},x_{2}^{U}] [x3L,x3U][x_{3}^{L},x_{3}^{U}]
GMAC (gain 2) [0.504,1.20][0.504,1.20] [0.0178,0.182][0.0178,0.182] [0.248,0.0250][-0.248,-0.0250]
No Measurements [0.000852,113][0.000852,113] [0.0179,0.182][0.0179,0.182] [0.248,0.0251][-0.248,-0.0251]
No Constraints (gain 2) [0.350,1.96][0.350,1.96] [0.0919,0.472][-0.0919,0.472] [0.502,0.564][-0.502,0.564]

5 Conclusions

This work has considered the problem of bounded-error state estimation for nonlinear systems. Using recent work for computing the reachable sets of constrained dynamic systems, we proposed a method that uses measurements in a novel way to construct interval state estimates. Numerical examples demonstrate that the method is effective in practice and can improve on existing approaches. The proposed approach also benefits from its reliance on interval arithmetic and its implementation in numerical libraries, which reduces the amount of manual analysis of a specific problem one must perfom. Finally, we proposed a new way of calculating the gain matrix and showed that it can produce asymptotically exact interval estimates in ideal cases.

References

  • [1] Olivier Bokanowski, Nicolas Forcadel, and Hasnaa Zidani. Reachability and minimal times for state constrained nonlinear problems without any controllability assumption. SIAM Journal on Control and Optimization, 48(7):4292–4316, 2010.
  • [2] Benoît Chachuat. MC++: A Versatile Library for McCormick Relaxations and Taylor Models. http://www.imperial.ac.uk/people/b.chachuat/research.html, 2015.
  • [3] Stanislav Chebotarev, Denis Efimov, Tarek Raïssi, and Ali Zolghadri. Interval observers for continuous-time lpv systems with l1/l2 performance. Automatica, 58:82–89, 2015.
  • [4] Denis Efimov and Tarek Raïssi. Design of interval observers for uncertain dynamical systems. Automation and Remote Control, 77(2):191–225, 2016.
  • [5] Denis Efimov, Tarek Raïssi, Stanislav Chebotarev, and Ali Zolghadri. Interval state observer for nonlinear time varying systems. Automatica, 49(1):200–205, 2013.
  • [6] Gary W. Harrison. Dynamic models with uncertain parameters. In X .J. R. Avula, editor, Proceedings of the First International Conference on Mathematical Modeling, volume 1, pages 295–304. University of Missouri Rolla, 1977.
  • [7] Stuart M. Harwood and Paul I. Barton. Efficient polyhedral enclosures for the reachable set of nonlinear control systems. Mathematics of Control, Signals, and Systems, 28(1):8, 2016.
  • [8] Stuart M. Harwood and Paul I. Barton. Affine relaxations for the solutions of constrained parametric ordinary differential equations. Optimal Control Applications and Methods, 2017.
  • [9] Stuart M. Harwood, Joseph K Scott, and Paul I. Barton. Bounds on reachable sets using ordinary differential equations with linear programs embedded. IMA Journal of Mathematical Control and Information, 33(2):519–541, 2016.
  • [10] Alan C. Hindmarsh, Peter N. Brown, Keith E. Grant, Steven L. Lee, Radu Serban, Dan E. Shumaker, and Carol S. Woodward. SUNDIALS: Suite of Nonlinear and Differential/Algebraic Equation Solvers. ACM Transactions on Mathematical Software, 31(3):363–396, 2005.
  • [11] L. Jaulin. Nonlinear bounded-error state estimation of continuous-time systems. Automatica, 38(6):1079–1082, 2002.
  • [12] R. E. Kalman. A New Approach to Linear Filtering and Prediction Problems. Journal of Basic Engineering, 82:35–45, 1960.
  • [13] Hassan K. Khalil. Nonlinear Systems. Prentice-Hall, Upper Saddle River, NJ, second edition, 1996.
  • [14] O. Knüppel. PROFIL/BIAS — A fast interval library. Computing, 53(3):277–287, Sep 1994.
  • [15] A. B. Kurzhanski, I. M. Mitchell, and P. Varaiya. Optimization techniques for state-constrained control and obstacle problems. Journal of Optimization Theory and Applications, 128(3):499–521, 2006.
  • [16] A. B. Kurzhanski and P. Varaiya. Ellipsoidal techniques for reachability under state constraints. SIAM Journal on Control and Optimization, 45(4):1369–1394, January 2006.
  • [17] J. D. Lambert. Numerical Methods for Ordinary Differential Systems: The Initial Value Problem. John Wiley & Sons, New York, 1991.
  • [18] N. Meslem and N. Ramdani. Interval observer design based on nonlinear hybridization and practical stability analysis. International Journal of Adaptive Control and Signal Processing, 25:228–248, 2011.
  • [19] N. Meslem, N. Ramdani, and Y. Candau. Using hybrid automata for set-membership state estimation with uncertain nonlinear continuous-time systems. Journal of Process Control, 20:481–489, 2010.
  • [20] M. Moisan and O. Bernard. Interval observers for non-montone systems. Application to bioprocess models. 16th IFAC World Congress, 16:43–48, 2005.
  • [21] M. Moisan and O. Bernard. An interval observer for non-monotone systems: application to an industrial anaerobic digestion process. 10th International IFAC Symposium on Computer Applications in Biotechnology, 1:325–330, 2007.
  • [22] Ramon E. Moore, R. Baker Kearfott, and Michael J. Cloud. Introduction to Interval Analysis. SIAM, Philadelphia, 2009.
  • [23] Siegfried M. Rump. INTLAB — INTerval LABoratory. In Tibor Csendes, editor, Developments in Reliable Computing, pages 77–104. Springer Netherlands, Dordrecht, 1999.
  • [24] Joseph K. Scott and Paul I. Barton. Bounds on the reachable sets of nonlinear control systems. Automatica, 49(1):93–100, 2013.
  • [25] Gilbert Strang. Linear Algebra and its Applications. Thomson Brooks/Cole, fourth edition, 2006.
  • [26] Rihab El Houda Thabet, Tarek Raïssi, Christophe Combastel, Denis Efimov, and Ali Zolghadri. An effective method to interval observer design for time-varying systems. Automatica, 50(10):2677–2684, 2014.
  • [27] Wolfgang Walter. Differential and Integral Inequalities. Springer, New York, 1970.