This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Adaptive Control of Time-Varying Parameter Systems with Asymptotic Tracking

Omkar Sudhir Patil, Runhan Sun, Shubhendu Bhasin and Warren E. Dixon Omkar Sudhir Patil, Runhan Sun, and Warren E. Dixon are with the Department of Mechanical and Aerospace Engineering, University of Florida, Gainesville FL 32611-6250 USA. Email: {patilomkarsudhir,runhansun,wdixon}@ufl.edu.Shubhendu Bhasin is with the Department of Electrical Engineering, Indian Institute of Technology Delhi, New Delhi, India (e-mail: [email protected]). This research is supported in part by NSF award number 1762829, Office of Naval Research Grant N00014-13-1-0151, AFOSR award number FA9550-18-1-0109 and FA9550-19-1-0169. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the sponsoring agency.
Abstract

A continuous adaptive control design is developed for nonlinear dynamical systems with linearly parameterizable uncertainty involving time-varying uncertain parameters. The key feature of this design is a robust integral of the sign of the error (RISE)-like term in the adaptation law which compensates for potentially destabilizing terms in the closed-loop error system arising from the time-varying nature of uncertain parameters. A Lyapunov-based stability analysis ensures asymptotic tracking, and boundedness of the closed-loop signals.

I Introduction

Adaptive control of nonlinear dynamical systems with time-varying uncertain parameters is an open and practically relevant problem. It has been well established that traditional gradient-based update laws can compensate for constant unknown parameters yielding asymptotic convergence. Moreover, the development of robust modifications of such adaptive update laws result in uniformly ultimately bounded (UUB) results for slowly varying parametric uncertainty using a Lyapunov-based analysis, under the assumption of bounded parameters and their time-derivatives (cf. [1]).

More recent results focus on tracking and parameter estimation performance improvement, though still limited to a UUB result, using various adaptive control approaches for systems with unknown time-varying parameters. One such approach involves a fast adaptation law [2], where a matrix of time-varying learning rates is utilized to improve the tracking and estimation performance under a finite excitation condition. Another approach uses a set-theoretic control architecture [3, 4, 5] to reject the effects of parameter variation, while restricting the system error within a prescribed performance bound. While the aforementioned approaches can potentially yield improved transient response, the results still yield UUB error systems.

Motivation exists to obtain asymptotic convergence of the tracking error to zero, despite the time-varying nature of the uncertain parameters. Robust adaptive control approaches such as [6] yield asymptotic adaptive tracking for systems with time-varying uncertain parameters; however, such approaches exploit high-gain feedback based on worst-case uncertainty, rather than an adaptive control approach that scales to compensate for the uncertainty without using worst-case gains. In [7], the iterative learning control result in [6] is extended to yield asymptotic tracking for systems with periodic time-varying parameters with known periodicity.

To the best of our knowledge, an asymptotic tracking result has not been achieved for a generalized class of nonlinear systems with unknown time-varying parameters, where the parameters are not necessarily periodic. Asymptotic tracking is difficult to achieve for the time-varying parameter case because the time-derivative of the parameter acts like an unknown exogenous disturbance in the parameter estimation dynamics, which is difficult to cancel with an adaptive update law in a Lyapunov-based stability analysis.

To illustrate this problem, consider the scalar dynamical system111Note that the system (1)(\text{\ref{eq:illussystem}}) is considered only for illustrative purpose. This paper presents result for a general system with a vector state and a linearly parameterizable uncertainty with time-varying parameters.

x˙(t)=a(t)x(t)+u(t),\dot{x}(t)=a(t)x(t)+u(t), (1)

with the controller u(t)=kx(t)a^(t)x(t)u(t)=-kx(t)-\hat{a}(t)x(t), where kk is a positive constant gain, a(t)a(t) is the unknown time-varying parameter, a^(t)\hat{a}(t) is the parameter estimate of a(t)a(t) and the parameter estimation error a~(t)\tilde{a}(t) is defined as a~(t)a(t)a^(t)\tilde{a}(t)\triangleq a(t)-\hat{a}(t). The traditional stability analysis approach for such problems is to consider the Lyapunov function candidate V(x(t),a~(t))=12x2(t)+12γa~2(t)V(x(t),\tilde{a}(t))=\frac{1}{2}x^{2}(t)+\frac{1}{2\gamma}\tilde{a}^{2}(t), where γ\gamma is a positive constant gain. The given definitions and controller yield the following time-derivative of the candidate Lyapunov function: V˙(t)=kx2(t)+a~(t)x2(t)+a~(t)γ(a˙(t)a^˙(t))\dot{V}(t)=-kx^{2}(t)+\tilde{a}(t)x^{2}(t)+\frac{\tilde{a}(t)}{\gamma}(\dot{a}(t)-\dot{\hat{a}}(t)). For the constant parameter case, i.e., a˙(t)=0\dot{a}(t)=0, the well-known adaptive update law a^˙(t)=γx2(t)\dot{\hat{a}}(t)=\gamma x^{2}(t) will cancel the cross term a~(t)x2(t)\tilde{a}(t)x^{2}(t) in V˙(t)\dot{V}(t). However, when the parameters are time-varying, it is unclear how to cancel or dominate a˙(t)\dot{a}(t) via an update law such that V˙(t)\dot{V}(t) becomes at least negative semi-definite. It would be desirable to have a sliding mode-like term based on a~(t)\tilde{a}(t) in the adaptation law, but a~(t)\tilde{a}(t) is unknown. Another approach could be to use a robust controller, e.g., u(t)=kx(t)a¯x(t)u(t)=-kx(t)-\bar{a}x(t), where a¯\bar{a} is a known constant upper bound on the norm of parameter |a(t)|\left|a(t)\right|, or an adaptive robust controller which involves certainty equivalence in terms of the unknown bound a¯\bar{a}. Either of these approaches would yield an asymptotic tracking result (cf., [6]), but, as stated earlier, these approaches are based on a high-gain worst case scenario, rather than an adaptive control approach that scales to compensate for the uncertainty without using worst-case gains.

A popular approach to design adaptive controllers for the time-varying parameter case is to consider robust modification of the update law and assume upper bounds of |a(t)|\left|a(t)\right| and a˙(t)\dot{a}(t) to obtain a UUB result. For instance, consider a standard gradient update law with sigma-modification [8], a^˙(t)=γx2(t)γσa^(t)\dot{\hat{a}}(t)=\gamma x^{2}(t)-\gamma\sigma\hat{a}(t), which yields V˙(t)=kx2(t)σa~2(t)+a~(t)(a˙(t)γ+σa(t))\dot{V}(t)=-kx^{2}(t)-\sigma\tilde{a}^{2}(t)+\tilde{a}(t)(\frac{\dot{a}(t)}{\gamma}+\sigma a(t)), implying a UUB result when the parameter a(t)a(t) and its time-derivative a˙(t)\dot{a}(t) are bounded. Moreover, the approaches developed in results such as [2] and [4] can be used to improve the transient response of the UUB error system.

The major challenge to achieve asymptotic stability is the derivative of the time-varying parameter term in the Lyapunov analysis, which is addressed in this paper with a Lyapunov-based design approach, that is inspired by the modular adaptive control approach in [9]. This approach includes higher order dynamics which appear after taking the time-derivative of (1). Since these higher order dynamics contain the time-derivative of the parameter estimate a^˙(t),\dot{\hat{a}}(t), it is possible to design a^˙(t)\dot{\hat{a}}(t) to facilitate the subsequent stability analysis. With this motivation, a continuous adaptive control algorithm is developed for nonlinear dynamical systems with linearly parameterized uncertainty involving time-varying parameters, where a semi-global asymptotic tracking result is achieved. A key feature of the proposed method is a robust integral of the sign of the error (RISE)-like (see [10, 11, 12, 9]) update law, i.e., the update law contains a signum function of the tracking error term multiplied by some desired regressor based terms. The update law also involves a projection algorithm to ensure that the parameter estimates stay within a bounded set. However, the projection algorithm introduces a potentially destabilizing term in the time-derivative of the Lyapunov function candidate, leading to an additional technical obstacle to obtain asymptotic tracking. This challenge is resolved by using an auxiliary term in the control input, which facilitates stability by providing a stabilizing term and canceling the aforementioned potentially destabilizing term in the time-derivative of the candidate Lyapunov function. With the proposed method, the closed-loop system dynamics have the same structure as previous RISE controllers [10, 11, 12, 9], for which the stability analysis tools are well established, yielding asymptotic convergence of the tracking error to zero, boundedness of the parameter estimation error, and boundedness of the closed-loop signals.

II Dynamic Model

Consider a control affine system with the nonlinear dynamics

x˙(t)=h(x(t),t)+d(t)+u(t),\dot{x}(t)=h(x(t),t)+d(t)+u(t), (2)

where x:[0,)nx:[0,\infty)\to\mathbb{R}^{n} denotes the state, h:n×[0,)nh:\mathbb{R}^{n}\times[0,\infty)\to\mathbb{R}^{n} denotes a continuously differentiable function, d:[0,)nd:[0,\infty)\to\mathbb{R}^{n} represents an exogenous disturbance acting on the system, and u:[0,)nu:[0,\infty)\to\mathbb{R}^{n} represents the control input. The function h(x(t),t)h(x(t),t) in (2)(\text{\ref{eq:dynamicsystem}}) is assumed to be linearly parameterized as

h(x(t),t)Yh(x(t),t)θf(t),h(x(t),t)\triangleq Y_{h}(x(t),t)\theta_{f}(t), (3)

where Yh:n×[0,)n×mY_{h}:\mathbb{R}^{n}\times[0,\infty)\to\mathbb{R}^{n\times m} is a known regression matrix, and θf:[0,)m\theta_{f}:[0,\infty)\to\mathbb{R}^{m} is a vector of time-varying unknown parameters.

The disturbance parameter vector d(t)d(t) can be appended to the θf(t)\theta_{f}(t) vector, yielding an augmented parameter vector θ:[0,)n+m\theta:[0,\infty)\to\mathbb{R}^{n+m} as

θ(t)[θf(t)d(t)],\theta(t)\triangleq\left[\begin{array}[]{c}\theta_{f}(t)\\ d(t)\end{array}\right], (4)

and the augmented regressor Y:n×[0,)n×(n+m)Y:\mathbb{R}^{n}\times[0,\infty)\to\mathbb{R}^{n\times(n+m)} can be designed as

Y(x(t),t)[Yh(x(t),t)In].Y(x(t),t)\triangleq\left[\begin{array}[]{cc}Y_{h}(x(t),t)&I_{n}\end{array}\right]. (5)

The parameterization in (4) and (5)(\text{\ref{eq:totparam}}) yields h(x(t),t)+d(t)=Y(x(t),t)θ(t)h(x(t),t)+d(t)=Y(x(t),t)\theta(t), so the dynamics in (2)(\text{\ref{eq:dynamicsystem}}) can be rewritten as

x˙(t)=Y(x(t),t)θ(t)+u(t).\dot{x}(t)=Y(x(t),t)\theta(t)+u(t). (6)
Assumption 1.

The time-varying augmented parameter θ(t)\theta(t) and its time-derivatives, i.e., θ˙(t)\dot{\theta}(t), θ¨(t)\ddot{\theta}(t) are bounded by known constants, i.e., θ(t)θ¯\left\|\theta(t)\right\|\leq\bar{\theta}, θ˙(t)ζ1,\left\|\dot{\theta}(t)\right\|\leq\zeta_{1}, and θ¨(t)ζ2\left\|\ddot{\theta}(t)\right\|\leq\zeta_{2}, where θ¯,ζ1,ζ2>0\bar{\theta},\zeta_{1},\zeta_{2}\in\mathbb{R}_{>0} are known bounding constants, and \left\|\cdot\right\| denotes the Euclidean norm.

III Control Design

III-A Control Objective

The objective is to design a controller such that the state tracks a smooth bounded reference trajectory, despite the time-varying nature of the uncertain parameters. The objective is quantified by defining the tracking error e:[0,)ne:[0,\infty)\to\mathbb{R}^{n} as222All function dependencies are suppressed equation (7)\left(\ref{eq:e}\right) onward; assume all variables to be time dependent unless stated otherwise.

exxd,e\triangleq x-x_{d}, (7)

where xd:[0,)nx_{d}:[0,\infty)\to\mathbb{R}^{n} is a reference trajectory.

Assumption 2.

The reference trajectory xd(t)x_{d}(t) is bounded and smooth, such that xd(t)x¯d\left\|x_{d}(t)\right\|\leq\bar{x}_{d}, x˙d(t)δ1\left\|\dot{x}_{d}(t)\right\|\leq\delta_{1}, and x¨d(t)δ2\left\|\ddot{x}_{d}(t)\right\|\leq\delta_{2}, where x¯d,δ1,δ2\bar{x}_{d},\delta_{1},\delta_{2}>0\in\mathbb{R}_{>0} are known bounding constants.

Substituting (6)(\text{\ref{eq:xdotsub1}}) into the time-derivative of (7)(\text{\ref{eq:e}}) yields

e˙=Yθ+ux˙d.\dot{e}=Y\theta+u-\dot{x}_{d}. (8)

To facilitate the subsequent analysis, a filtered tracking error r:[0,)nr:[0,\infty)\to\mathbb{R}^{n} is defined as

re˙+αe,r\triangleq\dot{e}+\alpha e, (9)

where α>0\alpha\in\mathbb{R}_{>0} is a constant control gain. Substituting (8)(\text{\ref{eq:edotsub1}}) into (9)(\text{\ref{eq:rdef}}) yields

r=Yθ+ux˙d+αe.r=Y\theta+u-\dot{x}_{d}+\alpha e. (10)

III-B Control and Update Law Development

From the subsequent stability analysis, the continuous control input is designed as

uYdθ^αe+x˙d+μ,u\triangleq-Y_{d}\hat{\theta}-\alpha e+\dot{x}_{d}+\mu, (11)

where YdY(xd(t),t)Y_{d}\triangleq Y(x_{d}(t),t) is the desired regression matrix, μ:[0,)n\mu:[0,\infty)\to\mathbb{R}^{n} is a subsequently defined auxiliary control term, and θ^:[0,)n+m\hat{\theta}:[0,\infty)\to\mathbb{R}^{n+m} denotes the parameter estimate of θ(t)\theta(t). Substituting the control input in (11)(\text{\ref{eq:controlinput}}) into the open-loop error system in (10)(\text{\ref{eq:r}}) yields the following closed-loop system

r=YθYdθ^+μ.r=Y\theta-Y_{d}\hat{\theta}+\mu. (12)

Adding and subtracting YdθY_{d}\theta in (12) yields

r=(YYd)θ+Ydθ~+μ,r=(Y-Y_{d})\theta+Y_{d}\tilde{\theta}+\mu, (13)

where θ~:[0,)n+m\tilde{\theta}:[0,\infty)\to\mathbb{R}^{n+m} denotes the parameter estimation error, i.e., θ~(t)θ(t)θ^(t)\tilde{\theta}(t)\triangleq\theta(t)-\hat{\theta}(t). Taking the time-derivative of (13)(\text{\ref{eq:rsub2}}) yields

r˙\displaystyle\dot{r} =(Y˙Y˙d)θ+(YYd)θ˙+Y˙dθ~+Ydθ˙Ydθ^˙+μ˙.\displaystyle=(\dot{Y}-\dot{Y}_{d})\theta+(Y-Y_{d})\dot{\theta}+\dot{Y}_{d}\tilde{\theta}+Y_{d}\dot{\theta}-Y_{d}\dot{\hat{\theta}}+\dot{\mu}. (14)

The control variables θ^˙(t)\dot{\hat{\theta}}(t) and μ˙(t)\dot{\mu}(t) now appear in the higher order dynamics in (14),(\text{\ref{eq:rdot}}), and these control variables are designed with the use of a continuous projection algorithm [13, Appendix E]. The projection algorithm constrains θ^(t)\hat{\theta}(t) to lie inside a bounded convex set ={θ(n+m)|θθ¯}\mathcal{B}=\{\theta\in\mathbb{R}^{(n+m)}|\left\|\theta\right\|\leq\bar{\theta}\} by switching the adaptation law to its component tangential to the boundary of the set \mathcal{B} when θ^(t)\hat{\theta}(t) reaches the boundary. A continuously differentiable convex function f:(n+m)f:\mathbb{R}^{(n+m)}\to\mathbb{R} is used to describe the boundaries of the bounded convex set \mathcal{B} such that f(θ(t))<0θ(t)<θ¯f(\theta(t))<0\,\,\mathbf{\forall}\,\left\|\theta(t)\right\|<\bar{\theta} and f(θ(t))=0θ(t)=θ¯f(\theta(t))=0\,\,\forall\,\left\|\theta(t)\right\|=\bar{\theta}. The adaptation law is then designed as

θ^˙\displaystyle\dot{\hat{\theta}} proj(Λ0(t))\displaystyle\triangleq\textrm{proj}(\Lambda_{0}(t))
={Λ0,||θ^||<θ¯(f(θ^))TΛ00Λ1,||θ^||θ¯(f(θ^))TΛ0>0,\displaystyle=\begin{cases}\Lambda_{0},&||\hat{\theta}||<\bar{\theta}\,\lor\,(\nabla f(\hat{\theta}))^{T}\Lambda_{0}\leq 0\\ \Lambda_{1},&||\hat{\theta}||\geq\bar{\theta}\,\land\,(\nabla f(\hat{\theta}))^{T}\Lambda_{0}>0,\end{cases} (15)

where ||θ^(0)||<θ¯,||\hat{\theta}(0)||<\bar{\theta}, \lor, \land denote the logical ‘or’, ‘and’ operators, respectively, \nabla represents the gradient operator, i.e., f(θ^)\nabla f(\hat{\theta})=[fϕ1fϕn+m]ϕ=θ^T\left[\begin{array}[]{ccc}\frac{\partial f}{\partial\phi_{1}}&\ldots&\frac{\partial f}{\partial\phi_{n+m}}\end{array}\right]_{\phi=\hat{\theta}}^{T}, and Λ0:[0,)n+m\Lambda_{0}:[0,\infty)\to\mathbb{R}^{n+m} and Λ1:[0,)n+m\Lambda_{1}:[0,\infty)\to\mathbb{R}^{n+m} are designed as333From Lemma 1 in the Appendix section, YdΓYdTY_{d}\Gamma Y_{d}^{T} is proven to be invertible, therefore it is reasonable to include (YdΓYdT)1(Y_{d}\Gamma Y_{d}^{T})^{-1} in the update law.

Λ0ΓYdT(YdΓYdT)1[βsgn(e)],\Lambda_{0}\triangleq\Gamma Y_{d}^{T}(Y_{d}\Gamma Y_{d}^{T})^{-1}\left[\beta\textrm{sgn}(e)\right], (16)
Λ1(Im+n(f(θ^))(f(θ^))T||f(θ^)||2)Λ0,\Lambda_{1}\triangleq\left(I_{m+n}-\frac{(\nabla f(\hat{\theta}))(\nabla f(\hat{\theta}))^{T}}{||\nabla f(\hat{\theta})||^{2}}\right)\Lambda_{0}, (17)

respectively. In (16)(\text{\ref{eq:lambda0}}) and (17)(\text{\ref{eq:lambda1}}), β>0\beta\in\mathbb{R}_{>0} is a constant gain, and Γ(n+m)×(n+m)\Gamma\in\mathbb{R}^{(n+m)\times(n+m)} is a positive-definite matrix with a block diagonal structure, i.e., Γ[Γ10m×n0n×mΓ2]\Gamma\triangleq\left[\begin{array}[]{cc}\Gamma_{1}&0_{m\times n}\\ 0_{n\times m}&\Gamma_{2}\end{array}\right], with Γ1m×m\Gamma_{1}\in\mathbb{R}^{m\times m}, Γ2n×n\Gamma_{2}\in\mathbb{R}^{n\times n}. The continuous auxiliary term μ(t)\mu(t), used in the control input in (11)(\text{\ref{eq:controlinput}}), acts as a stabilizing term in the Lyapunov analysis to account for the side effects of the projection, and is designed as a generalized solution to

μ˙{μ0,||θ^||<θ¯(f(θ^))TΛ00,μ1||θ^||θ¯(f(θ^))TΛ0>0,\dot{\mu}\triangleq\begin{cases}\mu_{0},&||\hat{\theta}||<\bar{\theta}\,\lor\,(\nabla f(\hat{\theta}))^{T}\Lambda_{0}\leq 0,\\ \mu_{1}&||\hat{\theta}||\geq\bar{\theta}\,\land\,(\nabla f(\hat{\theta}))^{T}\Lambda_{0}>0,\end{cases} (18)

where μ(0)=0,\mu(0)=0, and μ0:[0,)n\mu_{0}:[0,\infty)\to\mathbb{R}^{n} and μ1:[0,)n\mu_{1}:[0,\infty)\to\mathbb{R}^{n} are defined as μ0Kr\mu_{0}\triangleq-Kr and μ1μ0Yd(Λ0Λ1),\mu_{1}\triangleq\mu_{0}-Y_{d}\left(\Lambda_{0}-\Lambda_{1}\right), respectively. Substituting (15)\left(\ref{eq:adaptlaw}\right) and (18)\left(\ref{eq:riseterm}\right) in (14),\left(\ref{eq:rdot}\right), the closed-loop dynamics can be rewritten as

r˙=(Y˙Y˙d)θ+(YYd)θ˙+Y˙dθ~+Ydθ˙βsgn(e)Kr,\dot{r}=(\dot{Y}-\dot{Y}_{d})\theta+(Y-Y_{d})\dot{\theta}+\dot{Y}_{d}\tilde{\theta}+Y_{d}\dot{\theta}-\beta\,\textrm{sgn}(e)-Kr, (19)

for both cases, i.e., when ||θ^||<θ¯(f(θ^))TΛ00||\hat{\theta}||<\bar{\theta}\,\lor\,(\nabla f(\hat{\theta}))^{T}\Lambda_{0}\leq 0 or ||θ^||θ¯(f(θ^))TΛ0>0.||\hat{\theta}||\geq\bar{\theta}\,\land\,(\nabla f(\hat{\theta}))^{T}\Lambda_{0}>0. To facilitate the subsequent analysis, (19)(\text{\ref{eq:rdotsub2}}) can be rewritten as

r˙=N~+NBβsgn(e)Kre,\dot{r}=\widetilde{N}+N_{B}-\beta\,\textrm{sgn}(e)-Kr-e, (20)

where the variables N~:[0,)n\widetilde{N}:[0,\infty)\to\mathbb{R}^{n} and NB:[0,)nN_{B}:[0,\infty)\to\mathbb{R}^{n} are defined as

N~(Y˙Y˙d)θ+(YYd)θ˙+e,\widetilde{N}\triangleq(\dot{Y}-\dot{Y}_{d})\theta+(Y-Y_{d})\dot{\theta}+e,

and

NBYdθ˙+Y˙dθY˙dθ^,N_{B}\triangleq Y_{d}\dot{\theta}+\dot{Y}_{d}\theta-\dot{Y}_{d}\hat{\theta},

respectively. The Mean Value Theorem (MVT) can be used to develop the following upper bound on the term N~(t)\widetilde{N}(t)

||N~||ρ(||z||)||z||,||\widetilde{N}||\leq\rho(||z||)||z||, (21)

where z[rTeT]T2nz\triangleq\left[\begin{array}[]{cc}r^{T}&e^{T}\end{array}\right]^{T}\in\mathbb{R}^{2n} and ρ:2n\rho:\mathbb{R}^{2n}\to\mathbb{R} is a positive, globally invertible and non-decreasing function. By Assumption 1, Assumption 2, Corollary 1 in the Appendix, and the bounding effect of projection algorithm on θ^(t)\hat{\theta}(t), the term NB(t)N_{B}(t) and its time-derivative N˙B(t)\dot{N}_{B}(t) can be upper bounded by some constants γ1,\gamma_{1},γ2>0\gamma_{2}\in\mathbb{R}_{>0} as

||NB(t)||γ1,||N˙B(t)||γ2,||N_{B}(t)||\leq\gamma_{1},\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,||\dot{N}_{B}(t)||\leq\gamma_{2}, (22)

respectively.

IV Stability Analysis

Theorem 1.

The controller designed in (11)(\ref{eq:controlinput}) along with the adaptation laws designed in (15) and (18)(\text{\ref{eq:riseterm}}) ensure the closed-loop system is bounded and the tracking error e(t)0\left\|e(t)\right\|\to 0 as t,t\to\infty, provided that the gains α,β\alpha,\beta are selected such that the following condition is satisfied

β>γ1+γ2α.\beta>\gamma_{1}+\frac{\gamma_{2}}{\alpha}. (23)

Proof: Let 𝒟2n+1\mathcal{D}\subset\mathbb{R}^{2n+1} be an open connected set containing y(t)=0y(t)=0, where y:[0,)2n+1y:[0,\infty)\to\mathbb{R}^{2n+1} is defined as

y(t)[zT(t)P(t)]T.y(t)\triangleq\left[\begin{array}[]{cc}z^{T}(t)&\sqrt{P(t)}\end{array}\right]^{T}.

Let VL:𝒟×[0,)0V_{L}:\mathcal{D}\times[0,\infty)\to\mathbb{R}_{\geq 0} be a positive-definite candidate Lyapunov function defined as

VL(y(t),t)12rTr+12eTe+P,V_{L}(y(t),t)\triangleq\frac{1}{2}r^{T}r+\frac{1}{2}e^{T}e+P,

where P:[0,)P:[0,\infty)\to\mathbb{R} is a generalized solution to the differential equation

P˙(t)L(t),\text{$\dot{P}$}(t)\triangleq-L(t),\, (24)

where P(0)βi[=1]n|ei(0)|e(0)TNB(0)P(0)\triangleq\mathop{\beta\stackrel{{\scriptstyle[}}{{i}}=1]{n}{\sum}\left|e_{i}(0)\right|-e(0)^{T}N_{B}(0)} and

LrT(NBβsgn(e)).L\triangleq r^{T}(N_{B}-\beta\textrm{sgn}(e)). (25)
Remark 1.

Provided that the gain condition in (23)(\text{\ref{eq:gaincondition}}) is satisfied, P(t)0P(t)\geq 0.444See [10] for details. Hence it is valid to use P(t)P(t) in the candidate Lyapunov function as function of the variable P(t)\sqrt{P(t)}.

From (9)(\text{\ref{eq:rdef}}), (20) and (24)(\text{\ref{eq:Pdot}}), the differential equations describing the closed-loop system are

e˙\displaystyle\dot{e} =\displaystyle= rαe,\displaystyle r-\alpha e, (26)
r˙\displaystyle\dot{r} =\displaystyle= N~+NBβsgn(e)Kre,\displaystyle\widetilde{N}+N_{B}-\beta\,\textrm{sgn}(e)-Kr-e, (27)
P˙\displaystyle\dot{P} =\displaystyle= rT(NBβsgn(e)).\displaystyle-r^{T}(N_{B}-\beta\textrm{sgn}(e)). (28)

Let g:2n+1×[0,)2n+1g:\mathbb{R}^{2n+1}\times[0,\infty)\to\mathbb{R}^{2n+1} denote the right-hand side of (26)(\text{\ref{eq:edotsub2}})-(28)(\text{\ref{eq:Pdotsub1}}). Since g(y(t),t)g(y(t),t) is continuous almost everywhere, except in the set {(y(t),t)|e=0}\{(y(t),t)|e=0\}, an absolute continuous Filippov solution y(t)y(t) exists almost everywhere (a.e.), so that y˙(t)K[g](y(t),t)\dot{y}(t)\in K[g](y(t),t) a.e., except at the points in the set {(y(t),t)|e=0}\{(y(t),t)|e=0\}, where the Filippov set-valued map includes unique solutions. Using a generalized Lyapunov stability theory under the framework of Filippov solutions, a generalized time-derivative of the Lyapunov function VLV_{L} exists and V˙L(y,t)V~˙L(y,t)\dot{V}_{L}(y,t)\in\dot{\widetilde{V}}_{L}(y,t) , where

V~˙L(y,t)\displaystyle\dot{\widetilde{V}}_{L}(y,t) =\displaystyle= ξVL(y,t)ξTK[e˙Tr˙T12P12P˙]T\displaystyle\underset{\xi\in\partial V_{L}(y,t)}{\bigcap}\xi^{T}K\left[\begin{array}[]{ccc}\dot{e}^{T}&\dot{r}^{T}&\frac{1}{2}P^{-\frac{1}{2}}\dot{P}\end{array}\right]^{T} (30)
=\displaystyle= VLTK[e˙Tr˙T12P12P˙]T\displaystyle\nabla V_{L}^{T}K\left[\begin{array}[]{ccc}\dot{e}^{T}&\dot{r}^{T}&\frac{1}{2}P^{-\frac{1}{2}}\dot{P}\end{array}\right]^{T} (32)
\displaystyle\subset [eTrT2P12]×\displaystyle\left[\begin{array}[]{ccc}e^{T}&r^{T}&2P^{\frac{1}{2}}\end{array}\right]\times (36)
K[e˙Tr˙T12P12P˙]T,\displaystyle K\left[\begin{array}[]{ccc}\dot{e}^{T}&\dot{r}^{T}&\frac{1}{2}P^{-\frac{1}{2}}\dot{P}\end{array}\right]^{T},

where VL(y,t)\partial V_{L}(y,t) denotes Clarke’s generalized gradient [14]. Substituting (26)(\text{\ref{eq:edotsub2}})-(28)(\text{\ref{eq:Pdotsub1}}) into (36)(\text{\ref{eq:vltildedotsub2}}) yields

V~˙L\displaystyle\dot{\widetilde{V}}_{L} a.e.rT(N~+NBβsgn(e)Kre)\displaystyle\overset{a.e.}{\subset}r^{T}(\widetilde{N}+N_{B}-\beta\,\textrm{sgn}(e)-Kr-e)
+eT(rαe)rT(NBβsgn(e))\displaystyle+e^{T}(r-\alpha e)-r^{T}(N_{B}-\beta\textrm{sgn}(e)) (37)

where K[sgn(e)]=SGN(e)K\left[\textrm{sgn}(e)\right]=\textrm{SGN}(e) such that

SGN(ei)={{1},ei>0[1,1],ei=0{1},ei<0.\textrm{SGN}(e_{i})=\begin{cases}\left\{1\right\},&e_{i}>0\\ \left[-1,1\right],&e_{i}=0\\ \left\{-1\right\},&e_{i}<0.\end{cases}

Using (21)\left(\ref{eq:ntildebound}\right), the expression in (37)\left(\text{\ref{eq:vltildedotsub1}}\right) can be upper bounded as

V~˙La.e.ρ(z)zrKr2αe2.\dot{\widetilde{V}}_{L}\overset{a.e.}{\leq}\rho(\left\|z\right\|)\left\|z\right\|\left\|r\right\|-K\left\|r\right\|^{2}-\alpha e^{2}.

Using Young’s Inequality on ρ(z)zr\rho(\left\|z\right\|)\left\|z\right\|\left\|r\right\| yields ρ(z)zrρ2(z)z22+12r2\rho(\left\|z\right\|)\left\|z\right\|\left\|r\right\|\leq\frac{\rho^{2}(\left\|z\right\|)\left\|z\right\|^{2}}{2}+\frac{1}{2}\left\|r\right\|^{2}. Therefore,

V~˙L\displaystyle\dot{\widetilde{V}}_{L} a.e.\displaystyle\overset{a.e.}{\leq} ρ2(z)z22(K12)r2αe2\displaystyle\frac{\rho^{2}(\left\|z\right\|)\left\|z\right\|^{2}}{2}-(K-\frac{1}{2})\left\|r\right\|^{2}-\alpha e^{2} (38)
a.e.\displaystyle\overset{a.e.}{\leq} (λ3ρ2(z)2)z2,\displaystyle-\left(\lambda_{3}-\frac{\rho^{2}(\left\|z\right\|)}{2}\right)\left\|z\right\|^{2},

where λ3min{α,K12}>0\lambda_{3}\triangleq\min\{\alpha,K-\frac{1}{2}\}\in\mathbb{R}_{>0} is a known constant. The expression in (38)(\text{\ref{eq:VLtdot})} can be rewritten as

V˙La.e.cz2y𝒟,\dot{V}_{L}\overset{a.e.}{\leq}-c\left\|z\right\|^{2}\,\forall\,y\in\mathcal{D}, (39)

for some constant c>0c\in\mathbb{R}_{>0}, where

𝒟{y2n+1|yρ1(2λ3)}.\mathcal{D}\triangleq\left\{y\in\mathbb{R}^{2n+1}|\left\|y\right\|\leq\rho^{-1}\left(\sqrt{2\lambda_{3}}\right)\right\}.

In this region, λ3>ρ2(z)2\lambda_{3}>\frac{\rho^{2}(\left\|z\right\|)}{2}, so a constant cc satisfies (39), and larger values of λ3\lambda_{3} expand the size of 𝒟.\mathcal{D}. Furthermore, the relationship in (39)(\ref{eq:VLdot}) implies that VL(y(t),t),V_{L}(y(t),t)\in\mathcal{L}_{\infty}, hence e(t),r(t)e(t),r(t), P(t)P(t) \in\mathcal{L}_{\infty}. These facts along with the expression in (13)(\text{\ref{eq:rsub2}}), indicate that μ(t)\mu(t)\in\mathcal{L}_{\infty}. The parameter estimate θ^(t)\hat{\theta}(t)\in\mathcal{L}_{\infty} due to the projection operation. The state and its time-derivative, i.e., x(t),x˙(t)x(t),\dot{x}(t)\in\mathcal{L}_{\infty}, because e(t),r(t),xd(t),x˙d(t).e(t),r(t),x_{d}(t),\dot{x}_{d}(t)\in\mathcal{L}_{\infty}. Further the regression matrix Y(x(t),t)Y(x(t),t)\in\mathcal{L}_{\infty} since its a bounded function for a bounded argument x(t)x(t). Similarly, Yd(t)Y_{d}(t)\in\mathcal{L}_{\infty}, hence θ^˙\dot{\hat{\theta}}\in\mathcal{L}_{\infty} by Corollary 1. From the expression in (11), since θ^(t),e(t),x˙d(t),μ(t)\hat{\theta}(t),e(t),\dot{x}_{d}(t),\mu(t)\in\mathcal{L}_{\infty}, u(t)u(t)\in\mathcal{L}_{\infty}. Hence all the closed-loop signals are bounded.

Consider λ1y2VLλ2y2,\lambda_{1}\left\|y\right\|^{2}\leq V_{L}\leq\lambda_{2}\left\|y\right\|^{2}, where λ1,λ2>0\lambda_{1},\lambda_{2}\in\mathbb{R}_{>0}. To ensure zρ1(2λ3)\left\|z\right\|\leq\rho^{-1}(\sqrt{2\lambda_{3}}), it is sufficient to obtain the result from yρ1(2λ3)\left\|y\right\|\leq\rho^{-1}(\sqrt{2\lambda_{3}}). Since VLλ2y\sqrt{\frac{V_{L}}{\lambda_{2}}}\leq\left\|y\right\|, then VLλ2ρ1(2λ3)\sqrt{\frac{V_{L}}{\lambda_{2}}}\leq\rho^{-1}(\sqrt{2\lambda_{3}}), and VLV_{L} is non-increasing, so VL(t)VL(0)V_{L}(t)\leq V_{L}(0). Hence it sufficient to show that VL(0)λ2ρ1(2λ3)\sqrt{\frac{V_{L}(0)}{\lambda_{2}}}\leq\rho^{-1}(\sqrt{2\lambda_{3}}) to ensure that VLλ2ρ1(2λ3)\sqrt{\frac{V_{L}}{\lambda_{2}}}\leq\rho^{-1}(\sqrt{2\lambda_{3}}). Since λ1y(0)2VL(0)\lambda_{1}\left\|y(0)\right\|^{2}\leq V_{L}(0) implies y(0)VL(0)λ1λ2λ1ρ1(2λ3),\left\|y(0)\right\|\leq\sqrt{\frac{V_{L}(0)}{\lambda_{1}}}\leq\sqrt{\frac{\lambda_{2}}{\lambda_{1}}}\rho^{-1}(\sqrt{2\lambda_{3}}), so y𝒮{y(t)𝒟|y(t)λ2λ1ρ1(2λ3)}y\in\mathcal{S}\triangleq\left\{y(t)\in\mathcal{D}|y(t)\leq\sqrt{\frac{\lambda_{2}}{\lambda_{1}}}\rho^{-1}(\sqrt{2\lambda_{3}})\right\} is the region where y(0)y(0) should lie for guaranteed asymptotic stability. The gain condition λ3=min{α,K12}ρ2(λ1λ2y(0))2\lambda_{3}=\min\{\alpha,K-\frac{1}{2}\}\geq\frac{\rho^{2}\left(\sqrt{\frac{\lambda_{1}}{\lambda_{2}}}\left\|y(0)\right\|\right)}{2} needs to be satisfied according to the initial condition for asymptotic stability and the region of attraction can be made arbitrarily large to include any initial condition by increasing the gains α\alpha and KK accordingly. By the extension of LaSalle-Yoshizawa theorem for non-smooth systems in [14] and [15], cz(t)20c\left\|z(t)\right\|^{2}\to 0 and hence e0\left\|e\right\|\to 0 as tt\to\infty\,\,y(0)𝒮\forall\,\,y(0)\in\mathcal{S} , so the closed-loop error system is semi-globally asymptotically stable.

\blacksquare

V Conclusion

A continuous adaptive control design was presented to achieve semi-global asymptotic tracking for linearly parameterizable nonlinear systems with time-varying uncertain parameters. The key feature of this design is a RISE-like parameter update law along with a projection algorithm, which allows the system to compensate for potentially destabilizing terms in the closed-loop error system, arising due to the time-varying nature of parameters. Semi-global asymptotic tracking for the error system is guaranteed via a Lyapunov-based stability analysis. Future work will involve improvement of the parameter estimation performance of time-varying parameter systems and its extension to the system identification problem.

References

  • [1] P. Ioannou and J. Sun, Robust Adaptive Control.   Prentice Hall, 1996.
  • [2] J. E. Gaudio, A. M. Annaswamy, E. Lavretsky, and M. A. Bolender, “Parameter estimation in adaptive control of time-varying systems under a range of excitation conditions,” arXiv preprint arXiv:1911.03810, 2019.
  • [3] E. Arabi, B. C. Gruenwald, T. Yucelen, and N. T. Nguyen, “A set-theoretic model reference adaptive control architecture for disturbance rejection and uncertainty suppression with strict performance guarantees,” Int. J. Control, vol. 91, no. 5, pp. 1195–1208, 2018.
  • [4] E. Arabi and T. Yucelen, “Set-theoretic model reference adaptive control with time-varying performance bounds,” Int. J. Control, vol. 92, no. 11, pp. 2509–2520, 2019.
  • [5] E. Arabi, T. Yucelen, B. C. Gruenwald, M. Fravolini, S. Balakrishnan, and N. T. Nguyen, “A neuroadaptive architecture for model reference control of uncertain dynamical systems with performance guarantees,” Systems & Control Letters, vol. 125, pp. 37–44, 2019.
  • [6] Z. Qu and J. X. Xu, “Model-based learning controls and their comparisons using lyapunov direct method,” in Asian Journal of Control, vol. 4, No. 1, no. No. 1, Mar. 2002, pp. 99–110.
  • [7] J.-X. Xu, “A new periodic adaptive control approach for time-varying parameters with known periodicity,” IEEE Trans Autom. Contol, vol. 49, no. 4, pp. 579–583, Apr. 2004.
  • [8] P. A. Ioannou and P. V. Kokotovic, Eds., Adaptive Systems with Reduced Models, ser. Lecture Notes in Control and Information Sciences.   Springer Berlin Heidelberg, 1983, vol. 47, ch. 5. Adaptive control in the presence of disturbances, pp. 81–90.
  • [9] P. Patre, W. Mackunis, K. Dupree, and W. E. Dixon, “Modular adaptive control of uncertain Euler-Lagrange systems with additive disturbances,” IEEE Trans. Autom. Control, vol. 56, no. 1, pp. 155–160, 2011.
  • [10] B. Xian, D. M. Dawson, M. S. de Queiroz, and J. Chen, “A continuous asymptotic tracking control strategy for uncertain nonlinear systems,” IEEE Trans. Autom. Control, vol. 49, no. 7, pp. 1206–1211, 2004.
  • [11] C. Makkar, G. Hu, W. G. Sawyer, and W. E. Dixon, “Lyapunov-based tracking control in the presence of uncertain nonlinear parameterizable friction,” IEEE Trans. Autom. Control, vol. 52, pp. 1988–1994, 2007.
  • [12] P. M. Patre, W. Mackunis, C. Makkar, and W. E. Dixon, “Asymptotic tracking for systems with structured and unstructured uncertainties,” IEEE Trans. Control Syst. Technol., vol. 16, pp. 373–379, 2008.
  • [13] M. Krstic, I. Kanellakopoulos, and P. V. Kokotovic, Nonlinear and Adaptive Control Design.   New York, NY, USA: John Wiley & Sons, 1995.
  • [14] N. Fischer, R. Kamalapurkar, and W. E. Dixon, “LaSalle-Yoshizawa corollaries for nonsmooth systems,” IEEE Trans. Autom. Control, vol. 58, no. 9, pp. 2333–2338, Sep. 2013.
  • [15] R. Kamalapurkar, J. A. Rosenfeld, A. Parikh, A. R. Teel, and W. E. Dixon, “Invariance-like results for nonautonomous switched systems,” IEEE Trans. Autom. Control, vol. 64, no. 2, pp. 614–627, Feb. 2019.
Lemma 1.

Consider a positive-definite matrix Γ(n+m)×(n+m)\Gamma\in\mathbb{R}^{(n+m)\times(n+m)} such that Γ\Gamma has the block diagonal structure as Γ[Γ10m×n0n×mΓ2]\Gamma\triangleq\left[\begin{array}[]{cc}\Gamma_{1}&0_{m\times n}\\ 0_{n\times m}&\Gamma_{2}\end{array}\right], where Γ1m×m\Gamma_{1}\in\mathbb{R}^{m\times m} and Γ2n×n\Gamma_{2}\in\mathbb{R}^{n\times n}. The matrix Y(x(t),t)ΓYT(x(t),t)Y(x(t),t)\Gamma Y^{T}(x(t),t) is positive-definite, and hence invertible. Furthermore, the inverse of this matrix satisfies the property (Y(x(t),t)ΓYT(x(t),t))121λmin{Γ2}\left\|\left(Y(x(t),t)\Gamma Y^{T}(x(t),t)\right)^{-1}\right\|_{2}\leq\frac{1}{\lambda_{\textrm{min}}\left\{\Gamma_{2}\right\}}, where 2\left\|\cdot\right\|_{2} denotes the spectral norm and λmin{}\lambda_{\textrm{min}}\left\{\cdot\right\} denotes the minimum eigenvalue of {}\left\{\cdot\right\}.

Proof : Substituting the definitions for Y(x(t),t)Y(x(t),t) and Γ\Gamma in Y(x(t),t)ΓYT(x(t),t)Y(x(t),t)\Gamma Y^{T}(x(t),t) yields

Y(x(t),t)ΓYT(x(t),t)\displaystyle Y(x(t),t)\Gamma Y^{T}(x(t),t) =\displaystyle=
[Yh(x(t),t)In][Γ10m×n0n×mΓ2]\displaystyle\left[\begin{array}[]{cc}Y_{h}(x(t),t)&I_{n}\end{array}\right]\left[\begin{array}[]{cc}\Gamma_{1}&0_{m\times n}\\ 0_{n\times m}&\Gamma_{2}\end{array}\right] [Yh(x(t),t)In]\displaystyle\left[\begin{array}[]{c}Y_{h}(x(t),t)\\ I_{n}\end{array}\right]
=Yh(x(t),t)Γ1Yh(x(t),t)\displaystyle=Y_{h}(x(t),t)\Gamma_{1}Y_{h}(x(t),t) +Γ2.\displaystyle+\Gamma_{2}.

Since Γ\Gamma is selected to be a positive-definite matrix, the block matrices Γ1\Gamma_{1} and Γ2\Gamma_{2} are both positive-definite, so the first term Yh(x(t),t)Γ1Yh(x(t),t)Y_{h}(x(t),t)\Gamma_{1}Y_{h}(x(t),t) in this expression is positive semi-definite while the second term Γ2\Gamma_{2} is positive-definite, hence the sum of these two terms, i.e.,