This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Feasibility Analysis and Regularity Characterization
of Distributionally Robust Safe Stabilizing Controllers

Pol Mestres   Kehan Long   Nikolay Atanasov   Jorge Cortés The authors are with the Contextual Robotics Institute, UC San Diego (e-mails: {pomestre,k3long,natanasov,cortes}@ucsd.edu).The authors gratefully acknowledge support from NSF under grants IIS-2007141 and CCF-2112665.
Abstract

This paper studies the well-posedness and regularity of safe stabilizing optimization-based controllers for control-affine systems in the presence of model uncertainty. When the system dynamics contain unknown parameters, a finite set of samples can be used to formulate distributionally robust versions of control barrier function and control Lyapunov function constraints. Control synthesis with such distributionally robust constraints can be achieved by solving a (convex) second-order cone program (SOCP). We provide one necessary and two sufficient conditions to check the feasibility of such optimization problems, characterize their computational complexity and numerically show that they are significantly faster to check than direct use of SOCP solvers. Finally, we also analyze the regularity of the resulting control laws.

{IEEEkeywords}

Safety-critical control, control barrier functions, distributionally robust control synthesis.

1 Introduction

\IEEEPARstart

Recent years have seen increasing deployment of control systems and robots to aid transportation, warehouse management, and home automation. In these applications, it is crucial to implement controllers with provable safety and stability guarantees despite uncertainty in the system models and operational conditions. Recent work [1, 2, 3, 4, 5, 6] tackles this when some prior information about the uncertainty is known. Instead, here we rely on a line of work initiated in [7] that circumvents the need for knowledge about the uncertainty distribution and uses only uncertainty samples to formulate distributionally robust constraints for control synthesis. This approach is robust to distributional shift at deployment time and enjoys provable out-of-sample performance. However, it also introduces several challenges, which we focus on here: the characterization of the quality and number of uncertainty samples needed to guarantee the feasibility of the safety and stability constraints, and the study of the regularity properties of the resulting controllers.

Literature Review

Control Lyapunov functions (CLFs) [8] are a well-established tool to design stabilizing controllers for nonlinear systems. More recently, control barrier functions (CBFs) [9] have gained popularity as a tool to render a desired subset of the system state space safe. If the system is control affine, CLF and CBF constraints are linear in the control input and can be incorporated in a quadratic program (QP) [10] that, if feasible, can be solved efficiently to obtain control inputs guaranteeing safety and stability. Recent work has explored alternative optimization formulations when the system model is uncertain. Under the assumption that the uncertainty follows a Gaussian Process (GP) or satisfies worst-case bounds,[2, 11, 12, 1, 3, 5] formulate second-order cone constraints that can be used to design controllers achieving safe stabilization of the true system. The paper [4] gives sufficient conditions for the feasibility of such second-order cone constraints. Our work here is closely related to [7], which leverages ideas from distributionally robust optimization (DRO) [13, 14] to model the uncertainty. The DRO framework constructs an ambiguity set of probability distributions that contains the true (unknown) one with high confidence. Such ambiguity sets are constructed with only finitely many samples and are used to formulate distributionally robust versions of the control design problem.

Statement of Contributions

We study the problem of safe stabilization of control-affine systems under uncertainty. We assume that the distribution of the uncertainty is unknown and formulate a second-order cone program (SOCP) using distributionally robust versions of the CLF and CBF constraints constructed on the basis of uncertainty samples. Our first contribution is the derivation of a necessary condition and two sufficient conditions for the feasibility of the optimization problem. We characterize the computational complexity of these conditions and show that, for a large number of samples, it is significantly smaller than solving the SOCP directly, which makes them useful to efficiently check whether the problem is feasible without having to solve it. Our first sufficient condition is dependent on the quality of the uncertainty samples but is limited to a single control objective. Our second sufficient condition is only dependent on the number of samples but can be used for any number of constraints. Our final contribution shows that the solution of this distributionally robust optimization problem is point-Lipschitz, and hence continuous, which means that solutions of the closed-loop system are guaranteed to exist and the controller obtained from it can be implemented without inducing chattering.

2 Preliminaries

We review distrib. robust chance-constrained programs and control Lyapunov and barrier functions under uncertainty.

2.1 Distributionally Robust Chance Constrained Programs

Given a random vector 𝝃\boldsymbol{\xi} following distribution \mathbb{P}^{*} supported on set Ξk\Xi\subseteq^{k} and a closed convex set 𝒵n\mathcal{Z}\subset^{n}, let G:𝒵×ΞG:\mathcal{Z}\times\Xi\to define a probabilistic constraint G(z,𝝃)0G(z,\boldsymbol{\xi})\leq 0. We are interested in satisfying this constraint with a prescribed confidence 1ϵ1-\epsilon, with ϵ(0,1)\epsilon\in(0,1), while minimizing a convex objective function c:𝒵c:\mathcal{Z}\to. To achieve this111We denote by >0\mathbb{Z}_{>0}, and ≥0 the set of positive integers, real, and nonnegative real numbers, resp. We denote by 0n\textbf{0}_{n} the nn-dimensional zero vector. We write 𝒮\partial\mathcal{S} for the boundary of the set 𝒮\mathcal{S}. Given N>0N\in\mathbb{Z}_{>0}, we denote [N]={1,,N}[N]=\{1,\ldots,N\}. Given xnx\in^{n}, x\left\lVert x\right\rVert denotes the Euclidean norm of xx. For xx\in, we define (x)+=max(x,0)(x)_{+}=\max(x,0). A function β:0\beta:_{\geq 0}\to is of class 𝒦\mathcal{K}_{\infty} if β(0)=0\beta(0)=0, β\beta is strictly increasing and limtβ(t)=\lim\limits_{t\to\infty}\beta(t)=\infty. A function V:nV:^{n}\to is positive definite if V(0)=0V(0)=0 and V(x)>0V(x)>0 for all x0x\neq 0, and proper in a set Γ\Gamma if {xΓ:V(x)c}\{x\in\Gamma\;:\;V(x)\leq c\} is compact for any c0c\geq 0. Given an m×nm\times n matrix AA and two integers i,ji,j such that 1i<jm1\leq i<j\leq m, Ai:jA_{i:j} denotes the (ji+1)×n(j-i+1)\times n matrix obtained by selecting the rows from ii to jj of AA. A function f:nqf:^{n}\to^{q} is point-Lipschitz at a point x0nx_{0}\in^{n} if there exists a neighborhood UU of x0x_{0} and a constant Lx0>0L_{x_{0}}>0 such that f(x)f(x0)Lx0xx0\left\lVert f(x)-f(x_{0})\right\rVert\leq L_{x_{0}}\left\lVert x-x_{0}\right\rVert for all xUx\in U., define the chance-constrained program:

minz𝒵c(z)\displaystyle\min\limits_{z\in\mathcal{Z}}c(z) (1)
s.t. (G(z,𝝃)0)1ϵ.\displaystyle\ \mathbb{P}^{*}(G(z,\boldsymbol{\xi})\leq 0)\geq 1-\epsilon.

The feasible set of (1) is not convex in general. Nemirosvski and Shapiro [15, Section 2] propose a convex approximation of the feasible set of (1) by replacing the chance constraint with a conditional value-at-risk (CVaR\operatorname{CVaR}) constraint. CVaR of G(z,𝝃)G(z,\boldsymbol{\xi}) can be formulated as the following convex program:

CVaR1ϵ(G(z,𝝃)):=inft[ϵ1𝔼[(G(z,𝝃)+t)+]t].\displaystyle\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(G(z,\boldsymbol{\xi})):=\inf\limits_{t\in}[\epsilon^{-1}\mathbb{E}_{\mathbb{P}^{*}}[(G(z,\boldsymbol{\xi})+t)_{+}]\!-\!t]. (2)

The resulting problem

minz𝒵c(z)\displaystyle\min\limits_{z\in\mathcal{Z}}c(z) (3)
s.t. CVaR1ϵ(G(z,𝝃))0,\displaystyle\ \operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(G(z,\boldsymbol{\xi}))\leq 0,

is convex and its feasible set is contained in that of (1).

Both (1) and (3) assume that \mathbb{P}^{*} is known. Instead, suppose that it is unknown and we only have access to samples {𝝃i}i[N]\{\boldsymbol{\xi}_{i}\}_{i\in[N]} from \mathbb{P}^{*}. We describe a way of constructing a set of distributions that could have generated the samples. Let 𝒫p(Ξ)\mathcal{P}_{p}(\Xi) be the set of probability measures with finite pp-th moment supported on Ξ\Xi. Let ^N:=1Ni=1Nδ𝝃i\hat{\mathbb{P}}_{N}:=\frac{1}{N}\sum_{i=1}^{N}\delta_{\boldsymbol{\xi}_{i}} be the empirical distribution constructed from the samples {𝝃i}i=1N\{\boldsymbol{\xi}_{i}\}_{i=1}^{N}. Let WpW_{p} be the pp-Wasserstein distance [14, Definition 3.1] between two probability measures in 𝒫p(Ξ)\mathcal{P}_{p}(\Xi) and let Nr:={μ𝒫p(Ξ):Wp(μ,^N)r}\mathcal{M}_{N}^{r}:=\{\mu\in\mathcal{P}_{p}(\Xi)\;:\;W_{p}(\mu,\hat{\mathbb{P}}_{N})\leq r\} be the ball of radius rr centered at ^N\hat{\mathbb{P}}_{N}. We define a distributionally robust chance-constrained program:

minz𝒵c(z)\displaystyle\min\limits_{z\in\mathcal{Z}}c(z) (4)
s.t. infNr(G(z,𝝃)0)1ϵ.\displaystyle\inf_{\mathbb{P}\in\mathcal{M}_{N}^{r}}\mathbb{P}(G(z,\boldsymbol{\xi})\leq 0)\geq 1-\epsilon.

We can use CVaR to obtain a convex conservative approximation of (4):

minz𝒵c(z)\displaystyle\min_{z\in\mathcal{Z}}c(z) (5)
s.t. supNrCVaR1ϵ(G(z,𝝃))0.\displaystyle\sup_{\mathbb{P}\in\mathcal{M}_{N}^{r}}\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}}(G(z,\boldsymbol{\xi}))\leq 0.

If (5) is feasible, then (4) is also feasible [15, Section 2].

We say that a distribution \mathbb{P} is light-tailed if there exists a>0a>0 such that A:=𝔼[exp𝝃a]=Ξexp𝝃a(d𝝃)<A:=\mathbb{E}_{\mathbb{P}}[\exp{\left\lVert\boldsymbol{\xi}\right\rVert^{a}}]=\int_{\Xi}\exp{\left\lVert\boldsymbol{\xi}\right\rVert^{a}}\mathbb{P}(d\boldsymbol{\xi})<\infty. If \mathbb{P}^{*} is light-tailed, the following observation specifies how the radius of Nr\mathcal{M}_{N}^{r} should be selected so that the true distribution lies in the ball with high confidence.

Remark 2.1

(Choice of Wasserstein ball radius): If the true distribution \mathbb{P}^{*} is light-tailed, the choice of r=rN(ϵ¯)r=r_{N}(\bar{\epsilon}) given in [14, Theorem 3.5],

rN(ϵ¯)={(log(c1ϵ¯1)c2N)1max{k,2}ifNlog(c1ϵ¯1)c2,(log(c1ϵ¯1)c2N)1aelse,\displaystyle r_{N}(\bar{\epsilon})=\begin{cases}(\frac{\log(c_{1}\bar{\epsilon}^{-1})}{c_{2}N})^{\frac{1}{\max\{k,2\}}}\quad&\text{if}\ N\geq\frac{\log(c_{1}\bar{\epsilon}^{-1})}{c_{2}},\\ (\frac{\log(c_{1}\bar{\epsilon}^{-1})}{c_{2}N})^{\frac{1}{a}}\quad&\text{else},\end{cases} (6)

where c1,c2c_{1},c_{2} and aa are positive constants that only depend on aa, AA and kk (cf. [14, Theorem 3.4]), ensures that the ball NrN(ϵ¯)\mathcal{M}_{N}^{r_{N}(\bar{\epsilon})} contains \mathbb{P}^{*} with probability at least 1ϵ¯1-\bar{\epsilon}. Then, a solution zz^{*} of  (5) satisfies the constraint CVaR1ϵ(G(z,𝛏))0\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(G(z^{*},\boldsymbol{\xi}))\leq 0 with probability at least 1ϵ¯1-\bar{\epsilon}. Note that c1c_{1}, c2c_{2} and aa can be computed by knowing the class of distributions to which \mathbb{P}^{*} belongs to, without having actual knowledge of \mathbb{P}^{*}. If exact values are not known, but upper and lower bounds are, these can be used instead to compute an upper bound of rN(ϵ¯)r_{N}(\bar{\epsilon}). \bullet

Remark 2.2

(Choice of ϵ\epsilon): The parameter ϵ\epsilon determines the confidence level 1ϵ1-\epsilon for constraint satisfaction. Throughout the paper, we assume ϵ1N\epsilon\leq\frac{1}{N}, albeit results are valid generally, with explicit expressions becoming more involved. \bullet

2.2 Distributionally Robust Safety and Stability

The notions of CLF [8] and CBF [9] can be used to design controllers in uncertainty-free systems that enforce stability and safety, resp. Here we extend these notions for systems with uncertainty in the dynamics. Consider a nominal model FF and a linear combination of kk perturbations,

x˙\displaystyle\dot{x} =(F(x)+j=1kWj(x)ξj)u¯,\displaystyle=(F(x)+\sum_{j=1}^{k}W_{j}(x)\xi_{j})\underaccent{\bar}{u}, (7)

where for 1jk1\leq j\leq k, Wj(x)n×(m+1)W_{j}(x)\in^{n\times(m+1)} denotes known model perturbations, and ξj\xi_{j}\in denotes the corresponding unknown weight, and u¯=[1;u]𝒰¯:={1}×m\underaccent{\bar}{u}=[1;u]\in\underaccent{\bar}{\Uc}:=\{1\}\times^{m}. We let 𝝃=[ξ1,ξ2,,ξk]TΞk\boldsymbol{\xi}=[\xi_{1},\xi_{2},\ldots,\xi_{k}]^{T}\in\Xi\subseteq^{k}. We assume that 𝝃\boldsymbol{\xi} follows an unknown distribution \mathbb{P}^{*} but a set of samples {𝝃i}i=1N\{\boldsymbol{\xi}_{i}\}_{i=1}^{N} is available. We are interested in extending the notions of CLF [8] and CBF [9] for systems of the form (7). To do so, note that as shown in [7, Section IV], the CBF condition for a system of the form (7) and a function h:nh:^{n}\to reads as CBC(x,u¯,𝝃):=u¯Tqh(x)+u¯TRh(x)𝝃0\operatorname{CBC}(x,\underaccent{\bar}{u},\boldsymbol{\xi}):=\underaccent{\bar}{u}^{T}q_{h}(x)+\underaccent{\bar}{u}^{T}R_{h}(x)\boldsymbol{\xi}\geq 0, where the exact forms of qhq_{h} and RhR_{h} are given in [7, Section IV] and depend on hh and its gradient. Now, since 𝝃\boldsymbol{\xi} follows a distribution \mathbb{P}^{*}, we extend the definition of CBF by requiring that for all xx in the safe set, there exist u¯𝒰¯\underaccent{\bar}{u}\in\underaccent{\bar}{\Uc} such that

(CBC(x,u¯,𝝃)0)1ϵ.\displaystyle\mathbb{P}^{*}(\operatorname{CBC}(x,\underaccent{\bar}{u},\boldsymbol{\xi})\geq 0)\geq 1-\epsilon. (8)

The CLF condition for (7) takes a similar form and is written as CLC(x,u¯,𝝃)0\operatorname{CLC}(x,\underaccent{\bar}{u},\boldsymbol{\xi})\leq 0 (cf. [7, Section IV]). As shown in Section 2.1, CVaR\operatorname{CVaR} can be used as a convex approximation of (8) and its analogue with CLC\operatorname{CLC}. We use

CVaR1ϵ(CBC(x,u¯,𝝃))0,\displaystyle\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(\operatorname{CBC}(x,\underaccent{\bar}{u},\boldsymbol{\xi}))\geq 0, (9a)
CVaR1ϵ(CLC(x,u¯,𝝃))0,\displaystyle\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(\operatorname{CLC}(x,\underaccent{\bar}{u},\boldsymbol{\xi}))\geq 0, (9b)

as the distributionally robust analogues of the CLF and CBF conditions from [8] and [9], resp. The existence of a controller satisfying (9) implies the existence of a controller that makes the CLC (resp. the CBC) condition hold at every point with probability at least 1ϵ1-\epsilon, paving the way for the design of controllers that make the system stable (resp. safe) with arbitrarily high probability.

3 Problem Statement

Consider the system model in (7) with distributional uncertainty, meaning that the true distribution \mathbb{P}^{*} of the parameter 𝝃\boldsymbol{\xi} is unknown. We assume that the system admits a CLF and a CBF, which allow us to formulate the constraints (9). Given a nominal controller specified by a smooth function k¯:n𝒰¯\underaccent{\bar}{k}:^{n}\to\underaccent{\bar}{\mathcal{U}}, we would like to synthesize a controller closest to it that respects safety and stability constraints. Using (2), this problem can be written in general form as

minu¯𝒰¯u¯k¯(x)2\displaystyle\min\limits_{\underaccent{\bar}{u}\in\underaccent{\bar}{\mathcal{U}}}\left\lVert\underaccent{\bar}{u}-\underaccent{\bar}{k}(x)\right\rVert^{2} (10)
s.t. supNrinft[ϵ1𝔼[(Gl(x,u¯,𝝃)+t)+]t]0,l[M],\displaystyle\sup_{\mathbb{P}\in\mathcal{M}_{N}^{r}}\inf_{t\in}[\epsilon^{-1}\mathbb{E}_{\mathbb{P}}[(G_{l}(x,\underaccent{\bar}{u},\boldsymbol{\xi})+t)_{+}]-t]\leq 0,\ \forall l\in[M],

where M>0M\in\mathbb{Z}_{>0} and each Gl:n×𝒰¯×ΞG_{l}:^{n}\times\underaccent{\bar}{\Uc}\times\Xi\to is an affine function in u¯\underaccent{\bar}{u} and 𝝃\boldsymbol{\xi}, Gl(x,u¯,𝝃)=u¯Tql(x)+u¯TRl(x)𝝃G_{l}(x,\underaccent{\bar}{u},\boldsymbol{\xi})=\underaccent{\bar}{u}^{T}q_{l}(x)+\underaccent{\bar}{u}^{T}R_{l}(x)\boldsymbol{\xi}, for smooth functions ql:nm+1q_{l}:^{n}\to^{m+1} and Rl:n(m+1)×kR_{l}:^{n}\to^{(m+1)\times k}. With M=2M=2 and constraints corresponding to CBC\operatorname{CBC} and CLC\operatorname{CLC}, this corresponds to a stable and safe control synthesis problem. The case M=1M=1 with the constraint CBC\operatorname{CBC} corresponds to a distributionally robust version of a safety filter of k¯\underaccent{\bar}{k}.

Although the constraints in (10) are convex, the program is intractable due to the search of suprema over the Wasserstein set. Fortunately, [7, Proposition IV.1] shows that when Ξ=k\Xi=^{k} and p=1p=1, the following SOCP is equivalent to (10):

minu¯𝒰¯,y,t,siy\displaystyle\min_{\underaccent{\bar}{u}\in\underaccent{\bar}{\Uc},y\in,t\in,s_{i}\in}y (11a)
s.t. rRlT(x)u¯+1Ni=1Nsitϵ0,l[M],\displaystyle\quad r\left\lVert R_{l}^{T}(x)\underaccent{\bar}{u}\right\rVert+\frac{1}{N}\sum_{i=1}^{N}s_{i}-t\epsilon\leq 0,\ \forall l\in[M],~{} (11b)
siGl(x,u¯,𝝃i)+t,i[N],l[M],\displaystyle\quad s_{i}\geq G_{l}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i})+t,\ \forall i\in[N],\ \forall l\in[M],~{} (11c)
si0,i[N],\displaystyle\quad s_{i}\geq 0,\ \forall i\in[N],~{} (11d)
y+12(u¯k¯(x))2+(y1)2.\displaystyle\quad y+1\geq\sqrt{\left\lVert 2(\underaccent{\bar}{u}-\underaccent{\bar}{k}(x))\right\rVert^{2}+(y-1)^{2}}~{}. (11e)

We refer to (11) as the DRO-SOCP and take Ξ=k\Xi=^{k} and p=1p=1 Wasserstein distance throughout the paper.

A critical observation about problem (11) is that, in general, it might be infeasible, leading to controllers that are undefined.

Furthermore, even if the problem is feasible, the controller obtained from it might not be continuous, hence resulting in implementation problems (it might induce chattering behavior when implemented on physical systems) and theoretical problems (lack of existence of solutions of the closed-loop system). Hence, our goal in this paper is twofold. First, we derive conditions to ensure the feasibility of (11). Given the complexity of obtaining characterizations for the feasibility of such problems, we focus on identifying conditions that are easy to evaluate computationally as opposed to directly attempting to solve the optimization problem: either sufficient conditions, to quickly ensure feasibility, or necessary, to quickly discard it. Second, assuming that the problem (11) is feasible, we characterize the regularity properties of the resulting controller.

4 Feasibility Analysis

In this section, we study the feasibility properties of (11). We start by giving a necessary condition for its feasibility.

Proposition 4.1

(Necessary condition for feasibility of DRO-SOCP): Let ϵ(0,1N]\epsilon\in(0,\frac{1}{N}] and r>0r>0. For xnx\in^{n}, let

Q¯l(x)=rRl(x)2:(m+1)m×k,r¯l(x)=rRl(x)11×k,\displaystyle\bar{Q}_{l}(x)=rR_{l}(x)_{2:(m+1)}\in^{m\times k},\ \bar{r}_{l}(x)=rR_{l}(x)_{1}\in^{1\times k},
w¯l,i(x)=(ϵql(x)ϵRl(x)𝝃i)2:(m+1)m,\displaystyle\bar{w}_{l,i}(x)=(-\epsilon q_{l}(x)-\epsilon R_{l}(x)\boldsymbol{\xi}_{i})_{2:(m+1)}\in^{m},
v¯l,i(x)=(ϵql(x)ϵRl(x)𝝃i)1,\displaystyle\bar{v}_{l,i}(x)=(-\epsilon q_{l}(x)-\epsilon R_{l}(x)\boldsymbol{\xi}_{i})_{1}\in,
F¯l,i(x)=Q¯l(x)Q¯l(x)Tw¯l,i(x)w¯l,i(x)Tm×m,\displaystyle\bar{F}_{l,i}(x)=\bar{Q}_{l}(x)\bar{Q}_{l}(x)^{T}-\bar{w}_{l,i}(x)\bar{w}_{l,i}(x)^{T}\in^{m\times m},
J¯l,i(x)=r¯l(x)Q¯l(x)Tv¯l,i(x)w¯l,iT1×m,\displaystyle\bar{J}_{l,i}(x)=\bar{r}_{l}(x)\bar{Q}_{l}(x)^{T}-\bar{v}_{l,i}(x)\bar{w}_{l,i}^{T}\in^{1\times m},
H¯l,i(x)=(r¯lr¯lTv¯l,i2)(x)J¯l,i(x)J¯l,iT(x)F¯l,i(x))(m+1)×(m+1)\displaystyle\bar{H}_{l,i}(x)=\begin{pmatrix}\bar{r}_{l}\bar{r}_{l}^{T}-\bar{v}_{l,i}^{2})(x)&\bar{J}_{l,i}(x)\\ \bar{J}_{l,i}^{T}(x)&\bar{F}_{l,i}(x)\end{pmatrix}\in^{(m+1)\times(m+1)}

for l[M]l\in[M] and i[N]i\in[N]. Let λ¯l,i(x)\bar{\lambda}_{l,i}(x) be the minimum eigenvalue of F¯l,i(x)\bar{F}_{l,i}(x) and suppose Q¯l(x)Q¯l(x)T\bar{Q}_{l}(x)\bar{Q}_{l}(x)^{T} is invertible for all l[M]l\in[M]. If (11) is feasible, then for each l[M]l\in[M], there exists i[N]i\in[N] such that H¯l,i(x)\bar{H}_{l,i}(x) is not positive definite and one of the following holds:

  1. (i)

    λ¯l,i(x)<0\bar{\lambda}_{l,i}(x)\!<\!0,

  2. (ii)

    λ¯l,i(x)>0\bar{\lambda}_{l,i}(x)\!>\!0 and (v¯l,iw¯l,iTF¯l,i1(Q¯lr¯lTw¯l,iv¯l,i))(x)0\big{(}\bar{v}_{l,i}-\bar{w}_{l,i}^{T}\bar{F}_{l,i}^{-1}(\bar{Q}_{l}\bar{r}_{l}^{T}\!-\!\bar{w}_{l,i}\bar{v}_{l,i})\big{)}(x)\!\geq\!0,

  3. (iii)

    λ¯l,i(x)=0\bar{\lambda}_{l,i}(x)=0, and (v¯l,iw¯l,iT(Q¯lQ¯lT)1Q¯lr¯lT)(x)>0\big{(}\bar{v}_{l,i}-\bar{w}_{l,i}^{T}(\bar{Q}_{l}\bar{Q}_{l}^{T})^{-1}\bar{Q}_{l}\bar{r}_{l}^{T}\big{)}(x)>0.

Proof 4.2.

Note that (10) (and hence (11)) is equivalent to

minu¯𝒰¯u¯k¯(x)2\displaystyle\min\limits_{\underaccent{\bar}{u}\in\underaccent{\bar}{\Uc}}\left\lVert\underaccent{\bar}{u}-\underaccent{\bar}{k}(x)\right\rVert^{2} (12)
s.t.rRlT(x)u¯+inft[1Ni=1N(Gl(x,u¯,𝝃i)+t)+tϵ]0,\displaystyle\text{s.t.}\ r\left\lVert R_{l}^{T}(x)\underaccent{\bar}{u}\right\rVert+\inf_{t\in}\Big{[}\frac{1}{N}\sum_{i=1}^{N}(G_{l}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i})+t)_{+}-t\epsilon\Big{]}\leq 0,

for l{1,,M}l\in\{1,\dots,M\}, cf. [7, Proposition IV.1]. For (x,u¯)n×𝒰¯(x,\underaccent{\bar}{u})\in^{n}\times\underaccent{\bar}{\Uc}, the function Ax,u¯l(t)=1Ni=1N(Gl(x,u¯,𝛏i)+t)+tϵA_{x,\underaccent{\bar}{u}}^{l}(t)=\frac{1}{N}\sum_{i=1}^{N}(G_{l}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i})+t)_{+}-t\epsilon is a piecewise linear function in tt. Since ϵ1N\epsilon\leq\frac{1}{N}, it is decreasing for t<tl(x,u¯):=mini[N]Gl(x,u¯,𝛏i)t<t_{l}^{*}(x,\underaccent{\bar}{u}):=\min_{i\in[N]}-G_{l}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i}) and increasing for t>tl(x,u¯)t>t_{l}^{*}(x,\underaccent{\bar}{u}). Hence, it achieves its minimum at tl(x,u¯)t_{l}^{*}(x,\underaccent{\bar}{u}). Thus, (11) is feasible if and only if for all l[M]l\in[M] the following inequalities are simultaneously feasible:

rRlT(x)u¯+ϵu¯Tql(x)+ϵmaxi[N]u¯TRl(x)𝝃i0.\displaystyle r\left\lVert R_{l}^{T}(x)\underaccent{\bar}{u}\right\rVert+\epsilon\underaccent{\bar}{u}^{T}q_{l}(x)+\epsilon\max_{i\in[N]}\underaccent{\bar}{u}^{T}R_{l}(x)\boldsymbol{\xi}_{i}\leq 0. (13)

Note that, if for some l[M]l\in[M], the constraint rRlT(x)u¯+ϵu¯Tql(x)+ϵu¯TRl(x)𝛏i0r\left\lVert R_{l}^{T}(x)\underaccent{\bar}{u}\right\rVert+\epsilon\underaccent{\bar}{u}^{T}q_{l}(x)+\epsilon\underaccent{\bar}{u}^{T}R_{l}(x)\boldsymbol{\xi}_{i}\leq 0 is infeasible for all i[N]i\in[N], then (11) is infeasible. Note that this is only a sufficient, but not necessary, condition for infeasibility (or equivalently, a necessary, but not sufficient, condition for feasibility). The result follows from [2, Theorem 2], which characterizes the feasibility of a single second-order cone constraint.

Next, we state a sufficient condition for the feasibility of (11) in the case M=1M=1.

Proposition 4.3.

(Sufficient condition for feasibility of DRO-SOCP with one constraint): Let r>0r>0, M=1M=1, and 0<ϵ1N0<\epsilon\leq\frac{1}{N}. Given xnx\in^{n}, define

Q^(x)=(r+ϵmaxi[N]𝝃i)R1(x)2:(m+1)m×k,\displaystyle\hat{Q}(x)=(r+\epsilon\max_{i\in[N]}\left\lVert\boldsymbol{\xi}_{i}\right\rVert)R_{1}(x)_{2:(m+1)}\in^{m\times k},
r^(x)=(r+ϵmaxi[N]𝝃i)R1(x)11×k,\displaystyle\hat{r}(x)=(r+\epsilon\max_{i\in[N]}\left\lVert\boldsymbol{\xi}_{i}\right\rVert)R_{1}(x)_{1}\in^{1\times k},
w^(x)=ϵq1(x)2:(m+1)m,v^(x)=ϵq1(x)1,\displaystyle\hat{w}(x)=-\epsilon q_{1}(x)_{2:(m+1)}\in^{m},\ \hat{v}(x)=-\epsilon q_{1}(x)_{1}\in,
F^(x)=Q(x)Q(x)Tw(x)w(x)Tm×m,\displaystyle\hat{F}(x)=Q(x)Q(x)^{T}-w(x)w(x)^{T}\in^{m\times m},
J^(x)=r^(x)Q^(x)Tv^(x)w^(x)T1×m,\displaystyle\hat{J}(x)=\hat{r}(x)\hat{Q}(x)^{T}-\hat{v}(x)\hat{w}(x)^{T}\in^{1\times m},
H^(x)=((r^r^Tv^2)(x)J^(x)J^(x)TF^(x))(m+1)×(m+1).\displaystyle\hat{H}(x)=\begin{pmatrix}(\hat{r}\hat{r}^{T}-\hat{v}^{2})(x)&\hat{J}(x)\\ \hat{J}(x)^{T}&\hat{F}(x)\end{pmatrix}\in^{(m+1)\times(m+1)}.

Let λ^(x)\hat{\lambda}(x) be the minimum eigenvalue of F^(x)\hat{F}(x). Suppose that Q(x)Q(x)TQ(x)Q(x)^{T} is invertible, H^(x)\hat{H}(x) is not positive definite and one of the following holds:

  1. (i)

    λ^(x)<0\hat{\lambda}(x)<0,

  2. (ii)

    λ^(x)>0\hat{\lambda}(x)>0 and (v^w^TF^1(Q^r^Tw^v^)(x)0\big{(}\hat{v}-\hat{w}^{T}\hat{F}^{-1}(\hat{Q}\hat{r}^{T}-\hat{w}\hat{v}\big{)}(x)\geq 0,

  3. (iii)

    λ^(x)=0\hat{\lambda}(x)=0 and (v^w^T(Q^Q^T)1Q^r^T)(x)>0\big{(}\hat{v}-\hat{w}^{T}(\hat{Q}\hat{Q}^{T})^{-1}\hat{Q}\hat{r}^{T}\big{)}(x)>0.

Then, (11) is feasible at xx.

Proof 4.4.

By repeating an argument similar to the one in the proof of Proposition 4.1, (11) is feasible in the case M=1M=1 if and only if the following inequality is feasible:

rR(x)Tu¯+ϵu¯Tq(x)+ϵmaxi[N]u¯TR(x)𝝃i0.\displaystyle r\left\lVert R(x)^{T}\underaccent{\bar}{u}\right\rVert+\epsilon\underaccent{\bar}{u}^{T}q(x)+\epsilon\max_{i\in[N]}\underaccent{\bar}{u}^{T}R(x)\boldsymbol{\xi}_{i}\leq 0. (14)

Using the Cauchy-Schwartz inequality, the following inequality being feasible implies that (14) is feasible,

(r+ϵmaxi[N]𝝃i)R(x)Tu¯+ϵu¯Tq(x)0.\displaystyle(r+\epsilon\max_{i\in[N]}\left\lVert\boldsymbol{\xi}_{i}\right\rVert)\left\lVert R(x)^{T}\underaccent{\bar}{u}\right\rVert+\epsilon\underaccent{\bar}{u}^{T}q(x)\leq 0. (15)

If (15) is feasible, there exists u¯^\hat{\underaccent{\bar}{u}} such that ru¯^TR(x)+ϵu¯^Tq(x)+ϵu¯^TR(x)𝛏i0r\left\lVert\hat{\underaccent{\bar}{u}}^{T}R(x)\right\rVert+\epsilon\hat{\underaccent{\bar}{u}}^{T}q(x)+\epsilon\hat{\underaccent{\bar}{u}}^{T}R(x)\boldsymbol{\xi}_{i}\leq 0 for all i[N]i\in[N], and thus u¯^\hat{\underaccent{\bar}{u}} satisfies (14). The result follows by [2, Thm. 2].

Remark 4.5.

(More data leads to better feasibility guarantees): For a fixed rr, the addition of new data points (larger NN) implies that there are more chances that either of (i)-(iii) in Proposition 4.1 are satisfied for each l{1,,M}l\in\{1,\dots,M\}. Moreover, if \mathbb{P}^{*} is light-tailed, rN(ϵ¯)r_{N}(\bar{\epsilon}) decreases with NN. The choice r=rN(ϵ¯)r=r_{N}(\bar{\epsilon}) means that for each fixed i[N]i\in[N] and l[M]l\in[M], the feasible set of the inequality rRl(x)Tu¯+ϵu¯Tql(x)+ϵu¯TRl(x)𝛏i0r\left\lVert R_{l}(x)^{T}\underaccent{\bar}{u}\right\rVert+\epsilon\underaccent{\bar}{u}^{T}q_{l}(x)+\epsilon\underaccent{\bar}{u}^{T}R_{l}(x)\boldsymbol{\xi}_{i}\leq 0 increases, which from the proof of Proposition 4.1, also means that there are more chances that either of (i)-(iii) are met. Similarly, under the assumption that the norm of additional samples is upper bounded by maxi[N]𝛏i\max_{i\in[N]}\left\lVert\boldsymbol{\xi}_{i}\right\rVert, the choice r=rN(ϵ¯)r=r_{N}(\bar{\epsilon}) also leads to a larger feasible set of (15) and thus the sufficient condition in Proposition 4.3 has more chances of being satisfied. \bullet

We next give a sufficient condition for the feasibility of (11) with high probability for an arbitrary number of constraints.

Proposition 4.6.

(Sufficient condition for feasibility of DRO-SOCP): Let r>0r>0, ϵ(0,1)\epsilon\in(0,1) and ϵ¯(0,1)\bar{\epsilon}\in(0,1). Suppose that there exists a controller k^:n𝒰¯\hat{k}:^{n}\to\underaccent{\bar}{\Uc} and non-negative functions Sl:n0S_{l}:^{n}\to_{\geq 0} for l[M]l\in[M] satisfying

CVaR1ϵ(Gl(x,k^(x),𝝃))\displaystyle\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(G_{l}(x,\hat{k}(x),\boldsymbol{\xi})) Sl(x),l[M].\displaystyle\leq-S_{l}(x),\quad\forall l\in[M]. (16)

Moreover, suppose that \mathbb{P}^{*} is light-tailed and let rN(ϵ¯)r_{N}(\bar{\epsilon}) be defined as in (6). Let xnx\in^{n} be such that Rl(x)0\left\lVert R_{l}(x)\right\rVert\neq 0 for all l[M]l\in[M], and let B:n0B:^{n}\to_{\geq 0} be an upper bound on the norm of k^\hat{k}. Then, if

rN(ϵ¯)<minl[M]ϵSl(x)2Rl(x)B(x),\displaystyle~{}r_{N}(\bar{\epsilon})<\min\limits_{l\in[M]}\frac{\epsilon S_{l}(x)}{2\left\lVert R_{l}(x)\right\rVert B(x)}, (17)

(11) is strictly feasible at xx with probability at least 1ϵ¯1-\bar{\epsilon} for any rrN(ϵ¯)r\leq r_{N}(\bar{\epsilon}).

Proof 4.7.

Note that by definition, the first component of k^(x)\hat{k}(x) is 11 for all xnx\in^{n}. Hence, B(x)k^(x)1B(x)\geq\left\lVert\hat{k}(x)\right\rVert\geq 1 for all xnx\in^{n} so (17) is well-defined. Let t1t_{1}^{*}\in be such that

CVaR1ϵ(G1(x,u¯,𝝃))=1ϵ𝔼[(G1(x,u¯,𝝃)+t1)+]t1,\displaystyle\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(G_{1}(x,\underaccent{\bar}{u},\boldsymbol{\xi}))=\frac{1}{\epsilon}\mathbb{E}_{\mathbb{P}^{*}}[(G_{1}(x,\underaccent{\bar}{u},\boldsymbol{\xi})+t_{1}^{*})_{+}]-t_{1}^{*},

and define G^(x,𝛏)=1ϵ(G1(x,k^(x),𝛏)+t1)+t1\hat{G}(x,\boldsymbol{\xi})=\frac{1}{\epsilon}(G_{1}(x,\hat{k}(x),\boldsymbol{\xi})+t_{1}^{*})_{+}-t_{1}^{*}. Note that for any 𝛏,𝛏k\boldsymbol{\xi},\boldsymbol{\xi^{{}^{\prime}}}\in^{k},

|G^(x,𝝃)G^(x,𝝃)|1ϵR1(x)k^(x)𝝃𝝃,\displaystyle|\hat{G}(x,\boldsymbol{\xi})\!-\!\hat{G}(x,\boldsymbol{\xi}^{{}^{\prime}})|\!\leq\!\frac{1}{\epsilon}\left\lVert R_{1}(x)\right\rVert\!\cdot\!\left\lVert\hat{k}(x)\right\rVert\!\cdot\!\left\lVert\boldsymbol{\xi}-\boldsymbol{\xi^{{}^{\prime}}}\right\rVert, (18)

where we have used the fact that the operator ()+(\cdot)_{+} is Lipschitz with constant 1. Using (18) in [14, Theorem 3.2], we conclude that for any ^𝒫p(Ξ)\mathbb{\hat{P}}\in\mathcal{P}_{p}(\Xi), |𝔼(G^(x,𝛏))𝔼^(G^(x,𝛏))|1ϵR1(x)k^(x)W1(,^)|\mathbb{E}_{\mathbb{P}^{*}}\!(\hat{G}(x,\boldsymbol{\xi}))\!-\!\mathbb{E}_{\hat{\mathbb{P}}}(\hat{G}(x,\boldsymbol{\xi}))|\!\leq\!\frac{1}{\epsilon}\left\lVert R_{1}(x)\right\rVert\!\cdot\!\left\lVert\hat{k}(x)\right\rVert\!\cdot\!W_{1}(\mathbb{P}^{*}\!,\!\mathbb{\hat{P}}).

From (17), together with the fact that NrN(ϵ¯)\mathcal{M}_{N}^{r_{N}(\bar{\epsilon})} contains \mathbb{P}^{*} with probability at least 1ϵ¯1-\bar{\epsilon}, cf. Remark 2.1, and since the maximum Wasserstein distance between two distributions in NrN(ϵ¯)\mathcal{M}_{N}^{r_{N}(\bar{\epsilon})} is 2rN(ϵ¯)2r_{N}(\bar{\epsilon}), with probability at least 1ϵ¯1-\bar{\epsilon},

|CVaR1ϵ(G1(x,k^(x),𝝃))𝔼^(G^(x,𝝃))|<S1(x).\displaystyle|\!\operatorname{CVaR}_{1-\epsilon}^{\mathbb{P}^{*}}(G_{1}(x,\hat{k}(x),\boldsymbol{\xi}))\!-\!\mathbb{E}_{\hat{\mathbb{P}}}(\hat{G}(x,\boldsymbol{\xi}))|<S_{1}(x). (19)

for any ^NrN(ϵ¯)\mathbb{\hat{P}}\in\mathcal{M}_{N}^{r_{N}(\bar{\epsilon})}. By definition of CVaR\operatorname{CVaR}, cf. (2), for any ^𝒫p(Ξ)\mathbb{\hat{P}}\in\mathcal{P}_{p}(\Xi), CVaR1ϵ^(G1(x,k^(x),𝛏))𝔼^(G^(x,𝛏))\operatorname{CVaR}_{1-\epsilon}^{\hat{\mathbb{P}}}(G_{1}(x,\hat{k}(x),\boldsymbol{\xi}))\leq\mathbb{E}_{\hat{\mathbb{P}}}(\hat{G}(x,\boldsymbol{\xi})). Combining this with (19) and (16), we get that with probability at least 1ϵ¯1-\bar{\epsilon}, CVaR1ϵ^(G1(x,k^(x),𝛏))<0\operatorname{CVaR}_{1-\epsilon}^{\hat{\mathbb{P}}}(G_{1}(x,\hat{k}(x),\boldsymbol{\xi}))<0 for all ^NrN(ϵ¯)\mathbb{\hat{P}}\in\mathcal{M}_{N}^{r_{N}(\bar{\epsilon})}. This argument holds for l{2,,N}l\in\{2,\ldots,N\}, implying that k^(x)\hat{k}(x) is strictly feasible for (10) (and hence, (11)) with probability at least 1ϵ¯1-\bar{\epsilon} for any rrN(ϵ¯)r\leq r_{N}(\bar{\epsilon}).

Remark 4.8.

(Dependency of sufficient condition on slack terms): Condition (16) on the controller k^\hat{k} guarantees the satisfaction of the constraints in (10) with a slack term Sl(x)S_{l}(x) on the righthand side. Larger values of these slack terms mean that fewer samples are needed to satisfy (17). Moreover, for the constraints in (9), [4, Remark 2.3] shows how to obtain such functions SlS_{l}, even without knowledge of k^\hat{k}. \bullet

Remark 4.9.

(Applicability of the sufficient condition): Checking condition (17) does not require precise knowledge of k^\hat{k}, just an upper bound of its norm. In particular, if bounds on the control norm are included as constraints in (11), those can be used to construct BB. Moreover, unlike Proposition 4.3, condition (17) is agnostic to the samples {𝛏1,,𝛏N}\{\boldsymbol{\xi}_{1},\ldots,\boldsymbol{\xi}_{N}\} and instead solely depends on its number NN. Note that for each xnx\in^{n} with Rl(x)0\left\lVert R_{l}(x)\right\rVert\neq 0 for all l[M]l\in[M], if Sl(x)>0S_{l}(x)>0 for all l[M]l\in[M], there exists N^x\hat{N}_{x} such that condition (17) holds for all NN^xN\geq\hat{N}_{x}. This is because rN(ϵ¯)r_{N}(\bar{\epsilon}) is decreasing in NN and limNrN(ϵ¯)=0\lim_{N\to\infty}r_{N}(\bar{\epsilon})=0. The value N^x\hat{N}_{x} is state-dependent, larger for smaller values of ϵ\epsilon, Sl(x)S_{l}(x) and larger values of B(x)B(x). \bullet

Remark 4.10.

(Checking for (in)feasibility efficiently): A commonly used algorithm for solving SOCPs is the method in [16]. For an SOCP with rSr_{S} constraints and optimization variable of dimension nSn_{S}, it requires solving rS\sqrt{r_{S}} linear systems of dimension nSn_{S}, and hence has complexity 𝒪(rSnS3)\mathcal{O}(\sqrt{r_{S}}n_{S}^{3}), cf. [17]. Therefore, (11) has complexity 𝒪(MN(m+N)3)\mathcal{O}(\sqrt{MN}(m+N)^{3}). Instead, since checking the positive definiteness of a symmetric matrix of dimension nPn_{P} can be done by checking if its Cholesky factorization exists (which has complexity 𝒪(nP3)\mathcal{O}(n_{P}^{3})), the complexity of checking the condition in Proposition 4.1 is 𝒪(NMm3)\mathcal{O}(NMm^{3}). Hence, for large NN, it is much more efficient than solving the SOCP (11) directly. We also note that the scaling in MM for the complexity of the SOCP solver is more favorable than that of checking the necessary condition. On the other hand, the complexity of checking the sufficient condition in Proposition 4.3 reduces to finding a maximum of NN numbers (which has complexity linear in NN) and checking the positive definiteness of two symmetric matrices of dimension m+1m+1 and mm, resp. Hence, its complexity is 𝒪(N+m3)\mathcal{O}(N+m^{3}), which is also more efficient than solving the SOCP. Finally, note that the complexity of checking the conditions in Proposition 4.6 is constant in NN and mm, and is linear in MM due to the minimum in (17). Table 1 summarizes this complexity analysis. \bullet

Table 1: Complexity of SOCP solver versus the results in this section.
Method Necessary/Sufficient Complexity MM
Prop. 4.1 Necessary 𝒪(NMm3)\mathcal{O}(NMm^{3}) any
Prop. 4.3 Sufficient 𝒪(N+m3)\mathcal{O}(N+m^{3}) 11
Prop. 4.6 Sufficient 𝒪(M)\mathcal{O}(M) any
SOCP solver Necessary and sufficient 𝒪(NM(m+N)3)\mathcal{O}(\sqrt{NM}(m+N)^{3}) any

Proposition 4.1 provides necessary conditions for feasibility. If the conditions are not met, it is reasonable to gather more data for verifying feasibility without having to directly solve the program. Moreover, if the conditions in Propositions 4.3 and 4.6 are not met (which does not mean that (11) is not feasible), this might be an indication that more data is needed to certify feasibility, cf. Remarks 4.5 and 4.9.

5 Regularity Analysis

In this section, we show that the controller obtained by solving (11) is point-Lipschitz.

Proposition 5.1.

(Point-Lipschitzness of SOCP DRO): Let r>0r>0, 0<ϵ1N0<\epsilon\leq\frac{1}{N} and suppose RlR_{l} and qlq_{l} are twice continuously differentiable for all l[M]l\in[M]. Let u¯:nm\underaccent{\bar}{u}^{*}:^{n}\to^{m} be the function mapping xnx\in^{n} to the solution of (11) in u¯\underaccent{\bar}{u} at xx. If (10) is strictly feasible at x0nx_{0}\in^{n} (i.e., there exists a solution satisfying all the constraints strictly), then u¯\underaccent{\bar}{u}^{*} is point-Lipschitz at x0x_{0}.

Proof 5.2.

We first show the result for M=1M=1. Let :=argmaxi[N]G1(x0,u¯(x0),𝛏i)\mathcal{I}:=\mathrm{arg}\max_{i\in[N]}G_{1}(x_{0},\underaccent{\bar}{u}^{*}(x_{0}),\boldsymbol{\xi}_{i}), note that the set \mathcal{I} is dependent on x0x_{0}, but we omit this dependency to simplify the notation. Note also that since G1(x,u¯,𝛏i)G_{1}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i}) is continuous in xx and u¯\underaccent{\bar}{u} for all i[N]i\in[N], there exists a neighborhood 𝒩=𝒩x×𝒩u¯n×𝒰¯\mathcal{N}=\mathcal{N}_{x}\times\mathcal{N}_{\underaccent{\bar}{u}}\subset^{n}\times\underaccent{\bar}{\Uc} of (x0,u¯(x0))(x_{0},\underaccent{\bar}{u}^{*}(x_{0})) such that for all (x^,u¯^)𝒩(\hat{x},\hat{\underaccent{\bar}{u}})\in\mathcal{N}, there exists ix^,u¯^i_{\hat{x},\hat{\underaccent{\bar}{u}}}\in\mathcal{I} such that ix^,u¯^argmaxi[N]G1(x^,u¯^,𝛏i)i_{\hat{x},\hat{\underaccent{\bar}{u}}}\in\mathrm{arg}\max_{i\in[N]}G_{1}(\hat{x},\hat{\underaccent{\bar}{u}},\boldsymbol{\xi}_{i}). Recall from the proof of Proposition 4.1 that, for any x,u¯n×𝒰¯x,\underaccent{\bar}{u}\in^{n}\times\underaccent{\bar}{\Uc}, the function Ax,u¯(t):=1Ni=1N(G1(x,u¯,𝛏i)+t)+tϵA_{x,\underaccent{\bar}{u}}(t):=\frac{1}{N}\sum_{i=1}^{N}(G_{1}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i})+t)_{+}-t\epsilon attains its minimum at t(x,u¯):=maxi[N]G1(x,u¯,𝛏i)t^{*}(x,\underaccent{\bar}{u}):=\max_{i\in[N]}{G_{1}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i})}. Therefore, for (x^,u¯^)𝒩(\hat{x},\hat{\underaccent{\bar}{u}})\in\mathcal{N}, t(x^,u¯^)=G1(x^,u¯^,𝛏ix^,u¯^)t^{*}(\hat{x},\hat{\underaccent{\bar}{u}})=G_{1}(\hat{x},\hat{\underaccent{\bar}{u}},\boldsymbol{\xi}_{i_{\hat{x},\hat{\underaccent{\bar}{u}}}}).

For each ii\in\mathcal{I}, let u¯i:nm\underaccent{\bar}{u}_{i}^{*}:^{n}\to^{m} be defined as:

u¯i(x):=\displaystyle\underaccent{\bar}{u}_{i}(x):= minu¯𝒰¯u¯k¯(x)2\displaystyle\min\limits_{\underaccent{\bar}{u}\in\underaccent{\bar}{\Uc}}\left\lVert\underaccent{\bar}{u}-\underaccent{\bar}{k}(x)\right\rVert^{2} (20)
s.t.rR1(x)Tu¯+ϵG1(x,u¯,𝝃i)0.\displaystyle\text{s.t.}\ r\left\lVert R_{1}(x)^{T}\underaccent{\bar}{u}\right\rVert+\epsilon G_{1}(x,\underaccent{\bar}{u},\boldsymbol{\xi}_{i})\leq 0.

Note that since (10) is strictly feasible at x0x_{0}, there exists u¯~𝒰¯\tilde{\underaccent{\bar}{u}}\in\underaccent{\bar}{\Uc} such that rR1(x0)Tu¯~+maxi[N]ϵG1(x0,u¯~,𝛏i)<0r\left\lVert R_{1}(x_{0})^{T}\underaccent{\bar}{\tilde{u}}\right\rVert+\max_{i\in[N]}\epsilon G_{1}(x_{0},\tilde{\underaccent{\bar}{u}},\boldsymbol{\xi}_{i})<0. By continuity of R1R_{1} and G1G_{1} in xx, there exists a neighborhood 𝒩~x𝒩x\tilde{\mathcal{N}}_{x}\subset\mathcal{N}_{x} of x0x_{0} such that rR1(x)Tu¯~+ϵG1(x,u¯~,𝛏i)<0r\left\lVert R_{1}(x)^{T}\underaccent{\bar}{\tilde{u}}\right\rVert+\epsilon G_{1}(x,\tilde{\underaccent{\bar}{u}},\boldsymbol{\xi}_{i})<0 for all x𝒩~xx\in\tilde{\mathcal{N}}_{x} and ii\in\mathcal{I}. This implies that (20) is strictly feasible for any x𝒩~xx\in\tilde{\mathcal{N}}_{x}. Hence, by [4, Proposition 5.4], u¯i\underaccent{\bar}{u}_{i}^{*} is point-Lipschitz at x0x_{0} for each ii\in\mathcal{I}. Now, since for all y𝒩xy\in\mathcal{N}_{x} there exists ii\in\mathcal{I} such that u¯(y)=u¯i(y)\underaccent{\bar}{u}^{*}(y)=\underaccent{\bar}{u}_{i}^{*}(y), and 𝒩~x𝒩x\tilde{\mathcal{N}}_{x}\subset\mathcal{N}_{x}, it follows u¯(y)u¯(x0)=u¯i(y)u¯i(x0)γiyx0\left\lVert\underaccent{\bar}{u}^{*}(y)-\underaccent{\bar}{u}^{*}(x_{0})\right\rVert=\left\lVert\underaccent{\bar}{u}_{i}^{*}(y)-\underaccent{\bar}{u}_{i}^{*}(x_{0})\right\rVert\leq\gamma_{i}\left\lVert y-x_{0}\right\rVert for some γi>0\gamma_{i}>0. Now, by taking γ:=maxiγi\gamma:=\max_{i\in\mathcal{I}}\gamma_{i}, it follows that u¯(y)u¯(x0)γyx0\left\lVert\underaccent{\bar}{u}^{*}(y)-\underaccent{\bar}{u}^{*}(x_{0})\right\rVert\leq\gamma\left\lVert y-x_{0}\right\rVert for all y𝒩xy\in\mathcal{N}_{x} and hence u¯\underaccent{\bar}{u}^{*} is point-Lipschitz at x0x_{0}. The argument if M>1M>1 is analogous, defining a set l\mathcal{I}_{l} similar to \mathcal{I} for each l[M]l\in[M].

Proposition 5.1 implies in particular that uu^{*} is continuous at x0x_{0}. Note also that the strict feasibility assumption in Proposition 5.1 is satisfied with a prescribed probability if the hypothesis of Proposition 4.6 is satisfied.

6 Simulations

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 1: (a): Time complexity comparison between necessary condition verification (cf. Proposition 4.1) and SOCP solver along the robot trajectory. The label “undetermined” means that the necessary condition is met, from which we may not know if the problem is feasible or not. The label “precision” represents the ratio of instances where the necessary condition indicated infeasibility against the total number of instances where the SOCP was actually infeasible. (b): Time complexity of necessary condition verification and SOCP solver with increasing uncertainty samples (constraints). (c): Log-scaled time complexity comparison of two sufficient conditions (cf. Proposition 4.3 and Proposition 4.6) with the SOCP solver along the robot trajectory.

In this section, we evaluate our results in a ground-robot navigation example. We model the robot motion using unicycle kinematics and take a small distance a=0.05a=0.05 off the wheel axis, cf.[18] to obtain a relative-degree-one model:

[x˙1x˙2θ˙]=([0cos(θ)asin(θ)0sin(θ)acos(θ)001]+j=13Wj(x)ξj))[1vω],\begin{bmatrix}\dot{x}_{1}\\ \dot{x}_{2}\\ \dot{\theta}\end{bmatrix}=\Bigg{(}\begin{bmatrix}0&\cos(\theta)&-a\sin(\theta)\\ 0&\sin(\theta)&a\cos(\theta)\\ 0&0&1\end{bmatrix}+\sum_{j=1}^{3}W_{j}(x)\xi_{j})\Bigg{)}\begin{bmatrix}1\\ v\\ \omega\end{bmatrix},

where vv, ω\omega are the linear and angular velocity, and

W1(x)=[0.02000.02000.0100],W2(x)=[000000000.02],\displaystyle W_{1}(x)\!=\!\begin{bmatrix}0.02&0&0\\ 0.02&0&0\\ 0.01&0&0\end{bmatrix}\!\!,\,W_{2}(x)=\begin{bmatrix}0&0&0\\ 0&0&0\\ 0&0&-0.02\end{bmatrix}\!\!,\,
W3(x)=[00.02cos(θ)0.02asin(θ)00.02sin(θ)0.02acos(θ)000],\displaystyle W_{3}(x)\!=\!\begin{bmatrix}0&0.02\cos(\theta)&-0.02a\sin(\theta)\\ 0&0.02\sin(\theta)&0.02a\cos(\theta)\\ 0&0&0\end{bmatrix}\!\!,

represent the model perturbations in the drift, angular velocity, and orientation. We consider uncertainty samples: ξ1𝒩(0.5,1)\xi_{1}\sim\mathcal{N}(0.5,1), ξ2𝒰(1,1)\xi_{2}\sim\mathcal{U}(-1,1), and ξ3(2,0.2)\xi_{3}\sim\mathcal{B}(2,0.2), where 𝒩\mathcal{N}, 𝒰\mathcal{U}, \mathcal{B} denote normal, uniform, and beta distributions, resp. The optimization programs are solved using the Embedded Conic Solver in CVXPY [19] with an Intel i7 9700K CPU.

We first consider the problem of stabilizing the uncertain unicycle system to a goal position [x1,x2]=[7,7][x_{1}^{*},x_{2}^{*}]=[7,7] with initial state [0,0,0][0,0,0], so we take M=1M=1 in (10). At the initial state, the robot is assumed to have 33 samples {𝝃i}i=13\{\boldsymbol{\xi}_{i}\}_{i=1}^{3} and initial Wasserstein radius r=0.5r=0.5 with risk tolerance ϵ=0.01\epsilon=0.01. As the robot moves, each unsuccessful solver attempt prompts the collection of additional samples, and a corresponding reduction in the ambiguity radius as prescribed by (6). In all the figures presented, the xx-axis represents the simulation timestep, where each timestep is equivalent to 0.020.02 seconds, and the yy-axis denotes the time spent for carrying out the necessary and sufficient condition checks, as well as for running the solver at each timestep.

The time complexity, validity, and precision of Proposition 4.1 are explored in Fig. 1(a) and Fig. 1(b). Fig. 1(a) compares the time complexity of checking the necessary condition in Proposition 4.1 and of solving the corresponding SOCP along the whole robot trajectory. Notably, the SOCP becomes infeasible at around t=3t=3 s and more uncertainty samples are given until feasibility is regained. As expected, when Proposition 4.1 predicts the program is infeasible, such inference is consistently mirrored by the solver. Fig. 1(b) specifically emphasizes the time complexity during data collection stages. As the number of samples increases, the SOCP’s time complexity escalates at a much faster rate than the necessary condition verification, in agreement with Remark 4.10.

Fig. 1(c) compares the time complexity of solving the SOCP and of checking the sufficient conditions in Propositions 4.3 and 4.6. As expected, feasibility validation by either result ensures the actual feasibility of the program by the solver. Checking Proposition 4.3 is more time-consuming than checking Proposition 4.6, cf. Remark 4.10, but has greater accuracy in validating feasibility. Notably, both checks are significantly more efficient than solving the SOCP problem.

We also consider the safe stabilizing problem of the unicycle system. The stabilization goal is [x1,x2]=[5,5][x_{1}^{*},x_{2}^{*}]=[5,5] while the safety goal is to avoid a circular obstacle centered at [3,2][3,2] with radius 11. Fig. 2 compares the time complexity and conservativeness of Proposition 4.1 and Proposition 4.6 for the case M=2M=2 in (10). Proposition 4.1 is valid and requires significantly less time than solving the SOCP, while Proposition 4.6 is also valid and even more efficient.

Refer to caption
Figure 2: Time complexity comparison of necessary (cf. Proposition 4.1) and sufficient (cf. Proposition 4.6) conditions.

7 Conclusions

We studied the feasibility of SOCP problems whose solution provide safe stabilizing controllers for uncertain systems with no prior knowledge of the uncertainty distribution and only a finite number of samples available. We provided a necessary condition and two sufficient conditions to check feasibility, characterized their computational complexity, and showed through simulations their usefulness in practical scenarios. We also showed that the controller obtained by solving the SOCP is point-Lipschitz under fairly general conditions. Future work will consider leveraging the identified feasibility conditions to guide online data-gathering policies that aim to reduce uncertainty about the system dynamics.

References

  • [1] K. Long, V. Dhiman, M. Leok, J. Cortés, and N. Atanasov, “Safe control synthesis with uncertain dynamics and constraints,” IEEE Robotics and Automation Letters, vol. 7, no. 3, pp. 7295–7302, 2022.
  • [2] F. Castañeda, J. J. Choi, B. Zhang, C. J. Tomlin, and K. Sreenath, “Pointwise feasibility of Gaussian process-based safety-critical control under model uncertainty,” in IEEE Conference on Decision and Control, Austin, Texas, USA, 2021, pp. 6762–6769.
  • [3] A. J. Taylor, V. D. Dorobantu, S. Dean, B. Recht, Y. Yue, and A. D. Ames, “Towards robust data driven control synthesis for nonlinear systems with actuation uncertainty,” in IEEE Conf. on Decision and Control, Austin, Texas, USA, 2021, pp. 6469–6476.
  • [4] P. Mestres and J. Cortés, “Feasibility and regularity analysis of safe stabilizing controllers under uncertainty,” Automatica, 2023, submitted.
  • [5] A. Lederer, A. Begzadic, N. Das, and S. Hirche, “Safe learning-based control of elastic joint robots via control barrier functions,” arXiv 2212.00478, 2023.
  • [6] P. Seiler, M. Jankovic, and E. Hellstrom, “Control barrier functions with unmodeled input dynamics using integral quadratic constraints,” IEEE Control Systems Letters, vol. 6, pp. 1664–1669, 2022.
  • [7] K. Long, Y. Yi, J. Cortés, and N. Atanasov, “Safe and stable control synthesis for uncertain system models via distributionally robust optimization,” in American Control Conference, Jun. 2023, pp. 4651–4658.
  • [8] E. D. Sontag, Mathematical Control Theory: Deterministic Finite Dimensional Systems, 2nd ed., ser. TAM.   Springer, 1998, vol. 6.
  • [9] A. D. Ames, S. Coogan, M. Egerstedt, G. Notomista, K. Sreenath, and P. Tabuada, “Control barrier functions: theory and applications,” in European Control Conference, Naples, Italy, Jun. 2019, pp. 3420–3431.
  • [10] A. D. Ames, X. Xu, J. W. Grizzle, and P. Tabuada, “Control barrier function based quadratic programs for safety critical systems,” IEEE Transactions on Automatic Control, vol. 62, no. 8, pp. 3861–3876, 2017.
  • [11] F. Castañeda, J. J. Choi, B. Zhang, C. J. Tomlin, and K. Sreenath, “Gaussian process-based min-norm stabilizing controller for control-affine systems with uncertain input effects and dynamics,” in American Control Conference, New Orleans, LA, May 2021, pp. 3683–3690.
  • [12] K. Long, C. Qian, J. Cortés, and N. Atanasov, “Learning barrier functions with memory for robust safe navigation,” IEEE Robotics and Automation Letters, vol. 6, no. 3, pp. 4931–4938, 2021.
  • [13] A. Ben-Tal, L. E. Ghaoui, and A. Nemirovski, Robust Optimization, ser. Applied Mathematics Series. Princeton University Press, 2009.
  • [14] P. M. Esfahani and D. Kuhn, “Data-driven distributionally robust optimization using the Wasserstein metric: performance guarantees and tractable reformulations,” Mathematical Programming, vol. 171, no. 1-2, pp. 115–166, 2018.
  • [15] A. Nemirovski and A. Shapiro, “Convex approximations of chance constrained programs,” SIAM Journal on Optimization, vol. 17, no. 4, pp. 969–996, 2006.
  • [16] M. S. Lobo, L. Vandenberghe, S. Boyd, and H. Levret, “Applications of second-order cone programming,” Linear algebra and its applications, vol. 284, pp. 193–228, 1997.
  • [17] A.-L. Cholesky, “Sur la résolution numerique des systèmes d’équations linéaires,” Bulletin de la société des amis de la bibliothèque de l’École polytechnique, vol. 39, 2005.
  • [18] J. Cortés and M. Egerstedt, “Coordinated control of multi-robot systems: A survey,” SICE Journal of Control, Measurement, and System Integration, vol. 10, no. 6, pp. 495–503, 2017.
  • [19] S. Diamond and S. Boyd, “CVXPY: A Python-embedded modeling language for convex optimization,” Journal of Machine Learning Research, 2016.