This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Signal Temporal Logic Meets Convex-Concave Programming:
A Structure-Exploiting SQP Algorithm for STL Specifications

Yoshinari Takayama1, Kazumune Hashimoto2, and Toshiyuki Ohtsuka3 1Y. Takayama is with Laboratoire des Signaux et Systèmes, Université Paris-Saclay, CNRS, CentraleSupélec, Gif-sur-Yvette, France.[email protected]2K. Hashimoto is with the Graduate School of Engineering, Osaka University, Suita, Japan. [email protected]3T. Ohtsuka are with the Department of Systems Science, Graduate School of Informatics, Kyoto University, Kyoto, Japan. [email protected]This work was partially supported by JSPS KAKENHI under Grant JP22H01510 and 21K14184 and by JST CREST under Grant JPMJCR201.
Abstract

This study considers the control problem with signal temporal logic (STL) specifications. Prior works have adopted smoothing techniques to address this problem within a feasible time frame and solve the problem by applying sequential quadratic programming (SQP) methods naively. However, one of the drawbacks of this approach is that solutions can easily become trapped in local minima that do not satisfy the specification. In this study, we propose a new optimization method, termed CCP-based SQP, based on the convex-concave procedure (CCP). Our framework includes a new robustness decomposition method that decomposes the robustness function into a set of constraints, resulting in a form of difference of convex (DC) program that can be solved efficiently. We solve this DC program sequentially as a quadratic program by only approximating the disjunctive parts of the specifications. Our experimental results demonstrate that our method has a superior performance compared to the state-of-the-art SQP methods in terms of both robustness and computational time.

I Introduction

As technology becomes more and more powerful, autonomous systems––such as drones and self-driving cars––are required to execute increasingly complex control tasks. Signal temporal logic (STL) has been progressively used as a specification language for these complex robotic tasks in various situations, due to their expressiveness and closeness to natural language [1]. However, the resulting optimization problem is a mixed integer program (MIP) [2, 3], which can easily be computationally intractable and scales poorly with the number of integer variables. To avoid this complexity, recent work formulates the problem as nonlinear programs (NLP) and solve it by sequential quadratic programming (SQP) methods [4, 5]. Although these methods can find a locally optimal solution in a short time compared to the MIP-based methods in general, the major difficulty of this formulation is that the solution can get stuck in a local minimum far from the global optimum as this method approximates the highly non-convex programs to quadratic programs. Moreover, the popular formulation makes this method more inefficient because all nonconvex functions are compressed into the objective function––which results in a coarse approximation by the naive SQP method.

In this study, we address the aforementioned challenges by leveraging the structure of STL control problems. We introduce a novel optimization framework, termed the convex-concave procedure based sequential quadratic programming (CCP-based SQP), which is a variant of the SQP method that solves quadratic programs sequentially. However, unlike traditional SQP methods, this framework approximates the original program at each iteration differently. It is based on the iterative optimization approach of CCP [6, 7]. Stated differently, only the concave portion is linearized at each iteration, allowing the method to retain the complete information of the convex parts. This is in contrast to naive SQP, which can only retain second-order derivative information at each iteration. As a result, our approach was able to synthesize a trajectory that is not only more robust (in terms of robustness score) but is also produced in a shorter time than the naive SQP methods.

Refer to caption
Figure 1: Two-target scenario
Refer to caption
Figure 2: A tree description of the two-target formula with T=1T=1 and Td=1T_{d}=1

The remainder of this paper is organized as follows. We first provide the notation, preliminaries, and problem formulation in Section II. Then, Section III provides our robustness decomposition method that transforms the optimization problem to an efficient DC form. Section IV provides some properties of the resulting CCP-based SQP framework. Section V demonstrates the effectiveness of our method over a state-of-the-art naive SQP method.

II Preliminaries

\mathbb{R}, and \mathbb{Z} are defined as the fields of real and integer numbers, respectively. Given a,ba,b\in\mathbb{Z} with a<ba<b, [a,b][a,b] denotes a set of integers from aa to bb. True and false are denoted by \top and \perp. InI_{n} denotes the identity matrix of size nn. 0n×m0_{n\times m} denotes the zero matrix of size n×mn\times m.

II-A System description

Throughout this paper, we consider a discrete-time linear system:

xt+1=Axt+But,x_{t+1}=Ax_{t}+Bu_{t}, (1)

where An×n,Bn×m,A\in\mathbb{R}^{n\times n},B\in\mathbb{R}^{n\times m}, and xt𝒳n,ut𝒰x_{t}\in\mathcal{X}\subseteq\mathbb{R}^{n},u_{t}\in\mathcal{U}\subseteq m\mathbb{R}^{m} denote the state and input. We assume that 𝒳\mathcal{X} and 𝒰\mathcal{U} are a conjunction of polyhedra. Given an initial state x0𝒳x_{0}\in\mathcal{X} and a sequence of control inputs 𝒖=(u0,,uT1)\boldsymbol{u}=(u_{0},\ldots,u_{T-1}), the sequence of states 𝒙=(x0,,xT)\boldsymbol{x}=(x_{0},\ldots,x_{T}), which we call a trajectory, is uniquely generated.

II-B Specification by Signal temporal logic

In this study, we consider the specifications given in STL. STL is a predicate logic defined for specifying properties for continuous signals [1]. STL consists of predicates μ\mu that can be obtained through a function gμ()g^{\mu}(\cdot) as follows:

μ={ if gμ(𝒙)0 if gμ(𝒙)>0.\mu=\begin{cases}\top\text{ if }g^{\mu}(\boldsymbol{x})\leq 0\\ \perp\text{ if }g^{\mu}(\boldsymbol{x})>0.\end{cases} (2)

In this study, we consider affine functions for gμg^{\mu} of the form gμ=a𝖳xtb:𝒳g^{\mu}=a^{\mathsf{T}}x_{t}-b:\mathcal{X}\rightarrow\mathbb{R}, where ana\in\mathbb{R}^{n} and bb\in\mathbb{R}. In addition to standard boolean operators ,\wedge,\vee, in STL, temporal operators \square (always), \Diamond (eventually), and 𝑼\boldsymbol{U} (until) are also considered. Each temporal operator has an associated bounded time interval [t1,t2][t_{1},t_{2}] where 0t1<t2<0\leq t_{1}<t_{2}<\infty. We assume that STL formulae are written in negation normal form (NNF) without loss of generality [8]. STL formulae in NNF have negations only in front of the predicates [9] and we omit such negations from STL syntax. We refer to this negation-free STL as STL, which is defined by [10]:

φ:=μ|φ|φ|[t1,t2]φ|[t1,t2]φφ1𝑼[t1,t2]φ2.\displaystyle\varphi:=\mu|\vee\varphi|\wedge\varphi|\square_{\left[t_{1},t_{2}\right]}\varphi|\Diamond_{\left[t_{1},t_{2}\right]}\varphi\mid\varphi_{1}\boldsymbol{U}_{\left[t_{1},t_{2}\right]}\varphi_{2}.

The notion of robustness is a useful semantic defined for STL formulae, which is a real-valued function that describes how much a trajectory satisfies an STL formula. Let (𝒙,t)(\boldsymbol{x},t) denote a trajectory starting at timestep tt, i.e., (𝒙,t)=xt,xt+1,,xT(\boldsymbol{x},t)=x_{t},x_{t+1},...,x_{T}. The trajectory length TT should be selected large enough to calculate the satisfaction of a formula (see e.g., [9]). Given an STL formula φ\varphi, we define the reversed robustness with respect to a trajectory 𝒙\boldsymbol{x} and a time tt that can be obtained recursively according to the following quantitative semantics:

Definition 1

(reversed-robustness)

ρrevμ((𝒙,t))=gμ(xt)\displaystyle\rho_{\text{rev}}^{\mu}((\boldsymbol{x},t))=g^{\mu}\left(x_{t}\right) (3a)
ρrevφ1φ2((𝒙,t))=max(ρrevφ1((𝒙,t)),ρrevφ2((𝒙,t)))\displaystyle\rho_{\text{rev}}^{\varphi_{1}\wedge\varphi_{2}}((\boldsymbol{x},t))=\max\left(\rho_{\text{rev}}^{\varphi_{1}}((\boldsymbol{x},t)),\rho_{\text{rev}}^{\varphi_{2}}((\boldsymbol{x},t))\right) (3b)
ρrevφ1φ2((𝒙,t))=min(ρrevφ1((𝒙,t)),ρrevφ2((𝒙,t)))\displaystyle\rho_{\text{rev}}^{\varphi_{1}\vee\varphi_{2}}((\boldsymbol{x},t))=\min\left(\rho_{\text{rev}}^{\varphi_{1}}((\boldsymbol{x},t)),\rho_{\text{rev}}^{\varphi_{2}}((\boldsymbol{x},t))\right) (3c)
ρrev[t1,t2]φ((𝒙,t))=maxt[t+t1,t+t2](ρrevφ((𝒙,t)))\displaystyle\rho_{\text{rev}}^{\square_{\left[t_{1},t_{2}\right]\varphi}}((\boldsymbol{x},t))=\max_{t^{\prime}\in\left[t+t_{1},t+t_{2}\right]}\left(\rho_{\text{rev}}^{\varphi}\left(\left(\boldsymbol{x},t^{\prime}\right)\right)\right) (3d)
ρrev[t1,t2]φ((𝒙,t))=mint[t+t1,t+t2](ρrevφ((𝒙,t)))\displaystyle\rho_{\text{rev}}^{\Diamond_{\left[t_{1},t_{2}\right]\varphi}}((\boldsymbol{x},t))=\min_{t^{\prime}\in\left[t+t_{1},t+t_{2}\right]}\left(\rho_{\text{rev}}^{\varphi}\left(\left(\boldsymbol{x},t^{\prime}\right)\right)\right) (3e)
ρrevφ1𝑼[t1,t2]φ2((𝒙,t))=maxt[t+t1,t+t2](min([ρrevφ1((𝒙,t)),\displaystyle\rho_{\text{rev}}^{\varphi_{1}\boldsymbol{U}_{\left[t_{1},t_{2}\right]}\varphi_{2}}((\boldsymbol{x},t))=\max_{t^{\prime}\in\left[t+t_{1},t+t_{2}\right]}\biggl{(}\min\Bigl{(}\Bigl{[}\rho_{\text{rev}}^{\varphi_{1}}(\left(\boldsymbol{x},t^{\prime}\right)),
mint′′[t+t1,t](ρrevφ2((𝒙,t′′)))]))\displaystyle\min_{t^{\prime\prime}\in[t+t_{1},t^{\prime}]}\bigl{(}\rho_{\text{rev}}^{\varphi_{2}}((\boldsymbol{x},t^{\prime\prime}))\bigr{)}\Bigr{]}\Bigr{)}\biggr{)} (3f)

Note that we define the recursive semantics for the reversed robustness function, i.e., the minus of the original robustness ρorigφ\rho_{\text{orig}}^{\varphi} defined in the literature (see e.g., [11, Definition 2]) as ρrevφ=ρorigφ\rho_{\text{rev}}^{\varphi}=-\rho_{\text{orig}}^{\varphi}. We reverse the sign of the robustness functions ρrevμ\rho_{\text{rev}}^{\mu} of predicates μ\mu, and we interchange the max\max and min\min operators. The reversed-robustness function is introduced as a notational simplification within our robustness decomposition framework. This approach removes unfavorable minus sign symbols in the decomposition and establishes a correspondence between convex parts in the optimization and conjunctive operators in the specification, and similarly for concave parts. It is worth mentioning that this modification does not change all the properties of the original robustness as we simply reverse the signs in the definition of the robustness function.

STL can be represented as a tree [12, 13]. An example is shown below.

Example 1

This example deals with a specification called two-target specifications, which is borrowed from [12]. The formula of this specification is given as:

φ=[0,TTd]([0,Td]B1[0,Td]B2)[0,T]¬O[0,T]G.\varphi=\Diamond_{[0,T-T_{d}]}\left(\square_{[0,T_{d}]}B_{1}\vee\square_{[0,T_{d}]}B_{2}\right)\wedge\square_{[0,T]}\neg O\wedge\Diamond_{[0,T]}G. (4)

Figure 2 shows the graphical representation of the state trajectory satisfying the two-target specification with T=25,Td=5T=25,T_{d}=5. The robot must remain in one of the two regions, either B1B_{1} or B2B_{2}, for a dwelling time of TdT_{d}, and it must reach the goal region GG within the trajectory length of TT while avoiding the obstacle region OO.

Figure 2 shows a tree description of the two-target specification for reversed-robustness functions ρφ\rho^{\varphi} with T=1T=1 and Td=1T_{d}=1. While logically equivalent STL formulas can have different tree descriptions, we adopt the simplified form of the tree description, which is uniquely determined per STL formula (for details, see [12]). The tree nodes marked “LP” at the bottom represent linear predicates, while the top (outermost) node is marked “and.” The depth of this tree is four and the height of the predicates is either “and-or-and-predicate” or “and-or-predicate” regardless of the value of TT and TdT_{d}. Note that although this figure represents the specification with T=1,Td=1T=1,T_{d}=1 for visibility, the depth and height of the tree do not change even when we vary these parameters because changing them will only increase the number of predicates in each depth.

II-C Smooth approximation

The robustness function that results from (3) is not differentiable due to the max\operatorname{\max} and min\operatorname{\min} operators, and thus, we cannot calculate the gradients of this function. Recent papers such as [4] smooth these functions in order to apply gradient-based methods. However, because smoothing both the max\max and min\min of the function leads to a problem that deviates from the original, it is preferable to minimize the parts that are smoothed. Our method requires to smooth only min\min functions. This is because max\max functions in the robustness are rather decomposed in our method as shown in Section III. The popular smooth approximation of the max\max and min\min operators is by using the log-sum-exp (LSE) function. This approximation for min\min operators is given as:

min¯k(a1,,ar):=1klni=1rekai,\overline{\min}_{k}(a_{1},...,a_{r}):=-\frac{1}{k}\ln\sum_{i=1}^{r}e^{-ka_{i}}, (5)

where aia_{i}\in\mathbb{R} and kk is the smooth parameter. Now, we define a new robustness measure that will be used throughout this paper.

Definition 2

(Smoothed reversed-robustness) The smoothed reversed-robustness ρ¯revφ(𝐱)\overline{\rho}^{\varphi}_{\text{rev}}(\boldsymbol{x}) is defined by the quantitative semantics (3) where every min\min operator is replaced by min¯k\overline{\min}_{k}.

II-D Problem statement

This study considers the problem of synthesizing the trajectory that satisfies the STL specification in a maximally robust way. Given the system (1) and the specification φ\varphi and the initial state x0x_{0}, the optimization problem is as follows:

min𝒙,𝒖\displaystyle\min_{\boldsymbol{x},\boldsymbol{u}}\hskip 5.0pt ρ¯revφ(𝒙)\displaystyle\quad\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}) (6a)
s.t. xt+1=Axt+But\displaystyle x_{t+1}=Ax_{t}+Bu_{t} (6b)
xt𝒳,ut𝒰\displaystyle x_{t}\in\mathcal{X},u_{t}\in\mathcal{U} (6c)
ρ¯revφ(𝒙)0\displaystyle\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x})\leq 0 (6d)

Note that the reversed-robustness ρrevφ(𝒙)\rho_{\text{rev}}^{\varphi}(\boldsymbol{x}) is replaced with the smoothed reversed-robustness ρ¯revφ(𝒙)\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}) in (6a) and (6d). Thus, although a feasible solution of the non-smoothed problem always satisfies the specification (4), a feasible solution of (6) does not necessarily satisfy the specification (as it under-approximates the non-smoothed robustness). However, the approximation error can be made arbitrarily small by making the smooth parameter kk sufficiently large [4, 5].

III A structure-aware decomposition of STL formulae

In this section, we present our robustness decomposition framework. The goal of our robustness decomposition is to transform the problem in order to apply a convex-concave procedure (CCP)-based algorithm.

III-A Convex-concave procedure (CCP)

The convex-concave procedure [6, 7] is a heuristic method for finding a local optimum of a class of optimization problems called the difference of convex (DC) programs, which is defined as follows:

Definition 3

(Difference of convex (DC) programs)

minimizef0(𝒛)g0(𝒛)\displaystyle\operatorname{minimize}\quad f_{0}(\boldsymbol{z})-g_{0}(\boldsymbol{z}) (7a)
subject to fi(𝒛)gi(𝒛)0,i{1,,m},\displaystyle\text{subject to }\quad f_{i}(\boldsymbol{z})-g_{i}(\boldsymbol{z})\leq 0,\quad i\in\{1,\ldots,m\}, (7b)

where 𝐳h\boldsymbol{z}\in\mathbb{R}^{h} is the vector of hh optimization variables and the functions fi:hf_{i}:\mathbb{R}^{h}\rightarrow\mathbb{R} and gi:hg_{i}:\mathbb{R}^{h}\rightarrow\mathbb{R} are convex for i{0,,m}i\in\{0,\ldots,m\}. The DC problem (7) can also include equality constraints

pi(𝒛)=qi(𝒛),p_{i}(\boldsymbol{z})=q_{i}(\boldsymbol{z}), (8)

where pip_{i} and qiq_{i} are convex. These equality constraints are expressed as the pair of inequality constraints

pi(𝒛)qi(𝒛)0,qi(𝒛)pi(𝒛)0.\displaystyle p_{i}(\boldsymbol{z})-q_{i}(\boldsymbol{z})\leq 0,\quad q_{i}(\boldsymbol{z})-p_{i}(\boldsymbol{z})\leq 0. (9)

The basic idea of CCP is simple; at each iteration of the optimization of the DC program (7), we replace concave terms with a convex upper bound obtained by linearization and then solve the resulting convex problem using convex optimization algorithms. It can be shown that all of the iterates are feasible if the starting point lies within the feasible set and the objective value converges (possibly to negative infinity, see [6, 14] for proof).

Note that either of the functions pi(𝒛),qi(𝒛)p_{i}(\boldsymbol{z}),q_{i}(\boldsymbol{z}) in equality constraints (8) are also approximated by the CCP. This is because an equality constraint is transformed into two inequalities (9) and both convex functions can make either inequality concave. Therefore, we should avoid introducing equality constraints if possible.

III-B Robustness decomposition

Our decomposition framework alters the problem into an efficient form of a DC program considering the aforementioned features of CCP-based algorithms. This decomposition not only transforms the problem into a DC program but also enhances the efficiency of the overall algorithm. As a result, the resulting program approximates only the truly concave parts, and our approach results in a new SQP method.

To this end, we need to know whether each function in the program (6) is convex or concave. In this program, the robustness function ρ¯revφ\overline{\rho}_{\text{rev}}^{\varphi} is the only part whose curvature (convex or concave) cannot be determined. Thus, if the robustness function ρ¯revφ\overline{\rho}_{\text{rev}}^{\varphi}, which is a composite function consisting of max\max and min\min operators, can be rewritten as a combination of convex and concave functions, the entire program can be reduced to a DC program. The proposed framework achieves this by recursively decomposing the outermost operator of the robustness function until all the arguments of the functions are affine. As this study considers only affine functions for predicates gμ()g^{\mu}(\cdot), if the arguments of a function are all predicates, the convexity or concavity of the function can be determined by examining the operator preceding it, which may be either a max\max or min\min function. The recursive decomposition of a nonconvex robustness function into predicates corresponds to a decomposition of the robustness tree from the top to the bottom, in accordance with the tree structure.

Our method utilizes the idea of the epigraphic reformulation methods. This method transforms the convex (usually linear) cost function into the corresponding epigraph constraints. However, our approach is different from the normal epigraphic reformulation in the sense that we deal with nonconvex cost functions and we have to apply the reformulation recursively. Our robustness decomposition transforms (6) into a more efficient form of DC program. To demonstrate our main idea, for the remainder of this paper, we use the simplified two-target specification described in Example 1 as the specification φ\varphi.

As the robustness function ρ¯revφ\overline{\rho}_{\text{rev}}^{\varphi} is in the cost function, the first procedure is to decompose the function from the cost function into a set of constraints. As we consider the two-target specification, the outermost operator of the robustness function ρ¯revφ\overline{\rho}_{\text{rev}}^{\varphi} is max\max, i.e., ρ¯revφ=max(ρ¯revφ1,ρ¯revφ2,,ρ¯revφr)\overline{\rho}_{\text{rev}}^{\varphi}=\max(\overline{\rho}_{\text{rev}}^{\varphi_{1}},\overline{\rho}_{\text{rev}}^{\varphi_{2}},...,\overline{\rho}_{\text{rev}}^{\varphi_{r}}). We introduce a new variable sξs_{\xi}, and reformulate the program as follows.

min𝒙,𝒖,sξ\displaystyle\min_{\boldsymbol{x},\boldsymbol{u},s_{\xi}}\hskip 1.99997pt sξ\displaystyle s_{\xi} (10a)
s.t. xt+1=Axt+But\displaystyle x_{t+1}=Ax_{t}+Bu_{t} (10b)
xt𝒳,ut𝒰\displaystyle x_{t}\in\mathcal{X},u_{t}\in\mathcal{U} (10c)
sξ0\displaystyle s_{\xi}\leq 0 (10d)
ρ¯revφ1(𝒙)sξρ¯revφr(𝒙)sξ\displaystyle\overline{\rho}_{\text{rev}}^{\varphi_{1}}\left(\boldsymbol{x}\right)\leq s_{\xi}\dots\overline{\rho}_{\text{rev}}^{\varphi_{r}}\left(\boldsymbol{x}\right)\leq s_{\xi} (10e)

Note that the set of inequalities (10e) is equivalent to

max(ρ¯revφ1,ρ¯revφ2,,ρ¯revφr)sξ\max(\overline{\rho}_{\text{rev}}^{\varphi_{1}},\overline{\rho}_{\text{rev}}^{\varphi_{2}},...,\overline{\rho}_{\text{rev}}^{\varphi_{r}})\leq s_{\xi} (11)

although the original semantics of the variable sξs_{\xi} is the equality constraint:

max(ρ¯revφ1,ρ¯revφ2,,ρ¯revφr)=sξ.\max(\overline{\rho}_{\text{rev}}^{\varphi_{1}},\overline{\rho}_{\text{rev}}^{\varphi_{2}},...,\overline{\rho}_{\text{rev}}^{\varphi_{r}})=s_{\xi}. (12)

However, it is well known that this kind of max\max transformation is known to be equivalent, particularly for the convex case. Although (10) is nonconvex, we can also show that both (6) and (10) are equivalent (see [15] and [16, Section 8.3.4.4]). The proof of this equivalence is provided in Appendix -A for a later explanation (in particular (17)).

As each ρ¯revφi\overline{\rho}_{\text{rev}}^{\varphi_{i}} for i=1,,ri=1,\ldots,r is a robustness function, each inequality in (10e) can be restated as a constraint in one of the following two forms, depending on whether the outermost operator is max\max or min\min:

max(ρ¯revΦ1,,ρ¯revΦymax)sξ,\displaystyle\max(\overline{\rho}_{\text{rev}}^{\Phi_{1}},...,\overline{\rho}_{\text{rev}}^{\Phi_{y_{\max}}})\leq s_{\xi}, (13)
min¯(ρ¯revΨ1,,ρ¯revΨymin)sξ,\displaystyle\overline{\min}(\overline{\rho}_{\text{rev}}^{\Psi_{1}},...,\overline{\rho}_{\text{rev}}^{\Psi_{y_{\min}}})\leq s_{\xi}, (14)

where functions ρ¯revΦj(j{1,,ymax})\overline{\rho}_{\text{rev}}^{\Phi_{j}}(j\in\{1,...,y_{\max}\}) (resp. ρ¯revΨj(j{1,,ymin})\overline{\rho}_{\text{rev}}^{\Psi_{j}}(j\in\{1,...,y_{\min}\})) are robustness functions associated with Φj\Phi_{j} (resp. Ψj\Psi_{j}), which are the subformulas of φi(i{1,,r})\varphi_{i}(i\in\{1,...,r\}). Note that these arguments of the max\max and min\min functions are still not necessarily predicates. We continue the following steps until all the arguments in each function become the predicates.

For inequality constraints of the form (13), we transform it as follows:

ρ¯revΦ1sξ,ρ¯revΦ2sξ,,ρ¯revΦymaxsξ.\displaystyle\overline{\rho}_{\text{rev}}^{\Phi_{1}}\leq s_{\xi},\overline{\rho}_{\text{rev}}^{\Phi_{2}}\leq s_{\xi},...,\overline{\rho}_{\text{rev}}^{\Phi_{y_{\max}}}\leq s_{\xi}. (15)

This is of course equivalent to (13).

For constraints of the forms (14), we first check whether each argument of the min¯\overline{\min} function is a predicate or not. If not, we replace such an argument with a new variable. Because we consider the simplified two-target specification, the outermost operator of the function ρ¯revΨj(j{1,,ymin})\overline{\rho}_{\text{rev}}^{\Psi_{j}}(j\in\{1,...,y_{\min}\}) is max\max, i.e., ρ¯revΨj=max(ρ¯revψ1,,ρ¯revψh)\overline{\rho}_{\text{rev}}^{\Psi_{j}}=\max(\overline{\rho}_{\text{rev}}^{\psi_{1}},...,\overline{\rho}_{\text{rev}}^{\psi_{h}}). We first replace function ρ¯revΨj\overline{\rho}_{\text{rev}}^{\Psi_{j}} in (14) by a new variable snews_{\text{new}} as

min¯(ρ¯revΨ1,,snew,,ρ¯revΨymin)sξ.\overline{\min}(\overline{\rho}_{\text{rev}}^{\Psi_{1}},...,s_{\text{new}},...,\overline{\rho}_{\text{rev}}^{\Psi_{y_{\min}}})\leq s_{\xi}. (16)

Then, we add the following constraints:

max(ρ¯revψ1,,ρ¯revψh)snew.\max(\overline{\rho}_{\text{rev}}^{\psi_{1}},...,\overline{\rho}_{\text{rev}}^{\psi_{h}})\leq s_{\text{new}}. (17)

Although the inequality constraint (17) should be equality, this modification is justified by the similar discussion for (10e) even though the equivalence of this transformation is not immediately obvious (see the proof of Theorem 1).

All the inequalities in (15) and (17) are either in the max\max form (13) or the min\min form (14). Thus, subsequent to these steps, the constraints of the forms (13), (14) are decomposed into several constraints in the same forms again. We repeat this procedure until we reach the predicates, which is the bottom of the tree in Figure 2. When we finish the above recursive manipulations, (6) is transformed into an DC program as follows:

min𝒙,𝒖,sξ,𝒔max,𝒔min\displaystyle\min_{\boldsymbol{x},\boldsymbol{u},s_{\xi},\boldsymbol{s}_{\max},\boldsymbol{s}_{\min}}\hskip 1.99997pt sξ\displaystyle s_{\xi} (18a)
s.t. xt+1=Axt+But\displaystyle x_{t+1}=Ax_{t}+Bu_{t} (18b)
xt𝒳,ut𝒰\displaystyle x_{t}\in\mathcal{X},u_{t}\in\mathcal{U} (18c)
sξ0,\displaystyle s_{\xi}\leq 0, (18d)
ρ¯revλ1(1)s(1)max,,ρ¯revλymax(1)s(1)maxρ¯revλ1(v)s(v)max,,ρ¯revλymax(v)s(v)max}from maxfunctions\displaystyle\left.\begin{tabular}[]{l}$\overline{\rho}_{\text{rev}}^{\lambda^{(1)}_{1}}\leq s^{(1)}_{\max},...,\overline{\rho}_{\text{rev}}^{\lambda^{(1)}_{y_{\max}}}\leq s^{(1)}_{\max}$\\ $\quad\quad\vdots\hskip 71.13188pt\vdots$\\ $\overline{\rho}_{\text{rev}}^{\lambda^{(v)}_{1}}\leq s^{(v)}_{\max},...,\overline{\rho}_{\text{rev}}^{\lambda^{(v)}_{y_{\max}}}\leq s^{(v)}_{\max}$\\ \end{tabular}\right\}\begin{tabular}[]{l}\text{from }$\max$\\ \text{functions}\end{tabular} (18j)
min¯(ρ¯revμ1(1),,ρ¯revμymin(1))s(1)minmin¯(ρ¯revμ1(w),,ρ¯revμymin(w))s(w)min,}from minfunctions\displaystyle\left.\begin{tabular}[]{l}$\overline{\min}(\overline{\rho}_{\text{rev}}^{\mu^{(1)}_{1}},...,\overline{\rho}_{\text{rev}}^{\mu^{(1)}_{y_{\min}}})\leq s^{(1)}_{\min}$\\ $\quad\quad\vdots$\\ $\overline{\min}(\overline{\rho}_{\text{rev}}^{\mu^{(w)}_{1}},...,\overline{\rho}_{\text{rev}}^{\mu^{(w)}_{y_{\min}}})\leq s^{(w)}_{\min},$\\ \end{tabular}\right\}\begin{tabular}[]{l}\text{from }$\min$\\ \text{functions}\end{tabular} (18p)

where all the formulas λ1(),,λymax(),μ1(),,μymin()\lambda_{1}^{(\cdot)},...,\lambda_{y_{\max}}^{(\cdot)},\mu_{1}^{(\cdot)},...,\mu_{y_{\min}}^{(\cdot)} in (18j) and (18p) are predicates and 𝒔max=(smax(1),,smax(v))\boldsymbol{s}_{\max}=(s^{(1)}_{\max},...,s^{(v)}_{\max}) and 𝒔min=(smin(1),,smin(w))\boldsymbol{s}_{\min}=(s^{(1)}_{\min},...,s^{(w)}_{\min}). Note that, with some abuse of notation, some variables in 𝒔max,𝒔min\boldsymbol{s}_{\max},\boldsymbol{s}_{\min} in constraints (18j),(18p) may represent sξs_{\xi} and snews_{\text{new}}, and some arguments of the min¯\overline{\min} function in (18p) are not affine predicates but the new variables.

Important properties of this program (18) are stated below.

Proposition 1

Let 𝐳=(𝐱,𝐮,sξ,𝐬max,𝐬min)\boldsymbol{z}^{\prime}=(\boldsymbol{x}^{\prime},\boldsymbol{u}^{\prime},s^{\prime}_{\xi},\boldsymbol{s}^{\prime}_{\max},\boldsymbol{s}^{\prime}_{\min}) denote a feasible solution for program (18). Then, ρ¯revφ(𝐱)0\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime})\leq 0 holds.

Proof:

See Appendix -B. ∎

Proposition 2

If 𝐳\boldsymbol{z}^{\prime} is feasible for (18), (𝐱,𝐮)(\boldsymbol{x}^{\prime},\boldsymbol{u}^{\prime}) is feasible for (6).

Proof:

The inequality ρ¯revφ(𝒙)0\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime})\leq 0 holds from Proposition 1, which means that (6d) holds. Note that (6b) and (6c) are the same as (18b) and (18c). ∎

Theorem 1

The global optimum of two programs: (6) and (18) is the same.

Proof:

See Appendix -C. ∎

Refer to caption
(a) Computational times

Refer to caption

(b) Robustness scores
Figure 3: Computation time and robustness score of the three methods over every 5 horizons from T=50T=50 to 140140.

IV Properties of the subproblem

After we obtain the transformed program (18), we apply the CCP method described in Subsection III-A. We majorize the min¯\overline{\min} operator in (18p) at the current point of each iteration. In general, the resulting program becomes a convex program such as a second-order cone program (SOCP) and CCP has to solve the convex program sequentially. However, the subproblem we solve at each iteration is a linear program or a quadratic program, and not the general convex program as stated in the following theorem:

Theorem 2

The program (18) after the majorization by CCP is a linear program (LP). If we add the quadratic cost function, the program can be written as a convex quadratic program (QP).

Proof:

In the program (18), the cost function is sξs_{\xi} and the constraints including (18j) are all affine except for concave constraints (14). As CCP linearizes all the concave constraints, the program resulting from the majorization of (18) becomes an LP. ∎

Let 𝒫LP\mathcal{P}_{\text{LP}} denote this LP in the following. It is worth mentioning that inequalities in (18j) are all affine inequalities as we remove all the max\max operators in the end. As a corollary of the above theorem, our CCP-based method sequentially solves LP. If we add the quadratic cost function in (18a), our method becomes sequential quadratic programming (SQP).

Moreover, as all the constraints coming from the robustness decomposition framework in Section III do not have any equality constraints, the non-convex parts of (18) are only the inequality constraints (18p), which comes from min\min functions of the robustness function.

Proposition 3

The program (18) has no equality constraints except for the state equations (18b), which are affine constraints.

Hence, the CCP only approximates the truly concave portions of (18p) arising from the min\min functions in the robustness function. It is noteworthy that enforcing the constraints, such as (10e) and (17), as equality constraints would require the CCP to approximate them as described in Subsection III-A, thereby leading to undesirable approximations. In contrast, the proposed method minimizes the number of parts requiring approximation, which enhances the algorithm’s performance.

Furthermore, the feasible region of the program 𝒫LP\mathcal{P}_{\text{LP}} is interior to the feasible region of the program (18).

Proposition 4

A feasible solution of the linear program 𝒫LP\mathcal{P}_{\text{LP}} is a feasible solution to the program (18).

Proof:

This is because the first-order approximations of CCP at each step are global over-estimators; i.e., fi(𝒛)gi(𝒛)fi(𝒛)gi(𝒛())+gi(𝒛())(𝒛()𝒛)f_{i}(\boldsymbol{z})-g_{i}(\boldsymbol{z})\leq f_{i}(\boldsymbol{z})-g_{i}(\boldsymbol{z}_{(\cdot)})+\nabla g_{i}(\boldsymbol{z}_{(\cdot)})^{\top}(\boldsymbol{z}_{(\cdot)}-\boldsymbol{z}) where 𝒛()\boldsymbol{z}_{(\cdot)} is the current point of variable 𝒛\boldsymbol{z}. ∎

Finally, as a direct consequence of Propositions 1 and 4, the solution obtained by our CCP-based SQP method can always satisfy the specification, which is the most important property to guarantee safety.

Theorem 3

Let 𝐱′′\boldsymbol{x}^{\prime\prime} denote a feasible trajectory of the program 𝒫LP\mathcal{P}_{\text{LP}}. Then 𝐱′′\boldsymbol{x}^{\prime\prime} always satisfies the STL specification φ\varphi, i.e., 𝐱′′φ\boldsymbol{x}^{\prime\prime}\vDash\varphi in the limit kk\rightarrow\infty.

Proof:

Based on Propositions 1 and 4, ρ¯revφ(𝒙′′)0\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime\prime})\leq 0 holds. The limit as kk\rightarrow\infty results in the smoothed reversed-robustness ρ¯revφ(𝒙′′)\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime\prime}) being equivalent to the original reversed-robustness ρrevφ(𝒙′′)\rho_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime\prime}). Therefore, ρrevφ(𝒙′′)0\rho_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime\prime})\leq 0 as kk\rightarrow\infty, which proves the statement. ∎

In practice, we select a sufficiently large kk to minimize the approximation error.

V Numerical Experiments

We illustrate the advantages of our proposed method through the two-target scenarios with varying time horizons ranging from T=50T=50 to 140140. All experiments were conducted on a 2020 MacBook Air with an Apple M1 processor and 8 GB of RAM111The source code is available at https://github.com/yotakayama/STLCCP.. The state xtx_{t} and input utu_{t} are defined as xt=[px,py,p˙x,p˙y]𝖳,ut=[p¨x,p¨y]𝖳x_{t}=\left[p_{x},p_{y},\dot{p}_{x},\dot{p}_{y}\right]^{\mathsf{T}},u_{t}=\left[\ddot{p}_{x},\ddot{p}_{y}\right]^{\mathsf{T}}, where pxp_{x} is the horizontal position of the robot and pyp_{y} is the vertical position. As for the system dynamics, we use a double integrator, i.e., a system (1) with the matrix

A\displaystyle A =[I2I202×2I2],B=[02×2I2].\displaystyle=\left[\begin{array}[]{ll}I_{2}&I_{2}\\ 0_{2\times 2}&I_{2}\end{array}\right],\quad B=\left[\begin{array}[]{c}0_{2\times 2}\\ I_{2}\end{array}\right]. (23)

We implemented our algorithm in Python using CVXPY [17] as the interface to the optimizer. We chose the GUROBI(ver. 10)’s QP-solver with the default option. We took the penalty CCP [7] to solve the problem, which can admit the violation of constraints during the optimization procedures by penalizing the violation with variables. We introduced this penalty variable only for the concave constraints because the CCP only approximates such concave parts. Additionally, we added a small quadratic cost function. The weights on the penalty variables and the quadratic cost functions were 50.0 and 0.001, respectively. The resulting cost function (18a) is represented as:

sξ+50.0i=0w(si)+0.001t=0T(xt𝖳Qxt+ut𝖳Rut),s_{\xi}+50.0\sum_{i=0}^{w}(s_{i})+0.001\sum_{t=0}^{T}(x_{t}^{\mathsf{T}}Qx_{t}+u_{t}^{\mathsf{T}}Ru_{t}), (24)

where Q=diag(0,0,1,1),R=I2Q=\operatorname{diag}(0,0,1,1),R=I_{2} are positive semidefinite symmetric matrices, and sis_{i} for i{1,,w}i\in\{1,...,w\} (ww is the number of constraints in (18p)) is the penalty variable added to the r.h.s. of each concave constraints (18p) satisfying si0s_{i}\leq 0. The initial state is fixed as x0=[2.0,2.0,0,0]x_{0}=[2.0,2.0,0,0], and the bounds on the state and input variables are 𝒳=[0.0,0.0,1.0,1.0]𝖳<xt<[10.0,10.0,1.0,1.0]𝖳\mathcal{X}=[0.0,0.0,-1.0,-1.0]^{\mathsf{T}}<x_{t}<[10.0,10.0,1.0,1.0]^{\mathsf{T}} and 𝒰=[0.2,0.2]𝖳<ut<[0.2,0.2]𝖳\mathcal{U}=[-0.2,-0.2]^{\mathsf{T}}<u_{t}<[0.2,0.2]^{\mathsf{T}}, respectively.

We compared our method with a popular MIP-based method and a state-of-the-art NLP-based method. For the MIP-based approach, we formulated the problem as a MIP using the encoding framework in [11] and solved the problem with the GUROBI solver (GUROBI is often the fastest MIP solver). However, for the NLP-based method, we used the naive sequential quadratic programming (SQP) approach [18] using the smoothing method proposed in [5]. All the parameters of the two methods were the default settings.

Figure 3 compares convergence time and robustness score of the three methods from T=50T=50 to 140140. The SNOPT-NLP method (orange) failed to output a feasible trajectory at 3 horizons (T=55,90,95T=55,90,95) out of the 19 horizons, which are shown by the orange cross marker ×\color[rgb]{1,.5,0}\times in the left plot, and their robustness scores have been cut off in the right plot as they are -\infty. In contrast, the proposed method (blue) failed only 2 times out of the 95 total trials (5 trials at each horizon). The GUROBI-MIP method (green) is truncated at the horizon T=100T=100 due to the computation time 7371.37371.3 s, which exceeds the range of the plotted region. Therefore, subsequent results for longer horizons are not displayed for the sake of visibility. The plots of our method are the averages of different trials varying the initial values of variables, whereas the other two plots are from one trial. All the scores in the right plot are calculated using the original robustness function.

In summary, our proposed method outperformed the state-of-the-art SNOPT-NLP method with respect to both the robustness score and the computation time. Specifically, as shown in the left plot, the computation time of our proposed method remained at a consistently low level even when the horizon increased. In contrast, the other two methods displayed volatility or were infeasibly slow, particularly above T=100T=100. Furthermore, the right plot demonstrates that our proposed method did not compromise the robustness score to shorten the computation time. On the contrary, our method consistently achieved a higher robustness score than the SNOPT-NLP method. At some horizons, our method achieved a score of 0.50.5, with the global optimum obtained by the MIP-based method. Notably, the plot of our method represents the average of all the trials, indicating that the feasible trajectories our method outputs are all the global optimum. In contrast, the SNOPT-NLP method sometimes generated failure trajectories with a robustness score of -\infty in some horizons (at T=55,90,95T=55,90,95), and all of the robustness scores (orange) were lower than those of the proposed method (blue).

Finally, Figure 4 shows the trajectories generated by the proposed method with T=50T=50. Our method produced a range of satisfactory trajectories, each dependent on the initial values of the system’s variables.

Refer to caption

Figure 4: Example of trajectories generated by the proposed method with 5 random initial values for T=50T=50.

VI Conclusion

In this study, we have introduced the CCP-based SQP algorithm for control problems involving STL specifications leveraging the inherent structures of STL. The subproblem of our proposed method is an efficient quadratic subprogram, achieved by approximating solely the concave constraints of the program that correspond to the disjunctive nodes of the STL tree.

Future work includes extending the proposed method to non-affine predicates, nonlinear systems, and non-smooth cases. While this paper focused on linear systems to establish an efficient SQP approach, the same methodology can be applied to tackle more complex cases.

References

  • [1] G. E. Fainekos and G. J. Pappas, “Robustness of temporal logic specifications for continuous-time signals,” Theoretical Computer Science, vol. 410, no. 42, pp. 4262–4291, 2009.
  • [2] S. Karaman and E. Frazzoli, “Vehicle routing problem with metric temporal logic specifications,” in IEEE Conference on Decision and Control, pp. 3953–3958, 2008.
  • [3] V. Raman, A. Donzé, M. Maasoumy, R. M. Murray, A. Sangiovanni-Vincentelli, and S. A. Seshia, “Model predictive control with signal temporal logic specifications,” in IEEE Conference on Decision and Control, pp. 81–87, 2014.
  • [4] Y. V. Pant, H. Abbas, and R. Mangharam, “Smooth operator: Control using the smooth robustness of temporal logic,” in IEEE Conference on Control Technology and Applications, pp. 1235–1240, 2017.
  • [5] Y. Gilpin, V. Kurtz, and H. Lin, “A smooth robustness measure of signal temporal logic for symbolic control,” IEEE Control Systems Letters, vol. 5, no. 1, pp. 241–246, 2021.
  • [6] T. Lipp and S. Boyd, “Variations and extension of the convex–concave procedure,” Optimization and Engineering, vol. 17, no. 2, pp. 263–287, 2016.
  • [7] X. Shen, S. Diamond, Y. Gu, and S. Boyd, “Disciplined convex-concave programming,” in IEEE Conference on Decision and Control (CDC), pp. 1009–1014, 2016.
  • [8] S. Sadraddini and C. Belta, “Formal synthesis of control strategies for positive monotone systems,” IEEE Transactions on Automatic Control, vol. 64, no. 2, pp. 480–495, 2019.
  • [9] ——, “Robust temporal logic model predictive control,” in Annual Allerton Conference on Communication, Control, and Computing, pp. 772–779, 2015.
  • [10] C. Baier and J.-P. Katoen, Principles of Model Checking.   MIT Press, 2008.
  • [11] C. Belta and S. Sadraddini, “Formal methods for control synthesis: An optimization perspective,” Annual Review of Control, Robotics, and Autonomous Systems, vol. 2, no. 1, pp. 115–140, 2019.
  • [12] V. Kurtz and H. Lin, “Mixed-integer programming for signal temporal logic with fewer binary variables,” IEEE Control Systems Letters, vol. 6, pp. 2635–2640, 2022. Their source code is available at https://stlpy.readthedocs.io.
  • [13] D. Sun, J. Chen, S. Mitra, and C. Fan, “Multi-agent motion planning from signal temporal logic specifications,” IEEE Robotics and Automation Letters, 2022.
  • [14] B. K. Sriperumbudur and G. R. G. Lanckriet, “On the convergence of the concave-convex procedure,” in Advances in Neural Information Processing Systems, vol. 22, 2009.
  • [15] L. Lindemann and D. V. Dimarogonas, “Robust control for signal temporal logic specifications using discrete average space robustness,” Automatica, vol. 101, pp. 377–387, 2019.
  • [16] G. C. Calafiore and L. El Ghaoui, Optimization Models.   Cambridge University Press, 2014.
  • [17] S. Diamond and S. Boyd, “CVXPY: A python-embedded modeling language for convex optimization,” Journal of Machine Learning Research, vol. 17, no. 1, p. 2909–2913, 2016.
  • [18] P. E. Gill, W. Murray, and M. A. Saunders, “SNOPT: An SQP algorithm for large-scale constrained optimization,” SIAM Review, vol. 47, pp. 99–131, 2005.

-A Equivalence between the two programs: (6) and (10)

We prove the following equivalences.

Proposition 5

Both formulations (6) and (10) are equivalent in the sense that the global optimum is the same.

Proof:

We demonstrate that if (𝒙,𝒖)(\boldsymbol{x}^{*},\boldsymbol{u}^{*}) is optimal for (6), then (𝒙,𝒖,sξ=ρ¯revφ(𝒙))(\boldsymbol{x}^{*},\boldsymbol{u}^{*},s_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}^{*}\right)) is optimal for (10). Firstly, note that for any feasible solution (𝒙,𝒖)(\boldsymbol{x},\boldsymbol{u}) of (6), (𝒙,𝒖,sξ=ρ¯revφ(𝒙))(\boldsymbol{x},\boldsymbol{u},s_{\xi}=\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x})) is feasible for (10) as (10e) becomes an identity by this substitution sξ=ρ¯revφ(𝒙)s_{\xi}=\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}\right). Thus, (𝒙,𝒖,sξ=ρ¯revφ(𝒙))(\boldsymbol{x}^{*},\boldsymbol{u}^{*},s_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}^{*}\right)) is also feasible for (10).

Suppose this solution (𝒙,𝒖,sξ=ρ¯revφ(𝒙))(\boldsymbol{x}^{*},\boldsymbol{u}^{*},s_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}^{*}\right)) is not optimal for (10). Then, there exists another feasible solution (𝗑,𝗎,𝗌ξ)(\mathsf{x}^{*},\mathsf{u}^{*},\mathsf{s}^{*}_{\xi}) with a better objective, i.e., such that 𝗌ξ<sξ=ρ¯revφ(𝒙)\mathsf{s}^{*}_{\xi}<s_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{*}). Since this feasible solution also satisfies (10e), that is, ρ¯revφ(𝗑)𝗌ξ\overline{\rho}_{\text{rev}}^{\varphi}\left(\mathsf{x}^{*}\right)\leq\mathsf{s}^{*}_{\xi}, the inequality ρ¯revφ(𝗑)<ρ¯revφ(𝒙)\overline{\rho}_{\text{rev}}^{\varphi}\left(\mathsf{x}^{*}\right)<\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}^{*}\right) holds. Considering that (𝗑,𝗎)(\mathsf{x}^{*},\mathsf{u}^{*}) is feasible for (6) (as one of the inequality conditions in (10e) must become the equality condition when the cost function sξs_{\xi} takes a local minimum), the inequality ρ¯revφ(𝗑)<ρ¯revφ(𝒙)\overline{\rho}_{\text{rev}}^{\varphi}\left(\mathsf{x}^{*}\right)<\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}^{*}\right) contradicts the optimality of 𝒙\boldsymbol{x}^{*} for (6). Therefore, (𝒙,𝒖,𝒔ξ=ρ¯revφ(𝒙))(\boldsymbol{x}^{*},\boldsymbol{u}^{*},\boldsymbol{s}_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}\left(\boldsymbol{x}^{*}\right)) is optimal for (10), and the global optimum is the same. ∎

-B Proof of Proposition 1

Proof:

We present a proof sketch. Let us first consider the initial replacement of the robustness function ρ¯revΨi(𝒙)\overline{\rho}_{\text{rev}}^{\Psi_{i}}(\boldsymbol{x}^{\prime}) with a new variable snews_{\text{new}}. This transformation converts inequality (14) into two inequalities, (16) and (17). Note that (17) represents the inequality

ρ¯revΨi(𝒙)snew,\overline{\rho}_{\text{rev}}^{\Psi_{i}}(\boldsymbol{x}^{\prime})\leq s_{\text{new}}, (25)

although this inequality sign should be an equality if snews_{\text{new}} were a substituting variable. Nevertheless, due to the fact that min¯\overline{\min} is an increasing function with respect to each argument, we have the inequality

min¯(ρ¯revΨ1,,ρ¯revΨi,,ρ¯revΨymin)min¯(ρ¯revΨ1,,snew,,ρ¯revΨymin),\overline{\min}(\overline{\rho}_{\text{rev}}^{\Psi_{1}},...,\overline{\rho}_{\text{rev}}^{\Psi_{i}},...,\overline{\rho}_{\text{rev}}^{\Psi_{y_{\min}}})\leq\overline{\min}(\overline{\rho}_{\text{rev}}^{\Psi_{1}},...,s_{\text{new}},...,\overline{\rho}_{\text{rev}}^{\Psi_{y_{\min}}}),

Combining this inequality with (16), we conclude that inequality (14) holds even after the transformation with snews_{\text{new}}. Similar analogies can be used for all subsequent transformations of non-convex robustness functions with new variables that follow the same pattern. Thus, ultimately, the original inequalities in (10e) hold. By this inequality and (18d), we have ρ¯revφ(𝒙)0\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime})\leq 0. ∎

From a tree perspective, this proof demonstrates that by replacing the robustness of any subformula with a new variable that is greater than or equal to the original value (as shown in (25)), the parent node’s robustness score, due to the monotonicity of the max\max and min\min functions, also becomes greater than or equal to the original value (as shown in (14)). This property extends further to the values of the grandparents and subsequent nodes, ultimately resulting in the cost function snews_{\text{new}}, representing the top-node value, being greater than the robustness value of φ\varphi.

-C Proof of Theorem 1

We first prove the following proposition, which states the converse of Proposition 2:

Proposition 6

Let Φparent(i)\Phi^{(i)}_{\text{parent}} denotes the parent node of formulas Φ1(i),,Φymax(i)\Phi^{(i)}_{1},...,\Phi^{(i)}_{y_{\max}} for i{1,,v}i\in\{1,...,v\}, and Ψparent(i)\Psi^{(i)}_{\text{parent}} denotes the parent node of formulas Ψ1(i),,Ψymax(i)\Psi^{(i)}_{1},...,\Psi^{(i)}_{y_{\max}} for i{1,,w}i\in\{1,...,w\}. Let (𝐱,𝐮)(\boldsymbol{x}^{*},\boldsymbol{u}^{*}) denote a feasible solution of (6), then the following solution 𝐳\boldsymbol{z}^{*} is feasible for (18).

𝒛:=(𝒙,𝒖,sξ=ρ¯revφ(𝒙),𝒔max=(ρ¯revΦparent(1)(𝒙),\displaystyle\boldsymbol{z}^{*}:=(\boldsymbol{x}^{*},\boldsymbol{u}^{*},s_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{*}),\boldsymbol{s}_{\max}^{*}=(\overline{\rho}_{\text{rev}}^{\Phi^{(1)}_{\text{parent}}}(\boldsymbol{x}^{*}),
,ρ¯revΦparent(v)(𝒙)),𝒔min=(ρ¯revΨparent(1)(𝒙),,ρ¯revΨparent(w)(𝒙)))\displaystyle...,\overline{\rho}_{\text{rev}}^{\Phi^{(v)}_{\text{parent}}}(\boldsymbol{x}^{*})),\boldsymbol{s}_{\min}^{*}=(\overline{\rho}_{\text{rev}}^{\Psi^{(1)}_{\text{parent}}}(\boldsymbol{x}^{*}),...,\overline{\rho}_{\text{rev}}^{\Psi^{(w)}_{\text{parent}}}(\boldsymbol{x}^{*}))) (26)
Proof:

When all the inequalities in (18j) and (18p) becoming identity equations, the feasible region of (18) becomes identical to that of (6), i.e., S=Proj(SDC)S=\operatorname{Proj}(S_{\text{DC}}) where SS and SDCS_{\text{DC}} denote the feasible region of (6) and (18) respectively, and the operator Proj\operatorname{Proj} denotes the projection mapping by the substitution 𝒛()\boldsymbol{z}^{*}(\cdot) from (𝒙,𝒖,sξ,𝒔max,𝒔min)(\boldsymbol{x},\boldsymbol{u},s_{\xi},\boldsymbol{s}_{\max},\boldsymbol{s}_{\min}) to (𝒙,𝒖)(\boldsymbol{x},\boldsymbol{u}). Therefore, the feasible region of (6) is included in that of (18). ∎

We now show the following optimality equivalence:

Proposition 7

If a feasible solution (𝐱,𝐮\boldsymbol{x}^{*},\boldsymbol{u}^{*}) is optimal for (6), then 𝐳\boldsymbol{z}^{*} in (6) is optimal for (18).

Proof:

From Proposition 6, if 𝒛\boldsymbol{z}^{*} is not the optimal solution for (18), there exists another feasible solution (𝒙,𝒖,sξ,𝒔max,𝒔min)(\boldsymbol{x}^{\prime},\boldsymbol{u}^{\prime},s_{\xi}^{\prime},\boldsymbol{s}_{\max}^{\prime},\boldsymbol{s}_{\min}^{\prime}), with a better objective value, i.e., such that sξ<sξ=ρ¯revφ(𝒙)s_{\xi}^{\prime}<s_{\xi}^{*}=\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{*}). As in the proof of Proposition 1, ρ¯revφ(𝒙)sξ\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime})\leq s_{\xi}^{\prime}. Therefore, ρ¯revφ(𝒙)<ρ¯revφ(𝒙)\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{\prime})<\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{*}). However, as Proposition 2 shows that 𝒙\boldsymbol{x}^{\prime} is also feasible for (6), this inequality contradicts the fact that 𝒙\boldsymbol{x}^{*} is optimal for (6). Therefore, 𝒛\boldsymbol{z}^{*} is optimal for (18). ∎

From this proposition, the global optimum of the two programs (6) and (18) is the same, that is, ρ¯revφ(𝒙)\overline{\rho}_{\text{rev}}^{\varphi}(\boldsymbol{x}^{*}). Moreover, we can also prove analogously that if a feasible solution 𝒛\boldsymbol{z}^{\prime} is optimal for (18), then (𝒙,𝒖)(\boldsymbol{x}^{\prime},\boldsymbol{u}^{\prime}) is optimal for (6).