Scaling Robust Optimization for Multi-Agent Robotic Systems: A Distributed Perspective
Abstract
This paper presents a novel distributed robust optimization scheme for steering distributions of multi-agent systems under stochastic and deterministic uncertainty. Robust optimization is a subfield of optimization which aims in discovering an optimal solution that remains robustly feasible for all possible realizations of the problem parameters within a given uncertainty set. Such approaches would naturally constitute an ideal candidate for multi-robot control, where in addition to stochastic noise, there might be exogenous deterministic disturbances. Nevertheless, as these methods are usually associated with significantly high computational demands, their application to multi-agent robotics has remained limited. The scope of this work is to propose a scalable robust optimization framework that effectively addresses both types of uncertainties, while retaining computational efficiency and scalability. In this direction, we provide tractable approximations for robust constraints that are relevant in multi-robot settings. Subsequently, we demonstrate how computations can be distributed through an Alternating Direction Method of Multipliers (ADMM) approach towards achieving scalability and communication efficiency. Simulation results highlight the performance of the proposed algorithm in effectively handling both stochastic and deterministic uncertainty in multi-robot systems. The scalability of the method is also emphasized by showcasing tasks with up to 100 agents. The results of this work indicate the promise of blending robust optimization, distribution steering and distributed optimization towards achieving scalable, safe and robust multi-robot control.
I Introduction
Multi-agent optimization and control problems are emerging increasingly often in a vast variety of robotics applications such as multi-vehicle coordination [16, 43], swarm robotics [37], cooperative manipulation [13] and others. As the size and complexity of such systems has been rapidly escalating, a substantial requirement for developing scalable and distributed algorithms for tackling these problems has been developing as well [35, 38, 46]. In the meantime, accompanying such algorithms with safety and robustness guarantees in environments with substantial levels of uncertainty is also a highly desirable attribute for securing the safe operation of all agents.
Uncertainty has been primarily modeled as stochastic noise in a significant portion of the multi-agent control literature. Some indicative approaches involve multi-agent LQG control [1, 24, 39], distributed covariance steering [33, 34], multi-agent path integral control [14, 42] and learning-based stochastic optimal control methods [21, 26], among others. Nevertheless, such approaches would typically fall short in scenarios where the underlying disturbances might be exogenous interventions, environmental factors, modelling inaccuracies, etc., that cannot be properly represented as stochastic signals [5]. Such problems create the necessity of finding robust optimal policies that would guarantee feasibility for all possible values of such disturbances within some bounds.
Robust optimization is a subfield of optimization that addresses the problem of discovering an optimal solution that remains robust feasible for all possible realizations of the problem parameters within a given uncertainty set [5, 6]. Although robust optimization has found significant successful applications in many areas such as power systems [8], supply chain networks [29], communication systems [10], etc., there have only been limited attempts for addressing problems relevant to robotics [19, 23, 30, 40]. The field of robust control is a also closely related area with long history in addressing disturbances from predefined sets [12, 45]. However, as this area is mainly focused on synthesis techniques for stabilization and performance maintenance, merging such methods with distributed optimization and steering state distributions for multi-agent systems is not as straightforward. On the contrary, as robust optimization allows for more flexibility in designing such approaches, the current work is focused on leveraging the latter for safe multi-agent control.
Another promising area that has recently emerged in the context of safe stochastic control is the one of steering the state distributions of systems under predefined target distributions - mostly known as covariance steering [3, 11, 22]. Such methods have found several successful applications in path planning [25], trajectory optimization [4, 44], robotic manipulation [20], flight control [7, 31], multi-robot control [33, 34] and others. More recently, a convex optimization framework for steering state distributions and effectively addressing both stochastic noise and bounded deterministic disturbances was presented [19]. In a similar direction, a data-driven robust optimization-based method was recently proposed in [28], including estimating the noise realization from collected data. Nevertheless, as such approaches result in high-dimensional semidefinite programming (SDP) problems, the computational demands for addressing multi-agent control tasks have remained prohibitive.
The scope of this work is to provide a scalable robust optimization framework for steering distributions of multi-agent systems to specific targets under stochastic and deterministic uncertainty. Towards this direction, we present a distributed algorithm for safe and robust multi-robot control whose effectiveness and scalability are verified through simulation experiments. The specific contributions of this paper can be listed as follows.
-
•
We combine elements from the areas of robust optimization, distribution steering and distributed computation architectures for deriving a scalable and robust approach for multi-agent control. To our best knowledge, this is the first work to fuse these areas into a single methodology.
-
•
We circumvent computational expensiveness by providing tractable approximations of robust optimization constraints that are relevant for multi-robot control problems.
-
•
We present a distributed algorithm based on an Alternating Direction Method of Multipliers (ADMM) approach for multi-agent robust distribution steering.
-
•
We provide simulation results that demonstrate the effectiveness of the proposed algorithm in handling stochastic and deterministic uncertainty. The scalability of the method is also highlighted by showing multi-robot tasks with up to 100 agents.
The remaining of this paper is organized as follows. In Section II, we state the multi-agent problem to be considered in this work. The proposed distributed robust optimization framework is presented in Section III. An analysis of performance and scalability is demonstrated through simulation experiments in Section IV. Finally, the conclusion of this paper, as well as future research directions, are provided in Section V.
II Problem Statement
In this section, we present the multi-agent robust optimization problem to be addressed in this work. Initially, we familiarize the reader with the notation used throughout this paper. Subsequently, we present the basic multi-agent problem setup and the dynamics of the agents. The two main characterizations of underlying uncertainty, i.e., deterministic and stochastic, are then formulated. Then, we specify the control policies of the agents and the robust constraints that are considered in our framework. Finally, the exact formulation of the multi-agent problem is provided.
II-A Notation
The integer set is denoted with . The space of matrices that are symmetric positive (semi)-definite, i.e., () is denoted with (). The inner product of two vectors is denoted with . The -norm of a vector is defined as . The weighted norm is also defined for any . In addition, the Frobenius norm of a matrix is given by . With , we denote the vertical concatenation of a series of vectors . Furthermore, the probability of an event is denoted with . Given a random variable (r.v.) , its expectation and covariance are denoted with and , respectively. The cardinality of a set is denoted as . Finally, given a set , the indicator function is defined as , if , and , otherwise.
II-B Problem Setup
Let us consider a multi-agent system of agents defined by the set . All agents may be subject to diverse dynamics, inter-agent interactions, and exogenous deterministic and stochastic uncertainty. Each agent has a set of neighbors defined by . Furthermore, for each agent , we define the set of agents that consider as a neighbor through .
II-C Dynamics
We consider the following discrete-time linear time-varying dynamics for each agent ,
(1) |
where is the state, is the control input of agent at time , and is the time horizon. The terms represent exogenous deterministic disturbances, while the terms refer to stochastic disturbances, with their exact mathematical representations presented in Section II-D. Each initial state , is of the following form
(2) |
where is the known part, while and refer to the unknown exogenous deterministic and stochastic parts, respectively. Finally, are the dynamics matrices of appropriate dimensions.
For convenience, let us also define the sequences , , and . The dynamics (1) can then be rewritten in a more compact form as
(3) |
where the matrices are provided in Section I of the Supplementary Material (SM).
II-D Characterizations of Uncertainty
As previously mentioned, we consider the existence of both exogenous deterministic and stochastic uncertainties. These two types of uncertainty are formally described as follows.
Exogenous deterministic uncertainty
The unknown deterministic disturbances typically stem from external factors that cannot be accurately represented through stochastic signals. This creates a requirement to formulate optimization problems and achieve optimal policies that are feasible in a robust sense for all possible values of such disturbances in a bounded uncertainty set. Examples of such uncertainty sets include ellipsoids, polytopes, and others [5]. In this work, we assume the sequence to be lying inside an ellipsoidal uncertainty set defined as follows
(4) | ||||
where , , and is the uncertainty level. Ellipsoidal sets are used to model uncertainty in various robust control applications [27]. Further, ellipsoid sets have simple geometric structure, and more complex sets can be approximated using ellipsoids.
Stochastic uncertainty
The unknown stochastic component is assumed to be a Gaussian random vector with mean , and covariance .
II-E Control Policies
Towards addressing both types of uncertainty, we consider affine purified state feedback policies [5, 19] for all agents. Let us first introduce the disturbance-free states , whose dynamics are given by
(5) | ||||
(6) |
Subsequently, we can define the purified states , whose dynamics are given by
(7) | ||||
or more compactly by
(8) |
where . We consider the following affine purified state feedback control policy for each agent ,
(9) |
where are the feed-forward control inputs and are feedback gains on the purified states. The above policies can be written in compact form as
(10) |
where
II-F Robust Constraints
Next, we introduce the types of constraints that are considered in this work. We highlight that the following constraints must be satisfied robustly for all possible values of lying inside the uncertainty sets .
Robust linear mean constraints
The first class of constraints we consider are robust linear constraints on the expectation of the state sequence of agent ,
(11) |
where , . In a robotics setting, such constraints could typically represent bounds on the mean position, velocity, etc.
Robust nonconvex norm-of-mean constraints
Furthermore, we consider nonconvex inequality constraints of the form
(12) |
where , , . Such constraints are typically useful for capturing avoiding obstacles, forbidden regions, etc. In addition, we extend to the inter-agent version of these constraints which is given by
(13) |
where , , . This class of constraints typically represents collision avoidance constraints between different robots.
Robust linear chance constraints
Subsequently, we consider robust chance constraints on the states of the agents,
(14) |
where is the probability threshold. It should be noted that these constraints are also affected by the stochastic noise since they probabilistic constraints on the entire state and not only on the state mean.
Robust covariance constraints
Additionally, we consider the following constraints on the covariance matrices of the state sequences,
(15) |
where is a prespecified allowed upper bound on the state covariance.
II-G Problem Formulation
Finally, we present the mathematical problem formulation for the multi-agent robust optimization problem under both stochastic and bounded deterministic uncertainties.
Problem 1 (Multi-Agent Robust Optimization Problem).
For all agents , find the robust optimal policies , such that
(16) | ||||
where each cost penalizes the control effort of agent .
III Distributed Robust Optimization under Deterministic and Stochastic Uncertainty
In this section, we propose a distributed robust optimization framework for solving the multi-agent problem presented in Section II. Initially, we provide computationally tractable approximations for the robust constraints in Section III-A. Subsequently, Section III-B presents the new transformed problem to be solved based on the constraint approximations. Finally, in Section III-C, we present a distributed algorithm for solving the transformed problem in a scalable manner.
III-A Reformulation of Semi-infinite Constraints
We start by analyzing the effect of the disturbances on the system dynamics. In particular, since the robust constraints (11) - (13) depend on , let us first examine the expression for given by
(17) |
since (derivation can be referenced from Section II of the SM). It is important to emphasize that the mean state relies only on the deterministic - and not the stochastic - disturbances. For convenience, we also define
(18) |
as well as the matrix such that , , and . Subsequently, we proceed with providing tractable approximations for the robust constraints presented in Section II.
III-A1 Robust linear mean constraints
Let us consider a single constraint from the set of linear constraints (11), given by
(19) |
Note that the above constraint can be reformulated as
(20) |
The following proposition provides the exact optimal value of the maximization problem in the LHS of (20).
Proposition 1.
The maximum value of when the uncertainty vector lies in the set is given by
(21) |
Proof.
Provided in Section II of the SM. ∎
Using Proposition 1, an equivalent tractable version of the constraint (20) can be given as follows
(22) |
Similar to the maximum value case, the minimum value can be provided as
(23) |
Based on the above observations, we introduce the following additional form of robust linear mean constraints
(24) |
with and , which would imply the following constraints
(25) |
The constraint form (24) allows us to decouple the mean constraint in terms of and , such that the expectation of the disturbance-free state can be controlled using the feed-forward control , and the expectation of the state mean component dependent on the uncertainty can be controlled using the feedback control gain .
III-A2 Robust nonconvex norm-of-mean constraints
Subsequently, we consider the robust nonlinear constraint (12). We present two approaches for deriving the equivalent/approximate tractable constraints to the above constraint.
Initial Approach
The robust constraint in (12) can be rewritten as
(26) |
Although the problem in the LHS is convex, the equivalent tractable constraints to the above, obtained by applying S-lemma [19], might not be convex due to the non-convexity of the overall constraint (details can be referenced from Section III of the SM). Thus, we first linearize the constraint (12) with respect to control parameters around the nominal values . Through this approach, we obtain the following constraint
(27) | ||||
where and are provided in Section III of the SM. The equivalent tractable constraints to the above constraint can be given as [19] -
(28) | ||||
The derived equivalent tractable constraints can be used in cases where the number of robust non-convex constraints of the defined form is relatively small. However, in applications that involve a significant number of such constraints (e.g., collision avoidance constraints), this formulation might prove to be significantly expensive computationally. For instance, consider an obstacle avoidance scenario, where each agent needs to enforce the constraints of the above form for each time step. Then, each agent in an environment with obstacles would have number of constraints of the form (28). Each constraint of the form (28) involves a matrix of size (when , ). Thus, the computational cost of the above formulation increases exponentially with the number of obstacles, time horizon, and dimension of the disturbances. In order to overcome this issue, we propose a second approach for deriving equivalent/approximate tractable versions of the constraint (12).
Reducing computational complexity
Let us first write the constraint (12) in terms of as follows -
(29) |
In this approach, we consider the following tighter approximation of the above constraint
(30) |
which can be rewritten as
(31) |
Let us now introduce a slack variable such that we split the above constraint into the following system of two constraints.
(32) | |||
(33) |
Note however that the constraint (32) is non-convex. We address this by linearizing the constraint around the nominal trajectory . In addition, the constraint (33) is still a semi-infinite constraint; thus, we need to derive its equivalent tractable constraints. Since the LHS in the constraint (33) involves a non-convex optimization problem, deriving the equivalent tractable constraints is NP-hard. Hence, we derive tighter approximations of the constraint (33) in the following proposition.
Proposition 2.
The tighter approximate constraints to the semi-infinite constraint (33) are given by
(34) | |||
(35) |
where is row of .
Proof.
Provided in Section III of the SM. ∎
In this formulation, only the constraints (35) include . This formulation is useful in the case where there are multiple non-convex constraints of the form (12) at , which have the same matrix. In such a case, we only need to enforce the constraint (35) once per each unique matrix in the constraints. For instance, let us again consider the collision avoidance example. Since the collision constraints (both obstacle and inter-agent) are applied to the agent’s position, the matrix remains the same in all such constraints, irrespective of the number of obstacles. Therefore, it is only the amount of constraints of the form (34) that would increase with the number of obstacles.
Subsequently, we proceed with considering the inter-agent robust non-convex constraints (13), which can be rewritten as
(36) |
where , . The uncertainty involved for the variable would be . Further, it should be noted that if belong to the defined ellipsoidal sets , respectively, then the uncertainty vector can also be modeled by an ellipsoidal set of the defined form (II-D). Thus, the constraint (36) has a similar form as (12), and applying the initial approach would result in semi-definite constraints that are of the size . This might be particularly disadvantageous in a large-scale multi-agent setup, which might involve a significant number of neighbor agents. Thus, we derive equivalent/approximate tractable constraints using the second approach.
For that, let us first rewrite a single constraint corresponding to a neighbor agent of the agent () from the set of constraints (13) as follows.
(37) |
We then consider the following tighter approximation to the above constraint
(38) |
and similarly with the previous case, we introduce a slack variable , through which we can rewrite the above constraint as follows
(39) | ||||
(40) |
The non-convex constraint (39) is addressed by linearizing it around the nominal value . Similar to the previous case, the constraint (40) is semi-infinite, and deriving its equivalent tractable constraints is NP-hard. Thus, we derive the tighter approximations of the constraint (40).
Proposition 3.
The tighter approximate constraints to the semi-infinite constraint (40) are given by
(41) | |||
(42) | |||
(43) |
where are the rows of the matrices respectively.
III-A3 Robust linear chance constraints
III-A4 Robust covariance constraints
The covariance of the state of an agent is given as follows (derivation can be referenced from Section II of SM)
(46) |
It should be observed that the covariance does not depend on the deterministic uncertainty but only on the covariance of the stochastic noise. Now, let us rewrite the constraint (15) in terms of as follows
(47) |
The above constraint is non-convex and can be reformulated into the following equivalent convex constraint by using the Schur complement,
(48) |
where .
III-B Problem Transformation
Now that we have derived the computationally tractable constraints that could be applied to a multi-robot setting, let us define the multi-agent optimization problem using the reformulated constraints.
Problem 2 (Multi-Agent Robust Optimization Problem II).
For all agents , find the robust optimal policies , such that
(49) | ||||
We are interested in solving the above problem in a multi-robot setting. Hence, we consider the robust non-convex constraints to represent obstacle and inter-agent collision avoidance constraints and are applied at each time step . As mentioned in the previous subsection, all the collision avoidance constraints involve the same matrix in the constraint. Thus, in the subsequent sections, we consider in the constraints (32), (34), (35), (39), (41), (42) to be . Further, in this case, the constraints (35) and (42) are the same.
III-C Distributed Algorithm
Now that we have formulated the robust multi-agent trajectory optimization problem 2, we provide a distributed architecture for solving it.
III-C1 Problem Setup
In this subsection, we provide an overview of steps taken to reformulate the problem 2 into a structure that would allow us to solve it in a distributed manner. We start by considering the inter-agent constraints (39) and (41). To enforce these inter-agent constraints, each agent needs of each of its neighbor agents . For convenience, let us assume and define the sequences , . We can rewrite problem 2 in a simplified way as
(50a) | ||||
(50b) |
where | |||
To solve the problem in a distributed manner, we first introduce the copy variables at each agent corresponding to the variables of each of its neighbor agents . Each agent then uses these copy variables to enforce the inter-agent constraints. However, these copy variables should capture the actual value of their corresponding variables. Thus, we need to enforce a consensus between the actual values of the variables for each agent and their corresponding copy variables , respectively at each agent to which agent is a neighbor of (). Hence, the above problem can be rewritten in terms of the copy variables as follows -
(51a) | |||
(51b) | |||
(51c) | |||
(51d) |
Next, in order to use consensus ADMM (CADMM) [9], we introduce the global variables such that the constraints (51c), (51d) can be rewritten as follows.
(52) | ||||
(53) |
Then, we define the local variables of each agent as , . For convenience, we also define the global variable sequences , . Using the above-defined variables, problem (51) can be rewritten as
(54) | ||||
III-C2 Distributed Approach
Now that we have reformulated the problem to the form of (54), we can use CADMM to solve it in a distributed manner. The variables constitute the first block, and constitute the second block of the CADMM. In each iteration of CADMM, the Augmented Lagragian (AL) of the problem is minimized with respect to each block in a sequential manner followed by a dual update. The AL of the problem (54) is given by
(55) | ||||
where , are dual variables, and , are penalty parameters.


Now, the update steps in an iteration of CADMM applied to the problem (54) are as follows.
Local variables update (Block - I)
The update step for the local variables is given by
Note that there is no coupling between the local variables of different agents. Thus, the above update step can be performed in a decentralized manner where each agent solves its sub-problem to update the local variables .
Global variables update (Block - II)
Next, the global variables are updated through
Let us define , as sequence of dual vectors given as , , such that the above update can be rewritten in a closed form for each agent as follows (details are provided in Section III of the SM)
(56) | ||||
where . After the local variables update, each agent would receive the updated variable from all the agents to which it is a neighbor of (). Using these received local variables, each agent updates the global variables as per the update step (56).
Dual variables update
Finally, the dual variables are updated through the following rules
(57) | ||||
After the global variables update, each agent would receive the updated global variables from all its neighbor agents . Using these received variables, the agent updates the dual variables as per the update step (57). It should be noted that in each update step, all the agents do the update in parallel. Further, the agents need not exchange other system information (such as dynamics parameters or uncertainty set parameters).
IV Simulation Results
In this section, we provide simulation results of the proposed method applied to a multi-agent system with underlying stochastic and deterministic uncertainties. In Section IV-A, the effectiveness of the proposed robust constraints is analyzed under different scenarios. In Section IV-B, the scalability of the proposed distributed framework is illustrated. The agents have 2D double integrator dynamics with the exact dynamic matrices and the uncertainty information provided in Section IV of the SM. In all experiments, we set the time horizon to . All experiments are also illustrated through the supplementary video111https://youtu.be/yrDwPwg4WFI.
IV-A Effectiveness of Proposed Robust Constraints
We start by validating the effectiveness of the proposed robust constraints by considering different cases of underlying uncertainties. First, we analyze the constraints that are only on the expectation of the state. Subsequently, we analyze the robust chance constraints and covariance constraints, which are dependent on both the deterministic and stochastic disturbances.
IV-A1 With deterministic disturbances






Here, we compare the results of applying the proposed distributed framework for two cases: robust and non-robust. With non-robust, we refer to the case of finding the optimal control trajectory by only considering standard optimization constraints instead of explicitly accounting for the deterministic uncertainties through robust optimization constraints (robust case). Fig. 1 discloses a two-agent scenario with terminal mean and obstacle avoidance constraints. The plots demonstrate trajectories of the two agents for different realizations of the disturbances. As shown in Fig. 1a, in the robust case, each agent successfully avoids the obstacle, and its disturbance-free state mean reaches the target terminal mean with uncertainty in its terminal state mean lying inside the set target bounds (represented by black boxes). On the contrary, for the non-robust case in Fig. 1b, the agents might collide with the obstacle in the center, and the uncertainty in the terminal state mean of each agent might lie outside the target bounds. In Fig. 2, we present a scenario of 20 agents in a circle formation, where each agent must reach the opposite side while avoiding collisions with other agents and an obstacle in the center. Fig. 2a and 2b show realizations for the robust and non-robust case. Clearly, the robust approach is able to successfully steer the agents to their targets while avoiding collisions, in contrast with the non-robust case where the agents might collide with the obstacle. In Fig. 2c - 2f, we show snapshots of the robust trajectories of the agents for different time instants, demonstrating that the agents do not violate any of the constraints during the tasks. Further, it is worth noting that in the robust case, the agents not only avoid collision but also the uncertainty in the state mean of an agent is reduced when it approaches an obstacle or another agent.




IV-A2 With both deterministic and stochastic disturbances



In this subsection, we demonstrate the effectiveness of the proposed framework with robust chance constraints and covariance constraints. For that, we compare two cases, one where the constraints that are only on the expectation of the state are active (‘deterministic case’), and another where all the robust constraints, including the robust chance and covariance constraints, are active (‘mixed case’). Further, it should be noted that when both disturbances are present, the mean trajectory differs from the actual trajectory.
We present two scenarios of a two-agent case with different levels of deterministic uncertainty controlled using but with the same stochastic uncertainty characterized by initial state covariance to be , and covariance of each stochastic noise component to be for each agent . Fig. 3 discloses the first scenario with deterministic uncertainty parameterized by for both agents. It can be observed from Fig. 3a that the mean trajectories obtained by the proposed method in the deterministic case satisfy the inter-agent and the terminal mean constraints. However, Fig. 3b shows that in the presence of stochastic noise, a collision between the agents is highly probable, and the agents may not be able to reach the targets within the set bounds. Fig. 3c presents the case where we enforce robust chance constraints on the agents’ positions to stay within the blue dashed lines. As shown, most of the realizations of the trajectories lie within the specified bounds of the chance constraints, which ensures the safe operation of the agents despite the stochastic uncertainty. However, the uncertainty in the terminal state of the agents is not controlled by the applied robust chance constraints. We address this using covariance constraints. Fig 3d presents the scenario with additional terminal covariance constraints with the target covariance set to be for each agent . It can be observed that the uncertainty in the terminal state is reduced.
Fig. 4 presents the second scenario of the two-agent case that involves a higher deterministic uncertainty parameterized by for both agents. Fig. 4b shows that the uncertainty in the terminal state of the agents might lie outside the target uncertainty bounds (indicated by black boxes). To address this, we can use covariance steering as shown in the previous scenario, or robust chance constraints. We use robust chance constraints in this scenario. Fig. 4c shows results for the mixed case when robust chance constraints are applied on the terminal state to stay within the target uncertainty bounds represented by the blue dashed box.
Further, it is interesting to note the following two observations. From Fig. 3a and 3d, it should be observed that uncertainty in the state trajectory is reduced to be smaller than the uncertainty in the mean in the deterministic case by adding covariance constraints. From Fig. 4a and 4b, it should be observed that the actual trajectories of the agents in the deterministic case avoid the obstacle without enforcing the robust chance constraints. This is because the feedback control gain is used to control the deviations in the state due to the deterministic and stochastic uncertainties. Thus, controlling the deviation in the state due to one type of uncertainty also controls the deviation due to another. This might be useful in cases where one of the uncertainties (deterministic or stochastic) is significantly higher than the other. In such a case, we need not explicitly apply constraints to control the effect of the least pronounced uncertainty.
IV-B Scalability of Proposed Distributed Framework
In our experiments, we noticed that the centralized method with robust terminal mean and initial approach obstacle avoidance constraints alone would fail for a time horizon longer than 15 and a number of agents greater than 3. Fig. 5 shows the result of the proposed distributed framework applied to a large-scale scenario with 100 agents, which verifies the scalability of the proposed approach. The constraints considered are robust terminal mean and inter-agent constraints. Further, Fig. 6 reports the computational times of the proposed framework for an increasing number of agents for scenarios with robust terminal mean and inter-agent constraints. All simulations were carried out in Matlab2022b [17] environment using CVX [15] as the modeling software and MOSEK as the solver [2] on a system with an Intel Corei9-13900HX.
V Conclusion
In this paper, we merge principles from the areas of robust optimization, distribution steering and distributed optimization towards obtaining a scalable safe multi-agent control method under both stochastic and deterministic uncertainties. In particular, we present a distributed robust optimization method that exploits computationally tractable approximations of robust constraints and the separable nature of ADMM in order to significantly reduce the computational demands that such approaches would typically have when applied to multi-agent systems. The results verify that the proposed framework is able to effectively address both types of uncertainty and scale for multi-robot systems with up to a hundred agents.
In future work, we wish to extend the applicability of the proposed method and explore additional robust optimization formulations that would be relevant for robotics problems. In particular, since the current method considers linear time-varying dynamics, it would be possible to extend for nonlinear systems through sequential linearization schemes [32, 34]. Additionally, we intend to further explore suitable robust optimization formulations for multi-agent robotics and investigate connections with shooting methods [18, 41]. Finally, incorporating the current framework into hierarchical distributed optimization schemes [36] would be a promising direction for further increasing the scalability of the method.


Acknowledgments
This work was supported by the ARO Award W911NF2010151. Augustinos Saravanos acknowledges financial support by the A. Onassis Foundation Scholarship.
References
- Alvergue et al. [2016] Luis D Alvergue, Abhishek Pandey, Guoxiang Gu, and Xiang Chen. Consensus control for heterogeneous multiagent systems. SIAM Journal on Control and Optimization, 54(3):1719–1738, 2016.
- ApS [2019] MOSEK ApS. The MOSEK optimization toolbox for MATLAB manual. Version 9.0., 2019. URL http://docs.mosek.com/9.0/toolbox/index.html.
- Bakolas [2018] Efstathios Bakolas. Finite-horizon covariance control for discrete-time stochastic linear systems subject to input constraints. Automatica, 91:61–68, 2018.
- Balci et al. [2022] Isin M Balci, Efstathios Bakolas, Bogdan Vlahov, and Evangelos A Theodorou. Constrained covariance steering based tube-mppi. In 2022 American Control Conference (ACC), pages 4197–4202. IEEE, 2022.
- Ben-Tal et al. [2006a] Aharon Ben-Tal, Stephen Boyd, and Arkadi Nemirovski. Extending scope of robust optimization: Comprehensive robust counterparts of uncertain problems. Math. Program., 107:63–89, 06 2006a. doi: 10.1007/s10107-005-0679-z.
- Ben-Tal et al. [2006b] Aharon Ben-Tal, Stephen Boyd, and Arkadi Nemirovski. Extending scope of robust optimization: Comprehensive robust counterparts of uncertain problems. Mathematical Programming, 107(1-2):63–89, 2006b.
- Benedikter et al. [2022] Boris Benedikter, Alessandro Zavoli, Zhenbo Wang, Simone Pizzurro, and Enrico Cavallini. Convex approach to covariance control with application to stochastic low-thrust trajectory optimization. Journal of Guidance, Control, and Dynamics, 45(11):2061–2075, 2022.
- Bertsimas et al. [2013] Dimitris Bertsimas, Eugene Litvinov, Xu Andy Sun, Jinye Zhao, and Tongxin Zheng. Adaptive robust optimization for the security constrained unit commitment problem. IEEE Transactions on Power Systems, 28(1):52–63, 2013. doi: 10.1109/TPWRS.2012.2205021.
- Boyd et al. [2011] Stephen Boyd, Neal Parikh, Eric Chu, Borja Peleato, and Jonathan Eckstein. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends® in Machine Learning, 3(1):1–122, 2011. ISSN 1935-8237. doi: 10.1561/2200000016.
- Cai et al. [2020] Yuanxin Cai, Zhiqiang Wei, Ruide Li, Derrick Wing Kwan Ng, and Jinhong Yuan. Joint trajectory and resource allocation design for energy-efficient secure uav communication systems. IEEE Transactions on Communications, 68(7):4536–4553, 2020. doi: 10.1109/TCOMM.2020.2982152.
- Chen et al. [2015] Yongxin Chen, Tryphon T Georgiou, and Michele Pavon. Optimal steering of a linear stochastic system to a final probability distribution, part i. IEEE Trans. Automat. Contr., 61(5):1158–1169, 2015.
- Dorato [1987] Peter Dorato. A historical review of robust control. IEEE Control Systems Magazine, 7(2):44–47, 1987.
- Feng et al. [2020] Zhi Feng, Guoqiang Hu, Yajuan Sun, and Jeffrey Soon. An overview of collaborative robotic manipulation in multi-robot systems. Annual Reviews in Control, 49:113–127, 2020.
- Gómez et al. [2016] Vicenç Gómez, Sep Thijssen, Andrew Symington, Stephen Hailes, and Hilbert Kappen. Real-time stochastic optimal control for multi-agent quadrotor systems. In Proceedings of the International Conference on Automated Planning and Scheduling, volume 26, pages 468–476, 2016.
- Grant and Boyd [2014] Michael Grant and Stephen Boyd. CVX: Matlab software for disciplined convex programming, version 2.1. http://cvxr.com/cvx, March 2014.
- Hu et al. [2020] Junyan Hu, Parijat Bhowmick, Farshad Arvin, Alexander Lanzon, and Barry Lennox. Cooperative control of heterogeneous connected vehicle platoons: An adaptive leader-following approach. IEEE Robotics and Automation Letters, 5(2):977–984, 2020. doi: 10.1109/LRA.2020.2966412.
- Inc. [2022] The MathWorks Inc. Matlab version: 9.13.0 (r2022b), 2022. URL https://www.mathworks.com.
- Jacobson and Mayne [1970] David H Jacobson and David Q Mayne. Differential dynamic programming. Number 24. Elsevier Publishing Company, 1970.
- Kotsalis et al. [2021] Georgios Kotsalis, Guanghui Lan, and Arkadi S. Nemirovski. Convex optimization for finite-horizon robust covariance control of linear stochastic systems. SIAM Journal on Control and Optimization, 59(1):296–319, 2021. doi: 10.1137/20M135090X. URL https://doi.org/10.1137/20M135090X.
- Lee et al. [2022] Jaemin Lee, Efstathios Bakolas, and Luis Sentis. Hierarchical task-space optimal covariance control with chance constraints. IEEE Control Systems Letters, 6:2359–2364, 2022.
- Lin et al. [2021] Yiheng Lin, Guannan Qu, Longbo Huang, and Adam Wierman. Multi-agent reinforcement learning in stochastic networked systems. Advances in neural information processing systems, 34:7825–7837, 2021.
- Liu et al. [2022] Fengjiao Liu, George Rapakoulias, and Panagiotis Tsiotras. Optimal covariance steering for discrete-time linear stochastic systems. arXiv preprint arXiv:2211.00618, 2022.
- Lorenzen et al. [2019] Matthias Lorenzen, Mark Cannon, and Frank Allgöwer. Robust mpc with recursive model update. Automatica, 103:461–471, 2019.
- Nourian et al. [2012] Mojtaba Nourian, Peter E. Caines, Roland P. Malhame, and Minyi Huang. Mean field LQG control in leader-follower stochastic multi-agent systems: Likelihood ratio based adaptation. IEEE Transactions on Automatic Control, 57(11):2801–2816, 2012. doi: 10.1109/TAC.2012.2195797.
- Okamoto and Tsiotras [2019] Kazuhide Okamoto and Panagiotis Tsiotras. Optimal stochastic vehicle path planning using covariance steering. IEEE Robotics and Automation Letters, 4(3):2276–2281, 2019. doi: 10.1109/LRA.2019.2901546.
- Pereira et al. [2022] Marcus A Pereira, Augustinos D Saravanos, Oswin So, and Evangelos A. Theodorou. Decentralized Safe Multi-agent Stochastic Optimal Control using Deep FBSDEs and ADMM. In Proceedings of Robotics: Science and Systems, New York City, NY, USA, June 2022. doi: 10.15607/RSS.2022.XVIII.055.
- Petersen et al. [2000] Ian R. Petersen, Valery A. Ugrinovskii, and Andrey V. Savkin. Introduction, pages 1–18. Springer London, London, 2000. ISBN 978-1-4471-0447-6. doi: 10.1007/978-1-4471-0447-6˙1. URL https://doi.org/10.1007/978-1-4471-0447-6_1.
- Pilipovsky and Tsiotras [2023] Joshua Pilipovsky and Panagiotis Tsiotras. Data-driven robust covariance control for uncertain linear systems. arXiv preprint arXiv:2312.05833, 2023.
- Pishvaee et al. [2011] Mir Saman Pishvaee, Masoud Rabbani, and Seyed Ali Torabi. A robust optimization approach to closed-loop supply chain network design under uncertainty. Applied mathematical modelling, 35(2):637–649, 2011.
- Quintero-Pena et al. [2021] Carlos Quintero-Pena, Anastasios Kyrillidis, and Lydia E Kavraki. Robust optimization-based motion planning for high-dof robots under sensing uncertainty. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 9724–9730. IEEE, 2021.
- Rapakoulias and Tsiotras [2023] George Rapakoulias and Panagiotis Tsiotras. Discrete-time optimal covariance steering via semidefinite programming. arXiv preprint arXiv:2302.14296, 2023.
- Ridderhof et al. [2019] Jack Ridderhof, Kazuhide Okamoto, and Panagiotis Tsiotras. Nonlinear uncertainty control with iterative covariance steering. In 2019 IEEE 58th Conference on Decision and Control (CDC), pages 3484–3490, 2019. doi: 10.1109/CDC40024.2019.9029993.
- Saravanos et al. [2021] Augustinos D Saravanos, Alexandros Tsolovikos, Efstathios Bakolas, and Evangelos Theodorou. Distributed Covariance Steering with Consensus ADMM for Stochastic Multi-Agent Systems. In Proceedings of Robotics: Science and Systems, Virtual, July 2021. doi: 10.15607/RSS.2021.XVII.075.
- Saravanos et al. [2022] Augustinos D Saravanos, Isin M Balci, Efstathios Bakolas, and Evangelos A Theodorou. Distributed model predictive covariance steering. arXiv preprint arXiv:2212.00398, 2022.
- Saravanos et al. [2023a] Augustinos D. Saravanos, Yuichiro Aoyama, Hongchang Zhu, and Evangelos A. Theodorou. Distributed differential dynamic programming architectures for large-scale multiagent control. IEEE Transactions on Robotics, 39(6):4387–4407, 2023a. doi: 10.1109/TRO.2023.3319894.
- Saravanos et al. [2023b] Augustinos D Saravanos, Yihui Li, and Evangelos Theodorou. Distributed Hierarchical Distribution Control for Very-Large-Scale Clustered Multi-Agent Systems. In Proceedings of Robotics: Science and Systems, Daegu, Republic of Korea, July 2023b. doi: 10.15607/RSS.2023.XIX.110.
- Schranz et al. [2020] Melanie Schranz, Martina Umlauft, Micha Sende, and Wilfried Elmenreich. Swarm robotic behaviors and current applications. Frontiers in Robotics and AI, page 36, 2020.
- Shah et al. [2022] Kunal Shah, Annie Schmidt, Grant Ballard, and Mac Schwager. Large scale aerial multi-robot coverage path planning. Field Robotics, 2(1), 2022.
- Shamma [2008] Jeff Shamma. Cooperative control of distributed multi-agent systems. John Wiley & Sons, 2008.
- Taylor et al. [2021] Andrew J Taylor, Victor D Dorobantu, Sarah Dean, Benjamin Recht, Yisong Yue, and Aaron D Ames. Towards robust data-driven control synthesis for nonlinear systems with actuation uncertainty. In 2021 60th IEEE Conference on Decision and Control (CDC), pages 6469–6476. IEEE, 2021.
- Von Stryk and Bulirsch [1992] Oskar Von Stryk and Roland Bulirsch. Direct and indirect methods for trajectory optimization. Annals of operations research, 37(1):357–373, 1992.
- Wan et al. [2021] Neng Wan, Aditya Gahlawat, Naira Hovakimyan, Evangelos A. Theodorou, and Petros G. Voulgaris. Cooperative path integral control for stochastic multi-agent systems. In 2021 American Control Conference (ACC), pages 1262–1267, 2021. doi: 10.23919/ACC50511.2021.9482942.
- Xiao et al. [2021] Wei Xiao, Christos G Cassandras, and Calin A Belta. Bridging the gap between optimal trajectory planning and safety-critical control with applications to autonomous vehicles. Automatica, 129:109592, 2021.
- Yin et al. [2022] Ji Yin, Zhiyuan Zhang, Evangelos Theodorou, and Panagiotis Tsiotras. Trajectory distribution control for model predictive path integral control using covariance steering. In 2022 International Conference on Robotics and Automation (ICRA), pages 1478–1484. IEEE, 2022.
- Zhou and Doyle [1998] Kemin Zhou and John Comstock Doyle. Essentials of robust control, volume 104. Prentice hall Upper Saddle River, NJ, 1998.
- Zhu et al. [2021] Pingping Zhu, Chang Liu, and Silvia Ferrari. Adaptive online distributed optimal control of very-large-scale robotic systems. IEEE Transactions on Control of Network Systems, 8(2):678–689, 2021. doi: 10.1109/TCNS.2021.3097306.
Supplementary Material
In the following, we provide detailed derivations, proofs and experiment data that have been omitted in the main paper.
VI System Dynamics Compact Form
The system dynamics (1) can be written in a more compact form as follows
(58) |
where
(59a) | |||
(59b) | |||
(59c) | |||
(59d) |
with
(60) |
VII Robust Linear Constraints
VII-A Expectation of State
Let us rewrite the state equation for an agent as follows
Substituting the control as (from (10)) in the above state equation, we get the following.
(61) |
Further, using (8), we can rewrite the above equation as follows
(62) | ||||
which can be written in compact form as follows -
(63) | ||||
Now, the expectation of the state can be given as
(64) | ||||
Finally, using the fact that , we get
(65) |
VII-B Covariance of State
VII-C Proposition 1 Proof
The expression can be rewritten as
(71a) | ||||
(71b) |
Let us now consider the Lagrangian for the minimization problem inside (71b), given by
(72) |
where is the Lagrange multiplier corresponding to the inequality constraint . The above problem is convex, and the constraint set is non-empty - since lies in the interior of the feasible set because . Thus, the problem satisfies Slater’s condition, and we could find the minimizer using KKT conditions as follows
(73a) | |||
(73b) | |||
(73c) | |||
(73d) |
Let us first simplify the condition , as
(74) |
which leads to
(75) |
It can be observed that cannot be equal to the zero vector since it is the coefficient of in the objective function. Thus, since we also have that , then the RHS of the above equation is non-zero, which implies that both and need to be non-zero. Therefore, we obtain
(76) |
Further, from the condition (73b) and using , we get
(77) |
which through (76) gives
(78) |
The above relationship leads to
(79) |
since . The optimal value is then given by
(80a) | ||||
(80b) | ||||
(80c) | ||||
(80d) |
Therefore, we obtain
(81a) | ||||
(81b) |
VIII Robust Non-convex Constraints and Distributed Architecture
VIII-A Initial Approach for robust nonconvex constraint
Let us rewrite the robust nonconvex norm-of-mean constraint (26) as follows
(82) |
which can be written in terms of as
(83) | ||||
where
(84a) | ||||
(84b) | ||||
(84c) |
The constraint (83) could be reformulated through the following equivalent set of constraints (see [19]),
(85a) | |||
(85b) |
Nevertheless, the above constraint would still be a nonconvex constraint, as , , and are nonlinear in , . To address this issue, we first linearize the constraint (82) with respect to around the nominal , to obtain the following affine robust constraint
(86) | ||||
where
(87a) | |||
(87b) | |||
(87c) |
VIII-B Proposition 2 Proof
Let us initially rewrite the constraint (33) as follows
(88) |
The first step is to construct an upper bound for . Using Proposition 1, we obtain the following two equalities,
(89) | ||||
(90) |
where is row of , for all . By combining (89) and (90) for every row of , we can arrive to the following relationship,
(91a) | ||||
(91b) | ||||
(91c) |
Next, by introducing a variable such that
(92) |
for every , we obtain
(93) |
Therefore, we can consider the following tighter approximation of the constraint (88), given by
(94) |
VIII-C Proposition 3 Proof
The proof of Proposition 3 is similar to the one of Proposition 2. First, let us rewrite constraint (40) as follows
(95) |
The bound on each row of , with , can be given as follows
(96a) | |||
(96b) |
where
(97) | ||||
Using Proposition 1, we have
(98a) | |||
(98b) |
where are the rows of the matrices respectively. By combining (96) and (98), we obtain
(99) | ||||
As in the proof of Proposition 2, let us now introduce the variables such that
(100) | |||
(101) |
for every . Then, we obtain
(102) |
so we can consider the following tighter approximation to the constraint (95),
(103) |
VIII-D Distributed Architecture Global Updates
The global update steps are given by
(104) | |||
This update can be decoupled for each agent as follows
(105) | ||||
To find the minimum on the RHS, let us differentiate the RHS with respect to the global variables and equate it to zero. For , we get
(106) | ||||
which leads to
(107) | ||||
The update for can be derived similarly.
IX Simulation Data
IX-A System Parameters
Each agent is subject to the following 2D double integrator dynamics,
for . Further, we enforce that . We consider for all agents. In each experiment, we use as the uncertainty level. The cost matrices are set to be , and .
IX-B ADMM Parameters
The penalty parameters are set as , , with maximum ADMM iterations set to be 200. Further, we consider the following primal and dual residual conditions as the convergence criteria for ADMM at the iteration. The primal residual condition is given by
(108) |
while the dual residual one is
Finally, we set both residual thresholds .
IX-C Constraint Parameters
For both inter-agent constraints and obstacle avoidance constraints, we use . The collision distance threshold for inter-agent and obstacle avoidance constraints are set at 0.25, 0.5 respectively, for the experiments (1, 3, 4). For experiment (2), both inter-agent and obstacle collision distance threshold are set at 0.25. Further, in the experiments (3, 4) involving chance constraints, the probability threshold of chance constraints is set to be .