Improved Hierarchical ADMM for Nonconvex Cooperative Distributed Model Predictive Control
Abstract
Distributed optimization is often widely attempted and innovated as an attractive and preferred methodology to solve large-scale problems effectively in a localized and coordinated manner. Thus, it is noteworthy that the methodology of distributed model predictive control (DMPC) has become a promising approach to achieve effective outcomes, e.g., in decision-making tasks for multi-agent systems. However, the typical deployment of such distributed MPC frameworks would lead to the involvement of nonlinear processes with a large number of nonconvex constraints. To address this important problem, the development and innovation of a hierarchical three-block alternating direction method of multipliers (ADMM) approach is presented in this work to solve this nonconvex cooperative DMPC problem in multi-agent systems. Here firstly, an additional slack variable is introduced to transform the original large-scale nonconvex optimization problem. Then, a hierarchical ADMM approach, which contains outer loop iteration by the augmented Lagrangian method (ALM) and inner loop iteration by three-block semi-proximal ADMM, is utilized to solve the resulting transformed nonconvex optimization problem. Additionally, it is analytically shown and established that the requisite desired stationary point exists for convergence in the algorithm. Finally, an approximate optimization stage with a barrier method is then applied to further significantly improve the computational efficiency, yielding the final improved hierarchical ADMM. The effectiveness of the proposed method in terms of attained performance and computational efficiency is demonstrated on a cooperative DMPC problem of decision-making process for multiple unmanned aerial vehicles (UAVs).
Index Terms:
Alternating direction method of multipliers (ADMM), distributed optimization, model predictive control (MPC), multi-agent system, collision avoidance.I Introduction
With the rapid development of communication and computation technologies, cooperative decision making of large-scale multi-agent systems has become a promising trend in the development of automated systems [1, 2, 3]. Nevertheless, the significant challenge involved in cooperative decision making, in handling the coupling conditions among the interconnected agents with an appropriate guarantee of efficient computation still remains. In the existing literature, a significant amount of effort has been put into distributed control and coordinated networks for large-scale interconnected multi-agent systems, with the objective of optimizing the global cost while satisfying the required safety constraints collectively. Essentially for such approaches, the control and optimization algorithms deployed are distributed in structure, and the deployment utilizes only the local observations and information [4, 5]. With the appropriate communication technology in place, the agents are developed to suitably make decisions while attempting to satisfy all of the constraints or requirements invoked [6, 7]. On the other hand, a remarkable alternate development that utilizes an approach termed as distributed model predictive control (DMPC) has also been shown to display great promise to be effectively used to handle the input and state constraints with multiple objectives; and possibly also applicable to this important problem of cooperative decision making of large-scale multi-agent systems [8]. In addition, via the DMPC framework, there is great potential that the computational burden can be significantly relieved by decomposing the centralized system into several interconnected subsystems (with also a similar situation of availability of knowledge of necessary information among the connected agents).
For the DMPC problem with interconnected agents, the promising optimization approach of the alternating direction method of multipliers (ADMM) algorithm has the potential for very effective utilization. In this context, it can be noted that a large number of research works have demonstrated that the ADMM methodology has the capability to rather effectively determine the optimal solution to many challenging optimization problems, such as distributed optimization [9], decentralized control [10, 11], and statistical learning problems [12, 13]. The key idea of the ADMM approach is to utilize a decomposition coordination principle, wherein local sub-problems with smaller sizes are solved, and then the solutions of these subproblems can be coordinated to obtain the solution to the original problem. Notably, the ADMM solves the sub-problems by optimizing blocks of variables, such that efficient distributed computation can be attained. The convergence of the ADMM with two blocks has been investigated for convex optimization problems in [14, 15], and typical applications of such two-block ADMM approaches in distributed convex optimization problems are reported extensively in [16, 17]. Several researchers have also made rather noteworthy extensions from this two-block convex ADMM to multi-block convex ADMM in [18, 19]. However, although seemingly evident as a possibility with great potential, there has not been much substantial work on innovating the ADMM framework for nonconvex DMPC. Classical ADMM is known only to admit a linear convergence rate for convex optimization problems.
For nonconvex optimization problems, certain additional conditions and assumptions need to be considered to guarantee the required convergence [20, 21, 22, 23, 24, 25, 26]. More specifically, the work in [20] studies the two-block ADMM with one of the blocks defined as the identity matrix; and it also states that the objective function and the domain of the optimization variable need to be lower semi-continuous. Yet further, a nonconvex Bregman-ADMM is proposed in [21] by introducing a Bregman divergence term to the augmented Lagrangian function to promote the descent of certain potential functions. Moreover, the work in [22] proposes two proximal-type variants of ADMM with -stationarity conditions to solve the nonconvex optimization problem with affine coupling constraints under the assumption that the gradient of the objective function is Lipschitz continuous. Additionally, in [23], a multi-block ADMM is presented to minimize a nonconvex and nonsmooth objective function subject to certain coupling linear equality constraints; and the work in [24] presents an ADMM-type algorithm for nonconvex optimization problem in the application environment of modern machine learning problems, though requiring the assumption that the penalty parameter should be large enough. Besides, an augmented Lagrangian based alternating direction inexact newton (ALADIN) method is introduced in [27] to solve the optimization problems with a separable nonconvex objective function and coupled affine constraints.
Noting all these additional conditions and assumptions that typically can arise, it is the situation where the ADMM approach cannot yet be directly applied to solve a large-scale constrained nonconvex program without further relaxation or reformulation. Also, the ADMM with more than two blocks is typically implemented with sequential optimization; and while a direct parallelization of a multi-block ADMM formulation is also sometimes considered, the aforementioned approach has the disadvantage that it is typically less likely to converge [28, 29]. In order to address all these limitations, a hierarchical three-block ADMM approach is utilized in our paper to solve the nonconvex optimization problem that arises for decision making in multi-agent systems. Firstly, an additional slack variable is innovatively introduced to transform the original large-scale nonconvex optimization problem, such that the intractable nonconvex coupling constraints are suitably related to the distributed agents. Then, the approach with a hierarchical ADMM that contains the outer loop iteration by the augmented Lagrangian method (ALM), and inner loop iteration by three-block semi-proximal ADMM, is utilized to solve the resulting transformed nonconvex optimization problem. Next, the approximate optimization with a barrier method is then applied to improve the computational efficiency. Finally, a multi-agent system involving decision making for multiple unmanned aerial vehicles (UAVs) is utilized to demonstrate the effectiveness of the proposed method in terms of attained performance and computational efficiency.
The remainder of this paper is organized as follows. Section II describes a multi-agent system that contains multiple interconnected agents equipped with sensors and communication devices, and also formulates a large-scale nonconvex optimization problem for decision making. Section III gives a more compact form of the optimization problem, which is further transformed by introducing slack variables. In section IV, the hierarchical ADMM is used to solve the transformed optimization problem. Next, an improved version of such hierarchical ADMM is presented to accelerate the computation process based on the barrier method in Section V. Then, in Section VI, a multi-UAV system is used as an example to show the effectiveness and computational efficiency of the hierarchical ADMM and improved hierarchical ADMM. Finally, pertinent conclusions of this work are given in Section VII.
II Problem Formulation
The following notations are used in the remaining content. denotes the set of real matrices with rows and columns, means the set of -dimensional real column vectors. The symbol and mean that the matrix is positive definite and positive semi-definite, respectively. and mean that vector is element-wisely greater and no less than the vector . and denote the transpose of the matrix and vector . represents the -dimensional identity matrix; and denote the -dimensional all-one vector and the -by- all-one matrix, respectively; and represent the -dimensional all-zero vector and the -by- all-zero matrix, respectively. The operator denotes the Frobenius inner product, i.e., for all ; the operator is the Frobenius norm of matrix . denotes the Kronecker product and denotes the Hadamard product (Schur product). denotes a block diagonal matrix with diagonal entries ; denotes a diagonal matrix with diagonal entries . and mean the sets of positive integers and , respectively. denotes the set of positive integers . The operator means the concatenation of the vector for all , i.e., .
II-A Modelling of Interconnected Agents
In order to represent the constraint or information topology of multiple interconnected agents, an undirected graph can be utilized. The node set denotes the agents, and the edge set denotes the coupling constraints (information flow) between two interconnected agents, which is defined as
(1) |
where and are the number of agents and coupling constraints in the multi-agent system, respectively, are the position vectors of the th agent and th agent, and mean the maximum communication distance and minimum safe distance between two agents, respectively. Therefore, based on the communication topology, an adjacency matrix of the network (denoted by ) can be obtained, which is a square symmetric matrix. The elements of indicate whether pairs of vertices are adjacent/connected or not in the graph. Therefore, the neighbor nodes of the th agent are the corresponding indexes of nonzero elements of the th row in , which is denoted by .
II-B Problem Description
Each connected agent in the set has its origin and target state, which means each agent needs to achieve its task by coordinating with other connected agents in . Moreover, the information of the connected agents (neighbors) is necessary for each agent to make decisions, due to the requirement of the communication topology (1). Since the communication between connected agents occurs during the whole process, the present state information of neighbors needs to be conveyed to the th agent to make decisions. Thus, the cooperative task can be formulated as a cooperative DMPC problem for each agent . Then, given the initial state , the DMPC problem for the th agent at timestamp can be written as
(2a) | |||||
where the subscript means the corresponding variable of the th agent, and denote the state and control input vectors of the th agent, respectively. are the state matrix and control input matrix, respectively. denotes the initial time stamp and is the prediction horizon. Notice that (2a) is the objective function for the state and control variables; (II-B) denotes the kinematics of the agents; (II-B) and (II-B) denote the bound constraints of state and control input vectors, () and () are the upper (lower) bound of state and control input vectors, respectively; (II-B) denotes the coupling constraints of two interconnected agents (the th agent and the th agent with ) and denotes the set of the coupling constraints. The cost function of the th agent is a summation of the stage cost term and the terminal cost term , i.e.,
where the initial state is known. The stage cost function is defined as
where is the reference state the th agent tries to follow, and and are the diagonal weighting matrices for the state and control input vectors, respectively. The terminal cost function is
where is the diagonal weighting matrix for the terminal state vector.
In a specific multi-agent system, can be specified as
where is the locating matrix to extract the position vector from the state vector , i.e., .
Assumption 1.
in the prediction horizon is time-invariant.
III Problem Transformation
III-A Problem Reformulation
First of all, the decision variable is set as
It is straightforward that , where is the dimension of position in state vector and , and . For all and , the coupling constraints , , , can be equivalently expressed as
where is the number of the neighbours of the th agent, i.e., where is the cardinality operator for the set . This means the th agent has neighbors, and there are inequalities for each . Thus, the coupling constraint , , , denotes inequalities for all and .
Define a known vector . Notice that the position vector of all neighbors of the th agent are known due to the existence of communication. Thus, the coupling constraints can be rewritten as
with the mapping characterized by
where the matrix is a matrix to extract all position vectors from the decision variable , and . Besides, the mapping is given by
Therefore, the large-scale cooperative optimization problem (2a) can be reformulated as
(7) | |||||
where ; is the weighting matrix for the optimization variable ; is the reference vector, i.e., . Due to the fact that the agents’ dynamic models are linear time-invariant during the optimization horizon , we can rewrite (II-B) as the equality constraint in (III-A), where
and with a known state vector at current timestamp . Notice that and . Additionally, (II-B) and (II-B) can be rewritten as in (III-A), where
Remark 1.
It is well-known that increasing the prediction horizon can enlarge the domain of attraction of the MPC controller. However, this will increase the computational burden. Weighting the terminal cost can also enlarge the domain of attraction of the MPC controller without the occurrence of the terminal constraints; thus, the stabilizing weighting factor of a given initial state can be included, which has been proved in [30] that the asymptotic stability of the nonlinear MPC controller can be achieved.
III-B Optimization Problem Transformation
Due to the fact that the direct use of the conventional ADMM to the nonconvex optimization problem (III-A) directly cannot ensure the convergence, problem transformation is significant in the constrained nonconvex optimization problem [22]. Thus, a slack variable is introduced to the problem (III-A), and then the transformed optimization problem is given by
After introducing , the linear coupling constraints in (III-A) change from 2 blocks ( and ) to 3 blocks (, and ). Since the block about is an identity matrix whose image is the whole space, there always exists an such that is satisfied, given any and . In addition, the constraint can be treated separately from the equality constraint . Therefore, we can consider the constraints in the problem (III-B) by separating them into two groups of constraints. The first group considers the equality constraint and the nonconvex constraint involved by the indicator function and , and the other group deals with the constraint . As for the first sub-problem which does not include the constraint , some existing techniques of the ADMM can be employed. On the other hand, the constraint can be transformed by dualizing the constraint with and adding a quadratic penalty . Then the problem (III-B) can be rewritten as
where is the dual parameter and is the penalty parameter. Then, the ALM is used to solve the second sub-problem.
The augmented Lagrangian function of problem (III-B) is
where is the dual parameter of constraint and is the penalty parameter. Then, the first-order optimality conditions at a stationary solution are given by
(10a) | |||
where and denote the normal cone of the set and with respect to and , respectively, and the superscript means the variable in the th iteration.
Remark 2.
A solution that satisfies the optimality conditions (10a) may not necessarily satisfy the primal feasibility of the constraint . However, ALM can prompt the slack variable to zero through updating its dual parameter .
Based on Remark 2, the update of helps to prompt to zero with .
IV Hierarchical ADMM
For the problem (III-B), we treat it into two groups. The first one is to solve the problem (III-B) and update in sequence with keeping and constant, and the other group is to update the parameter and . In order to clarify the two groups, the iteration number of the first group (inner iteration) is indexed by , and the one of the second group (outer iteration) is .
1. Update :
2. Update :
In order to obtain a close-form solution to the problem (IV), we define as a given positive semidefinite linear operator. Then, it follows that
Then, the optimality condition to the sub-problem of updating is given by
Setting where is the largest eigenvalue of the matrix , we have
Then, it follows that
Lemma 1.
The projection operator with respect to the convex set can be expressed as
(11) |
where is the sub-differential operator, and is any arbitrary number.
Proof.
Define a finite dimensional Euclidean space equipped with an inner product and its induced norm such that . For any , there exists such that . Then, it follows that
The projection operator is
(12) |
Due to the fact that the optimization problem (12) is strictly convex, the sufficient and necessary condition of (12) is
(13) |
Since (13) is equivalent to (12) and the projection onto a convex set is unique, the operator is a point-to-point mapping. This completes the proof of Lemma 1. ∎
Based on Lemma 1, can be determined as
where the projection operator means the projection of onto a closed convex set .
Lemma 2.
With , then for any , the projection of onto is given by
(14) |
According to Lemma 2, can be easily obtained.
3. Update :
4. Update :
Assumption 2.
Given , and , the -update can find a stationary solution such that
for all .
Lemma 3.
For all , the following conditions holds:
Proof.
Theorem 1.
Proof.
With Assumption 2, we have
Besides, we also have
Note that the equality in (IV) holds because of the following claim:
In addition, the inequality in (IV) can be obtained from Lemma 3. Now, the descent of and updates has been proved. Next, we will show the descent of and updates.
Remarkably, the first equality holds due to (IV) and the following derivation:
To prove the inequality in (IV), we define a function . The gradient of is due to (III-B). It follows that , as is convex. Therefore, the inequality in (IV) holds.
Under the condition , the descent of with updating can be proved, which means the augmented Lagrangian function is sufficient descent.
Since the function , a part of the augmented Lagrangian function, is Lipschitz differentiable with , and is bounded by , it is obvious that is lower bounded [28]. Thus, the solution of the inner loop of this transformed problem (III-B) will converge to the stationary point . This completes the proof of Theorem 1. ∎
5. Update and : Then, in the outer iterations, the dual variable is updated. In order to drive the slack variable to zero, the penalty parameter is multiplied by a ratio if the currently computed from the inner iterations does not decrease enough from the previous outer iteration , i.e., .
where the projection operator means the projection of onto a closed convex set .
Lemma 4.
Given a hypercube set with and , for any , the projection of onto the hypercube is given by
(15) |
where the subscript means the th component in vector , and .
Theorem 2.
Proof.
Assume that converges to the stationary point . We have and .
For the case where is bounded, we have and . For the case where is not bounded, based on the optimality condition (10a), it follows that
(18) |
Then, we have by taking the limitation of (18) to zero, as and . Since we have in both of the two cases, the optimality condition with respect to , i.e., (2), can be satisfied. Besides, the optimality condition with respect to and , i.e., (16a) and (2), can also be satisfied by taking on (10a) and (III-B), respectively. This completes the proof of Theorem 2. ∎
Remark 3.
Our algorithm explicitly demonstrates how to apply the hierarchical ADMM in the multi-agent system with the presence of nonconvex collision-avoidance constraints. The closed-form solution of every variable is provided. In this approach, a proximal term is introduced to the two-level algorithm presented in [28] and a more compact convergence condition is proposed, i.e., . Therefore, the Theorem 1 in our paper proves the convergence of the inner iteration under the condition . Theorem 2 is proved under different assumptions, compared with the algorithm proposed in [28]. Thus, we believe our algorithm and theorems provide more details about the hierarchical ADMM algorithms for the multi-gent systems.
The stopping criterion of this ADMM algorithm is given by
(19a) | |||
where .
To summarize the above developments, the pseudocode of this algorithm is presented in Algorithm 1. After obtaining each agent’s control input and executing the control input, the agents will communicate with their neighbors to obtain the updated position information of their neighbors. Then, this process will be repeated until finishing the designed task.
V Improved Hierarchical ADMM
In the hierarchical ADMM approach, the minimization of in each iteration is computationally expensive, and it requires nonlinear programming (NLP) solvers. Therefore, the improved minimization of is required to improve the computational efficiency by approximating and solving this inexact minimization problem.
The original -minimization problem can be expressed as
where
The KKT conditions of this -minimization problem (V) are
Notice that the hierarchical ADMM algorithm requires complete minimization of in each inner iteration. However, this is neither desirable due to the computational cost nor practical, as any NLP solver finds only a point that approximately satisfies the KKT optimality conditions. In the hierarchical ADMM algorithm, we propose the use of the interior-point method to solve the minimization problem (V). The interior-point method employs double-layer iterations to find the solution. In the outer iteration, a barrier technique is used to convert the inequality constraints into an additional term in the objective; the stationary points of resulting barrier problems converge to true stationary points as the barrier parameter converges to 0. In the inner iteration, a proper search method is used to obtain the optimum of the barrier problem. Since both the interior-point algorithm and the hierarchical ADMM algorithm have a double-layer structure, we consider matching these two layers.
In order to accelerate the computation process of solving (V), the barrier method can be employed to transform the inequality constraints to additional terms in the objective function. For example, in the -th outer loop, the function can be added with a barrier term based on the barrier method, where is the barrier parameter with . Hence, the barrier augmented Lagrangian in the -th outer loop can be written as
(22) |
If the -minimization step returns by minimizing with respect to , then the inner iterations lead to the decrease of , which also matches the KKT conditions of the original optimization problem. Therefore, the outer iterations can find an approximate stationary point of the original problem with the decrease of the barrier parameter . Note that only is used as the inequality constraint for the purpose of demonstration.
Therefore, it remains to minimize , which is given by
where .
Lemma 5.
The gradient of with respect to is a linear mapping with respect to .
Proof.
Set and define such that . Then, the gradient of the barriered augmented Lagrangian function is
A closed-form algebraic solution of problem (V) is challenging to determine. However, since the gradient is available, some well-known gradient-based numerical methods, e.g., Barzilai-Borwein, Polak-Ribiere, and Fletcher-Reeves, can be used to calculate the optimal . Therefore, the -update step in Algorithm 1 can be replaced by solving an unconstrained optimization problem (V) with the known gradient (V).
Usually, iterative algorithms for NLP need to be called when solving the minimization problem (V), and always searching for a highly accurate solution in each ADMM iteration will result in an excessive computational cost. It is thus desirable to solve the optimization subproblem (V) in ADMM inexactly when the dual variables are yet far from the optimum, i.e., to allow to be chosen such that
(23) |
In this paper, we use externally shrinking and summable sequence of the absolute errors, i.e.,
(24) |
such as a summable absolute error sequence can effectively reduce the total number of subroutine iterations throughout the ADMM algorithm. Such a criterion is a constructive one, rendered to guarantee the decrease of a quadratic distance between the intermediate solutions and the optimum .
The above approximate NLP solution is realizable by NLP solvers where the tolerances of the KKT conditions are allowed to be specified by the user, e.g., the IPOPT solver. The pseudocode of the detailed algorithm is shown in Algorithm 2.
In this method, a nonlinear programming problem without constraints is solved with inequality constraints handled by a fixed barrier constant throughout inner iterations. The barrier constant decreases across outer iterations. In the hierarchical ADMM, the NLP solver needs to solve an inequality-constrained NLP in each inner iteration. When the NLP solver, i.e., interior-point algorithm, is called for NLP with inequalities, the barrier constant will go through an iteration inside the solver, and thus every inner iteration will have more loops of internal iteration. By this barrier Lagrangian construction, we integrate the inner loop of the IPOPT with the inner loop of the hierarchical ADMM algorithm. In addition, in the improved hierarchical ADMM, each inner iteration only solves NLP, and an approximate solution is reached without the need of convergence. The accuracy, namely the tolerances of the errors from the KKT conditions, can be controlled when using the interior-point algorithm. Therefore, these tolerances can be set adaptively throughout the iterations (for instance, the tolerances of the 1st, 5th, and 20th iteration can be set to 0.1, 0.01, and 0.001, respectively). By this adaptive tolerance setting, we integrate the outer loop of the IPOPT with the inner loop of the hierarchical ADMM. Therefore, the improved hierarchical ADMM gives less iteration number and computation time than the hierarchical ADMM.
VI Illustrative Example
In order to show the effectiveness of the hierarchical ADMM and improved hierarchical ADMM, a multi-UAV system is presented as an application test platform for demonstration of the proposed decision-making framework. The simulations are performed in Python 3.7 environment with processor Intel Xeon(R) E5-1650 v4 @ 3.60GHz. All nonlinear problems are solved by the open source IPOPT solver.
VI-A Dynamic Model
For each quadcopter system, the system model can be expressed in the following state-space representation:
where and are state and control input vector for the agents, respectively. In this dynamics model, rotor thrusts of each rotor are chosen as control input vector, i.e., . is a vector comprising the position in the -, -, and -dimension. Additionally, the velocity vector is represented by . denotes a vector in terms of the roll, pitch, and yaw angles. The angular velocity vector is represented by . Also, , is the control input to balance the gravity of the agents, is the gravitational acceleration, is the mass of the quadcopter, denotes the moment of inertia of the quadcopter, and denote the rotation matrices of the quadcopter. In addition, is the control input matrix in the above state-space representation.
Here, (VI-A) is nonlinear and time-variant, which can be represented as a pseudo-linear form by using state-dependent coefficient factorization [31]. The state-space expression is
(25) |
where is the sampling time interval. Since and are dependent on the current state , this state-space representation is in a pseudo-linear form, and then we can suitably consider the system matrices to be constant during the prediction horizon. More details of the parameter settings in (VI-A) and (25) are presented in [32].
VI-B Optimization and Simulation Results
The dynamic model of each agent in the multi-UAV system is shown in (25). All the agents are set evenly distributed in a circle with the radius of m and m, and each agent needs to reach its corresponding point in this same sphere. Therefore, a large number of potential collision possibilities are set up in the simulation task, and our objective is to avoid these designed potential collisions among these agents and make all agents reach their desired destinations. Here, a distributed MPC with a quadratic objective function is designed to realize the path tracking and collision avoidance tasks. Set the number of agents . The safety distance is m. The dimensions are . In addition, the agents are connected with all of the remaining agents. The adjacency matrix of the communication network is the difference between the upper triangular matrix of the square one matrix and the identity matrix , i.e., . In this adjacency matrix , each entry represents whether there is the connection between the th agent and the th agent. If , there is a communication connection between the two agents; otherwise, there is no connection between the two agents. For example, the th agent will get communication with the agents. The velocity of the agents is confined into [-5, 5] m/s, and the upper and lower bounds of angles in three dimensions are and . The upper bound of control input variable is N and its lower bound is . Set the prediction horizon with the sampling time s. The weighting matrices and for all . Each element in and are set as and , respectively. and are set as 0.9 and 1.1. The initial tolerances are defined as respectively, and . The terminated tolerances for outer iterations , , are , respectively. The maximum iteration number is set to .
The reference trajectories and the resulted trajectories computed by the improved hierarchical ADMM are shown in Fig. 1 and Fig. 2. The initial positions of the 8 agents are distributed uniformly in a circle, whose radius equals 2 m, and the z-axis is 0. In this simulation task, the 8 agents need to move towards their corresponding points, which are the central symmetric points of their initial position, respectively. For example, the agents with initial position (0,2,0) m and () m needs to reach (0,-2,0) m and () m. In the two figures, each agent is represented by a different color. These lines in different colors denote the resulted trajectories of the 8 agents, respectively. From the two figures, it is obvious that all agents can effectively avoid their neighbors and maintain the safety distance. On the other hand, for the hierarchical ADMM, a similar performance of collision avoidance can be attained.


In order to show the bound of the control inputs, we only choose the control input results of the 1st agent as an example for the clarity of this figure, as shown in Fig. 3. The 4 dimensions of the control inputs are bounded in the predefined range N.

The distances between each pair of the 8 agents are shown in Fig. 4. The solid gray line in this figure represents the safety distance all of the agents need to maintain, and the other colored lines denote the distance of each pair of agents with respect to time. It shows that the distance between agents are no smaller than the safety distance.

The comparison of computational efficiency between our proposed hierarchical ADMM and improved hierarchical ADMM can be illustrated, according to Fig. 5 and Fig. 6. Fig. 5 demonstrates the comparison of iteration numbers the inner loop and outer loop require in both hierarchical ADMM and improved hierarchical ADMM. It is straightforward to observe that the improved hierarchical ADMM can effectively reduce the iteration numbers of both the inner loop and outer loop, compared with the hierarchical ADMM. The comparison of computation time for all inner loop iterations between hierarchical ADMM and improved hierarchical ADMM is shown in Fig. 6. From this figure, it can be observed that the computation time is significantly reduced by using the improved version.


Centralized MPC | Penalty Dual Decomposition [33] | Hierarchical ADMM | Improved Hierarchical ADMM | |
Average outer iterations | - | 4.26 | 2.12 | 1.03 |
Average inner iterations | - | 17.36 | 6.38 | 1.12 |
2.21E-5 | 3.32E-5 | 2.54E-5 | 1.47E-5 | |
Cost value | 253.12 | 315.35 | 294.23 | 245.36 |
Stopping criterions | - | 1E-6, 7.52E-2, 2.32E-3 | 1E-6, 9.38E-2, 3.5E-3 | 1E-6, 3.28E-2, 2.64E-3 |
Time | 1.47s | 0.22s | 0.16s | 0.09s |
In order to show the effectiveness of the proposed improved hierarchical ADMM, the penalty dual decomposition method [33] and the centralized MPC method are used as the comparison methods. The centralized MPC method means we solve the centralized MPC problem by using a NLP solver, and the penalty dual decomposition method is also a double-loop iterative algorithm [33]. The inner loop of the penalty dual decomposition method is used to solve a nonconvex nonsmooth augmented Lagrangian problem via block-coordinate-descent-type methods, and the outer loop of this algorithm is to update the dual variables and/or the penalty parameter [33]. Table I illustrates the advantages of our proposed algorithms (hierarchical ADMM and improved hierarchical ADMM) over the penalty dual decomposition method and the centralized MPC method.
In order to demonstrate the performance in terms of computational efficiency, the comparison of computation time per step (including all outer loop iterations and inner loop iterations) under the hierarchical ADMM and the improved hierarchical ADMM, and the computation time per step under the centralized MPC (solved by the IPOPT solver) are shown in Fig. 7. From this figure, the computational efficiency of the ADMM for solving large-scale problems can be clearly observed. Moreover, the improved hierarchical ADMM is capable of further improving the efficiency, as compared with the hierarchical ADMM. With the increasing number of agents, the number of nonlinear constraints increases rapidly, which leads to difficulty in solving the problem. Therefore, the computation time is increasing with the increase in the number of agents , as shown in Fig. 8. However, when the centralized MPC solver is used, the growth rate of the computation time is much higher than the ones when the hierarchical ADMM and improved hierarchical ADMM are used, as shown in Fig. 8, which also demonstrates the efficiency of the two proposed methods. Besides, the efficiency of the improved hierarchical ADMM can also be proved in this figure.


VII Conclusion
In this paper, a large-scale nonconvex optimization problem for multi-agent decision making is investigated and solved by the hierarchical ADMM approach. For the multi-agent system, an appropriate DMPC problem is formulated and presented to handle the constraints and achieve the collective objective. Firstly, an innovation using appropriate slack variables is deployed, yielding a resulting transformed optimization problem, which is used to deal with the nonconvex characteristics of the problem. Then, the hierarchical ADMM is applied to solve the transformed problem in parallel. Next, the improved hierarchical ADMM is proposed to increase the computational efficiency and decrease the number of inner iterations. Additionally, it is analytically shown that the appropriate desired stationary point exists for the procedures of the hierarchical stages for algorithmic convergence. Finally, comparative simulations are conducted using a multi-UAV system as the test platform, which further illustrates the effectiveness and efficiency of the proposed methodology.
References
- [1] X. He, T. Huang, J. Yu, C. Li, and Y. Zhang, “A continuous-time algorithm for distributed optimization based on multiagent networks,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 49, no. 12, pp. 2700–2709, 2017.
- [2] H. Hong, W. Yu, G. Wen, and X. Yu, “Distributed robust fixed-time consensus for nonlinear and disturbed multiagent systems,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 47, no. 7, pp. 1464–1473, 2016.
- [3] M. Ye, G. Wen, S. Xu, and F. L. Lewis, “Global social cost minimization with possibly nonconvex objective functions: An extremum seeking-based approach,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2020.
- [4] J. Ma, S.-L. Chen, C. S. Teo, A. Tay, A. Al Mamun, and K. K. Tan, “Parameter space optimization towards integrated mechatronic design for uncertain systems with generalized feedback constraints,” Automatica, vol. 105, pp. 149–158, 2019.
- [5] S. Yang, Q. Liu, and J. Wang, “Distributed optimization based on a multiagent system in the presence of communication delays,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 47, no. 5, pp. 717–728, 2016.
- [6] ——, “A multi-agent system with a proportional-integral protocol for distributed constrained optimization,” IEEE Transactions on Automatic Control, vol. 62, no. 7, pp. 3461–3467, 2016.
- [7] D. Yuan, S. Xu, and H. Zhao, “Distributed primal–dual subgradient method for multiagent optimization via consensus algorithms,” IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 41, no. 6, pp. 1715–1724, 2011.
- [8] E. Camponogara and L. B. De Oliveira, “Distributed optimization for model predictive control of linear-dynamic networks,” IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans, vol. 39, no. 6, pp. 1331–1338, 2009.
- [9] M. Hong, “A distributed, asynchronous, and incremental algorithm for nonconvex optimization: an ADMM approach,” IEEE Transactions on Control of Network Systems, vol. 5, no. 3, pp. 935–945, 2017.
- [10] J. Ma, Z. Cheng, X. Zhang, M. Tomizuka, and T. H. Lee, “On symmetric Gauss-Seidel ADMM algorithm for guaranteed cost control with convex parameterization,” arXiv preprint arXiv:2001.00708, 2020.
- [11] ——, “Optimal decentralized control for uncertain systems by symmetric Gauss-Seidel semi-proximal ALM,” IEEE Transactions on Automatic Control, DOI: 10.1109/TAC.2021.3052768, 2021.
- [12] Y. Hu, E. C. Chi, and G. I. Allen, “ADMM algorithmic regularization paths for sparse statistical machine learning,” in Splitting Methods in Communication, Imaging, Science, and Engineering. Springer, 2016, pp. 433–459.
- [13] J. Ding, Y. Gong, C. Zhang, M. Pan, and Z. Han, “Optimal differentially private ADMM for distributed machine learning,” arXiv preprint arXiv:1901.02094, 2019.
- [14] B. He and X. Yuan, “On the convergence rate of the Douglas-Rachford alternating direction method,” SIAM Journal on Numerical Analysis, vol. 50, no. 2, pp. 700–709, 2012.
- [15] R. D. Monteiro and B. F. Svaiter, “Iteration-complexity of block-decomposition algorithms and the alternating direction method of multipliers,” SIAM Journal on Optimization, vol. 23, no. 1, pp. 475–507, 2013.
- [16] E. Wei and A. Ozdaglar, “Distributed alternating direction method of multipliers,” in 2012 IEEE 51st IEEE Conference on Decision and Control (CDC). IEEE, 2012, pp. 5445–5450.
- [17] A. Makhdoumi and A. Ozdaglar, “Convergence rate of distributed ADMM over networks,” IEEE Transactions on Automatic Control, vol. 62, no. 10, pp. 5082–5095, 2017.
- [18] B. He, M. Tao, and X. Yuan, “Alternating direction method with Gaussian back substitution for separable convex programming,” SIAM Journal on Optimization, vol. 22, no. 2, pp. 313–340, 2012.
- [19] T. Lin, S. Ma, and S. Zhang, “On the global linear convergence of the ADMM with multiblock variables,” SIAM Journal on Optimization, vol. 25, no. 3, pp. 1478–1497, 2015.
- [20] G. Li and T. K. Pong, “Global convergence of splitting methods for nonconvex composite optimization,” SIAM Journal on Optimization, vol. 25, no. 4, pp. 2434–2460, 2015.
- [21] H. Wang and A. Banerjee, “Bregman alternating direction method of multipliers,” in Advances in Neural Information Processing Systems, 2014, pp. 2816–2824.
- [22] B. Jiang, T. Lin, S. Ma, and S. Zhang, “Structured nonconvex and nonsmooth optimization: algorithms and iteration complexity analysis,” Computational Optimization and Applications, vol. 72, no. 1, pp. 115–157, 2019.
- [23] Y. Wang, W. Yin, and J. Zeng, “Global convergence of ADMM in nonconvex nonsmooth optimization,” Journal of Scientific Computing, vol. 78, no. 1, pp. 29–63, 2019.
- [24] X. Wang, J. Yan, B. Jin, and W. Li, “Distributed and parallel ADMM for structured nonconvex optimization problem,” IEEE Transactions on Cybernetics, 2019.
- [25] Z. Cheng, J. Ma, X. Zhang, C. W. de Silva, and T. H. Lee, “ADMM-based parallel optimization for multi-agent collision-free model predictive control,” arXiv preprint arXiv:2101.09894, 2021.
- [26] J. Ma, Z. Cheng, X. Zhang, M. Tomizuka, and T. H. Lee, “Alternating direction method of multipliers for constrained iterative LQR in autonomous driving,” arXiv preprint arXiv:2011.00462, 2020.
- [27] B. Houska, J. Frasch, and M. Diehl, “An augmented lagrangian based algorithm for distributed nonconvex optimization,” SIAM Journal on Optimization, vol. 26, no. 2, pp. 1101–1127, 2016.
- [28] K. Sun and X. A. Sun, “A two-level distributed algorithm for nonconvex constrained optimization,” arXiv preprint arXiv:1902.07654, 2019.
- [29] W. Tang and P. Daoutidis, “Fast and stable nonconvex constrained distributed optimization: The ELLADA algorithm,” arXiv preprint arXiv:2004.01977, 2020.
- [30] D. Limón, T. Alamo, F. Salas, and E. F. Camacho, “On the stability of constrained mpc without terminal constraint,” IEEE Transactions on Automatic Control, vol. 51, no. 5, pp. 832–836, 2006.
- [31] X. Zhang, J. Ma, S. Huang, Z. Cheng, and T. H. Lee, “Integrated planning and control for collision-free trajectory generation in 3D environment with obstacles,” in Proceedings of International Conference on Control, Automation and Systems, 2019, pp. 974–979.
- [32] X. Zhang, J. Ma, Z. Cheng, S. Huang, S. S. Ge, and T. H. Lee, “Trajectory generation by chance-constrained nonlinear MPC with probabilistic prediction,” IEEE Transactions on Cybernetics, DOI: 10.1109/TCYB.2020.3032711, 2020.
- [33] Q. Shi and M. Hong, “Penalty dual decomposition method for nonsmooth nonconvex optimization-Part I: Algorithms and convergence analysis,” IEEE Transactions on Signal Processing, vol. 68, pp. 4108–4122, 2020.