Multi-Agent Path Finding with Prioritized Communication Learning
Abstract
Multi-agent pathfinding (MAPF) has been widely used to solve large-scale real-world problems, e.g., automation warehouses. The learning-based, fully decentralized framework has been introduced to alleviate real-time problems and simultaneously pursue optimal planning policy. However, existing methods might generate significantly more vertex conflicts (or collisions), which lead to a low success rate or more makespan. In this paper, we propose a PrIoritized COmmunication learning method (PICO), which incorporates the implicit planning priorities into the communication topology within the decentralized multi-agent reinforcement learning framework. Assembling with the classic coupled planners, the implicit priority learning module can be utilized to form the dynamic communication topology, which also builds an effective collision-avoiding mechanism. PICO performs significantly better in large-scale MAPF tasks in success rates and collision rates than state-of-the-art learning-based planners.
I INTRODUCTION
With the rapid development of low-cost sensors and computing devices, more and more manufacturing application scenarios can support the concurrent control of large-scale automated guided vehicles (AGVs) [1, 2]. To achieve efficient large-scale transportation via AGVs, many efforts have been devoted to multi-agent path finding, which aims to plan paths for all agents [3, 4, 5, 6, 7, 8, 9]. A critical constraint is that the agents must follow their paths concurrently without collisions [10]. Although classic “optimal” planners [3, 4, 5, 6] can guarantee the completeness of the solution and collision-free under certain assumptions, these planners have to re-plan for new scenarios, which cannot adequately satisfy the real-time requirements of realistic tasks.


Recently, many learning-based methods [7, 8, 9] were proposed to solve the above issues. Unlike classic planners that calculate a complete path for each agent based on global information, learning-based methods will learn a planning policy for each agent, which only employs local observations to decide one-step or limited length of paths. These learning-based methods generally model the MAPF problem as a multi-agent reinforcement learning (MARL) problem. The collision-free path planning can further be approximated in two ways. The first is to introduce post-processing techniques to avoid collisions by adjusting each agent’s individual paths [8], however this scheme might be extremely time-consuming. In a more “soft” manner, the second way is penalizing collisions by hand-crafted reward shaping [7, 9], which do not assure that the solutions are collision-free. But, its apparent shortcoming is that collision-free planning cannot usually be achieved. From the experimental perspective, the agents following the obtained planning paths will frequently collide with each other, thereby increasing the length of the planned paths than classic planners due to their decentralized framework, e.g., PRIMAL [7] and ODrM* [11] in Figure 1. More importantly, each agent make independent decisions based on its local observations without any global information, which could lead to a deadlock state [12].
The key reason, that existing learning-based MAPF algorithms cannot achieve better collision avoidance and real-time simultaneously, lies in their fully decentralized framework. For each agent, classic planners usually employ the full or partial information of others to achieve collision-free. Therefore, this paper utilizes the decentralized path planning and centralized collision avoiding framework, and introduces the communication-based MARL to design our learning-based method. However, one challenge is that a collision avoiding driven communication protocol becomes more challenging to obtain through this end-to-end communication learning scheme than classic planners [13]. Prioritized MAPF algorithms [3, 14, 15], which usually assign a total priority to each agent and plan a path for each agent from high-priority to low-priority, are among the most efficient state-of-the-art MAPF methods for collision avoidance [5, 16, 17]. However, they determine a predefined total priority ordering of the agents, and might lead to lousy quality solutions or even fail to find any solutions [14]. This motivates us to introduce the priority learning skill into communication learning scheme to learn a novel communication protocol with the collision alleviating ability.

In this paper, we propose the PrIoritized COmmunication Learning method (PICO), to incorporate the implicit planning priority into the communication topology within the communication-based MARL framework. Specifically, PICO includes two phases, i.e., implicit priority learning phase and prioritized communication learning phase as shown in Figure 2. In the implicit priority learning, PICO aims to build an auxiliary imitation learning task to predict the local priority of each agent by imitating classical coupled planner (e.g., ODrM*). The information diffusion path will be determined by these implicit priorities, while the obtained asymmetric information and different implicit constraints will drive the agents to plan the nearly collision-free paths. Besides benefiting to collision avoiding, the priorities can further be utilized to establish the collision avoiding driven communication protocol. In the prioritized communication learning, the obtained local priorities is used to generate a time-varying communication topology consists of agent clusters, where these priorities are considered as the weights of an ad-hoc routing protocol. Each agent can learn its collision-reducing policy via the received messages, which are considered as the environment’s perception and are encoded by the collision-avoiding communication topology.
The main contributions are summarized as follows:
1) The decentralized path planning and centralized collision avoiding scheme is utilized via the communication-based MARL framework, to avoid collision more efficiently;
2) The auxiliary imitation learning task is introduced to conduct implicit priority learning based on classic coupled planners’ local priorities;
3) The structured communication protocol with the collision alleviating ability is constructed by integrating the learnt priorities into communication learning scheme;
4) Our proposed PICO performs significantly better in large-scale multi-agent path finding tasks in both success rate and collision rate than state-of-the-art learning-based planners.
II RELATED WORK
II-A Learning-based Multi-Agent Path Finding
Many works employed MARL or deep learning (DL) methods to solve the MAPF problem. The typical PRIMAL method [7] introduced MARL firstly, and combined behavior cloning from a dynamically coupled planner to accelerate the training procedure. Further the PRIMALc [18] extended PRIMAL from 2D to 3D search space. PRIMALc introduced “agent modeling” technique to assist path planning by predicting the actions of other agents in a decoupled manner. Recently, PRIMAL2 [9] extended PRIMAL to the lifelong MAPF (LMAPF) scenario, in which each agent will be immediately assigned a new goal once upon reaching their current goal. [8] proposed the MATS method which employed the multi-step ahead tree-search strategy in single-agent reinforcement learning (SARL) and imitation learning scheme to fit the results of the tree search strategy to solve the MAPF problem. BitString [19] and DHC [20] introduced the communication-based learning scheme based on end-to-end training procedure, but under fixed communication topology.
II-B Prioritized Multi-Agent Path Finding
Prioritized planning [21] can be considered as a decoupled approach, where agents are ordered by predefined priorities. Fixed predefined priorities are assigned to all agents globally while all collisions are resolved according to these priors before the movements begin. Specially, the priorities can be assigned arbitrarily [3, 22, 23] or handcrafted [21]. Heuristic methods can also be utilized to assign priorities, e.g., [24, 25, 26]. Further, some local orderings can assign temporary priorities to some agents in order to resolve collisions on the fly. This kind of method requires all agents to follow their assigned paths, while if an impasse is reached the priorities are assigned dynamically to determine which agent to wait [27, 28]. In order to improve the solution quality, many existing works attempt to elaborately explore the space of all total priority orderings. Some works explored this space randomly by generating several total priorities ordering as part of a hill-climbing scheme [23], instead of ergodically which can be enumerated for only up to agents [28]. Recently, CBSw/P [14] enumerated the sub-space (of all local priority orderings) as part of a conflict-driven combinatorial search framework.
III PRELIMINARIES
Multi-Agent Path Finding. Classical MAPF can be formalized as follows. The input to a MAPF problem with agents is a tuple where denotes an undirected graph, and maps an agent to a source and goal vertex. Each agent can either move to an adjacent vertex or wait at its current vertex. A sequence of actions is a single-agent plan for agent iff executing this sequence of actions in results in being at , that is, iff . The solution is a set of single-agent plans, one for each agent. The overarching goal of MAPF solvers is to find a solution that can be executed without collisions. Let be a pair of single-agent plans, a vertex conflict between and occurs iff according to their plans, the agents are planned to occupy the same vertex at the same timestep. This paper uses one of the most common functions, makespan which is defined as , for evaluating a solution as in classical MAPF [10].
Partially Observable Stochastic Game. POSG [29] can be denoted as a tuple where is the number of agents, is the agent space, is the state space, is the action space of agent , is the joint action space, is the state transition probability function, is the observation space of agent , is the joint observation space, is the observation emission probability function, and is the reward function of agent . The objective of each agent is to maximize its expected total return during the game.
IV THE PROPOSED PICO METHOD
In this section, we propose our prioritized communication learning method, to incorporate the implicit planning priority into the communication topology learning procedure within the communication-based MARL framework. As shown in Figure 2, PICO includes two phases, i.e., the implicit priority learning and the prioritized communication learning. These two phases are executed randomly during training procedure.
IV-A Implicit Priority Learning
The implicit priority learning phase aims to imitatively learn implicit priorities from classical coupled planner. In this paper, we utilize the ODrM* as the targeted classical coupled planner, which is considered as the state-of-the-art MAPF method [11]. Further, we introduce the behavior cloning [7, 8, 9, 18] to enhance the training procedure efficiency from the perspectives of both convergence speed and training performance. To emphasis, the behavior cloning shares the same low-level parameters with the implicit priority learning model, while this mutual restriction will benefit to the effectiveness of both implicit priority learning and behavior cloning [30, 31]. As follows, we will introduce the detailed procedure of this phase, while training datasets need to be constructed first.
Imitation Dataset Construction: Each agent’s observation and action spaces are the similar as previous work [7, 9]. The training sets , for the implicit priority learning and the behavior cloning training respectively can be constructed based on the solutions of the classical coupled planner, i.e., the ODrM*. In every priority learning episode, given the -th randomly generated MAPF instance , the ODrM* method will obtain a solution, i.e., the set of all single-agent plans (). By assuming the makespan of the solution is , all position-position pairs () can be calculated through unfolding all single-agent plans . At each timestep , current positions and corresponding goals of all agents can be combined to construct , i.e.,
where denotes the observations constructed by () within the same scheme as [7, 9], and denotes the actions obtained by calculating the difference between two consecutive positions .
Further, we construct the ground-truth implicit priorities of each observation according to the following steps. Based on the observations of each agent with its goal , the next optimal action set is established through a greedy strategy (the next optimal action is the action that can touch the goal at the fastest speed without considering other agents). If the agent is at the goal, then the optimal action set is empty. Considering the optimal action set and the output action of ODrM*, the implicit priority of can be calculated via following two rules: i) the action is “no-movement”: the priority is set to be high () if the optimal action set is empty, otherwise is set to be low (); ii) the action is “movement”: the priority is set to be high () if the action is in the optimal action set, otherwise is low (). The dataset is constructed accordingly,







Network Structure and Training Procedure: With the obtained training dataset, we further establish the implicit priority learning model and the behavior cloning model structures as in Figure 3. These two neural networks share the same low-layer parameters with different output-heads. It is worth noting that messages from other agents are set to be the input of the neural network in Figure 3. The corresponding messages are generated through the outputs (the “message” output-head in the Figure 3) of other agents. The objective agents of sending the obtained messages are determined by the dynamic communication topology generated in the following prioritized communication learning phase.
The above-mentioned neural network is parameterized by and denoted as . The output message, predicted priority and prediction action are denoted as , and respectively. In the training procedure111The predicted priorities are utilized to construct the communication topology in the prioritized communication learning phase. To strengthen the stability of the communication learning procedure, PICO introduces a similar target network as in [32] and [33]. The target network is updated with a slow-moving average of the online network., the loss function is set to be
(1) |
where denotes the penalty parameter and
where , and represent the action space’s size, the -th element in and respectively.
For the multi-agent learning problem, one of the key challenges is to encourage each agent to be selfless, while it might be detrimental to the immediate return maximization. This challenge is usually denoted as social dilemma [34] and exists conspicuously in MAPF [12]. Many algorithms have been proposed to alleviate this problem in multi-agent learning [35, 36, 37, 38], however rarely few works pay attention to MAPF problem. PICO method introduces additional blocking prediction tasks together with message exchanging to alleviate the social dilemma similar with the learning-based PRIMAL [7]. Furthermore, an on-goal prediction task is introduced to establish an accurate perception of whether the agent has reached the goal. Ablation studies show that the blocking prediction task can obtain better results when paired with the on-goal prediction task. These two tasks can provide more supervision signals from different perspectives for both implicit priority learning and behavior cloning learning, which also can help to learn better representations of the local observations.
IV-B Prioritized Communication Learning
Dynamic Prioritized Communication Topology: In order to achieve proper integration of implicit priority and communication learning, PICO introduces the classic ad-hoc routing protocol, i.e., cluster-based routing protocol (CBRP [39]) into the overall framework. CBRP employs the “weight” attributes attached to each agent to construct a cluster-based communication topology, which can be dynamically updated once the agent weights change. The high-weight agent becomes the center of the cluster, receives messages from all low-weight agents in the cluster, and broadcasts it to all low-weight agents after post-processing. Agents with low weights become affiliated members and make decisions based on the messages broadcast from the cluster‘s central agent. PICO chooses CBRP as the fundamental communication topology construction algorithm in MAPF because as long as the artificially defined weights in CBRP are replaced by the obtained implicit priorities, the integration of priority and communication learning can be properly realized.
Specifically, CBRP usually takes a cluster radius to establish the structured communication topology. Each member agent can confirm whether others have larger weights or contain central agents within its receptive area. If no such agent is found, this agent is elected as a central agent; otherwise, its role is kept. Meanwhile, each central agent checks whether other centrals exist in radius . If no such agent is found or the found centrals’ weights are smaller than its weights, its role is kept; otherwise, it downgrades to a member agent. After a sufficient number of rounds, all agents are separated into clusters with one central agent in each cluster. This cluster-based communication topology inherits the classic planners’ idea of avoiding planning only in the agent’s neighborhood but also achieves the centralization-decentralization trade-off in MARL. The illustration diagram of CBRP is shown in Figure 4.

Network Structure and Training Procedure: PICO employs the A3C [40] as the backbone of communication-based MARL. The observation space is the same as the implicit priority learning phase. We employ the different action space from the prioritized communication learning phase for the reason that the agent may execute invalid actions. If there are static obstacles, other agents, or beyond the map’s boundary at the position reached by an agent’s following action, such an action is called invalid. Actions are sampled only from valid actions in training. Previous results in [41, 12] show that this technique enables a more stable training procedure, compared with giving negative rewards to agents for selecting invalid moves. Moreover, the reward function is same as [7]. The network structure is consistent with previous phase as shown in Figure 3. In the training procedure, the parameters of “priority”, “blocking”, and “on-goal” three output-heads are fixed. The predicted value by the “value” output-head is denoted as . The parameters are updated to minimize the Bellman error between the predicted value and the discounted return, i.e.,
where denotes the timestep of , is the mini-batch size, and is the discount factor. An approximation of the advantage function by bootstrapping the value function is used to update the policy. An entropy term is further introduced to encourage exploration and discourage premature convergence [42]. The loss function of this prioritized communication learning phase can be denoted as
where and denotes a small positive entropy weight.
V NUMERICAL EXPERIMENTS
In this section, we conduct the numerical experiments based on the modified gridworld environment of the Asprilo benchmark [43], which is commonly adopted by classic or learning-based MAPF solvers. First, some fundamental settings are discussed in the following.
Environments. In training stage, the sizes of the gridworld are randomly selected as (twice as likely), , or , when each episode starts. The number of agents is fixed at . The obstacle density is randomly selected from a triangular distribution between and . In testing stage, the gridworld size is fixed at , the number of agents varies among , , , and . The obstacle density is selected from to with an equal interval .
Measurement | Description |
---|---|
Collision with agents (CA) | Number of vertex conflicts with other agents. |
Collision with obstacles (CO) | Number of collisions with obstacles. |
Success rate (SR) | Number of successful solutions. |
Makespan (MS) | As defined in Section 2.1 |
Collision rate (CR) | CA / MS |
Total moves (TM) | Number of non-still actions taken by all agents. |
Measurements. To evaluate the performance from various aspects, the average values of randomly generated cases/environments on measurements are shown in Table I.
Baseline Methods and Training Details. Baseline methods include classic planners and learning-based (or MARL-based) methods. Based on the results shown in recent works [7, 9], we selects ODrM* with inflation factor as the classic planner baseline. In addition, we choose PRIMAL [7] and DHC [20] as learning-based baselines. For ODrM*, a timeout of seconds is used to match previous results [44]. For PRIMAL and DHC, the agents plan individual paths are for up to timesteps. The training procedure is performed at the same device as baselines. See Table II for more details.
Name | Value | Name | Value | Name | Value | Name | Value |
parallel envs | 12 | episode length | 256 | lr decay | 5e-5 | policy reg | L2 |
steps/update | 128 | batch size | 128 | policy reg coef | 1e-3 | value reg | L2 |
batch handling | shuffle | value loss | MSE | value reg coef | 1.0 | value grad clip | 10*num of agents |
imitation loss | CrossEntropy | discount | 0.95 | policy grad clip | 0.5 | temperture | 0.5 |
optimizer | Nadam | lr | 2e-4 | warm up episodes | 2500 | CBRP neighbors | 11 |
moment | 0.9 | epsilon | 1e-7 | CBRP waiting time | 1 | LSTM embedding | 512 |
8 Agents (0%,10%,20%,30% Obstacle Densities) | ||||||||||||||||||||||||
Methods | CA | CO | SR | MS | CR | TM | ||||||||||||||||||
ODrM* [11] | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 100.0 | 100.0 | 100.0 | 100.0 | 25.55 | 23.95 | 25.86 | 29.45 | 0.0 | 0.0 | 0.0 | 0.0 | 112.75 | 108.83 | 117.49 | 117.38 |
PRIMAL [7] | 1.94 | 3.02 | 3.03 | 5.98 | 0.0 | 0.01 | 0.01 | 0.0 | 93.0 | 90.0 | 48.0 | 15.0 | 34.86 | 62.69 | 148.59 | 234.22 | 0.06 | 0.05 | 0.02 | 0.03 | 221.3 | 223.01 | 345.09 | 565.11 |
DHC [20] | 1.66 | 2.72 | 3.74 | 4.17 | 0.0 | 0.0 | 0.0 | 0.0 | 91.0 | 87.0 | 55.0 | 11.0 | 33.7 | 63.0 | 136.89 | 241.55 | 0.05 | 0.04 | 0.03 | 0.02 | 271.56 | 239.66 | 401.48 | 638.62 |
PICO-heuristic | 2.81 | 2.7 | 3.68 | 3.4 | 0.0 | 0.0 | 0.0 | 0.0 | 90.0 | 86.0 | 50.0 | 10.0 | 33.57 | 59.81 | 141.94 | 236.99 | 0.05 | 0.04 | 0.03 | 0.02 | 248.19 | 252.88 | 451.09 | 650.62 |
PICO | 0.59 | 0.62 | 1.31 | 2.32 | 0.0 | 0.0 | 0.0 | 0.0 | 100.0 | 96.0 | 55.0 | 25.0 | 27.45 | 41.81 | 134.7 | 204.83 | 0.02 | 0.02 | 0.01 | 0.01 | 123.96 | 142.69 | 290.13 | 463.41 |
16 Agents (0%,10%,20%,30% Obstacle Densities) | ||||||||||||||||||||||||
Methods | CA | CO | SR | MS | CR | TM | ||||||||||||||||||
ODrM* [11] | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 100.0 | 100.0 | 100.0 | 99.0 | 26.54 | 27.06 | 28.51 | 33.47 | 0.0 | 0.0 | 0.0 | 0.0 | 216.35 | 224.4 | 228.96 | 265.0 |
PRIMAL [7] | 6.59 | 8.29 | 11.63 | 17.64 | 0.02 | 0.04 | 0.06 | 0.06 | 92.0 | 88.0 | 50.0 | 3.0 | 57.47 | 71.82 | 176.16 | 249.37 | 0.12 | 0.11 | 0.07 | 0.07 | 481.63 | 510.11 | 765.89 | 1396.23 |
DHC [20] | 7.28 | 8.95 | 11.75 | 18.7 | 0.0 | 0.06 | 0.11 | 0.13 | 94.0 | 88.0 | 48.0 | 5.0 | 54.02 | 71.19 | 169.18 | 242.06 | 0.13 | 0.12 | 0.07 | 0.08 | 476.51 | 477.0 | 822.23 | 1503.58 |
PICO-heuristic | 5.93 | 9.71 | 12.28 | 17.54 | 0.0 | 0.04 | 0.04 | 0.28 | 92.0 | 91.0 | 44.0 | 2.0 | 54.07 | 69.49 | 183.38 | 251.16 | 0.12 | 0.14 | 0.07 | 0.07 | 416.37 | 456.39 | 721.84 | 1538.51 |
PICO | 2.98 | 3.98 | 4.96 | 8.0 | 0.0 | 0.02 | 0.02 | 0.03 | 100.0 | 95.0 | 57.0 | 7.0 | 30.83 | 49.06 | 144.67 | 240.15 | 0.1 | 0.08 | 0.03 | 0.03 | 251.39 | 298.55 | 526.4 | 1291.71 |
32 Agents (0%,10%,20%,30% Obstacle Densities) | ||||||||||||||||||||||||
Methods | CA | CO | SR | MS | CR | TM | ||||||||||||||||||
ODrM* [11] | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 100.0 | 100.0 | 100.0 | 91.0 | 29.09 | 29.67 | 32.88 | 38.93 | 0.0 | 0.0 | 0.0 | 0.0 | 441.34 | 462.07 | 505.13 | 530.07 |
PRIMAL [7] | 26.23 | 30.48 | 47.27 | 98.33 | 0.0 | 0.36 | 1.56 | 2.07 | 92.0 | 72.0 | 9.0 | 0.0 | 53.91 | 108.02 | 244.76 | 256.0 | 0.47 | 0.28 | 0.19 | 0.38 | 958.39 | 1094.29 | 2226.83 | 3431.49 |
DHC [20] | 27.04 | 34.79 | 49.29 | 95.51 | 0.02 | 0.65 | 4.28 | 4.67 | 92.0 | 62.0 | 3.0 | 0.0 | 48.76 | 135.78 | 242.67 | 256.0 | 0.52 | 0.26 | 0.21 | 0.37 | 957.06 | 1144.55 | 2009.0 | 3742.41 |
PICO-heuristic | 27.49 | 31.01 | 49.02 | 93.5 | 0.01 | 0.64 | 3.76 | 5.18 | 91.0 | 63.0 | 0.0 | 0.0 | 49.45 | 121.05 | 256.0 | 256.0 | 0.56 | 0.27 | 0.19 | 0.36 | 1144.46 | 1179.09 | 2236.79 | 3723.63 |
PICO | 14.8 | 20.62 | 36.28 | 83.38 | 0.0 | 0.21 | 1.28 | 1.63 | 100.0 | 75.0 | 19.0 | 0.0 | 38.09 | 96.5 | 224.54 | 256.0 | 0.4 | 0.22 | 0.15 | 0.33 | 550.87 | 774.01 | 1712.7 | 3175.57 |
64 Agents (0%,10%,20%,30% Obstacle Densities) | ||||||||||||||||||||||||
Methods | CA | CO | SR | MS | CR | TM | ||||||||||||||||||
ODrM* [11] | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | 100.0 | 96.0 | 79.0 | 17.0 | 31.82 | 31.87 | 29.82 | 7.53 | 0.0 | 0.0 | 0.0 | 0.0 | 916.53 | 936.43 | 844.11 | 202.05 |
PRIMAL [7] | 115.79 | 171.31 | 341.95 | 634.7 | 0.11 | 2.27 | 8.04 | 26.08 | 75.0 | 7.0 | 0.0 | 0.0 | 111.27 | 241.73 | 256.0 | 256.0 | 1.04 | 0.71 | 1.34 | 2.48 | 2418.58 | 3679.7 | 6611.44 | 9156.88 |
DHC [20] | 106.78 | 146.92 | 312.92 | 622.58 | 0.25 | 12.75 | 38.37 | 145.89 | 72.0 | 0.0 | 0.0 | 0.0 | 108.76 | 256.0 | 256.0 | 256.0 | 0.98 | 0.57 | 1.22 | 2.43 | 2121.45 | 3261.68 | 6091.9 | 8407.14 |
PICO-heuristic | 112.99 | 142.03 | 303.23 | 621.68 | 0.25 | 12.75 | 38.37 | 145.89 | 70.0 | 3.0 | 0.0 | 0.0 | 108.76 | 243.36 | 256.0 | 256.0 | 1.03 | 0.58 | 1.18 | 2.42 | 2201.45 | 3246.68 | 6162.9 | 8338.14 |
PICO | 90.95 | 128.4 | 279.61 | 591.14 | 0.36 | 8.76 | 38.36 | 130.27 | 83.0 | 13.0 | 0.0 | 0.0 | 94.36 | 224.88 | 256.0 | 256.0 | 0.96 | 0.55 | 1.09 | 2.31 | 1472.92 | 2621.25 | 5342.04 | 7713.69 |

Results Analysis. Table III shows six measurement comparisons of PICO and baselines in a square gridworld with a size of , different obstacle densities (, , , ), and the number of agents (, , , ). The collision rates between PICO agents in various scenarios are significantly lower than the baselines. It can directly found from these experimental results that PICO has better generalization. PICO is trained only in the -agents scenario, but PICO can still maintain good performance in a new environment that is expanded to times the number of agents.
Ablation Studies. To further analyze the importance of priority and the effectiveness of implicit priority learning, we conduct ablation studies to verify the impact of different priorities on the algorithm’s performance. Specifically, we have designed a greedy priority allocation strategy. At each timestep , we extract the current Manhattan distance to the goal and denote it as . At the same time, for each pair of agents with the same goal distance , if the initial goal distance of the agent is farther from than , then has a higher priority. Therefore, for any agent , the priority is calculated by . If the two agents have the same highest priority, one of them will be randomly selected to be the high-level agent. This ablation version of PICO that uses the above greedy priority strategy is denoted as PICO-heuristic. The comparison results are also shown in Table III. The experimental results well verify the importance of priority in communication learning and the effectiveness of implicit priority learning.
Diving Into the Path Planning. We visually analyze the cluster composition of PICO during path planning. Specifically, we count the changes of the cluster numbers formed by CBRP based on implicit priority during PICO’s path planning, while the results are shown in Figure 5. In the beginning, agents are generally uniformly distributed in the environment, so there will be more clusters formed; As the path planning progresses, the agents begin to intersect each other more, and the number of clusters will be reduced accordingly; Finally, the agents are closer to the goals, which makes the agents farther apart again. We can find that the farther away from the target, the more likely it is to become a central agent. This figure also shows more diverse central agent selection strategies.
VI CONCLUSION
In this paper, we propose the prioritized communication learning method to incorporate the implicit planning priority into the communication topology within the communication-based MARL framework. PICO makes a trade-off between decentralized path planning but (nearly) centralized collision avoiding planning, and achieves better collision avoidance and real-time performance simultaneously. Experiments show that PICO performs significantly better in large-scale pathfinding tasks in both success rates and collision rates than state-of-the-art learning-based planners.
ACKNOWLEDGMENT
This work was supported in part by the National Key Research and Development Program of China (No. 2020AAA0107400), NSFC (No. 12071145), STCSM (No. 19ZR141420), the Open Research Projects of Zhejiang Lab (NO.2021KE0AB03), and the Fundamental Research Funds for the Central Universities.
References
- [1] M. Rubenstein, A. Cornejo, and R. Nagpal, “Programmable self-assembly in a thousand-robot swarm,” Science, vol. 345, pp. 795–799, 2014.
- [2] A. Howard, L. Parker, and G. Sukhatme, “Experiments with a large heterogeneous mobile robot team: Exploration, mapping, deployment and detection,” The International Journal of Robotics Research, vol. 25, pp. 431–447, 2006.
- [3] D. Silver, “Cooperative pathfinding,” in AIIDE, 2005.
- [4] J. V. D. Berg, S. Guy, M. Lin, and D. Manocha, “Reciprocal n-body collision avoidance,” in ISRR, 2009.
- [5] G. Sharon, R. Stern, A. Felner, and N. R. Sturtevant, “Conflict-based search for optimal multi-agent pathfinding,” Artificial Intelligence, vol. 219, pp. 40–66, 2012.
- [6] A. Felner, J. Li, E. Boyarski, H. Ma, L. Cohen, T. K. S. Kumar, and S. Koenig, “Adding heuristics to conflict-based search for multi-agent path finding,” in ICAPS, 2018.
- [7] G. Sartoretti, J. Kerr, Y. Shi, G. Wagner, T. K. S. Kumar, S. Koenig, and H. Choset, “Primal: Pathfinding via reinforcement and imitation multi-agent learning,” IEEE Robotics and Automation Letters, vol. 4, pp. 2378–2385, 2019.
- [8] Y. Zhang, Y. Qian, Y. Yao, H. Hu, and Y. Xu, “Learning to cooperate: Application of deep reinforcement learning for online AGV path finding,” in AAMAS, 2020.
- [9] M. Damani, Z. Luo, E. Wenzel, and G. Sartoretti, “Primal2: Pathfinding via reinforcement and imitation multi-agent learning - lifelong,” IEEE Robotics and Automation Letters, vol. 6, pp. 2666–2673, 2021.
- [10] R. Stern, N. R. Sturtevant, A. Felner, S. Koenig, H. Ma, T. T. Walker, J. Li, D. Atzmon, L. Cohen, T. K. S. Kumar, E. Boyarski, and R. Barták, “Multi-agent pathfinding: Definitions, variants, and benchmarks,” in SOCS, 2019.
- [11] C. Ferner, G. Wagner, and H. Choset, “ODrM* optimal multirobot path planning in low dimensional search spaces,” in ICRA, 2013.
- [12] G. Sartoretti, Y. Wu, W. Paivine, T. K. S. Kumar, S. Koenig, and H. Choset, “Distributed reinforcement learning for multi-robot decentralized collective construction,” in DARS, 2018.
- [13] R. Lowe, J. Foerster, Y. Boureau, J. Pineau, and Y. Dauphin, “On the pitfalls of measuring emergent communication,” in AAMAS, 2019.
- [14] H. Ma, D. Harabor, P. Stuckey, J. Li, and S. Koenig, “Searching with consistent prioritization for multi-agent path finding,” in SOCS, 2019.
- [15] N. R. Sturtevant and M. Buro, “Improving collaborative pathfinding using map abstraction,” in AIIDE, 2006.
- [16] P. Velagapudi, K. Sycara, and P. Scerri, “Decentralized prioritized planning in large multirobot teams,” in IROS, 2010.
- [17] K. Wang and A. Botea, “MAPP: a scalable multi-agent path planning algorithm with tractability and completeness guarantees,” Journal of Artificial Intelligence Research, vol. 42, pp. 55–90, 2011.
- [18] L. Zhiyao and G. Sartoretti, “Deep reinforcement learning based multi-agent pathfinding,” Technical Report., 2020.
- [19] B. Freed, G. Sartoretti, and H. Choset, “Simultaneous policy and discrete communication learning for multi-agent cooperation,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 2498–2505, 2020.
- [20] Z. Ma, Y. Luo, and H. Ma, “Distributed heuristic multi-agent path finding with communication,” in ICRA, 2021.
- [21] E. Michael and T. Lozano-Perez, “On multiple moving objects,” Algorithmica, vol. 2, no. 1, pp. 477–521, 1987.
- [22] C. W. Warren, “Multiple robot path coordination using artificial potential fields,” in ICRA, 1990.
- [23] M. Bennewitz, W. Burgard, and S. Thrun, “Finding and optimizing solvable priority schemes for decoupled path planning techniques for teams of mobile robots,” Robotics and Autonomous Systems, vol. 41, pp. 89–99, 2002.
- [24] J. V. D. Berg and M. Overmars, “Prioritized motion planning for multiple robots,” in IROS, 2005.
- [25] S. Buckley, “Fast motion planning for multiple moving robots,” in ICRA, 1989.
- [26] C. Ferrari, E. Pagello, J. Ota, and T. Arai, “Multirobot motion coordination in space and time,” Robotics and Autonomous Systems, vol. 25, pp. 219–229, 1998.
- [27] P. A. O’Donnell and T. Lozano-Perez, “Deadlock-free and collision-free coordination of two robot manipulators,” in ICRA, 1989.
- [28] K. Azarm and G. Schmidt, “Conflict-free motion of multiple mobile robots based on decentralized motion planning and negotiation,” in ICRA, 1997.
- [29] E. A. Hansen, D. S. Bernstein, and S. Zilberstein, “Dynamic programming for partially observable stochastic games,” in AAAI, 2004.
- [30] S. Levine and V. Koltun, “Learning complex neural network policies with trajectory optimization,” in ICML, 2014.
- [31] S. Levine and P. Abbeel, “Learning neural network policies with guided policy search under unknown dynamics,” in NeurIPS, 2014.
- [32] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. A. Riedmiller, A. Fidjeland, G. Ostrovski, S. Petersen, C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran, D. Wierstra, S. Legg, and D. Hassabis, “Human-level control through deep reinforcement learning,” Nature, vol. 518, pp. 529–533, 2015.
- [33] J.-B. Grill, F. Strub, F. Altch’e, C. Tallec, P. H. Richemond, E. Buchatskaya, C. Doersch, B. A. Pires, Z. Guo, M. G. Azar, B. Piot, K. Kavukcuoglu, R. Munos, and M. Valko, “Bootstrap your own latent: A new approach to self-supervised learning,” in NeurIPS, 2020.
- [34] Y. Shoham and K. Leyton-Brown, Multiagent systems: Algorithmic, game-theoretic, and logical foundations. Cambridge University Press, 2008.
- [35] A. Peysakhovich and A. Lerer, “Consequentialist conditional cooperation in social dilemmas with imperfect information,” in AAAI Workshops, 2018.
- [36] ——, “Prosocial learning agents solve generalized stag hunts better than selfish ones,” in AAMAS, 2018.
- [37] E. Hughes, J. Z. Leibo, M. Phillips, K. Tuyls, E. A. Duéñez-Guzmán, A. Castañeda, I. Dunning, T. Zhu, K. R. McKee, R. Koster, H. Roff, and T. Graepel, “Inequity aversion improves cooperation in intertemporal social dilemmas,” in NeurIPS, 2018.
- [38] N. Jaques, A. Lazaridou, E. Hughes, Çaglar Gülçehre, P. A. Ortega, D. Strouse, J. Z. Leibo, and N. D. Freitas, “Social influence as intrinsic motivation for multi-agent deep reinforcement learning,” in ICML, 2019.
- [39] M. Rezaee and M. Yaghmaee, “Cluster based routing protocol for mobile ad hoc networks,” INFOCOM, vol. 8, no. 1, pp. 30–36, 2009.
- [40] V. Mnih, A. P. Badia, M. Mirza, A. Graves, T. Lillicrap, T. Harley, D. Silver, and K. Kavukcuoglu, “Asynchronous methods for deep reinforcement learning,” in ICML, 2016.
- [41] O. Vinyals, T. Ewalds, S. Bartunov, P. Georgiev, A. Vezhnevets, M. Yeo, A. Makhzani, H. Küttler, J. Agapiou, J. Schrittwieser, J. Quan, S. Gaffney, S. Petersen, K. Simonyan, T. Schaul, H. V. Hasselt, D. Silver, T. Lillicrap, K. Calderone, P. Keet, A. Brunasso, D. Lawrence, A. Ekermo, J. Repp, and R. Tsing, “Starcraft II: A new challenge for reinforcement learning,” ArXiv, vol. abs/1708.04782, 2017.
- [42] T. Haarnoja, A. Zhou, P. Abbeel, and S. Levine, “Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor,” in ICML, 2018.
- [43] V. Nguyen, P. Obermeier, T. C. Son, T. Schaub, and W. Yeoh, “Generalized target assignment and path finding using answer set programming,” in IJCAI, 2017.
- [44] G. Wagner and H. Choset, “Subdimensional expansion for multirobot path planning,” Artificial Intelligence, vol. 219, pp. 1–24, 2015.
- [45] L. V. D. Maaten, “Accelerating t-sne using tree-based algorithms,” J. Mach. Learn. Res., vol. 15, pp. 3221–3245, 2014.