Greedy UnMixing for Q-Learning in Multi-Agent Reinforcement Learning
Abstract
This paper introduces Greedy UnMix (GUM) for cooperative multi-agent reinforcement learning (MARL). Greedy UnMix aims to avoid scenarios where MARL methods fail due to overestimation of values as part of the large joint state-action space. It aims to address this through a conservative Q-learning approach through restricting the state-marginal in the dataset to avoid unobserved joint state action spaces, whilst concurrently attempting to unmix or simplify the problem space under the centralized training with decentralized execution paradigm. We demonstrate the adherence to Q-function lower bounds in the Q-learning for MARL scenarios, and demonstrate superior performance to existing Q-learning MARL approaches as well as more general MARL algorithms over a set of benchmark MARL tasks, despite its relative simplicity compared with state-of-the-art approaches.
1 Introduction
In multi-agent reinforcement learning, the nature of the joint state action space leads to complexity in estimation, as the size of the joint action spaces grow exponentially with the number of agents. In order to constrain and learn policies effectively, we leverage approaches within offline reinforcement learning, which is used to avoid overestimation of values induced by distribution shift between the learned policy and experiences of the multiple agents. In order to scale out reinforcement learning, especially in adherence to constraints on agent observability and communication in multi-agent environments, decentralized policies are used so that agents may only act on their local observations. Previous work, and this work, leverage centralized training with decentralized execution (CTDE) [19, 26, 21, 22, 14], where independent agents can access additional state information, (including other agents observation and actions) that is unavailable during policy inference.
However the key question remains - is there a way to determine the usefulness of this additional state information? Is there a way to effectively “unmix” this additional state information, removing the need to learn these interactions in the first place? We propose a method for concurrently simplifying the problem space, and learning a conservative estimate for the joint value function, which allows for practical simplification of the joint state action space in MARL settings. We demonstrate this idea in a novel Q-learning algorithm called Greedy UnMix (GUM). First, GUM leverages conservative estimation attempts to avoid poor performance due to issues with bootstrapping out of distribution actions [10, 11] and overfitting, which arises from erroneous optimistic value function estimates. In order to achieve this, conservative estimates of the value function could be used to address the overestimation problem. In our experiments, we demonstrate empirically the importance of providing a conservative estimate on the decentralized policies and providing a conservative estimate on the mixer network is required in order to avoid over optimistic estimations. Furthermore, adaptive control of the regularization effect in the conservative estimate between the mixing and independent policies is necessary in order to effectively unmix and ensemble the Q-learning estimates to stabilize learning. We evaluate GUM over several benchmark multi-agent environments, and find that it achieves superior results to other Q-learning MARL approaches, and is competitive with state of the art MARL approaches, whilst maintaining the simplicity of Q-learning MARL approaches.
2 Preliminaries
2.0.1 Dec-POMDP:
We assume a Dec-POMDP for agents , where is the state space of the environment [16]. At each timestep, , every agent , which chooses an action which forms the joint action . At each time step , the next state is drawn from state transition function . A transition yields a collaborative reward , and denotes the discount factor. We consider partially observable settings, where each agent does not have full state, but samples observations according to observation function . We denote the action observation history for an agent to be , on which it conditions its stochastic policy . The joint policy is based on the action-value function: , where which is the discounted reward. The goal of the problem is to find the optimal action value function . Q-learning methods train the Q-function by iteratively applying the Bellman optimality operator . Actor-critic methods alternate between computing through iterating the Bellman operator, , where is the transition matrix coupled with the policy, and improving the policy by updating it towards actions that maximize the expected Q-value.
2.0.2 Centralized Training Decentralized Execution (CTDE):
In this setting we further consider the scenario where training is centralized, and execution is decentralized; that is, during training, the learning algorithm has access to the action-observation histories of all agents and the full state, however each agent can only condition on its own local action-observation history during decentralized execution.
2.0.3 Overcoming Bootstrap Accumulation Error in Reinforcement Learning:
Overcoming bootstrap accumulation error is a key focus in offline learning (or batch reinforcement learning), whereby the temporal difference leverages a bootstrapped target network, which leads to overly optimistic Q function estimations when attempting to extrapolate unseen or out of distribution actions. These approaches leverage large datasets of experience in order to train policies without active interaction with the environment. This differs from imitation learning, which assumes optimal data, as the dataset of experiences may be sub-optimal or even random! Approaches to resolve this avoid actions which lie outside the observed data distribution [10, 6], or modeling the distribution of actions which can reduce overestimation through providing a lower-bound estimate [11, 5, 4]. Ensemble approaches have also been explored for stabilising Q-functions [6, 1].
2.0.4 Multi-Agent Reinforcement Learning:
Within the MARL approaches which focus on the CTDE paradigm, one approach is independent Q-learning [23] whereby each agent consists of its own Q-learning policy. Approaches which leverage joint state space are mostly composed through the construction of the Independent Global Max problem [21], whereby a Q-learner with parameter sharing is learned for each agent, and then composed with a mixing function. This alters the Q-learning problem with an objective to leverage a mixer function. In value decomposition approach [22], each individual value function is combined in a linear fashion; QMIX [19] combines the value functions through a neural network with monotonic constraints, and QTRAN demonstrates how arbitrary value factorization can occur, leading to computationally intractable constraints in learning the joint state action spaces in the MARL setting.
Another approach to MARL problems is through a policy gradient framework or actor critic framework. These typically operate through each agent having its own policy, but the critic may be composed in different ways. One approach is for each agent to have its own policy and critic network, in which experience is shared across multiple agents [2]; in the deterministic policy gradient setting, the critic is composed through constructing a joint state-action space [14]. The use of a mixer network similar to QMIX has been explored in which a mixer network is used as a critic with adaptive entropy regularization [26].
3 Methodology
In this section we develop Greedy UnMixing (GUM) algorithm, such that the expected value of the policy under the learned Q-function lower-bounds its true value. Naively we could use independent Q-learning [23] approach and apply it to conservative Q-learning approach. We further expand this through applying lower-bound to the joint Q-learning mixer function in the multi-agent setting and provide an approach to unmix the joint Q-learning and independent Q-learning procedure.
3.1 Independent Conservative Q-Learning
Conservative Q-learning can be used in off-policy scenario, and is formulated through the addition of a conservative policy regularizer which uses quantile regression Q-learning as the base algorithm so that a distribution of values can be learned to calculate the action distribution [4]. The CQL framework can be applied in MARL framework in the same way as how DQN is applied in IQL algorithm - that is, each agent learns its own independent Conservative Q-learning policy conditioned on its own partial observation [23]. It operates through the addition of a penalty which minimizes the expected Q-value under a particular distribution of state-action pairs (can be constructed through the restricting to match the state-marginal in the dataset), which can be used to match the state-marginal in the dataset, denoted by . We define the dataset by , so that on all states let denote the empirical behavior policy at the partially observed state. This is further improved through the addition of Q-value maximization term , with a trade-off factor for the penalty . This iterative update per each individual agent is shown below based on the CQL algorithm [11]:
(1) |
As each agent leverages CQL independently, then each agent conservatively estimates Q-values independently from each other, which allows for theoretical guarantees to hold. This leads to the natural question; under what conditions the appropriate value function factorization would still retain conservative Q-value guarantees which will provide the lower-bounds of the policy?
3.2 Conservative Monotonic Value Function Factorization
QMIX leverages monotonic value function factorization, in which the mixer hypernetwork consists of monotonic transformations of each agent Q function, . The naive approach is to regularize with respect to the whole hypernetwork, in a similar fashion as the Conservative IQL approach in the previous Section 3.1, the iterative update, where the full joint state is , and the joint action is :
(2) |
This approach would regularize the weights in both the mixer hypernetwork and the individual agent policy networks.
However, this is not entirely necessary, as regularizing only the individual agent policies would also provide a lower bound on the Q function estimation; the effect of this will be regularising the policy weights but not the weights in the monotonic hypernetwork. As the mixer network is not used directly in the execution of the policy, it only affects the weights updating in the training step. We will explore the effect of regularizing the mixer network, or the policy network, or both concurrently in the ablation studies.
In consideration of the independent global max formulation, to guarantee that the condition holds, we similarly apply the CQL regularizer to the function. Replacing with the mixer function would adhere to the theoretical considerations due to the monotonic nature of QMIX.
3.3 Greedy UnMixing
The construction of Greedy UnMixing (GUM) takes into account the manner in which conservative QMIX is constructed, with both regularizer on the mixer network and individual network. In determining the updates of the mixer network or the individual policy networks, we consider the nature in which they are updated. Can we bias the weights towards considering individual information only and purposely neglect the coordination aspect of the mixer network, thereby “unmixing” the network?
In order to unmix, we use a learnable weight to control the regularization strength for the independent and mixer networks. Intuitively, the “better” the estimate the stronger the strength of the regularizer. In order to construct this, we apply a soft constraint on the regularizer trade-off factor, such that the larger the relative regularizer strength, the smaller its corresponding component has on the neural network weights update. To simplify notation, we represent the regularizer term from Equation 1 to be and regularizer term from Equation 2 to be , with the trade-off factors and respectively and is diagrammatically shown in Figure 1. Then the iterative update is provided as
(3) |

We explored GUM, in which regularizer trade-off weights were used to implicitly to unmix the cooperative and independent Q functions. Next we explore the effects for explicitly unmixing the two components through an ensemble interpretation of GUM as a neural network architecture selection problem.
3.4 Ensemble Learning Interpretation of Greedy UnMix
The construction and choice of neural network architectures have been explored as a boosting problem with learning guarantees. AdaNet framework explores neural architecture selection through iteratively proposing candidate networks with hard selection proposals [3], or soft selection proposals through application of a learnable trade-off factor [20]. Through the learnable trade-off , the GUM algorithm can be interpreted as a neural network architecture selection problem at training time between the mixer network and individual agent networks. This ensemble can be applied implicitly, as suggested in Section 3.3 or explicitly as a stacking ensemble [25] of the individual agent policies and the mixer function. Rather than leveraging a learned parameter for neural network selection problem, an alternative approach is to simply take the minimum ensemble as the solution, which leads to a hard assignment in the choice of the Q function ensemble rather than a soft assignment. This approach is a variety of ensemble Q-learning approaches [13, 6, 10], which is used to stabilize training in the context of homogeneous target networks, as opposed to this scenario where it is between the individual agent policies and the mixer network. Then the Adanet mixer is a linear combination of the individual agent Q function, and the monotonic function factorization, , where and are learnable parameters by gradient ascent. The iterative update is given by
(4) |
Another interpretation of the AdaNet neural network selection setting, is that the base learner is the independent Q-learning policies, which are then boosted (ensemble) from the subsequent mixer network. This interpretation is shown in Figure 2 and also addresses the overestimation concerns in reinforcement learning, through penalizing the additional weight inversely relative to the regularizer strength. There have been several approaches leveraging ensemble of Q functions which assume independence and their analysis results would not apply here either [6, 10]. In the boosting setting, our approach can be interpreted as a variation of AdaNet [3], in the bushy construction, whereby the independent Q-learning model is the base policy, and the mixer is the boosted network module. Rather than having hard update or stopping rule, we leverage a soft choice as a shrinkage parameter [20]. In this setting, the networks are explicitly combined rather than implicitly through the regularizer.
We present the two variations of GUM, one which allows for implicit unmixing through learning the trade-off factors through an inverse weighting scheme, and the other approach in which the combination of the mixer and independent policies are explicitly learned, summarized in algorithm 1.
3.4.1 Theoretical Considerations of GUM.
GUM is an extension of conservative Q-learning, and in the implicit regularizer formulation, extends the theoretical results from [8, 17, 11] to multi-agent setting which are related to prior work on safe policy improvement [12, 18].
Theorem 3.1
Proof: We can set the modified objective in Equation 3 and compute the Q-function update induced in the exact tabular setting. Let denote an iteration of policy evaluation, so that at iteration the objective - Equation 3 - is optimized. To demonstrate that Equation 3 results in a lower bound estimate of the true policy, we consider the value function update:
For any particular agent, if , we first consider considering (dropping the for brevity) , we have the following derivation which follows the result from CQL Theorem 3.2 [11]:
Then is non-negative, as each independent agent regularizer is additive. A similar argument is applied for the joint state action mixer component .
As and is non-negative, then each value iterate of GUM algorithm incurs some underestimation . This result can be generalized to incorporate sampling error in the non-tabular setting in the actor-critic setting based on prior work [11].

4 Experiments
![]() |
![]() |
![]() |
![]() |
In this section we describe experiments we use to empirically justify our algorithm and approach. We use several reinforcement learning benchmarks [24, 14, 7], shown in Figure 3.
Cooperative Pong is a multi-agent game of pong from “Butterfly” environments [24], where the goal is for both agents (paddles) to keep the ball in play. The game is over if the ball goes out of bounds from either the left or right edge of the screen. In this setting the action space is discrete, and the paddles can either move up or down. To make learning more challenging the right paddle is tiered cake-shaped by default, and the observations of each agent are its own half of the screen. The agents receive a positive, fixed, combined reward based on successful completion of a game frame, otherwise a negative reward is given, and the game ends.
Pistonball is a physics based cooperative game from “Butterfly” environments [24] where the goal is move the a ball to the left wall of the game border by activating vertically moving pistons. Each piston agent’s observation is an RGB image of is local surrounding area. The agents have a discrete action space, and can choose to stay still, move down or move up. Pistons must learn highly coordinated emergent behavior to achieve an optimal policy for the environment. Each agent gets a reward based on the overall movement of the ball, and how much the ball moved in a left-wards direction if it was in the agent’s local observation.
Pursuit is an environment from the “SISL” set of environments [7], where our agents are control of pursuer agents, which aim to capture randomly controlled evader agents. Evader agents are only removed from the environment when pursuers fully surround an evader. Agents receive a small reward when they touch an evader, and a large reward when they successfully surround and remove an evader agent. In this scenario, each pursuer agent observes a local RGB grid, and operate in a discrete action space.
Waterworld is a simulation of archea trying to navigate and surviving their environment from the “SISL” set of environments [7]. The goal of the agent is to control the archea to simultaneously avoid poison and find food. The original environment consists of a continuous action space, indicating their horizontal and vertical thrust within the environment. We instead modify this to a discrete action space, in order to assess Q-learning approaches. Within this environment, the agent’s observation is based on range limited sensors to detect neighboring entities, indicated by black lines in the right-most image in Figure 3.
Reference and Spread are environments consisting of agents and landmarks from the “MPE” set of environments [14]. In Reference, each agent wants to get closer to their target landmark, which is known only by other agents. The agents are rewarded by the distance to their target landmark. In Spread Agents must learn to cover each landmark without colliding with other agents. The agents are rewarded based on how far the closest agent is to each landmark, and are penalized if they collide with other agents. For both environments, the agent’s observation space consists of distance to every landmark, and the action space consists of the movement action. In addition the Reference environment has an additional actions to communicate with other agents.
4.1 Architecture Overview and Implementation Details
Across all environments, we consistently preprocessed the inputs as suggested in the PettingZoo benchmarks [24]. We evaluated on 1000 evaluation episode rollouts (separate from the train distributions) every training iteration and used the average score and variance for the plots and tables, shown in Figure 5
4.2 Hyper-parameters
Hyperparameters | Value |
---|---|
Policy Hidden Sizes | [64, 64, 64] |
Mixer Hidden Sizes | [32, 32, 32] |
Policy Hidden Activation | ReLU |
Target Network | 0.005 |
Learning Rate | 0.0003 |
Batch Size | 256 |
Replay Buffer Size | 1000000 |
Number of pretraining steps | 1000 |
Steps per Iteration | 1000 |
Discount | 0.99 |
Reward Scale | 1 |
The policies leverage the default hyper-parameters based on the official implementation of LICA, QMIX, IQL and their accompanying code based on the pymarl library [19, 26, 21]. SEAC was re-implemented in rlkit framework with the same hyperparameters proposed in their paper, but using Soft Actor-Critic instead of A2C. Similarly, MADDPG extended the rlkit implementation of DDPG to support multi-agent setting. In order to facilitate discrete action spaces gumbel softmax [9, 15] is used as proposed by the approach taken by by SEAC [2]. Similar to the usage in the pymarl library, parameter sharing was used in all models in order to improve the general performance of the agents.
Our particular implementation of GUM leverages the existing approach of Conservative Q-learning, which regularizes through an entropy interpretation of the off-policy CQL regularizer, which is framed as
where the distribution over data is modeled after the previous policy, as suggested and used in the official CQL implementation [11].
4.3 Ablations
![]() ![]() |
In this section, we explore the effects of different variations of our algorithm on the performance of the agent in order to determine which algorithm generalizes best. We list the “GUM” algorithm as described in Section 3.3, where the trade-off factor for the regularizer is fixed, “GUM-lagrange” to be the GUM algorithm where the trade-off factor for the regularizer is learned, “GUM-adanet” to use an explicit stacking with a fixed trade-off factor for the CQL regularizer. We contrast our approach CQMIX as described in Section 3.2, which does not attempt to unmix the independent Q function and the mixing function, with a fixed trade-off factor for the regularizer and learned trade-off factor in algorithms labeled “CQMIX” and “CQMIX-lagrange” respectively. Finally to demonstrate the effectiveness of the algorithm is not related to the ensembling of agent Q policies with the mixing function, we use ensembling schemes involving the minimum or mean used in other RL approaches [13, 6, 10] labeled “QMIX-min” and “QMIX-mean” respectively.
From the results, we observe that the GUM family of algorithms are generally the best performing algorithms. Surprisingly, explicitly stacking the networks together has strong performance in half of the environments (pistonball, pong, pursuit) but is inferior to a naive application of CQL to QMIX for the other environments (reference, spread, waterworld). For the other algorithms, it is clear that the naive application of CQL to QMIX and ensembling of QMIX algorithms is inferior to our GUM algorithm across a wide range of MARL tasks. This justifies our neural architecture choices and construction empirically. We hypothesize this is due to the dependent nature of the mixer hypernetwork and the independent Q-learning networks. In the same way target networks or Polyak averaging is used to minimise the update rate across the networks. Similar approaches can be used to learn the ensemble weighting, which appears to lead to degenerative cases. The inverse weighting scheme explicitly tempers the weights directly, based on the greedy weighting scheme. We leave exploring variations of avoiding this, as an exercise for future, perhaps through this mechanism, the networks may learn to unmix themselves without the need for inverse weighing scheme.
4.4 Experimental Results
![]() ![]() |
We demonstrate that GUM is comparable to existing state of the art approaches, and performs well against models which use cooperation [19] and entropy exploration [26, 2]. The GUM models leverage “GUM-lagrange” where the trade-off factors , are learned implicitly.
From the charts in Figure 5, our approach is clearly superior or comparable with existing state of the art approaches in both Q-learning and actor-critic approaches for multi-agent reinforcement learning across a range of benchmark MARL tasks. The ability to train and effectively unmix the agent and the mixer network in a CTDE framework enables efficient training and representation in the GUM framework and suggests that GUM can effectively capture the nuances between independent agents and cooperative behavior. This is best observed when compared with LICA which captures implicit credit assignment in the actor-critic framework [26]. This suggests that our approach is competitive with actor-critic approaches in MARL whist maintaining the simplicity in neural network architecture and training as Q-learning approaches which may allow GUM to scale to large-scale problems effectively.
5 Conclusion and Future Work
In this paper we explored Greedy UnMix (GUM), which is a dynamic ensemble technique for MARL in CTDE framework. We have explored and justified the construction of our network empirically through our experiments, and theoretically through linkages with frameworks for neural network architecture search. Being able to learn the effects of joint or individual policies is essential in learning underlying concepts effectively in cooperative multi-agent RL problems. GUM simplifies the joint state action space in the multi-agent setting through learning Q functions in a conservative fashion, and through weighing the trade-off between independent policies and mixer networks either implicitly or explicitly which leads to more purposeful coordination. Promising empirical results on a wide range of MARL benchmark environments and the simplicity of MARL Q-learning frameworks demonstrate the effectiveness of the algorithm in practice. For future work, we aim to extend this approach to the actor-critic class of algorithms and explore further theoretical properties of our architecture selection process in order to address effective ways to train MARL algorithms.
References
- [1] Rishabh Agarwal, Dale Schuurmans, and Mohammad Norouzi. An optimistic perspective on offline reinforcement learning. In International Conference on Machine Learning, 2020.
- [2] Filippos Christianos, Lukas Schäfer, and Stefano V Albrecht. Shared experience actor-critic for multi-agent reinforcement learning. Advances in Neural Information Processing Systems, 2020.
- [3] Corinna Cortes, Xavier Gonzalvo, Vitaly Kuznetsov, Mehryar Mohri, and Scott Yang. AdaNet: Adaptive structural learning of artificial neural networks. In International Conference on Machine Learning, 2017.
- [4] Will Dabney, Mark Rowland, Marc Bellemare, and Rémi Munos. Distributional reinforcement learning with quantile regression. In Proceedings of the AAAI Conference on Artificial Intelligence, 2018.
- [5] Scott Fujimoto, Edoardo Conti, Mohammad Ghavamzadeh, and Joelle Pineau. Benchmarking batch deep reinforcement learning algorithms. arXiv preprint arXiv:1910.01708, 2019.
- [6] Scott Fujimoto, David Meger, and Doina Precup. Off-policy deep reinforcement learning without exploration. In International Conference on Machine Learning, pages 2052–2062, 2019.
- [7] Jayesh K Gupta, Maxim Egorov, and Mykel Kochenderfer. Cooperative multi-agent control using deep reinforcement learning. In International Conference on Autonomous Agents and Multiagent Systems, pages 66–83. Springer, 2017.
- [8] Thomas Jaksch, Ronald Ortner, and Peter Auer. Near-optimal regret bounds for reinforcement learning. J. Mach. Learn. Res., 11:1563–1600, August 2010.
- [9] Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-softmax. In International Conference on Learning Representations, 2018.
- [10] Aviral Kumar, Justin Fu, Matthew Soh, George Tucker, and Sergey Levine. Stabilizing off-policy q-learning via bootstrapping error reduction. In Advances in Neural Information Processing Systems, pages 11784–11794, 2019.
- [11] Aviral Kumar, Aurick Zhou, George Tucker, and Sergey Levine. Conservative q-learning for offline reinforcement learning. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 1179–1191, 2020.
- [12] Romain Laroche, Paul Trichelair, and Remi Tachet Des Combes. Safe policy improvement with baseline bootstrapping. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 3652–3661. PMLR, 09–15 Jun 2019.
- [13] Timothy P Lillicrap, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. Continuous control with deep reinforcement learning. In International Conference on Learning Representations, 2016.
- [14] Ryan Lowe, Yi Wu, Aviv Tamar, Jean Harb, OpenAI Pieter Abbeel, and Igor Mordatch. Multi-agent actor-critic for mixed cooperative-competitive environments. In Advances in Neural Information Processing Systems, pages 6379–6390, 2017.
- [15] Chris J. Maddison, Andriy Mnih, and Yee Whye Teh. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables. In International Conference on Learning Representations, 2018.
- [16] Frans A. Oliehoek and Christopher Amato. A Concise Introduction to Decentralized POMDPs. Springer Publishing Company, Incorporated, 1st edition, 2016.
- [17] Ian Osband and Benjamin Van Roy. Why is posterior sampling better than optimism for reinforcement learning? In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 2701–2710, International Convention Centre, Sydney, Australia, 06–11 Aug 2017. PMLR.
- [18] Marek Petrik, Mohammad Ghavamzadeh, and Yinlam Chow. Safe policy improvement by minimizing robust baseline regret. In Proceedings of the 30th International Conference on Neural Information Processing Systems, NIPS’16, page 2306–2314, Red Hook, NY, USA, 2016. Curran Associates Inc.
- [19] Tabish Rashid, Mikayel Samvelyan, Christian Schroeder De Witt, Gregory Farquhar, Jakob Foerster, and Shimon Whiteson. Qmix: monotonic value function factorisation for deep multi-agent reinforcement learning. International Conference on Machine Learning, 2018.
- [20] Chapman Siu. Residual networks behave like boosting algorithms. In 2019 IEEE International Conference on Data Science and Advanced Analytics (DSAA), pages 31–40. IEEE, 2019.
- [21] Kyunghwan Son, Daewoo Kim, Wan Ju Kang, David Earl Hostallero, and Yung Yi. Qtran: Learning to factorize with transformation for cooperative multi-agent reinforcement learning. International Conference on Machine Learning, 2019.
- [22] Peter Sunehag, Guy Lever, Audrunas Gruslys, Wojciech Marian Czarnecki, Vinicius Zambaldi, Max Jaderberg, Marc Lanctot, Nicolas Sonnerat, Joel Z Leibo, Karl Tuyls, et al. Value-decomposition networks for cooperative multi-agent learning based on team reward. International Foundation for Autonomous Agents and Multiagent Systems, pages 2085–2087, 2018.
- [23] Ming Tan. Multi-agent reinforcement learning: independent versus cooperative agents. In Proceedings of the Tenth International Conference on International Conference on Machine Learning, pages 330–337. Morgan Kaufmann Publishers Inc., 1993.
- [24] Justin K Terry, Benjamin Black, Mario Jayakumar, Ananth Hari, Luis Santos, Clemens Dieffendahl, Niall L Williams, Yashas Lokesh, Ryan Sullivan, Caroline Horsch, and Praveen Ravi. Pettingzoo: Gym for multi-agent reinforcement learning. arXiv preprint arXiv:2009.14471, 2020.
- [25] David H. Wolpert. Stacked generalization. Neural Networks, 5(2):241–259, 1992.
- [26] Meng Zhou, Ziyu Liu, Pengwei Sui, Yixuan Li, and Yuk Ying Chung. Learning implicit credit assignment for multi-agent actor-critic. Advances in Neural Information Processing Systems, 2020.