Phasic Diversity Optimization for Population-Based Reinforcement Learning
Abstract
Reviewing the previous work of diversity Reinforcement Learning, diversity is often obtained via an augmented loss function, which requires a balance between reward and diversity. Generally, diversity optimization algorithms use Multi-armed Bandits algorithms to select the coefficient in the pre-defined space. However, the dynamic distribution of reward signals for MABs or the conflict between quality and diversity limits the performance of these methods. We introduce the Phasic Diversity Optimization (PDO) algorithm, a Population-Based Training framework that separates reward and diversity training into distinct phases instead of optimizing a multi-objective function. In the auxiliary phase, agents with poor performance diversified via determinants will not replace the better agents in the archive. The decoupling of reward and diversity allows us to use an aggressive diversity optimization in the auxiliary phase without performance degradation. Furthermore, we construct a dogfight scenario for aerial agents to demonstrate the practicality of the PDO algorithm. We introduce two implementations of PDO archive and conduct tests in the newly proposed adversarial dogfight and MuJoCo simulations. The results show that our proposed algorithm achieves better performance than baselines.
I INTRODUCTION
Deep Reinforcement Learning (RL) has been successfully applied to robotic grasping, obstacle avoidance, navigation, and automatic driving. The essence of RL is the exploitation of known knowledge and the exploration of unknown environments. Whereas exploitation refers to policy improvement from environmental interaction, exploration refers to the actions sampled from the conditional distribution.
Recent studies have shown great interest in achieving precise aerial tracking and maneuver avoidance through RL. Agents need enough exploration to escape from sub-optimal policies. ”Natural selection and survival of the fittest”. The Population-Based Training (PBT, [1]) provides an elegant way to balance exploration and exploitation, that is, to train agents of a set of diverse behaviors asynchronously and accelerate training by copying the information of the elite in the population. However, the expensive cost of large-scale distributed training [2] accounts for the limitation of the number of learners. On the other hand, the small-scale training or inappropriate exploitation frequency leads to rapid convergence to local optima. Therefore, maintaining the diversity of the small-scale population becomes popular, and modeling diversity in a unified and principled way is still an open issue in the field of RL [3, 4]. For example, the diversity of most algorithms is obtained from the expectation of some distance (or divergence) between pairwise policies regardless of the differences in norms between different pairwise similarities.
The determinantal point process (DPP, [5, 6]) provides a solution for efficient diversity modeling and prevents the clustering phenomenon by a high-order joint optimization. The diversity is derived from a determinant of a positive semidefinite (PSD) matrix that defines a global measure of similarity between pairs of policies. And the matrices should be full rank for effective exploration. If the kernel matrix is differentiable, we can make the behaviors of agents repel each other by maximizing probabilities characterized by the determinant.
Determinants show great potential in modeling diversity. But in the face of real-world problems, there are still many obstacles that need to be overcome to improve population diversity. A minor problem is the high computational complexity of the determinants. The determinant may still be zero when exploitation occurs in population although it comes from a PSD matrix. A more urgent problem is to find a balance between reward and diversity. The mainstream algorithms [6, 7] regard the adaptive exploration as an online learning problem, and reduce the manual process through Multi-armed Bandits algorithms (MABs) [8, 9]. Unfortunately, the so-called equilibrium of reward and diversity may not exist due to the complex environmental dynamics or the non-stationary distribution of reward signals for MABs in the training process. the conflict between reward and diversity makes the MABs unable to eliminate the impact of exploration on performance.
In this paper, we mainly focus on small-scale population-based RL and address the above problems. We consider several existing probability measures as kernels to characterize the similarities between pairwise stochastic policies and construct the matrices of DPPs. To satisfy the restrictive constraint on Cholesky decomposition, we advocate the use of a regularization to improve population diversity so that the optimization can also be performed when the original matrix is singular, which is common in PBT. To avoid the conflict between reward and diversity, we introduce an archive and an auxiliary phase into the PBT framework. In our proposed Phasic Diversity Optimization (PDO) algorithm, the archive can be a MAP-Elites[10] grid or a fitness priority queue without a Behavior Descriptor (BD). In the auxiliary phase, we diversify the potential agents from the archive to fill the blank area of the grid or eliminate the worst agents in the archive. To demonstrate the practicality of the proposed algorithm, we introduce the task for a dogfight agent, considering modeling the problem as a Markov Decision Process (MDP) for end-to-end training and evaluation. Experiments on different tasks show that PDO can diversify agents without harming reward.
II RELATED WORK
Diversity RL. Previous works[11, 12, 13, 14] have fully discussed the importance of population diversity from different perspectives. DIAYN[11] updates the skill of agents to visit diverse states. Reward-Randomized PG[12] perturbs the reward space of the original environment. TrajeDi[13] and MEP[14] generalize the JSD and generates diverse policies through trajectory-based objectives. Diversity-Inducing PG [15] and Reward-Switching PG [16] obtain the policies of different behaviors in incremental ways. The above methods are proved to be effective in where diversity can be distinguished by different states, reward functions, and action distributions.
High-performance diversity. A common way to encourage exploration is to maximize an entropy regularization[17, 18] which increases the variance of action distribution and are not directly related to the improvement of performance with increasing the complexity of convergence. The idea of Diversity-Driven Exploration Strategy (Div, [19]) shares a common intuition with the P3S[20], where both exploitation and exploration may improve performance. DvD[6] and EDO-CS[7], as typical applications of adaptive exploration, using MABs, may still cause long-term performance damage due to complex reasons. The parameters of the above methods need to be fine-tuned.
Quality-Diversity (QD) RL. QD is a hot topic in Evolutionary Computation, in which MAP-Elites[10] as the typical algorithm using BDs to map agents into the archive. Different cells represent different behaviors, and each cell stores one agent with the best fitness. Compared with RL, the Genetic Algorithm can more effectively diversify agents but have less sample efficiency. Therefore, algorithms such as QD-PG [21] and PGA-MAP-Elites [2] combining QD and TD3[22] are becoming mainstream. In contrast, our PDO uses joint optimization and can deploy any on-policy algorithm.
Air Combat Based on Deep RL. Recently, Deep RL has been applied to autonomously perform air combat [23, 24, 25, 26, 27]. [23] trains an agent in a custom three-dimensional environment that takes actions from a set of 15 discrete maneuvers and achieves victory over human opponents. [24] reaches its goal in an optimal path using TD3[22] + HER[28]. [25] presents a state stacking method for noisy RL environments as a noise reduction. And [26, 27] apply RL methods to air combat, but not take into account the population diversity.
III PRELIMINARIES
III-A Stochastic Policy Gradient
We consider RL in a MDP denoted as a tuple , where is the state space, is the action space, is the state transition probability, is the reward function and is the discount factor. A stochastic policy , parameterized by a vector , assigns a probability of an action given a state. The policy’s goal is to maximize the accumulated reward by maximizing the expected discounted return:
(1) |
where trajectory is generated by sampling actions according to the policy and states according to the dynamics .
III-B Diversity via Determinants
Determinantal point processes are probabilistic models of repulsion where distributions of diverse sets are characterized by determinants. From a geometric point of view, the determinant measures the directed volume of the vectors in the hyperplane space.
The Diversity via Determinants (DvD, [6]) algorithm maximizes the volume to boost exploration. Formally, the population diversity parameterized by is defined as:
(2) |
where is the dimension of embedding, is a similarity kernel function, and is the policy embedding [29]. If the entries of the kernel matrix are measurements of similarity between pairs of policies, the similar policies are unlikely to appear together. And DvD encourages diversity by introducing an explicit diversity criterion:
(3) |
where is the trade-off between reward and diversity. The is selected by MABs, which maintain the posterior distribution of cumulative reward over arms. Specifically, each arm represents a coefficient , and these arms can be selected by Thompson sampling (TS, [8]) or Upper Confidence Bound (UCB, [9]) algorithm to maximize reward in limited sampling times. The reward signal for MABs comes from a Bernoulli distribution of whether the performance of the best agent has been improved after optimization with corresponding coefficients. Finally, the arm with the higher reward will be selected more times and helps improve population performance.
IV PHASIC DIVERSITY OPTIMIZATION ALGORITHM
IV-A Two-phase Opimization
Several motivations drive us to propose the Phasic Diversity Optimization (PDO) algorithm. First, if coefficients come from an inappropriate space, it will affect the performance of the agent for a long time even if MABs abandon pulling these arms during training. Moreover, MABs need a stationary model of reward distribution, which is non-trivial because of the fluctuating returns, skills learned by agents, and even replication dynamics. Second, diversity may conflict with reward. We want to avoid multi-objective optimization, which may both lose quality and fail to diversify. Third, if the diversity criterion for multi-objective optimization is joint loss, additional gradient transfer overhead is required between processes. The serialized training process grows the wall-clock time in practice.
The idea of PDO is simple. It separates the training into two phases. In the first phase, we perform policy iterations and decide whether to store them in the archive based on the evaluation results. In the second (auxiliary) phase, we elect the best agents from the archive, improve their population diversity via determinants, and then evaluate them to decide whether to store them in the archive. The pseudo-code is given in Algorithm 1. Note that the two phases can run in parallel.
The PDO algorithm improves population diversity in a way that is not harmful to reward while using as few learners as possible to diversify agents. It prevents performance degradation by archive backups. If the agents generated in the auxiliary phase are not diverse enough, they will only be compared with their previous selves and will be permanently discarded if they get worse. If the archive has a Behavior Descriptor (BD, denoted ), the agents may be mapped to another cell because of repulsion and compare performance with the elite. The motivation to use BD is weakened by PDO, and if we are not experts in behavior modeling, we can use the reward-priority queues as the storage media for agents, which can be easily implemented with heaps.
Each iteration of the PDO performs two reads of the archive. The first read occurs at the exploitation of the archive, where we uniformly sample an agent from the archive to eliminate the current worst agent, with the aim of selecting targets for the policy iterations. The second read occurs to select the top agents with the aim of selecting target for the diversity iterations. We can also use the clustering-based selection[7] algorithm to select agents for the auxiliary phase, which is different from their original motivation.
IV-B Learning to Score Population Diversity
From the perspective of probability theory, we propose a differentiable method to estimate the similarity between two stochastic policy distributions and form the PSD matrix .
Let denotes the distance metrics of any two stochastic distributions and . The distance between two stochastic policies and defined as:
(4) |
Definition 1 (Differentiable Similarity Estimation).
The similarity between two stochastic policies and can be estimated by the approximate kernel function :
(5) |
where is a differentiable function.
The Differentiable Similarity Estimation (DSE) allows us to sample the visited states from trajectories of mixed average policies, and estimate the similarity by finite sampling instead of the intractable integral in (4). We can easily map the value to by the function if is a symmetric and bounded metric. Then the policy can be updated by applying the chain rule to (5) w.r.t. the parameter :
(6) |
In discrete action spaces, a symmetric and bounded distance between two probability measures and and its corresponding function can be the Jensen–Shannon divergence (JSD): and , where .
However, JSD has no closed-form solution for two Gaussian distributions. In continuous action spaces, we can use the Euclidean norm as the distance function, defined by the -Wasserstein distance (WDs) [30]:
(7) |
where is the set of all couplings (joint distributions) between and on .
Equation (7) is inspired by the optimal transport problem (OTP) where we want to transform from with minimum cost. In the rest of the paper, we only consider the -Wasserstein distance because the closed-form [31] formula is available when and are multi-variate Gaussian distributions. If and , we have:
(8) |
In particular, if the and are uncorrelated multivariate normal distributions, the covariance matrices and are diagonal and (8) can be further simplified as:
(9) |
It only measures the distance between the mean values of two distributions if we ignore the second term in (9). And can be the Radial Basis Function (RBF) kernel with length-scale :
(10) |
which is the same as the kernel function for the deterministic action embedding. Note that the result of WDs is highly correlated with the norm of , so we use variance normalization to eliminate hyperparameter .
IV-C Surrogate Determinant Regularization
We introduce a joint diversity loss based on determinants. Since the entries of the matrix are computed by differentiable approximate kernel function, we can compute the gradient of the determinant directly by automatic differentiation.
Lemma 1.
The gradient of the determinant of positive definite matrix w.r.t. equals:
(11) |
Theorem 1.
Consider a positive semidefinite matrix with , is positive definite, and the Cholesky decomposition of is available as: where is an invertible lower (or upper) triangular matrix.
Note, due to the existence of PBT’s exploitation mechanism, the network weights of a policy may be copied by another. Consequently, some two rows (or columns) in the matrix are linearly correlated, the determinant is zero, and the in (11) may encounter numerical instability (e.g., maximum likelihood estimation). In this case, will degenerate into a PSD matrix which makes the decomposition impossible (or the gradient is zero). To meet the conditions in Theorem 1, we use surrogate matrix :
(12) |
where is the smoothing parameter and is identity matrix.
Lemma 2.
The surrogate determinant does not change the repulsive of the original one. Lemma 2 implies that if all the pairs of different policies in the population are not perfectly similar, and the kernel matrix is positive definite the condition of Theorem 1 is satisfied. Therefore, we derive the determinant from the principal diagonal of the lower triangular matrix via Cholesky decomposition. Then is replaced by a positive definite matrix , and the objective in the auxiliary phase is given by:
(13) |
The gradient of objective in (13) w.r.t. parameter is given by:
(14) |

V SIMULATION RESULTS

V-A 2-Agent Dogfight Environment
In this section, we present a competition scenario that consists of 2 adversarial agents and a simplified aerodynamic three-dimensional world. An agent needs control of 4 volumes (Brake and Throttle, Elevator, Roll, and Rudder) to lock onto the target while evading its lock. The state space contains the position, velocity, and posture information of both agents. We set expert policy for the opponent, which means that the agent needs to learn evasive maneuvers and precision tracking skills in a highly dynamic environment.
We set , and rewards for locking, being locked, and out of bounds respectively. Each episode terminates at: (1) the number of steps exceeds 3000. (2) Any agent is out of bounds. (3) Any agent is locked for more than 1000 steps. Considering the lock requires the target to be within the agent’s line of sight (a 10-degree cone and within 1 km) which is a harsh condition, we use a dense reward at every step during the learning process (exclude from evaluation) for all algorithms. The expert is programmed to control the elevator and rudder in order to minimize the antenna train angle and the distance from the red agent. It decelerate and apply brakes when the aspect angle is below 30 degrees and the distance is less than 3 km. The above action takes a definite value of 0.9, and the remaining values are uniformly sampled from to improve randomness. We consider the following techniques:
-
•
Proximal Policy Optimization (PPO, [32]). The base on-policy algorithm with one learner.
-
•
Population Based Training of Neural Networks (PBT, [1]). It removes the auxiliary stage of the PDO algorithm, which involves parallel learners and periodic exploitation through a performance-based priority queue. This is considered as one of the ablation settings for the PDO algorithm.
-
•
PDO. The two phases are reward optimization and diversity optimization respectively. We perform several diversity iterations in each auxiliary phase.
- •
-
•
DSE-UCB. Similar to DvD, but the trade-off is selected by the Upper Confidence Bound[9].
-
•
EDO-CS. Similar to DSE-UCB, but it selects the agents for policy iterations through clustering-based selection[7].
We employ PPO[32] and use cumulative rewards of the 10-episodes-average to evaluate agents after 25 policy iterations for all PPO variants. Each learner trains 3e6 steps. We set the population size and run with 5 random seeds for all Populations-Based methods, and provide MAP-Elites archives for the PDO, PBT[1], DvD[6], and DSE-UCB[9] algorithms. They uniformly sample the solution from the archives and replace the worst performing agent with model weights, optimizer, rewards and state normalization information after 6e5 steps. In DSE[9], the sample size is 256, and we use the -Wasserstein distance and RBF kernel without the Frobenius norm of the covariance matrix because we use deterministic evaluation. We set for MABs that are the same as those in paper[6, 7]. The number of diversity iterations is 20 in the auxiliary phase. The BD is defined as the average values of the Elevator and Roll controlled by the agent, and the archive size is (10, 10). We visualize the archives from one of the seeds, as shown in Fig. 1.
PBT | DvD | DSE-UCB | EDO-CS | PDO | |
Max Fitness | 968 | 575 | 634 | 682 | 972 |
QD-score | 21032 | 30043 | 35724 | 23731 | 39201 |
Coverage | 20.2 | 33.0 | 40.0 | 23.2 | 50.8 |
We can observe from Table I111Max Fitness refers to the cumulative rewards of best agents. QD-score refers to the total sum of rewards offset by minimum rewards across all agents in the archive. Coverage refers to the total number of agents in the archive. that among the all compared methods, the PDO algorithm exhibits the highest maximum fitness, coverage, and quality diversity scores. In terms of the maximum fitness metric, the PDO algorithm slightly outperformed the PBT[1] algorithm.
A more advanced inquiry would be whether the strategies learned by the PDO are diverse, despite its high coverage. Fig. 2 displays reproducible (under the same random seed and expert policy setting) trajectory segments of the agent piloting the red aircraft and engaging the blue target. The four red trajectories correspond to the agents from the (0,4), (3,5), (5,4), and (5,5) cells in the PDO algorithm archive is shown in Fig. 1. We can observe that the agents in the bright-colored cells of the PDO algorithm-generated MAP-Elites archive have learned effective attacking strategies, while the agents in the dark-colored cells have not yet learned effective attacking strategies. Thus, the agents in different cells exhibit different flight trajectories, indicating their learned behaviors and strategies are diverse.
V-B Performance on MuJoCo Environments

We examine the performance of PDO on standard MuJoCo[33] environments from OpenAI Gym[34]. The goal of agents in MuJoCo is to move forward by controlling the joints. In order to demonstrate the advantage of PDO, we consider the following methods incorporating parallel techniques:
We set the same hyperparameters and PPG for all parallel learners. For all tasks, we strive to make minimal changes to PPG’s hyper-parameters compared to PPO’s [32] paper, and we only reduced the number of data usage of actors. We use learners with 5 random seeds for all Populations-Based methods. We use the 10-episodes-average cumulative rewards to evaluate agents after 10 policy iterations. The capacity of the priority queue is 10. The epoch of diversity iterations is 5 in the auxiliary phase. The other hyper-parameters are the same as described in V-A.
Fig. 3 shows the results of different variants of PPG algorithm on four MuJoCo tasks. We can observe that the population-based algorithm demonstrates significantly better performance compared to the baseline the PPG algorithm. The final cumulative reward of the PDO algorithm surpasses other diversity-enhanced RL techniques in the Ant-v3, Hopper-v3, and Humanoid-v3 environments, while approaching the performance of the best algorithm in the Walker-v3 environment.

V-C Ablation Studies
We conduct ablation experiments on PDO and performed sensitivity analysis on different levels of diversity during the auxiliary phase. Each experiment runs with 5 random seeds and we record the mean and standard deviation. When PDO removes the auxiliary phase, it becomes the PBT algorithm. Fig. 4 illustrates this concept, where Subfig. 3(a)-(d) display the curves of maximum fitness, minimum fitness, coverage range, and quality diversity score for different levels of diversity. The curves represent the mean values, while the shaded areas indicate the standard deviation.
From the four subfigures in Fig. 4, we can conclude that the PDO algorithm significantly outperforms the ablated PBT algorithm in terms of coverage range and QD-score indicators of the archive. Subfig. 4(a) infers that the PDO algorithm does not overlook performance due to excessive diversification. Subfig. 4(b) implies a negative correlation between the quantity of diversity and the minimum fitness, indicating that a higher degree of diversity iterations may result in the learning of less effective strategies.
VI CONCLUSIONS
In this paper, we propose PDO algorithm to achieve the decoupling of reward and diversity. We extend the DvD algorithm to a more general form compatible with PBT. We practically apply the proposed framework to the well-known policy gradient algorithms[32, 35]. Experiments on our dogfight task and MuJoCo show that our simple modification on PBT is effective and brings further improvement to the performance.
ACKNOWLEDGMENT
This work was supported in part by National Key Research and Development Program of China (2022ZD0210500/2021ZD0112400), the National Natural Science Foundation of China under Grants 62102058/62272081/61972067/62332019, the Distinguished Young Scholars Funding of Dalian (No. 2022RJ01), and the open funding of State Key Laboratory of Structural Analysis for Industrial Equipment.
References
- [1] Max Jaderberg, Valentin Dalibard, Simon Osindero, Wojciech M. Czarnecki, Jeff Donahue, Ali Razavi, Oriol Vinyals, Tim Green, Iain Dunning, Karen Simonyan, Chrisantha Fernando, and Koray Kavukcuoglu. Population based training of neural networks. arXiv preprint arXiv:1711.09846, 2017.
- [2] Olle Nilsson and Antoine Cully. Policy gradient assisted map-elites. In Francisco Chicano and Krzysztof Krawiec, editors, GECCO ’21: Genetic and Evolutionary Computation Conference, Lille, France, July 10-14, 2021, pages 866–875. ACM, 2021.
- [3] Yaodong Yang, Jun Luo, Ying Wen, Oliver Slumbers, Daniel Graves, Haitham Bou-Ammar, Jun Wang, and Matthew E. Taylor. Diverse auto-curriculum is critical for successful real-world multiagent learning systems. In AAMAS ’21, pages 51–56, 2021.
- [4] Félix Chalumeau, Raphael Boige, Bryan Lim, Valentin Macé, Maxime Allard, Arthur Flajolet, Antoine Cully, and Thomas Pierrot. Neuroevolution is a competitive alternative to reinforcement learning for skill discovery. arXiv preprint arXiv:2210.03516, 2022.
- [5] Alex Kulesza and Ben Taskar. Determinantal point processes for machine learning. Found. Trends Mach. Learn., 5(2-3):123–286, 2012.
- [6] Jack Parker-Holder, Aldo Pacchiano, Krzysztof M Choromanski, and Stephen J Roberts. Effective diversity in population based reinforcement learning. In Advances in Neural Information Processing Systems, volume 33, pages 18050–18062, 2020.
- [7] Yutong Wang, Ke Xue, and Chao Qian. Evolutionary diversity optimization with clustering-based selection for reinforcement learning. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022.
- [8] William R Thompson. On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika, 25(3-4):285–294, 1933.
- [9] Peter Auer, Nicolò Cesa-Bianchi, and Paul Fischer. Finite-time analysis of the multiarmed bandit problem. Mach. Learn., 47(2-3):235–256, 2002.
- [10] Jean-Baptiste Mouret and Jeff Clune. Illuminating search spaces by mapping elites. arXiv preprint arXiv:1504.04909, 2015.
- [11] Benjamin Eysenbach, Abhishek Gupta, Julian Ibarz, and Sergey Levine. Diversity is all you need: Learning skills without a reward function. In 7th International Conference on Learning Representations, 2019.
- [12] Zhenggang Tang, Chao Yu, Boyuan Chen, Huazhe Xu, Xiaolong Wang, Fei Fang, Simon Shaolei Du, Yu Wang, and Yi Wu. Discovering diverse multi-agent strategic behavior via reward randomization. In 9th International Conference on Learning Representations, 2021.
- [13] Andrei Lupu, Brandon Cui, Hengyuan Hu, and Jakob N. Foerster. Trajectory diversity for zero-shot coordination. In Proceedings of the 38th International Conference on Machine Learning, volume 139, pages 7204–7213. PMLR, 2021.
- [14] Rui Zhao, Jinming Song, Haifeng Hu, Yang Gao, Yi Wu, Zhongqian Sun, and Yang Wei. Maximum entropy population based training for zero-shot human-ai coordination. arXiv preprint arXiv:2112.11701, 2021.
- [15] Muhammad A. Masood and Finale Doshi-Velez. Diversity-inducing policy gradient: Using maximum mean discrepancy to find a set of diverse policies. In Sarit Kraus, editor, Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, pages 5923–5929. ijcai.org, 2019.
- [16] Zihan Zhou, Wei Fu, Bingliang Zhang, and Yi Wu. Continuously discovering novel strategies via reward-switching policy optimization. In International Conference on Learning Representations, 2022.
- [17] Volodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy P. Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. Asynchronous methods for deep reinforcement learning. In Proceedings of the 33nd International Conference on Machine Learning, volume 48, pages 1928–1937, 2016.
- [18] Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the 35th International Conference on Machine Learning, volume 80, pages 1856–1865. PMLR, 2018.
- [19] Zhang-Wei Hong, Tzu-Yun Shann, Shih-Yang Su, Yi-Hsiang Chang, Tsu-Jui Fu, and Chun-Yi Lee. Diversity-driven exploration strategy for deep reinforcement learning. In Advances in Neural Information Processing Systems 31, pages 10510–10521, 2018.
- [20] Whiyoung Jung, Giseung Park, and Youngchul Sung. Population-guided parallel policy search for reinforcement learning. In 8th International Conference on Learning Representations, 2020.
- [21] Thomas Pierrot, Valentin Macé, Félix Chalumeau, Arthur Flajolet, Geoffrey Cideron, Karim Beguir, Antoine Cully, Olivier Sigaud, and Nicolas Perrin-Gilbert. Diversity policy gradient for sample efficient quality-diversity optimization. In Jonathan E. Fieldsend and Markus Wagner, editors, GECCO ’22: Genetic and Evolutionary Computation Conference, Boston, Massachusetts, USA, July 9 - 13, 2022, pages 1075–1083. ACM, 2022.
- [22] Scott Fujimoto, Herke van Hoof, and David Meger. Addressing function approximation error in actor-critic methods. In Proceedings of the 35th International Conference on Machine Learning, volume 80, pages 1582–1591. PMLR, 2018.
- [23] Q. Yang, J. Zhang, G. Shi, J. Hu, and Y. Wu. Maneuver decision of uav in short-range air combat based on deep reinforcement learning. IEEE Access, 8, 2019.
- [24] Muhammed Murat Özbek and Emre Koyuncu. Reinforcement learning based air combat maneuver generation. CoRR, abs/2201.05528, 2022.
- [25] Ahmet Semih Tasbas, Safa Onur Sahin, and Nazim Kemal Ure. Reinforcement learning based self-play and state stacking techniques for noisy air combat environment. CoRR, abs/2303.03068, 2023.
- [26] Haiyin Piao, Yue Han, Hechang Chen, Xuanqi Peng, Songyuan Fan, Yang Sun, Chen Liang, Zhimin Liu, Zhixiao Sun, and Deyun Zhou. Complex relationship graph abstraction for autonomous air combat collaboration: A learning and expert knowledge hybrid approach. Expert Syst. Appl., 215:119285, 2023.
- [27] Zhixiao Sun, Huahua Wu, Yandong Shi, Xiangchao Yu, Yifan Gao, Wenbin Pei, Zhen Yang, Haiyin Piao, and Yaqing Hou. Multi-agent air combat with two-stage graph-attention communication. Neural Comput. Appl., 35(27):19765–19781, 2023.
- [28] Marcin Andrychowicz, Alex Ray Filip Wolski, Jonas Schneider, Rachel Fong, Peter Welinder, Bob McGrew, Josh Tobin, Pieter Abbeel, and Wojciech Zaremba. Hindsight experience replay. 2017.
- [29] Aldo Pacchiano, Jack Parker-Holder, Yunhao Tang, Krzysztof Choromanski, Anna Choromanska, and Michael I. Jordan. Learning to score behaviors for guided policy optimization. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pages 7445–7454. PMLR, 2020.
- [30] Martín Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein GAN. arXiv preprint arXiv:1701.07875, 2017.
- [31] Victor M. Panaretos and Yoav Zemel. Statistical aspects of wasserstein distances. Annual Review of Statistics ant Its Application, 6(1), 2019.
- [32] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- [33] Emanuel Todorov, Tom Erez, and Yuval Tassa. Mujoco: A physics engine for model-based control. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 5026–5033. IEEE, 2012.
- [34] Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. Openai gym. arXiv preprint arXiv:1606.01540, 2016.
- [35] Karl Cobbe, Jacob Hilton, Oleg Klimov, and John Schulman. Phasic policy gradient. In Proceedings of the 38th International Conference on Machine Learning, volume 139, pages 2020–2027. PMLR, 2021.