by
Provably Robust Federated Reinforcement Learning
Abstract.
Federated reinforcement learning (FRL) allows agents to jointly learn a global decision-making policy under the guidance of a central server. While FRL has advantages, its decentralized design makes it prone to poisoning attacks. To mitigate this, Byzantine-robust aggregation techniques tailored for FRL have been introduced. Yet, in our work, we reveal that these current Byzantine-robust techniques are not immune to our newly introduced Normalized attack. Distinct from previous attacks that targeted enlarging the distance of policy updates before and after an attack, our Normalized attack emphasizes on maximizing the angle of deviation between these updates. To counter these threats, we develop an ensemble FRL approach that is provably secure against both known and our newly proposed attacks. Our ensemble method involves training multiple global policies, where each is learnt by a group of agents using any foundational aggregation rule. These well-trained global policies then individually predict the action for a specific test state. The ultimate action is chosen based on a majority vote for discrete action systems or the geometric median for continuous ones. Our experimental results across different settings show that the Normalized attack can greatly disrupt non-ensemble Byzantine-robust methods, and our ensemble approach offers substantial resistance against poisoning attacks.
1. Introduction
Background and Motivation: Reinforcement learning (RL) is a sequential decision-making procedure and can be modeled as a Markov decision process (MDP) (sutton2018reinforcement, ). Specifically, an agent in RL operates by taking actions according to a certain policy within a stochastic environment. The agent earns rewards for its actions and uses these rewards to enhance its policy. The ultimate goal of the agent is to learn the best possible policy by consistently engaging with the environment, aiming to maximize its cumulative rewards over the long term. Despite the advancements in current RL models, they are often data-intensive and face issues due to their limited sample efficiency (dulac2021challenges, ; fan2021fault, ). To tackle this challenge, a straightforward solution might be parallel RL (nair2015massively, ; mnih2016asynchronous, ). In parallel RL, agents share their trajectories with a central server to train a policy. However, this is often impractical due to high communication costs, especially for IoT devices (wang2020federated, ), and prohibited in applications like medical records (liu2020reinforcement, ) due to data sensitivity.
Toward this end, the federated reinforcement learning (FRL) (fan2021fault, ; khodadadian2022federated, ; jin2022federated, ; liu2019lifelong, ; gao2024federated, ; yuan2023federated, ) paradigm has been introduced as a solution to the problems faced by traditional parallel RL methods. In FRL, various agents work together to train a global policy under the guidance of a central server, all while keeping their raw trajectories private. Specifically, during each global training round, the central server shares the current global policy with all agents or a selected group. Agents then refine their local policies using the shared global policy and through interactions with their environment. Subsequently, agents send their local policy updates back to the server. Once the server receives these policy updates from the agents, it employs aggregation rule, like FedAvg (mcmahan2017communication, ), to merge these received policy updates and refine the global policy. Owing to its willingness to respect agents’ privacy, FRL has been widely deployed in real-world systems, such as robotics (kober2013reinforcement, ), autonomous driving (liang2022federated, ), and IoT network (wang2020federated, ).
While FRL has its merits, it is susceptible to poisoning attacks owing to its decentralized nature (fan2021fault, ). Such an attack might involve controlling malicious agents, who may either corrupt their local training trajectories (known as data poisoning attacks (fan2021fault, )), or intentionally send carefully crafted policy updates to the server (known as model poisoning attacks (fang2020local, ; shejwalkar2021manipulating, ; baruch2019little, )), with an aim to manipulate the ultimately learnt global policy. A seemingly direct defense against these poisoning attacks would be to implement existing federated learning (FL) based aggregation rules, such as Trimmed-mean (yin2018byzantine, ) and Median (yin2018byzantine, ), within the FRL context. Nevertheless, as subsequent experimental results will demonstrate, merely extending existing FL-based aggregation rules does not provide a satisfactory defense performance. This is because these rules, originally designed for FL, remain vulnerable to poisoning attacks (fang2020local, ; shejwalkar2021manipulating, ). Within the domain of FRL, a recently introduced Byzantine-robust aggregation rule, FedPG-BR (fan2021fault, ), has demonstrated exceptional robustness against existing advanced poisoning attacks (fang2020local, ; shejwalkar2021manipulating, ).
Our work: In this paper, we propose the first model poisoning attacks to Byzantine-robust FRL. In the attack we propose, the attacker deliberately crafts the policy updates on malicious agents so as to maximize the discrepancy between the aggregated policy updates before and after the attack. While a direct strategy might be to maximize the distance between the aforementioned policy updates (shejwalkar2021manipulating, ), this method only accounts for the magnitude of the aggregated policy update, neglecting its directionality. To address this challenge, we propose the Normalized attack, wherein the attacker strives to maximize the angular deviation between the aggregated policy updates pre and post-attack. Nevertheless, solving the reformulated optimization problem remains challenging, especially given that the existing robust aggregation rules like FedPG-BR (fan2021fault, ) are not differentiable. To tackle this issue, we introduce a two-stage approach to approximate the solution to the optimization problem. Specifically, in the first stage, we determine the optimal direction for the malicious policy updates, and in the second phase, we calculate the optimal magnitude for these malicious policy updates.
We subsequently propose an innovative ensemble FRL method that is provably secure against both existing attacks and our newly proposed Normalized attack. Within our proposed ensemble framework, we first leverage a deterministic method to divide agents into multiple non-overlapping groups by using the hash values of the agents’ IDs. Each group then trains a global policy, employing a foundational aggregation rule such as Median (yin2018byzantine, ) and FedPG-BR (fan2021fault, ), using the agents within its respective group. During the testing phase, given a test state , we deploy the well-trained multiple global policies to predict the action for state . Considering that the action space in FRL may be either discrete or continuous, we apply varying strategies to aggregate these predicted actions accordingly. Specifically, in a discrete action space, we select the action with the highest frequency as the final action. Conversely, if the action space is continuous, the final action is determined by calculating the geometric median (ChenPOMACS17, ) of the predicted actions. We theoretically prove that our proposed ensemble method will consistently predict the same action for the test state before and after attacks, provided that the number of malicious agents is below a certain threshold when the action space is discrete. In the context of a continuous space FRL system, we demonstrate that the distance between actions predicted by our ensemble approach, before and after the attack, is bounded, as long as the number of malicious agents is less than half of the total number of groups.
Our proposed Normalized attack and the proposed ensemble method have been thoroughly evaluated on three RL benchmark datasets. These include two discrete datasets, namely Cart Pole (barto1983neuronlike, ) and Lunar Lander (duan2016benchmarking, ), and one continuous dataset, Inverted Pendulum (barto1983neuronlike, ). We benchmarked against four existing poisoning attacks including Random action attack (fan2021fault, ), Random noise attack (fan2021fault, ), Trim attack (fang2020local, ), and Shejwalkar attack (shejwalkar2021manipulating, ). Furthermore, we employed six foundational aggregation rules for evaluation including FedAvg (mcmahan2017communication, ), Trimmed mean (yin2018byzantine, ), Median (yin2018byzantine, ), geometric median (ChenPOMACS17, ), FLAME (nguyen2022flame, ), and FedPG-BR (fan2021fault, ). Experimental findings illustrate that our proposed Normalized attack can remarkably manipulate non-ensemble-based methods (where a single global policy is learnt using all agents along with a particular foundational aggregation rule). Distinctively, within a non-ensemble context, our Normalized attack stands out as the exclusive poisoning attack that can target the FRL-specific aggregation rule. We further demonstrate that our proposed ensemble method can effectively defend against all considered poisoning attacks, including our Normalized attack. Notably, for all robust foundational aggregation rules, the test reward of our proposed ensemble method, even when under attack, closely mirrors that of the FedAvg in a non-attack scenario. Our main contributions can be summarized as follows:
-
•
We propose the Normalized attack, the first model poisoning attacks tailored to Byzantine-robust FRL.
-
•
We propose an efficient ensemble FRL method that is provably secure against poisoning attacks.
-
•
Comprehensive experiments highlight that our proposed Normalized attack can notably compromise non-ensemble-based robust foundational aggregation rules. Additionally, our proposed ensemble method shows significant capability in defending against both existing and our newly introduced poisoning attacks.
2. Preliminaries and Related Work
2.1. Federated Reinforcement Learning
A federated reinforcement learning (FRL) system (fan2021fault, ; khodadadian2022federated, ; jin2022federated, ) consists of agents and a central server collaborating to train a global policy. Each agent solves a local Markov decision process (MDP) (sutton2018reinforcement, ), defined as , with as the state space, the action space, the transition probability, the reward function, the discount factor, and the initial state distribution. In FRL, agent follows a policy that gives the probability of taking action in state . Through interactions with its environment, the agent generates a trajectory , starting from an initial state drawn from , with as the trajectory length. The cumulative reward is calculated as . Let denote a policy parameterized by , where is the dimension of . The distribution of agent ’s trajectories under is . For simplicity, we will refer to as . Agent evaluates the effectiveness of a policy by solving the following optimization problem:
(1) |
In FRL, the agents collaborate to train a global policy aimed at maximizing the total cumulative discounted reward. Thus, the optimization problem in FRL becomes FRL solves this problem in an iterative manner. Specifically, in each global training round , FRL performs the following three steps:
-
•
Step I: Global policy synchronization. The server distributes the current global policy to all agents or a selection of them.
-
•
Step II: Local policy updating. Each agent uses the current policy to sample a batch of trajectories , where , is the batch size. Subsequently, agent calculates a local policy update . For example, using the REINFORCE algorithm (williams1992simple, ), is calculated as:
(2) where is a constant. Then agent sends to the server.
-
•
Step III: Global policy updating. The server updates the global policy by aggregating local updates using :
(3) where is the learning rate.
FRL methods vary in their aggregation rules. For example, using FedAvg (mcmahan2017communication, ), the global policy is updated as: .
2.2. Poisoning Attacks to FRL
The distributed nature of FRL makes it susceptible to poisoning attacks (fang2020local, ; shejwalkar2021manipulating, ; fan2021fault, ; zhang2024poisoning, ; yin2024poisoning, ), where malicious agents manipulate local training data (data poisoning) or policy updates (model poisoning) to compromise the global policy. For example, in Random action attack (fan2021fault, ), agents act randomly without following a pattern. Model poisoning attacks include Random noise attack (fan2021fault, ), where agents send Gaussian noise as policy updates, Trim attack (fang2020local, ), which maximizes deviation in policy updates, and Shejwalkar attack (shejwalkar2021manipulating, ), which increases the distance between pre- and post-attack updates. While some studies (ma2023local, ; zhang2020adaptive, ) assume agents can manipulate environments or rewards, such scenarios are often impractical and are not considered in our paper.
2.3. Byzantine-robust Aggregation Rules
2.3.1. FL-based Aggregation Rules
In typical federated learning (FL), the server uses FedAvg (mcmahan2017communication, ) to aggregate local model updates111Note that in FL, we commonly refer to a “model update” rather than a “policy update”., but this method is vulnerable to poisoning attacks since even one malicious agent can skew the results. To counter such attacks, several Byzantine-resilient aggregation rules have been proposed (yin2018byzantine, ; nguyen2022flame, ; ChenPOMACS17, ; Blanchard17, ; cao2020fltrust, ; xie2019zeno, ; rajput2019detox, ; mozaffari2023every, ; pan2020justinian, ; zhang2022fldetector, ; cao2021provably, ; rieger2022deepsight, ; fang2024byzantine, ; fang2022aflguard, ; fang2025FoundationFL, ; yueqifedredefense, ). Examples include Median (yin2018byzantine, ), which computes the median for each dimension, and Trimmed-mean (yin2018byzantine, ), which removes extreme values before averaging. FLAME (nguyen2022flame, ) clusters agents based on cosine similarity, discarding suspicious updates and adding adaptive noise to the rest. Our proposed ensemble method differs from (cao2021provably, ) by addressing continuous action spaces, while their approach only supports categorical labels. We also provide theoretical evidence that an attacked agent behaves similarly to pre-attack conditions as long as malicious agents are fewer than half of the total groups.
2.3.2. FRL-based Aggregation Rules
The authors in (fan2021fault, ) proposed FedPG-BR to defend against poisoning attacks in FRL. Each training round, the server computes the vector median of local policy updates and marks an update as benign if it aligns in direction and magnitude with the median. It then averages these benign updates to form a policy update estimator. Additionally, the server samples trajectories to compute its own policy update. The final global policy update is obtained by combining the estimator and the server’s update using the stochastically controlled stochastic gradient (SCSG) (lei2017less, ) to reduce variance.
Direction | Magnitude | |
Trim attack (fang2020local, ) | ✗ | ✗ |
Shejwalkar attack (shejwalkar2021manipulating, ) | ✗ | ✓ |
Normalized attack | ✓ | ✓ |
2.3.3. Limitations of Existing Attacks and Defenses
Current poisoning attacks and defense strategies have limitations. The Trim attack (fang2020local, ) targets individual dimensions in linear aggregation rules like Trimmed-mean and Median (yin2018byzantine, ), ignoring the update’s overall direction. In contrast, the Shejwalkar attack (shejwalkar2021manipulating, ) considers the entire update but overlooks its direction. Table 1 compares these with our proposed Normalized attack. Additionally, applying FL-based aggregation rules in FRL leads to poor performance, as they remain vulnerable to known attacks (fang2020local, ; shejwalkar2021manipulating, ). Although FedPG-BR (fan2021fault, ) counters Trim and Shejwalkar attacks, our experiments show it is still vulnerable to our Normalized attack.
3. Problem Setting
Threat model: We adopt the threat model from (fan2021fault, ), where an attacker controls some malicious agents. These agents may poison their training trajectories or send random policy updates to the server. The attacker’s goal is to disrupt the global policy’s convergence or push it toward a bad optimum. In a full knowledge attack, the attacker knows all agents’ policy updates and the server’s aggregation rule. In a partial knowledge attack, the attacker only knows the malicious agents’ updates and the aggregation rule.
Defense objectives: We aim to propose a method that achieves the following two goals. I) Superior learning performance: In non-adversarial settings, the method should perform as well as FedAvg, achieving comparable test rewards when all agents are benign. II) Resilience: It should defend against both data and model poisoning attacks. Even under attacks, the final global policy should maintain test rewards similar to those of FedAvg in attack-free scenarios.
4. Our Attack
4.1. Attack as an Optimization Problem
In our proposed Normalized attack, the attacker crafts malicious policy updates to maximize the deviation between the aggregated updates before and after the attack. A simple way to achieve this is by maximizing the distance between the two updates, resulting in an optimization problem for each global training round:
(4) |
where denotes the -norm, and and represent the aggregated policy updates before and after the attack, respectively. However, Eq. (4) focuses only on the magnitude of the post-attack aggregated update, ignoring its direction. As a result, the original and attacked updates could align in the same direction. Since FedPG-BR (fan2021fault, ) evaluates both the direction and magnitude of policy updates, attackers must carefully craft updates to bypass this defense. To address this, we propose the Normalized attack, which maximizes the angular deviation between the original and attacked aggregated updates, rather than just their magnitude. The formulation of our Normalized attack is as follows:
(5) |
4.2. Solving the Optimization Problem
Solving Problem (5) is challenging because many aggregation rules, like FedPG-BR (fan2021fault, ), are non-differentiable. To overcome this, we use practical techniques to approximate the solution by determining the direction of malicious updates in Stage I, followed by calculating their magnitude in Stage II.
Stage I (Optimize the direction): We let be the set of malicious agents. Assume that the malicious policy update , , is the perturbed version of normalized benign policy update:
(6) |
where is an adjustment parameter and is a perturbation vector. Then we can reformulate Problem (5) as follows:
(7) |
Finding the optimal and simultaneously is also not trivial. In this paper, we fix the and turn to finding the optimal . For example, we can let , where means the average of local policy updates. After we fix , the optimization problem of Eq. (7) becomes the following:
(8) |
There exist multiple methods to determine . In this study, we adopt the subsequent way to compute . Specifically, in each global training round, if the value of increases, then we update as , otherwise we let . We repeat this process until the convergence condition satisfies, e.g., the difference of between two consecutive iterations is smaller than a given threshold.
Stage II (Optimize the magnitude): After obtaining the direction of malicious policy update , we proceed to demonstrate how to determine the magnitude of for . In particular, let represent the scaled policy update for malicious agent , where , and is the scaling factor. We then formulate the following optimization problem to determine the scaling factor :
(9) |
The way to compute is similar to that of . Specifically, if increases, we update as , otherwise . We repeat this process until the convergence condition is met, then malicious agent sends to the server.
Fig. 1 shows the impact of our Normalized attack. In each global round, the attacker maximizes the deviation between pre- and post-attack aggregated updates, causing the global policy to drift. Over multiple rounds, this drift leads the FRL system to converge to a suboptimal solution. Since RL loss functions are highly non-convex, with many local optima, the attack’s impact can be significant.
Note that we do not provide a theoretical analysis of our attack for the following reasons: In our Normalized attack, the attacker carefully crafts malicious updates to induce subtle deviations in the aggregated policy each round. These deviations are hard to detect but still degrade the model’s performance. Modeling them theoretically is challenging. As shown in prior works (fang2020local, ; shejwalkar2021manipulating, ), the true goal of an attack is its real-world impact, such as causing incorrect predictions or compromising security. While theory offers insights, practical performance better reflects real-world outcomes.


5. Our Defense
5.1. Overview
In our approach, we train multiple global policies instead of a single one, each using a foundational aggregation rule like Trimmed-mean or Median (yin2018byzantine, ) with different subsets of agents. During testing, the agent predicts an action using all trained policies. For discrete action spaces, the final action is chosen by majority vote, while for continuous spaces, it is determined by the geometric median (ChenPOMACS17, ). Fig. 2 illustrates the process for a discrete action space with six agents split into three groups, each training a global policy over rounds, resulting in policies , , and . Since the third group contains a malicious agent, is poisoned. During evaluation, given a test state , the three policies predict “UP”, “UP”, and “Down”. With a majority vote, the final action selected is “UP”.
5.2. Our Ensemble Method
In an FRL system with agents, our method divides them into non-overlapping groups deterministically, such as by hashing their IDs. Each group trains a global policy using its agents with an aggregation rule like Trimmed-mean, Median (yin2018byzantine, ), or FedPG-BR (fan2021fault, ). Let represent the policy learned by group after global rounds. At the end of training, we obtain global policies: . During testing, the agent independently executes the trained policies. At a test state , let denote the action taken by the agent using policy . With policies, the agent generates actions: . The final action at state is determined using an ensemble method, which varies based on whether the action space is discrete or continuous.
Discrete action space: If the action space is discrete, the agent’s action is determined by majority vote among the actions. Let represent the frequency of action at state , calculated as:
(10) |
where is the indicator function, which returns 1 if , and 0 otherwise. The final action at test state is the one with the highest frequency, calculated as:
(11) |
Continuous action space: For a continuous action space , we use the Byzantine-robust geometric median (ChenPOMACS17, ) to aggregate the actions. The final action at state is computed as:
(12) |
We use the geometric median (ChenPOMACS17, ) to aggregate the continuous actions instead of FedAvg or Trimmed-mean, as our experiments show that these methods are vulnerable to poisoning attacks.
Complete algorithm: Algorithm 1 in Appendix outlines the ensemble method during training. In Lines 6-12, each group trains its global policy in round . In Line 7, the server for group shares the current global policy with its agents, who refine their local policies and send updates back (Lines 8-10). Here, represents the agents in group . Finally, the server aggregates these updates to revise the global policy (Line 11). Algorithm 2 in Appendix summarizes the testing phase, where the final action is selected by majority vote for discrete actions (Line 5) or by geometric median for continuous actions (Line 7).
Complexity analysis: In our ensemble FRL approach, each agent participates in only one global training round over rounds. Thus, the computational cost per agent is .
5.3. Formal Security Analysis
In this section, we present the security analysis of our ensemble method. For discrete action spaces, we show that the predicted action at a test state remains unchanged despite poisoning attacks, as long as the number of malicious agents stays below a certain threshold. For continuous action spaces, we prove that the difference between actions predicted before and after an attack is bounded if malicious agents make up less than half of the groups.
Theorem 1 (Discrete Action Space).
Consider an FRL system with agents and a test state , where the action space is discrete. The agents are divided into non-overlapping groups based on the hash values of their IDs, and each group trains its global policy using an aggregation rule AR. Define actions and as those with the highest and second-highest frequencies for state , with ties resolved by selecting the action with the smaller index. Our ensemble method aggregates the actions using Eq. (11). Let and represent the actions predicted when all agents are benign and when up to agents are malicious, respectively. The condition for is:
(13) |
where and represent the pre-attack frequencies of actions and for state , respectively. The notation indicates that action has a smaller index than action . Then we have that:
(14) |
Proof.
The proof is in Appendix A.1. ∎
Theorem 2 (Continuous Action Space).
In a continuous action space FRL system with agents and a test state , the agents are divided into non-overlapping groups. If agents are malicious and , each group trains a global policy using an aggregation rule AR. Our ensemble method aggregates the continuous actions using Eq. (12). Let and be the actions predicted before and after the attack, respectively. The following holds:
(15) |
where , is the set of continuous actions before attack.
Proof.
The proof is in Appendix A.2. ∎
6. Evaluation
6.1. Experimental Setup
6.1.1. Datasets
We use the following three datasets from different domains, including two discrete datasets (Cart Pole (barto1983neuronlike, ), Lunar Lander (duan2016benchmarking, )), and one continuous dataset (Inverted Pendulum (barto1983neuronlike, )). Details of these datasets are provided in Appendix A.4.
6.1.2. Compared Poisoning Attacks
We compare our Normalized attack with one data poisoning attack (Random action attack (fan2021fault, )) and three model poisoning attacks (Random noise (fan2021fault, ), Trim (fang2020local, ), and Shejwalkar (shejwalkar2021manipulating, )). Details are in Appendix A.5.





6.1.3. Foundational Aggregation Rules
We consider the following state-of-the-art foundational aggregation rules, including five aggregation rules designed for FL (FedAvg (mcmahan2017communication, ), coordinate-wise trimmed mean (Trimmed-mean) (yin2018byzantine, ), coordinate-wise median (Median) (yin2018byzantine, ), geometric median (ChenPOMACS17, ) and FLAME (nguyen2022flame, )) and one aggregation rule designed for FRL (FedPG-BR (fan2021fault, )). Details are in Appendix A.6.
6.1.4. Evaluation Metric
We evaluate an FRL method’s robustness using test reward. For the non-ensemble approach, the test reward is the average reward from 10 sampled trajectories using the trained global policy. In our ensemble method, we also average rewards from 10 trajectories, but actions are predicted using the ensemble framework. A lower test reward indicates a more effective attack and weaker defense.
6.1.5. Parameter Settings
By default, we assume that there are 30 agents in total. Following (fan2021fault, ), we assume that of agents are malicious. For our proposed ensemble method, we partition the agents into disjoint groups. The batch sizes for Cart Pole, Lunar Lander, and Inverted Pendulum datasets are set to 16, 64, and 32, respectively. The learning rates for these three datasets are individually set to , , and . Furthermore, for each of the three datasets, every agent samples a total of , , and trajectories during the training phase, respectively. Regarding the policy architectures, we train a Categorical MLP for the Cart Pole and Lunar Lander datasets and a Gaussian MLP for the Inverted Pendulum dataset. The policy architectures are shown in Table 3 in Appendix. We assume all agents use the same discount factor , trajectory horizon , and . Our Normalized attack has parameters , and . The default value of these six parameters are shown in Table 4 in Appendix. FedPG-BR uses unique parameters, including mini-batch size , global sampling steps , variance bound , and confidence parameter . Detailed settings are in Table 5 in Appendix. We assume the attacker has full knowledge of the FRL system unless stated otherwise. Results are presented on the Cart Pole dataset by default. We compare our ensemble method with the non-ensemble approach. In the non-ensemble method, the server trains a single global model with all agents using a foundational aggregation rule from Section 6.1.3. In our ensemble method, agents are divided into groups, each training a global policy with the same aggregation rule by default.
6.2. Experimental Results
Normalized attack is effective against non-ensemble methods: Fig. 3 shows the results of different defenses under various attacks on Cart Pole dataset. The results on Lunar Lander and Inverted Pendulum datasets are shown in Fig. 7 and Fig. 8 in Appendix, respectively. Based on Fig. 3 and Figs. 7-8, it is evident that our proposed Normalized attack successfully targets the non-ensemble methods. For instance, in the Lunar Lander dataset, our Normalized attack reduces the test reward of the Median to -33.3 in the non-ensemble setting, compared to a reward of 219.3 when all agents are benign. Notably, our Normalized attack stands out as the sole method capable of substantially manipulating the non-ensemble FedPG-BR aggregation rule across all three datasets. For example, in the Cart Pole dataset, the test reward of non-ensemble-based FedPG-BR drops from 500 in the absence of an attack to 101.4 under our Normalized attack. However, existing attacks such as Trim attack and Shejwalkar attack achieve unsatisfactory attack performance. The reason is that the Trim attack solely takes into account each dimension of the policy update, neglecting the entirety of the update itself. Furthermore, the Shejwalkar attack ignores the direction of the policy update.
Our ensemble method is effective: From Fig. 3 and Figs. 7-8 (in Appendix), we observe that when all agents are benign, our ensemble framework achieves test rewards comparable to FedAvg without attacks across all datasets and Byzantine-robust aggregation rules, fulfilling the goal of “superior learning performance.” For example, in the Inverted Pendulum dataset, the Trimmed-mean rule within the ensemble framework achieves a test reward of 1000, matching FedAvg’s performance without attacks. However, non-ensemble Byzantine-robust aggregation rules remain vulnerable to poisoning attacks, including our Normalized attack. As shown in the figures, embedding these robust rules within our ensemble framework ensures defense against all considered attacks, achieving the “resilience” goal. For instance, in the Cart Pole dataset under the Normalized attack, FedPG-BR achieves a test reward of 101.4 in the non-ensemble setting but 500 within our ensemble framework. Similarly, for the Inverted Pendulum dataset, Trimmed-mean yields test rewards of 152.7 without ensemble but 1000 with it under the Random action attack. However, FedAvg, even within our ensemble framework, remains vulnerable to attacks due to its inherent lack of robustness.




Impact of the fraction of malicious agents: Fig. 4 shows the impact of the fraction of malicious agents on the robustness of our ensemble method on Cart Pole dataset. From Fig. 4, we observe that when a large fraction of agents are malicious, our ensemble framework can still tolerate all the poisoning attacks across all robust aggregation rules. For example, when 40% of agents are malicious, our ensemble method achieves similar test rewards with that of FedAvg without attack. However, as shown in Fig. 3, even when the fraction of malicious agents is 30%, existing robust aggregation rules under non-ensemble setting could be easily poisoned
Impact of total number of agents: Fig. 9 in Appendix shows the influence of varying total agent numbers on our ensemble method under various attacks, with the proportion of malicious agents set at 30% and the overall number of agents ranging from 30 to 90. The numbers of groups are set to 5, 7, 7, and 9 when the total agents are 30, 50, 70, and 90, respectively. We observe that our ensemble method remains robust when the total number of agents varies.
Impact of number of groups: Fig. 10 in Appendix shows the results of our ensemble method with different group numbers, with 30 agents, 30% of which are malicious. When there is only one group, the ensemble method becomes equivalent to the non-ensemble approach. For three Byzantine-robust methods under various attacks, the test rewards match those of FedAvg without attack when the group sizes are 3, 5, or 7.
Stage I | Stage II | Normalization | |
Variant I | ✓ | ✗ | ✓ |
Variant II | ✗ | ✓ | ✓ |
Variant III | ✓ | ✓ | ✗ |
Variant IV (default) | ✓ | ✓ | ✓ |


Impact of different perturbation vectors: Our Normalized attack uses a perturbation vector . Table 6 in Appendix lists three types: “uv”, “std”, and “sgn”. calculates the average of the local policy updates, and computes their standard deviation. “sgn” is our default perturbation vector. Fig. 11 in Appendix shows the results of FedAvg, Trimmed-mean, Median, and FedPG-BR under Normalized attacks with different vectors. “Normalized-uv” refers to the Normalized attack using the “uv” vector. We observe that FedPG-BR in the non-ensemble setting is particularly vulnerable to the “sgn” vector.
Different variants of Normalized attack: Our Normalized attack consists of two stages, with policy updates normalized during optimization. Table 2 outlines its variants. For example, Variant I skips Stage II, which optimizes the magnitude of malicious updates. Variant IV is our default attack. Fig. 5 illustrates the results of different variants of our Normalized attack. We observe from Fig. 5 that Variant IV achieves the most effective attack performance overall.
Results of partial knowledge attack: By default, we assume the attacker knows all agents’ policy updates. Here, we explore a more realistic scenario where the attacker only knows updates from malicious agents. In this partial knowledge attack, we use to estimate the pre-attack aggregated update, where is the update from malicious agent , and is the set of malicious agents. Fig. 12 in Appendix shows the results of the Trim, Shejwalkar, and our Normalized attacks. Even with partial knowledge, Byzantine-robust rules in non-ensemble settings remain vulnerable to poisoning. For example, FedPG-BR achieves a test reward of 242.6 under our Normalized attack.
Initiate the attack in the middle of the training phase: We assume by default that the attacker targets the FRL system from the start of training. Here, we explore a scenario where the attack begins midway through training. In the Cart Pole dataset, each agent samples 5,000 trajectories during training. Fig. 13 in Appendix shows the results of FedPG-BR, Trimmed-mean, and Median in the non-ensemble setting. Results for the ensemble framework are omitted since it remains robust against all attacks, even if initiated from the start, as seen in Fig. 3. From Fig. 13, we observe that current robust aggregation methods in non-ensemble settings are still vulnerable to poisoning, even when the attack starts mid-training.
Results of heterogeneous environment: In this part, we explore a heterogeneous environment setting where two agents execute the same action at the same state but receive different rewards. In our experiments, we introduce some noise generated from Gaussian distribution to the rewards to simulate the heterogeneous environment. Fig. 14 in Appendix presents the results. We observe that non-ensemble aggregation rules remain vulnerable to poisoning attacks in heterogeneous environments. For example, FedPG-BR achieves a test reward of less than 100 under our Normalized attack. However, our ensemble framework effectively defends against all considered attacks using robust aggregation rules.
Results of other foundational aggregation rules: Fig. 6 shows results using geometric median (ChenPOMACS17, ) and FLAME (nguyen2022flame, ) aggregation rules on the Cart Pole dataset. Note that within our ensemble framework, the server still uses the majority vote (action space in the Cart Pole dataset is discrete) to select the action during the testing phase. We observe that, in a non-ensemble setting, the geometric median and FLAME aggregation rules are susceptible to either existing poisoning attacks or our proposed Normalized attack. In contrast, our ensemble framework remains robust.
Results of our ensemble method when using other aggregation rules to combine the continuous actions: In our proposed ensemble framework, the server employs the geometric median aggregation rule to select the subsequent action during the testing phase when the action space is continuous. In this context, we investigate a scenario where the continuous actions are aggregated by the FedAvg or Trimmed-mean aggregation rules during the testing phase within our ensemble framework. The results are shown in Figs. 15-16 in Appendix, where the Inverted Pendulum dataset is considered (with a continuous action space). From Figs. 15-16, we observe that existing robust foundational aggregation rules like Trimmed-mean and Median are susceptible to poisoning attacks when our ensemble framework employs FedAvg or Trimmed-mean to predict the subsequent action in the testing phase. This vulnerability arises because FedAvg is not robust, and Trimmed-mean is a coordinate-wise aggregation rule, it only has the capability to filter out individual outlier parameters and fails to eliminate an entire policy update, even when detected as malicious.
7. Conclusion, Limitations, and Future Work
We introduced the first model poisoning attacks on Byzantine-robust FRL. Rather than increasing the distance between the aggregated policy updates before and after the attacks, our introduced Normalized attack strives to amplify the angular deviation between the two updates. Additionally, we proposed a unique ensemble method that is provably resistant to poisoning attacks under some mild assumptions. Comprehensive experimental findings demonstrated that our Normalized attack can significantly corrupt Byzantine-robust aggregation methods in non-ensemble configuration, and our ensemble approach effectively safeguards against poisoning attacks.
A limitation of our work is that our Normalized attack requires the attacker to be aware of the server’s aggregation rule. An intriguing avenue for future research would be to develop new attacks that do not necessitate such information. Our Normalized attack is limited to untargeted poisoning attacks, another interesting future work is to study targeted poisoning attacks (bagdasaryan2020backdoor, ; wang2020attack, ; xie2019dba, ) to FRL. Additionally, investigating security issues in multi-agent reinforcement learning (tan1993multi, ; bucsoniu2010multi, ; vinyals2019grandmaster, ; zhang2018fully, ; lin2020robustness, ; fang2024hardness, ) would be a fruitful area for further exploration.
Acknowledgements.
We thank the anonymous reviewers for their comments. This work was supported in part by NSF grant No. 2131859, 2125977, 2112562, 1937787, and ARO grant No. W911NF2110182.References
- [1] Eugene Bagdasaryan, Andreas Veit, Yiqing Hua, Deborah Estrin, and Vitaly Shmatikov. How to backdoor federated learning. In AISTATS, 2020.
- [2] Andrew G Barto, Richard S Sutton, and Charles W Anderson. Neuronlike adaptive elements that can solve difficult learning control problems. In IEEE transactions on systems, man, and cybernetics, 1983.
- [3] Gilad Baruch, Moran Baruch, and Yoav Goldberg. A little is enough: Circumventing defenses for distributed learning. In NeurIPS, 2019.
- [4] Peva Blanchard, El Mahdi El Mhamdi, Rachid Guerraoui, and Julien Stainer. Machine learning with adversaries: Byzantine tolerant gradient descent. In NeurIPS, 2017.
- [5] Lucian Buşoniu, Robert Babuška, and Bart De Schutter. Multi-agent reinforcement learning: An overview. In Innovations in multi-agent systems and applications-1, 2010.
- [6] Ricardo JGB Campello, Davoud Moulavi, and Jörg Sander. Density-based clustering based on hierarchical density estimates. In PAKDD, 2013.
- [7] Xiaoyu Cao, Minghong Fang, Jia Liu, and Neil Zhenqiang Gong. Fltrust: Byzantine-robust federated learning via trust bootstrapping. In NDSS, 2021.
- [8] Xiaoyu Cao, Jinyuan Jia, and Neil Zhenqiang Gong. Provably secure federated learning against malicious clients. In AAAI, 2021.
- [9] Yudong Chen, Lili Su, and Jiaming Xu. Distributed statistical machine learning in adversarial settings: Byzantine gradient descent. In POMACS, 2017.
- [10] Michael B Cohen, Yin Tat Lee, Gary Miller, Jakub Pachocki, and Aaron Sidford. Geometric median in nearly linear time. In STOC, 2016.
- [11] Yan Duan, Xi Chen, Rein Houthooft, John Schulman, and Pieter Abbeel. Benchmarking deep reinforcement learning for continuous control. In ICML, 2016.
- [12] Gabriel Dulac-Arnold, Nir Levine, Daniel J Mankowitz, Jerry Li, Cosmin Paduraru, Sven Gowal, and Todd Hester. Challenges of real-world reinforcement learning: definitions, benchmarks and analysis. In Machine Learning, 2021.
- [13] Xiaofeng Fan, Yining Ma, Zhongxiang Dai, Wei Jing, Cheston Tan, and Bryan Kian Hsiang Low. Fault-tolerant federated reinforcement learning with theoretical guarantee. In NeurIPS, 2021.
- [14] Minghong Fang, Xiaoyu Cao, Jinyuan Jia, and Neil Gong. Local model poisoning attacks to byzantine-robust federated learning. In USENIX Security Symposium, 2020.
- [15] Minghong Fang, Jia Liu, Neil Zhenqiang Gong, and Elizabeth S Bentley. Aflguard: Byzantine-robust asynchronous federated learning. In ACSAC, 2022.
- [16] Minghong Fang, Seyedsina Nabavirazavi, Zhuqing Liu, Wei Sun, Sundararaja Sitharama Iyengar, and Haibo Yang. Do we really need to design new byzantine-robust aggregation rules? In NDSS, 2025.
- [17] Minghong Fang, Zifan Zhang, Prashant Khanduri, Jia Liu, Songtao Lu, Yuchen Liu, Neil Gong, et al. Byzantine-robust decentralized federated learning. In CCS, 2024.
- [18] Minghong Fang, Zifan Zhang, Alvaro Velasquez, Jia Liu, et al. On the hardness of decentralized multi-agent policy evaluation under byzantine attacks. In WiOpt, 2024.
- [19] Yunfei Gao, Mingliu Liu, Xiaopeng Yuan, Yulin Hu, Peng Sun, and Anke Schmeink. Federated deep reinforcement learning based trajectory design for uav-assisted networks with mobile ground devices. In Scientific Reports, 2024.
- [20] Hao Jin, Yang Peng, Wenhao Yang, Shusen Wang, and Zhihua Zhang. Federated reinforcement learning with environment heterogeneity. In AISTATS, 2022.
- [21] Sajad Khodadadian, Pranay Sharma, Gauri Joshi, and Siva Theja Maguluri. Federated reinforcement learning: Linear speedup under markovian sampling. In ICML, 2022.
- [22] Jens Kober, J Andrew Bagnell, and Jan Peters. Reinforcement learning in robotics: A survey. In The International Journal of Robotics Research, 2013.
- [23] Lihua Lei and Michael Jordan. Less than a single pass: Stochastically controlled stochastic gradient. In AISTATS, 2017.
- [24] Xinle Liang, Yang Liu, Tianjian Chen, Ming Liu, and Qiang Yang. Federated transfer reinforcement learning for autonomous driving. In Federated and Transfer Learning, 2022.
- [25] Jieyu Lin, Kristina Dzeparoska, Sai Qian Zhang, Alberto Leon-Garcia, and Nicolas Papernot. On the robustness of cooperative multi-agent reinforcement learning. In IEEE Security and Privacy Workshops, 2020.
- [26] Boyi Liu, Lujia Wang, and Ming Liu. Lifelong federated reinforcement learning: a learning architecture for navigation in cloud robotic systems. In IEEE Robotics and Automation Letters, 2019.
- [27] Siqi Liu, Kay Choong See, Kee Yuan Ngiam, Leo Anthony Celi, Xingzhi Sun, and Mengling Feng. Reinforcement learning for clinical decision support in critical care: comprehensive review. In Journal of medical Internet research, 2020.
- [28] Evelyn Ma, Praneet Rathi, and S Rasoul Etesami. Local environment poisoning attacks on federated reinforcement learning. arXiv preprint arXiv:2303.02725, 2023.
- [29] H Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, et al. Communication-efficient learning of deep networks from decentralized data. In AISTATS, 2017.
- [30] Stanislav Minsker. Geometric median and robust estimation in banach spaces. In Bernoulli, 2015.
- [31] Volodymyr Mnih, Adria Puigdomenech Badia, Mehdi Mirza, Alex Graves, Timothy Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. Asynchronous methods for deep reinforcement learning. In ICML, 2016.
- [32] Hamid Mozaffari, Virat Shejwalkar, and Amir Houmansadr. Every vote counts: Ranking-based training of federated learning to resist poisoning attacks. In USENIX Security Symposium, 2023.
- [33] Arun Nair, Praveen Srinivasan, Sam Blackwell, Cagdas Alcicek, Rory Fearon, Alessandro De Maria, Vedavyas Panneershelvam, Mustafa Suleyman, Charles Beattie, Stig Petersen, et al. Massively parallel methods for deep reinforcement learning. arXiv preprint arXiv:1507.04296, 2015.
- [34] Thien Duc Nguyen, Phillip Rieger, Roberta De Viti, Huili Chen, Björn B Brandenburg, Hossein Yalame, Helen Möllering, Hossein Fereidooni, Samuel Marchal, Markus Miettinen, et al. Flame: Taming backdoors in federated learning. In USENIX Security Symposium, 2022.
- [35] Xudong Pan, Mi Zhang, Duocai Wu, Qifan Xiao, Shouling Ji, and Min Yang. Justinian’s gaavernor: Robust distributed learning with gradient aggregation agent. In USENIX Security Symposium, 2020.
- [36] Shashank Rajput, Hongyi Wang, Zachary Charles, and Dimitris Papailiopoulos. Detox: A redundancy-based framework for faster and more robust gradient aggregation. In NeurIPS, 2019.
- [37] Phillip Rieger, Thien Duc Nguyen, Markus Miettinen, and Ahmad-Reza Sadeghi. Deepsight: Mitigating backdoor attacks in federated learning through deep model inspection. In NDSS, 2022.
- [38] Virat Shejwalkar and Amir Houmansadr. Manipulating the byzantine: Optimizing model poisoning attacks and defenses for federated learning. In NDSS, 2021.
- [39] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018.
- [40] Ming Tan. Multi-agent reinforcement learning: Independent vs. cooperative agents. In ICML, 1993.
- [41] Emanuel Todorov, Tom Erez, and Yuval Tassa. Mujoco: A physics engine for model-based control. In IROS, 2012.
- [42] Oriol Vinyals, Igor Babuschkin, Wojciech M Czarnecki, Michaël Mathieu, Andrew Dudzik, Junyoung Chung, David H Choi, Richard Powell, Timo Ewalds, Petko Georgiev, et al. Grandmaster level in starcraft ii using multi-agent reinforcement learning. In Nature, 2019.
- [43] Hongyi Wang, Kartik Sreenivasan, Shashank Rajput, Harit Vishwakarma, Saurabh Agarwal, Jy-yong Sohn, Kangwook Lee, and Dimitris Papailiopoulos. Attack of the tails: Yes, you really can backdoor federated learning. In NeurIPS, 2020.
- [44] Xiaofei Wang, Chenyang Wang, Xiuhua Li, Victor CM Leung, and Tarik Taleb. Federated deep reinforcement learning for internet of things with decentralized cooperative edge caching. In IEEE Internet of Things Journal, 2020.
- [45] Ronald J Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. In Machine learning, 1992.
- [46] Chulin Xie, Keli Huang, Pin-Yu Chen, and Bo Li. Dba: Distributed backdoor attacks against federated learning. In ICLR, 2020.
- [47] Cong Xie, Sanmi Koyejo, and Indranil Gupta. Zeno: Distributed stochastic gradient descent with suspicion-based fault-tolerance. In ICML, 2019.
- [48] Yueqi Xie, Minghong Fang, and Neil Zhenqiang Gong. Fedredefense: Defending against model poisoning attacks for federated learning using model update reconstruction error. In ICML, 2024.
- [49] Dong Yin, Yudong Chen, Ramchandran Kannan, and Peter Bartlett. Byzantine-robust distributed learning: Towards optimal statistical rates. In ICML, 2018.
- [50] Ming Yin, Yichang Xu, Minghong Fang, and Neil Zhenqiang Gong. Poisoning federated recommender systems with fake users. In The Web Conference, 2024.
- [51] Zhenyuan Yuan, Siyuan Xu, and Minghui Zhu. Federated reinforcement learning for generalizable motion planning. In American Control Conference, 2023.
- [52] Kaiqing Zhang, Zhuoran Yang, Han Liu, Tong Zhang, and Tamer Basar. Fully decentralized multi-agent reinforcement learning with networked agents. In ICML, 2018.
- [53] Xuezhou Zhang, Yuzhe Ma, Adish Singla, and Xiaojin Zhu. Adaptive reward-poisoning attacks against reinforcement learning. In ICML, 2020.
- [54] Zaixi Zhang, Xiaoyu Cao, Jinyuan Jia, and Neil Zhenqiang Gong. Fldetector: Defending federated learning against model poisoning attacks via detecting malicious clients. In KDD, 2022.
- [55] Zifan Zhang, Minghong Fang, Jiayuan Huang, and Yuchen Liu. Poisoning attacks on federated learning-based wireless traffic prediction. In IFIP/IEEE Networking Conference, 2024.
Appendix A Appendix
Parameter | Dataset | ||
Cart Pole | Lunar Lander | Inverted Pendulum | |
Hidden weights | |||
Activation | RELU | Tanh | Tanh |
Output activation | Tanh |
A.1. Proof of Theorem 1
Given a test state , the action frequencies for actions and when up to agents are malicious are represented as and , respectively. Under the worst-case condition, for a specific group, if malicious agents are present, the global policy learnt by the group might predict action instead of at state . That is, will decrease by 1 and will increase by 1 after the attack. Moreover, given that up to agents can be malicious, a maximum of groups may include malicious agents. Then we have that:
(16) | |||
(17) |
In our proposed ensemble approach, when the test state is given, if the prediction of action still holds, then either Condition I or Condition II must be true:
(18) | Condition I: | |||
(19) | Condition II: |
where Condition II is true due to the assumption in Theorem 1 that if two actions possess the same action frequencies, the action with the smaller index is chosen.
A.2. Proof of Theorem 2
Let be the set of actions after attack. Since and are respectively the before-attack and after-attack aggregated policy updates, then is the geometric median of vectors . Based on Lemma A.3, let be defined as , and with the condition , one has that:
(21) |
which completes the proof.
A.3. Useful Technical Lemma
Lemma 1.
Let’s consider to be vectors in a Hilbert space, let represent a -approximation of their geometric median. This means that for , we have . Given any with the condition that and a real number , if the following condition satisfies:
(22) |
Then one has:
(23) |
where , . Ideally, the geometric median sets .
A.4. Datasets
Cart Pole [2]: The Cart Pole environment is a simulation of a cart with a pole attached to it by a hinge. The cart can move along a horizontal track, and the pole can swing freely in the air. The goal is to balance the pole on the cart by applying forces to the left or right of the cart. The action space is a discrete space representing the direction of the fixed force applied to the cart, where for pushing the cart to the left, and for pushing it to the right. A reward of +1 is added for every time step that the pole remains upright. The episode ends if the pole falls over more than 12 degrees from vertical, the episode length is greater than 500, or the cart moves more than 2.4 units from the center. Given that the maximum episode length is 500, the highest possible reward in this scenario should also be 500.
Lunar Lander [11]: The Lunar Lander environment is a simulation of a rocket landing on the moon. The rocket’s engines are controlled by choosing one of four actions: do nothing, fire left engine, fire main engine, or fire right engine. The action space, therefore, is a discrete space and can be represented as . The goal is to land safely on the landing pad without crashing or going out of bounds. A reward is obtained for every step that the rocket is kept upright, and a penalty for using the engines. The environment is stochastic, meaning that the initial state of the rocket is random within a certain range.
Inverted Pendulum [2]: The Inverted Pendulum is similar to the Cart Pole problem, which is another classic control problem where you have to balance a pole on a cart by applying forces to the left or right. Yet, it is different from the Cart Pole in several key aspects. First, the Inverted Pendulum is powered by the Mujoco physics simulator[41], which allows for more realistic and complex experiments, such as varying the effects of gravity. Second, the action space of the Inverted Pendulum is continuous. Thirdly, considering that the maximum episode duration is set at 1000, the utmost attainable reward for the Inverted Pendulum dataset is 1000.
A.5. Compared Poisoning Attacks
Random action attack [13]: Random action attack is a category of data poisoning attacks in which malicious agents intend to corrupt their local trajectories. In particular, every malicious agent chooses a random action regardless of the state.
Random noise attack [13]: Random noise attack is a kind of model poisoning attack. In each training round, a malicious agent draws each coordinate of its policy update from an isotropic Gaussian distribution with a mean of 0 and a variance of 1,000.
Trim attack [14]: This attack operates under the assumption that the server uses Trimmed-mean [49] or Median [49] as its aggregation rule, to combine the local policy updates sent from agents. Trim attack considers each dimension of policy update independently. Specifically, malicious agents intentionally manipulate their policy updates so that the aggregated policy update post-attack differs significantly from the one before the attack, for each dimension of policy updates.
Shejwalkar attack [38]: In the Shejwalkar attack, the attacker designs malicious local policy updates with the intent to enlarge the distance between the aggregated policy update before the attack and the one after the attack.
A.6. Foundational Aggregation Rules
FedAvg [29]: In FedAvg, once the server receives local policy updates from all agents, it calculates the global policy update by taking the average of these updates.
Coordinate-wise trimmed mean (Trimmed-mean) [49]: Upon receiving local policy updates, the server first discards the largest and smallest elements for each dimension, then computing the average of the remaining values, where is the trim parameter.
Coordinate-wise median (Median) [49]: In the Median aggregation rule, the server determines the aggregated global policy update by computing the coordinate-wise median from all received local policy updates.
Geometric median [9]: For the geometric median aggregation rule, the server computes the aggregated policy update by taking the geometric median of received local policy updates from all agents.
FLAME [34]: The FLAME method starts by computing the cosine similarity among agents’ local policy updates. It then employs clustering methods like HDBSCAN [6] to identify potentially malicious updates. To further reduce the impacts of poisoning attacks, it implements an adaptive clipping mechanism to adjust the local updates. Finally, the server adds noise to the aggregated policy update to obtain the final global update.
FedPG-BR [13]: In the FedPG-BR aggregation rule, the server first calculates the vector median of all received local policy updates. A local policy update is deemed malicious if it fars from the calculated vector median. To additionally minimize the policy update variance, the server independently samples some trajectories to compute a server policy update. Subsequently, the server leverages the Stochastically Controlled Stochastic Gradient (SCSG) [23] framework to update the global policy.
Parameter | Dataset | ||
Cart Pole | Lunar Lander | Inverted Pendulum | |
0.999 | 0.99 | 0.995 | |
500 | 1000 | 1000 | |
0 | |||
0.83 (decays at each iteration with factor ) | 1 (decays at each iteration with factor ) | 0.83 (decays at each iteration with factor ) | |
0.03 (decays at each iteration with factor ) | 0.02 (decays at each iteration with factor ) | 0.2 (decays at each iteration with factor ) |
Parameter | Dataset | ||
Cart Pole | Lunar Lander | Inverted Pendulum | |
4 | 8 | 12 | |
0.06 | 0.07 | 0.25 | |
0.6 | 0.6 | 0.6 |
uv | std | sgn (default) | |






























