Efficient Multi-Task Reinforcement Learning via Task-Specific Action Correction*
Abstract
Multi-task reinforcement learning (MTRL) demonstrate potential for enhancing the generalization of a robot, enabling it to perform multiple tasks concurrently. However, the performance of MTRL may still be susceptible to conflicts between tasks and negative interference. To facilitate efficient MTRL, we propose Task-Specific Action Correction (TSAC), a general and complementary approach designed for simultaneous learning of multiple tasks. TSAC decomposes policy learning into two separate policies: a shared policy (SP) and an action correction policy (ACP). To alleviate conflicts resulting from excessive focus on specific tasks’ details in SP, ACP incorporates goal-oriented sparse rewards, enabling an agent to adopt a long-term perspective and achieve generalization across tasks. Additional rewards transform the original problem into a multi-objective MTRL problem. Furthermore, to convert the multi-objective MTRL into a single-objective formulation, TSAC assigns a virtual expected budget to the sparse rewards and employs Lagrangian method to transform a constrained single-objective optimization into an unconstrained one. Experimental evaluations conducted on Meta-World’s MT10 and MT50 benchmarks demonstrate that TSAC outperforms existing state-of-the-art methods, achieving significant improvements in both sample efficiency and effective action execution.
Index Terms:
Multi-task reinforcement learning, robotic manipulation tasks, goal-oriented sparse rewards, Lagrangian methodI Introduction
Empowering generalist robots through reinforcement learning is one of the essential targets of robotic learning. Reinforcement learning (RL) with the assistance of neural networks has become a crucial methodology in various domains, such as gaming[1, 2, 3, 4], large language models[5] and real-world applications including robotics[6]. However, the majority of research in RL predominantly focuses on specific problem scenarios, prioritizing mastery of individual tasks through the learning of single policies, often at the expense of generalization. Multi-task reinforcement learning (MTRL), on the other hand, emerges as a promising approach to improve generalization by leveraging domain information obtained from related tasks.
MTRL naturally incorporates a curriculum, as it enables the learning of more manageable tasks to facilitate the teaching of more challenging tasks[7]. However, MTRL is prone to negative transfer[8], a phenomenon where the task-specific knowledge from a task can impede the overall learning process of other tasks. This phenomenon is also referred to as task conflict, which becomes more acute as the number of tasks increases. From the optimization standpoint, task conflict arises from conflicting gradients[9] between tasks, where the gradients move in opposite directions.
A multitude of methods have been developed to alleviate negative transfer and achieve efficient MTRL. Classical approaches include policy distillation[10, 11], explicit measurement of task relatedness[12, 13], and information sharing[14, 15], among others. Policy distillation involves training a smaller network structure to bring previous tasks to an expert level, thereby integrating multiple policies into a single policy. However, these methods increase the number of network parameters as they require separate networks for different tasks and an additional distillation step. Some researchers utilize validation loss on tasks[12] or causal influence[13] to determine better task groupings. One drawback of the aforementioned methods is the need for substantial computational resources. Often, calculating task correlations and adjusting learning methods can only be done through trial and error, leading to high costs. In MTRL, information sharing can be achieved through data sharing, parameters sharing, representations sharing, or behavior sharing. For instance, CDS[15] routes data based on task-specific data to improve information sharing. Similarly, a simple method proposed in [14] applies a zero reward to unlabeled data, facilitating data sharing in theory and practice. Parameter sharing through learning shared representations can effectively transfer knowledge between policies. For instance, a soft modularization method presented in [16] shares parameters by generating soft combinations of different modules. Similarly, AdaShare[17] and an automated multi-task learning algorithm in [18] adaptively determine the feature sharing mode across multiple tasks. However, these methods suffer from high computational complexity as they require dynamic exploration of network connectivity. Attention mechanisms can be leveraged to share representations, as proposed in[19, 20], which group task knowledge into sub-networks without the need for prior assumptions. CARE[21] leverages metadata to capture the shared structure among tasks. There is another significant MTRL work that focuses on the challenge of multi-objective optimization from a gradient perspective, aiming to reduce the impact of conflicting gradients by manipulating the gradients based on various criteria[9, 22]. However, these methods impose an additional computational burden. The aforementioned methods reduce or coordinate conflicts between tasks from the perspectives of representation, gradient, task grouping, etc., effectively achieving MTRL. However, they neglect the potential causes of task conflict. An agent solely focuses on task-specific information and details within each individual task, it may result in short-sightedness and hinder the generalization across tasks. By introducing goal-oriented sparse learning signals, it is possible to strike a balance between task-specific performance and generalization across tasks.
Imagine that humans are simultaneously learning a variety of related manipulation tasks, for instance, as depicted in Fig. 1. Within these tasks, there are several actions that exhibit similarity, such as approaching objects like doors or drawers and interacting with them. Humans have the ability to leverage previously learned behaviors when encountering a specific task, making slight adjustments based on the task’s goal. Furthermore, when humans are confronted with numerous tasks simultaneously, having detailed instructions and requirements for each task can lead to confusion and a sense of being overwhelmed, even though they can be helpful in completing individual tasks. In contrast, when each task has a clearly defined goal, humans naturally adopt a longer-term perspective regarding conflicts and priorities among tasks. The above human behaviors imply their recognition that current conflicts may not be necessary for achieving the overall objectives of the tasks. Our idea is inspired by this recognition.

In this paper, we propose a novel approach called Task-Specific Action Correction (TSAC) as a general and complementary method for MTRL. TSAC decomposes policy learning into two policies: a shared policy (SP) and an action correction policy (ACP). SP maximizes well-shaped and intensive rewards, which focus on task-specific information and accelerate the learning process. Its output actions, referred to as shared actions, are potentially short-sighted. On the other hand, ACP utilizes goal-oriented sparse rewards to generate a far-sighted edited action that can cross tasks. The goal-oriented sparse rewards which is sparse and strongly correlated with the completion of the objective. SP and ACP collaborate with each other, where SP provides a suboptimal policy that facilitates the training of ACP in the sparse rewards setting. ACP, in return, improves the overall performance. To balance the training of these two policies, we assign a virtual expected budget to the sparse rewards and use the Lagrangian method to dynamically adjust the weights of the loss in ACP. Our two-policy paradigm draws inspiration from works in safe reinforcement learning[23, 24, 25]. However, our approach differs significantly in terms of motivation and interpretation.
To implement our approach, we employ the Soft Actor-Critic algorithm[26] as the underlying reinforcement learning policy. It is worth noting that our approach is algorithm-agnostic and can be integrated with existing MTRL methods. Our experimental results demonstrate the efficiency and significance of the cooperation between the two policies and simply combining the two rewards do not yield comparable results. Moreover, our experiments conducted on the Metaworld benchmark[27] using MT10 and MT50 showcase significant improvements in both sample efficiency and final performance compared to previous state-of-the-art multi-task policies. In summary, our contributions are as follows:
(1) We propose Task-Specific Action Correction (TSAC), a general and complementary approach for MTRL, which decomposes policy learning into two policies, facilitating efficient MTRL. TSAC can be combined with any existing MTRL methods.
(2) We introduce goal-oriented sparse rewards to provide agents with a long-term perspective for handling task conflicts that arise from excessive focus on specific tasks’ details.
(3) We assign a virtual expected budget to the sparse rewards and utilize Lagrangian method to transform a constrained single-objective optimization into an unconstrained one. The Lagrangian multiplier dynamically adjust the loss weights in the two policy networks.
(4) We demonstrate the efficiency and significance of the cooperation between the two policies in MTRL, and show that TSAC achieves significant improvements in both sample efficiency and effective action execution.
II preliminaries
II-A Multi-Task Reinforcement Learning
We extend the single-task Markov Decision Process (MDP) problem to multi-task MDP for a agent under the framework of Contextual MDP (CMDP)[28]. CMDP is defined by a tuple . Here, can be viewd as a task set , where denotes the task and is the number of tasks. represents the shared state space , denotes the shared action space, and is the discount factor. State and action . is a fuction that maps a task to MDP parameters, such that . The transition probability , reward function and initial state distribution vary by each task. During training, tasks are sampled from a uniform distribution . The agent’s policy takes state as input and outputs action . The objective of the agent’s policy is to maximize the expected return , where and represent the state and the action at timestep .
II-B Soft Actor-Critic
In this paper, we adopt Soft Actor-Critic (SAC) [26] as the fundamental policy. As observed in [27], SAC, being an off-policy actor-critic algorithm with a strong exploration ability based on maximum entropy, exhibits superior performance.
The concept of SAC goes beyond merely maximizing cumulative rewards. It also introduces additional stochasticity to the policy. Thus, a regularization term that incorporates entropy is included in the reinforcement learning objective. The correction term added with entropy can be defined as follows:
(1) |
In this context, represents a regularization coefficient that controls the significance of the entropy term. denotes the entropy.
III method
In this section, we present a general and complementary approach named Task-Specific Action Correction (TSAC) to decompose policy learning across two policies: the shared policy (SP) and the action correction policy (ACP).
III-A Overall Structure of TSAC
As illustrated in Fig. 2, TSAC consists of a pair of cooperative policies. The first policy, called the shared policy (SP) and denoted as , optimizes the guiding dense rewards by proposing a preliminary action . However, this preliminary action may be shortsighted. The second policy, referred to as the action correction policy (ACP) and denoted as , corrects the preliminary action by providing an action correction to execute an effective action. The result action, denoted as , is then output to the environment, where represents an editing function. ACP is conditioned on SP’s output . Together, these two policies cooperate to improve sample efficiency and performance. For simplicity, we denote the overall composed policy as .

Motivation: TSAC decomposes policy learning into two subtasks that focus on guiding dense rewards or goal-oriented sparse rewards. This decomposition is motivated by the following considerations:
III-A1 Different Conflict Horizons
To achieve effective MTRL, most applications incorporate guiding dense rewards into each task. However, SP is myopic as it only focus on specific tasks’ details and overlooks whether the final goal is accomplished, despite the final goal being the most important aspect. SP only coordinates task conflicts from various tasks in the short term. In contrast, ACP aims to maximize goal-oriented sparse rewards, enabling an agent to adopt a long-term perspective and achieve generalization across tasks. With the assistance of SP, ACP has more opportunities to obtain goal-oriented sparse rewards, thus alleviating the challenge of learning sparse rewards in ACP.
III-A2 Efficient Exploration
From the perspective of SP, its action is altered by ACP. Instead of discouraging SP from taking suboptimal actions, ACP offers suggestions to improve the preliminary action, enabling SP to continue exploration in an effective and far-sighted manner. This guarded exploration leads to a better overall exploration policy because ACP is less likely to hinder SP’s exploration. From the perspective of ACP, it determines the action based on SP, enhancing its ability to explore. From an entropy perspective, the decomposition of policy learning into two policy networks introduces additional entropy.
III-B Goal-oriented sparse rewards
Manually designed dense rewards incorporate prior knowledge and effectively guide policy learning. However, the magnitude of reward values does not directly indicate the ability of a policy to accomplish tasks. Therefore, we have introduced goal-oriented sparse rewards, which are correlated with the completion of the objective. The goal-oriented sparse rewards of a task is characterized by an "-region" in state space, represented by:
(2) |
In this equation, denotes the goal-oriented sparse rewards of task , is the current state, denotes the goal state, is a function that maps the goal state and current state to a latent space, computing the distance between them. defines the reward value and set . represents a small distance threshold.
Following the description of CMDP (section II. A), the initial state distribution determines the probability density of an episode starting at state . For each transition from task at timestep , the environment produces a scalar . It is worth mentioning that the reward function is artificially designed and intensive, which we refer to as guiding dense rewards. Similarly, the environment produces a scalar as a goal-oriented sparse reward. Higher values for both the guiding dense rewards and the goal-oriented sparse rewards indicate better performance.
For each state , the guiding dense rewards state value of policy is denoted as . The guiding dense rewards state-action value is denoted as . Similarly, We define and for the goal-oriented sparse rewards.
We consider the MTRL objective from the perspective of guiding dense rewards and goal-oriented sparse rewards:
(3) |
MTRL can be viewed as a multi-objective optimization problem. The introduction of sparse rewards adds an additional objective to the MTRL framework, which increases the problem’s complexity. To convert multi-objective MTRL into single-objective MTRL, we assign a virtual expected budget to the sparse rewards. This allows us to write the MTRL objective as follows:
(4) |
In this equation, represents a virtual expected budget. Specifically for each time step, Eq.4 can be rewritten as:
(5) |
Here, denotes the expected budget specific to each step and it relates to through the equation . Notably, the expected budget represents an average target to be achieved rather than a strict enforcement.
To simplify the problem, we transform the multi-objective MTRL into a single-objective MTRL using the Lagrangian method. This method converts the constrained optimization problem Eq.(4) into an unconstrained one by introducing a multiplier :
(6) |
In this formulation, the weight of the goal-oriented sparse rewards combined with the guiding dense rewards is represented by . We solve this objective by using model-free MTRL algorithms.
III-C Objective function
We utilize an off-policy actor-critic approach to train the two policies. Given the overall policy , we use the typical Temporal Difference (TD)[29] backup to learn and , which are parameterized as and respectively. Here, represents the network parameters collectively for the two state-action values. When provided with and , the Bellman backup operator for the guiding dense rewards state-action value is expressed as:
(7) |
where represents the guiding dense rewards obtained from task . The backup operator for the goal-oriented sparse rewards state-action value is defined similarly with . Both and can be learned from transitions sampled from a replay buffer.
To achieve off-policy training of and , we convert Eq.(6) into a bi-level optimization surrogate. The formulation is presented as follows:
(8) |
Here, represents a replay buffer containing a historical marginal state distribution used to train the policies. The initial state distribution is employed to train . This distinction arises from the idea that when fine-tuning , we should primarily consider how well the policy satisfies the virtual expected budget starting with , rather than with some historical state distribution. Subsequently, We further transform the off-policy objective(Eq.(8),a) into two parts:
(9) |
In this modified formulation, the distance function quantifies the change from to . It is not necessarily proportional to as the editing function can introduce nonlinearity. ACP aims to maximize the goal-oriented sparse rewards while minimizing the distance between the actions before and after the correction. On the other hand, SP solely focus on maximizing the guiding dense rewards. Importantly, ACP modifies SP’s action, which aligns with the discussed motivation for efficient exploration. The training objective (Eq.(9),b) for ACP relies on a critic , which learns the expected future goal-oriented sparse rewards. Consequently, guided by , ACP explores actions with greater potential in long-term sequences.
III-D Action correction function and Distance function
In this section, we present the design for the action correction function and the distance function .
III-D1 Action correction function
We opt for the correction function to be primarily additive and simple, which helps to reduce training difficulty. Without loss of generality, we assume a bounded action space , and that both and are already within this space. Consequently, we define:
(10) |
Here, the element-wise min and max, along with the multiplication by and the clipping, ensure that . This means that SP retains full control over the final action and can overwrite the action if necessary. This is crucial because ACP faces challenges in learning an effective correction policy in the short horizon when SP still dominates the learning process. Although the additive operation is simple, the overall editing process is sufficiently general to encompass any modification.
This additive editing function is motivated by the goal of achieving sparsity. Policies are typically evaluated based on metrics such as success, which are only triggered for certain states. To explicitly incorporate this inductive bias, we adopt the additive correction function, which ensures that ACP learns a policy that maximizes metrics such as success based on SP. This simplifies the optimization landscape of ACP.
III-D2 Distance function
we utilize the hinge loss to compare the guiding dense rewards state-action values of and :
(11) |
Here, represents the critic and denotes the parameters. This loss yields zero if the edited action already attains a higher state-action value than the preliminary action . In such cases, only is optimized by . Otherwise, the inner part of (Eq.(9),b) is recovered as . Our distance function in the critic is more appropriate than distance in the action space, because we ultimately care about how the changes after the action is edited.
III-E Training process
To practically train the objectives, we employ stochastic gradient descent (SGD) simultaneously to Eq.(8)(b) and Eq.(9). The re-parameterization trick is utilized for both and to enable the application of SGD. To assess , a batch of rollout experiences following is provided. The gradient of (Eq.(8)(b)) is approximated as:
(12) |
Here, represents the virtual expected budget as defined in Eq.(5). Subsequent to each rollout, a batch of goal-oriented sparse rewards is collected, each reward is compared to , and the mean of the differences is used to adjust . This approximation enables the updating of using mini-batches of data, irather than waiting for complete episodes to conclude or relying on the often inaccurate estimated . Multiple parallel environments are employed to mitigate temporal correlation within the rollout batch data, which constitutes the data to be placed into the replay buffer.

The computational graph presenting Eq.(9) is illustrated in Fig. 3. Our approach is applicable to various goal-oriented sparse rewards and action distance functions. To encourage exploration, we integrate SAC[26], which incorporates the entropy terms of and in Eq.(9), dynamically adjusting their weights based on two entropy targets as described in [26].In our experiments, both SP and ACP are trained from scratch. The pseudocode for the overall TSAC is shown in the algorithm 1.
IV Experiments
In this section, we evaluate TSAC in the Meta-World multi-task RL environment[27] and use Meta-World’s MT10 and MT50 benchmarks. The MT10 and MT50 evaluation protocols consist of 10 and 50 tasks, respectively (shown in Fig. 4). We compare TSAC against several baseline methods and conduct ablation studies to verify the effectiveness of our method.
The first goal of our experimental evaluation is to assess Whether TSAC improves the performance of a multi-task agent. We compare the performance of TSAC in two different settings: a short horizon to evaluate its sample efficiency and a long horizon to measure its overall performance. For comparison, We select CARE[21], MT–SAC, Soft Modularization[16] and PCGrad[9] as our baselines.
Furthermore, we evaluate different action correction functions to identify which yields the best performance. The action correction functions considered are SP-dominated, ACP-dominated, equal, and Softclip (See Section IV-C for the definitions) .

IV-A Baselines
We will compare our method against the following baselines:
CARE: As a representations sharing method, representations are shared to learn how to compose them by leveraging additional information about each task.
MT–SAC: This approach directly applies the SAC algorithm to the multi-task setting. It utilizes a shared backbone with disentangled alphas.
Soft Modularization: As a parameters sharing method, Soft Modularization shares parameters and uses a routing network to softly combine all possible routes for each task.
PCGrad: It is a gradient manipulation method that projects a task’s gradient onto the normal plane of any other conflicting task. However, it has high time complexity and is not suitable for MT50 in the long horizon.
TSAC(ours): Our proposed method builds upon CARE, leveraging its representation-sharing module. In contrast, our approach is based on behavior sharing and utilizes goal-oriented sparse rewards.

Success——150K | MT10 | MT50 |
---|---|---|
TSAC(ours) | 0.390 0.115 | 0.362 0.051 |
CARE | 0.260 0.062 | 0.277 0.028 |
MT-SAC | 0.198 0.068 | 0.220 0.035 |
Soft-Mod | 0.180 0.108 | 0.151 0.040 |
PCGrad | 0.276 0.107 | 0.238 0.035 |
Success——MT10 | 0.8M | 1M | Best |
---|---|---|---|
TSAC(ours) | 0.762 0.109 | 0.722 0.052 | 0.827 0.038 |
CARE | 0.667 0.082 | 0.642 0.076 | 0.708 0.114 |
MT-SAC | 0.574 0.097 | 0.555 0.166 | 0.635 0.120 |
Soft Mod | 0.549 0.087 | 0.560 0.107 | 0.596 0.102 |
PCGrad | 0.655 0.150 | 0.644 0.090 | 0.708 0.114 |
Success——MT50 | 0.8M | 1M | Best |
---|---|---|---|
TSAC(ours) | 0.450 0.046 | 0.445 0.045 | 0.524 0.030 |
CARE | 0.418 0.057 | 0.395 0.031 | 0.497 0.035 |
MT-SAC | 0.381 0.044 | 0.390 0.045 | 0.431 0.046 |
Soft Mod | 0.155 0.033 | 0.162 0.034 | 0.207 0.051 |
IV-B Comparative evaluation
Fig. 5a shows the average success rate on the 10 tasks of the MT10 benchmark form Meta-world for TSAC, CARE, MT–SAC, Soft Modularization, and PCGrad. Since the success rate is a binary variable, it is noisy; therefore, the results were averaged across multiple seeds and the curves were smoothed. Mean and standard error are reported for each value.
We consider 1 million steps as a long horizon, and 150 thousand steps as a short horizon. The short horizon is utilized to observe the exploration ability of different methods, while the long horizon is used to visualize the performance of the method at various time points. It is worth noting that all methods are trained using SAC with disentangled alphas.
Table LABEL:table1 and Fig. 5a demonstrate that our method outperforms all the baselines on the short horizon in MT10. Additionally, Fig. 5c illustrates that even in MT50 our method still outperform all baselines. For comparison, it takes around 1 million steps for the Multi-task SAC agent to reach the accuracy that our TSAC agent achieves around 300 thousand steps, suggesting that our method is highly sample-efficient and exploration-efficient.
Table LABEL:table2 and LABEL:table3 as well as Fig. 5b and Fig. 5d depict that TSAC is able to learn a good policy on the long horizon. For MT10, TSAC performs best and reaches a top success rate of 0.827 during the training. Furthermore, sampling the success rate around 0.8 million and 1 million steps reveals that TSAC has the best performance. For MT50, conflicts between tasks become more acute. CARE stops learning around 600 thousand steps and performance begins to decline in the following training steps. Despite suffering from the conflicts between tasks, TSAC’s performance declines, but it still performs better than CARE. MT-SAC achieves smooth learning and has similar performance to that of CARE and TSAC at 0.8 million and 1 million steps. However, in terms of the best performance throughout training, MT–SAC is far inferior to TSAC, which outperforms all methods. Since the success metric is a binary variable and very noisy, the best performance is obtained by smoothing over the curve. Importantly, TSAC achieves average success rates of 0.450 and 0.445 on MT50 at 0.8 million and 1 million steps, surpassing the reported results from CARE, MT–SAC and Soft Modularization.
IV-C Ablation study
The result in the previous section suggests that TSAC is both sample efficient and exploration efficient. Furthermore, TSAC yields a good policy on long horizons. In this section, we further investigate the impact of the various action correction functions employed by TSAC and discuss the potential for improved performance.
In ablation study, We will consider four different functions :
SP dominated (ours): , SP’s action dominates the final action.
ACP dominated: , ACP’s action dominates the final action.
equal contribution: , SP and ACP contribute equally to the final action.
Softclip: , use softclip to smooth out the output action and bring in nonlinearity.

Fig. 6 illustrates that SP dominated action correction function outperforms other functions in producing the final action. However, when SP and ACP contribute equally to the final action, the agent encounters diffculties in learning a policy. This conflict relationship between SP and ACP presents challenges in optimizing each respective goal. Conversely, when either SP or ACP dominates the action correction funcion, the agent can successfully learn a well-performing policy. Comparatively, SP outperforms ACP due to ACP’s focus on optimizing the goal-oriented sparse rewards, which is challenging to train due to its sparsity. In contrast, the Softclip function exhibits a slight decrease in performance compared to our action correction function. This decline can be attributed to the introduction of nonlinearity, which further complicates and hampers the learning process. Consequently, we opt for a simple and linear operation instead of training a learnable network.
V CONCLUSION AND FURTHER WORK
In this work, we present the Task-Specific Action Correction (TSAC), a general and complementary MTRL method inspired by Safe Reinforcement Learning, that surpasses the several well-performed baselines on the MT10 and MT50 benchmark from Meta-World.
In this paper, we showed that our method is able to learn a high-performing policy and achieve significant improvements in both sample efficiency and final performance compared to previous state-of-the-art multi-task policies. Furthermore, TSAC is general and complementary enough to be integrated with existing methods like CARE, MT–SAC to improve them. Finally, we show the benifits of decomposing policy learning into two policies and the validity of introducing goal-oriented sparse rewards. TSAC still performs well with more difficult and diverse tasks (MT50).
In future work, We will introduce a pre-trained paradigm which policy is pretrained to maximize guiding dense rewards, and this policy is used as the initialization for SP. In this case, SP cannot be frozen because ACP continuously modifies its MDP. Instead, SP needs to be fine-tuned to adapt to the evolving actions of ACP. This pre-trained SP has the potential to accelerate the convergence of TSAC.
References
- [1] D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot et al., “Mastering the game of go with deep neural networks and tree search,” Nature, vol. 529, no. 7587, pp. 484–489, 2016.
- [2] D. Silver, J. Schrittwieser, K. Simonyan, I. Antonoglou, A. Huang, A. Guez, T. Hubert, L. Baker, M. Lai, A. Bolton et al., “Mastering the game of go without human knowledge,” Nature, vol. 550, no. 7676, pp. 354–359, 2017.
- [3] C. Berner, G. Brockman, B. Chan, V. Cheung, P. Dębiak, C. Dennison, D. Farhi, Q. Fischer, S. Hashme, C. Hesse et al., “Dota 2 with large scale deep reinforcement learning,” arXiv preprint arXiv:1912.06680, 2019.
- [4] O. Vinyals, I. Babuschkin, W. M. Czarnecki, M. Mathieu, A. Dudzik, J. Chung, D. H. Choi, R. Powell, T. Ewalds, P. Georgiev et al., “Grandmaster level in starcraft ii using multi-agent reinforcement learning,” Nature, vol. 575, no. 7782, pp. 350–354, 2019.
- [5] L. Ouyang, J. Wu, X. Jiang, D. Almeida, C. Wainwright, P. Mishkin, C. Zhang, S. Agarwal, K. Slama, A. Ray et al., “Training language models to follow instructions with human feedback,” Advances in Neural Information Processing Systems, vol. 35, pp. 27 730–27 744, 2022.
- [6] T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, and D. Wierstra, “Continuous control with deep reinforcement learning,” arXiv preprint arXiv:1509.02971, 2015.
- [7] L. Pinto and A. Gupta, “Learning to push by grasping: Using multiple tasks for effective learning,” in 2017 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2017, pp. 2161–2168.
- [8] S. Ruder, “An overview of multi-task learning in deep neural networks,” arXiv preprint arXiv:1706.05098, 2017.
- [9] T. Yu, S. Kumar, A. Gupta, S. Levine, K. Hausman, and C. Finn, “Gradient surgery for multi-task learning,” Advances in Neural Information Processing Systems, vol. 33, pp. 5824–5836, 2020.
- [10] A. A. Rusu, S. G. Colmenarejo, C. Gulcehre, G. Desjardins, J. Kirkpatrick, R. Pascanu, V. Mnih, K. Kavukcuoglu, and R. Hadsell, “Policy distillation,” arXiv preprint arXiv:1511.06295, 2015.
- [11] Y. Teh, V. Bapst, W. M. Czarnecki, J. Quan, J. Kirkpatrick, R. Hadsell, N. Heess, and R. Pascanu, “Distral: Robust multitask reinforcement learning,” Advances in Neural Information Processing Systems, vol. 30, 2017.
- [12] S. Liu, S. James, A. J. Davison, and E. Johns, “Auto-lambda: Disentangling dynamic task relationships,” arXiv preprint arXiv:2202.03091, 2022.
- [13] C. Fifty, E. Amid, Z. Zhao, T. Yu, R. Anil, and C. Finn, “Efficiently identifying task groupings for multi-task learning,” Advances in Neural Information Processing Systems, vol. 34, pp. 27 503–27 516, 2021.
- [14] T. Yu, A. Kumar, Y. Chebotar, K. Hausman, S. Levine, and C. Finn, “Conservative data sharing for multi-task offline reinforcement learning,” Advances in Neural Information Processing Systems, vol. 34, pp. 11 501–11 516, 2021.
- [15] T. Yu, A. Kumar, Y. Chebotar, K. Hausman, C. Finn, and S. Levine, “How to leverage unlabeled data in offline reinforcement learning,” in International Conference on Machine Learning. PMLR, 2022, pp. 25 611–25 635.
- [16] R. Yang, H. Xu, Y. Wu, and X. Wang, “Multi-task reinforcement learning with soft modularization,” Advances in Neural Information Processing Systems, vol. 33, pp. 4767–4777, 2020.
- [17] X. Sun, R. Panda, R. Feris, and K. Saenko, “Adashare: Learning what to share for efficient deep multi-task learning,” Advances in Neural Information Processing Systems, vol. 33, pp. 8728–8740, 2020.
- [18] P. Guo, C.-Y. Lee, and D. Ulbricht, “Learning to branch for multi-task learning,” in Proceedings of the 37th International Conference on Machine Learning. PMLR, 2020, pp. 3854–3863.
- [19] T. Bram, G. Brunner, O. Richter, and R. Wattenhofer, “Attentive multi-task deep reinforcement learning,” in Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2019, Würzburg, Germany, September 16–20, 2019, Proceedings, Part III. Springer, 2020, pp. 134–149.
- [20] G. Cheng, L. Dong, W. Cai, and C. Sun, “Multi-task reinforcement learning with attention-based mixture of experts,” IEEE Robotics and Automation Letters, 2023.
- [21] S. Sodhani, A. Zhang, and J. Pineau, “Multi-task reinforcement learning with context-based representations,” in International Conference on Machine Learning. PMLR, 2021, pp. 9767–9779.
- [22] B. Liu, X. Liu, X. Jin, P. Stone, and Q. Liu, “Conflict-averse gradient descent for multi-task learning,” Advances in Neural Information Processing Systems, vol. 34, pp. 18 878–18 890, 2021.
- [23] H. Yu, W. Xu, and H. Zhang, “Towards safe reinforcement learning with a safety editor policy,” Advances in Neural Information Processing Systems, vol. 35, pp. 2608–2621, 2022.
- [24] Y. Zhang, Q. Vuong, and K. Ross, “First order constrained optimization in policy space,” Advances in Neural Information Processing Systems, vol. 33, pp. 15 338–15 349, 2020.
- [25] Z. Qin, Y. Chen, and C. Fan, “Density constrained reinforcement learning,” in International Conference on Machine Learning. PMLR, 2021, pp. 8682–8692.
- [26] T. Haarnoja, A. Zhou, P. Abbeel, and S. Levine, “Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor,” in Proceedings of the 35th International Conference on Machine Learning. PMLR, 2018, pp. 1861–1870.
- [27] T. Yu, D. Quillen, Z. He, R. Julian, K. Hausman, C. Finn, and S. Levine, “Meta-world: A benchmark and evaluation for multi-task and meta reinforcement learning,” in Proceedings of the Conference on Robot Learning. PMLR, 2020, pp. 1094–1100.
- [28] A. Hallak, D. Di Castro, and S. Mannor, “Contextual markov decision processes,” arXiv preprint arXiv:1502.02259, 2015.
- [29] G. Tesauro et al., “Temporal difference learning and td-gammon,” Communications of the ACM, vol. 38, no. 3, pp. 58–68, 1995.