The Importance of Pessimism in Fixed-Dataset Policy Optimization
Abstract
We study worst-case guarantees on the expected return of fixed-dataset policy optimization algorithms. Our core contribution is a unified conceptual and mathematical framework for the study of algorithms in this regime. This analysis reveals that for naïve approaches, the possibility of erroneous value overestimation leads to a difficult-to-satisfy requirement: in order to guarantee that we select a policy which is near-optimal, we may need the dataset to be informative of the value of every policy. To avoid this, algorithms can follow the pessimism principle, which states that we should choose the policy which acts optimally in the worst possible world. We show why pessimistic algorithms can achieve good performance even when the dataset is not informative of every policy, and derive families of algorithms which follow this principle. These theoretical findings are validated by experiments on a tabular gridworld, and deep learning experiments on four MinAtar environments.
1 Introduction
We consider fixed-dataset policy optimization (FDPO), in which a dataset of transitions from an environment is used to find a policy with high return.222We use the term fixed-dataset policy optimization to emphasize the computational procedure; this setting has also been referred to as batch RL (Ernst et al., 2005; Lange et al., 2012) and more recently, offline RL (Levine et al., 2020). We emphasize that this is a well-studied setting, and we are simply choosing to refer to it by a more descriptive name. We compare FDPO algorithms by their worst-case performance, expressed as high-probability guarantees on the suboptimality of the learned policy. It is perhaps obvious that in order to maximize worst-case performance, a good FDPO algorithm should select a policy with high worst-case value. We call this the pessimism principle of exploitation, as it is analogous to the widely-known optimism principle (Lattimore & Szepesvári, 2020) of exploration.333The optimism principle states that we should select a policy with high best-case value.
Our main contribution is a theoretical justification of the pessimism principle in FDPO, based on a bound that characterizes the regret of a generic decision-maker which optimizes a proxy objective. We further demonstrate how this bound may be used to derive principled algorithms. Note that the core novelty of our work is not the idea of pessimism, which is an intuitive concept that appears in a variety of contexts; rather, our contribution is a set of theoretical results rigorously explaining how pessimism is important in the specific setting of FDPO. An example conveying the intuition behind our results can be found in Appendix G.1.
We first analyze a family of non-pessimistic naïve FDPO algorithms, which estimate the environment from the dataset via maximum likelihood and then apply standard dynamic programming techniques. We prove a bound which shows that the worst-case suboptimality of these algorithms is guaranteed to be small when the dataset contains enough data that we are certain about the value of every possible policy. This is caused by the outsized impact of value overestimation errors on suboptimality, sometimes called the optimizer’s curse (Smith & Winkler, 2006). It is a fundamental consequence of ignoring the disconnect between the true environment and the picture painted by our limited observations. Importantly, it is not reliant on errors introduced by function approximation.
We contrast these findings with an analysis of pessimistic FDPO algorithms, which select a policy that maximizes some notion of worst-case expected return. We show that these algorithms do not require datasets which inform us about the value of every policy to achieve small suboptimality, due to the critical role that pessimism plays in preventing overestimation. Our analysis naturally leads to two families of principled pessimistic FDPO algorithms. We prove their improved suboptimality guarantees, and confirm our claims with experiments on a gridworld.
Finally, we extend one of our pessimistic algorithms to the deep learning setting. Recently, several deep-learning-based algorithms for fixed-dataset policy optimization have been proposed (Agarwal et al., 2019; Fujimoto et al., 2019; Kumar et al., 2019; Laroche et al., 2019; Jaques et al., 2019; Kidambi et al., 2020; Yu et al., 2020; Wu et al., 2019; Wang et al., 2020; Kumar et al., 2020; Liu et al., 2020). Our work is complementary to these results, as our contributions are conceptual, rather than algorithmic. Our primary goal is to theoretically unify existing approaches and motivate the design of pessimistic algorithms more broadly. Using experiments in the MinAtar game suite (Young & Tian, 2019), we provide empirical validation for the predictions of our analysis.
The problem of fixed-dataset policy optimization is closely related to the problem of reinforcement learning, and as such, there is a large body of work which contains ideas related to those discussed in this paper. We discuss these works in detail in Appendix E.
2 How Proxy Objectives Impact Regret
Our central theoretical contribution is a novel bound for generalized decision-making. Theorem 1 shows how the regret incurred by a decision-maker who optimizes a proxy objective is governed by the quality of the proxy. We consider an abstract decision-making task, whose goal is to choose an element from some space to maximize some objective .
Theorem 1.
For any space , objective , and proxy objective ,
where and . Furthermore, this bound is tight.
This result reveals a formal connection between prediction tasks and decision-making tasks. When the proxy objective is an estimator of the true objective, a reduction in estimation error (prediction) leads to reduced regret (decision-making). Crucially, this relationship is not straightforward: there is an asymmetry between the impact of overestimation errors, i.e., for which , and underestimation errors, where . To see this, we consider each of its terms in isolation:
The term labeled (a) reflects the degree to which the proxy is accurate on a near-optimal . For any choice , (a1) captures its suboptimality, and (a2) captures its underestimation error. Since (a) takes an infimum over , this term will be small whenever there is at least one reasonable choice whose score is not very underestimated by the proxy. On the other hand, the term labeled (b) corresponds to the largest overestimation error on any . Because it consists of a supremum over , it will be small only when no choices at all are overestimated much. Even a single large overestimation can lead to significant regret.
Current algorithms for prediction tasks have not been designed with this asymmetry in mind. As a result, when the function learned by such an algorithm is used as a proxy objective for a decision-making task, we observe a characteristic behavior: the choice we select turns out to have been overestimated, and we experience high regret. The remainder of this paper explores this effect in detail on one specific formalization of a decision-making problem: FDPO.
3 FDPO Background
We anticipate most readers will be familiar with the concepts and notation, which is fairly standard in the reinforcement learning literature. In the interest of space, we relegate a full presentation to Appendix A. Here, we briefly give an informal overview of the background necessary to understand the main results.
We represent the environment as a Markov Decision Process (MDP), denoted . We assume without loss of generality that , and denote its expectation as . represents the start-state distribution. Policies can act in the environment, represented by action matrix , which maps each state to the probability of each state-action when following . Value functions assign some real value to each state. We use to denote the value function which assigns the sum of discounted rewards in the environment when following policy . A dataset contains transitions sampled from the environment. From a dataset, we compute the empirical reward and transition functions, and , and the empirical policy, .
An important concept for our analysis is the value uncertainty function, denoted , which returns a high-probability upper-bound to the error of a value function derived from dataset . Certain value uncertainty functions are decomposable by states or state-actions, meaning they can be written as the weighted sum of more local uncertainties. See Appendix B for more detail.
Our goal is to analyze the suboptimality of a specific class of FDPO algorithms, called value-based FDPO algorithms, which have a straightforward structure: they use a fixed-dataset policy evaluation (FDPE) algorithm to assign a value to each policy, and then select the policy with the maximum value. Furthermore, we consider FDPE algorithms whose solutions satisfy a fixed-point equation. Thus, a fixed-point equation defines a FDPE objective, which in turn defines a value-based FDPO objective; we call the set of all algorithms that implement these objectives the family of algorithms defined by the fixed-point equation.
FDPO represents a specific instantiation of the more general decision-making framework in Theorem 1, where , , and . Thus, from Theorem 1:
Corollary 1 (Value-based FDPO suboptimality bound).
Consider any value-based fixed-dataset policy optimization algorithm , with fixed-dataset policy evaluation subroutine . For any policy and dataset , denote . The suboptimality of is bounded by
4 Naïve Algorithms
The goal of this section is to paint a high-level picture of the worst-case suboptimality guarantees of a specific family of non-pessimistic approaches, which we call naïve FDPO algorithms. Informally, the naïve approach is to take the limited dataset of observations at face value, treating it as though it paints a fully accurate picture of the environment. The value function of a naïve algorithm is an example of an asymmetry-agnostic proxy objective, as discussed in Section 2.
Definition 1.
A naïve algorithm is any algorithm in the family defined by the fixed-point function
Various FDPE and FDPO algorithms from this family could be described; in this work, we do not study these implementations in detail, although we do give pseudocode for some implementations in Appendix D.1.
One example of a naïve FDPO algorithm which can be found in the literature is certainty equivalence (Jiang, 2019a). The core ideas behind naïve algorithms can also be found in the function approximation literature, for example in FQI (Ernst et al., 2005; Jiang, 2019b). Additionally, when available data is held fixed, nearly all existing deep reinforcement learning algorithms are transformed into naïve value-based FDPO algorithms. For example, DQN (Mnih et al., 2015) with a fixed replay buffer is a naïve value-based FDPO algorithm.
Theorem 2 (Naïve FDPO suboptimality bound).
Consider any naïve value-based fixed-dataset policy optimization algorithm . Let be any value uncertainty function. With probability at least , the suboptimality of is bounded with probability at least by
The infimum term is small whenever there is some reasonably good policy with low value uncertainty. In practice, this condition can typically be satisfied, for example by including expert demonstrations in the dataset. On the other hand, the supremum term will only be small if we have low value uncertainty for all policies – a much more challenging requirement. This explains the behavior of pathological examples, e.g. in Appendix G.1, where performance is poor despite access to virtually unlimited amounts of data from a near-optimal policy. Such a dataset ensures that the first term will be small by reducing value uncertainty of the near-optimal data collection policy, but does little to reduce the value uncertainty of any other policy, leading the second term to be large.
However, although pathological examples exist, it is clear that this bound will not be tight on all environments. It is reasonable to ask: is it likely that this bound will be tight on real-world examples? We argue that it likely will be. We identify two properties that most real-world tasks share: (1) The set of policies is pyramidal: there are an enormous number of bad policies, many mediocre policies, a few good policies, etc. (2) Due to the size of the state space and cost of data collection, most policies have high value uncertainty.
Given that these assumptions hold, naïve algorithms will perform as poorly on most real-world environments as they do on pathological examples. Consider: there are many more policies than there is data, so there will be many policies with high value uncertainty; naïve algorithms will likely overestimate several of these policies, and erroneously select one; since good policies are rare, the selected policy will likely be bad. It follows that running naïve algorithms on real-world problems will typically yield suboptimality close to our worst-case bound. And, indeed, on deep RL benchmarks, which are selected due to their similarity to real-world settings, overestimation has been widely observed, typically correlated with poor performance (Bellemare et al., 2016; Van Hasselt et al., 2016; Fujimoto et al., 2019).
5 The Pessimism Principle
“Behave as though the world was plausibly worse than you observed it to be.” The pessimism principle tells us how to exploit our current knowledge to find the stationary policy with the best worst-case guarantee on expected return. Being pessimistic means reducing overestimation errors, but incurring additional underestimation errors; thus, pessimistic approaches optimize a proxy objective which exploits the asymmetry described in Section 2.
We consider two specific families of pessimistic algorithms, the uncertainty-aware pessimistic algorithms and proximal pessimistic algorithms, and bound the worst-case suboptimality of each. These algorithms each include a hyperparameter, , controlling the amount of pessimism, interpolating from fully-naïve to fully-pessimistic. (For a discussion of the implications of the latter extreme, see Appendix G.2.) Then, we compare the two families, and see how the proximal family is simply a trivial special case of the more general uncertainty-aware family of methods.
5.1 Uncertainty-Aware Pessimistic Algorithms
Our first family of pessimistic algorithms is the uncertainty-aware (UA) pessimistic algorithms. As the name suggests, this family of algorithms estimates the state-wise Bellman uncertainty and penalizes policies accordingly, leading to a pessimistic value estimate and a preference for policies with low value uncertainty.
Definition 2.
An uncertainty-aware pessimistic algorithm, with a Bellman uncertainty function and pessimism hyperparameter , is any algorithm in the family defined by the fixed-point function
This fixed-point function is simply the naïve fixed-point function penalized by the Bellman uncertainty. This can be interpreted as being pessimistic about the outcome of every action. Note that it remains to specify a technique to compute the Bellman uncertainty function, e.g. Appendix B.1, in order to get a concrete algorithm. It is straightforward to construct algorithms from this family by modifying naïve algorithms to subtract the penalty term. Similar algorithms have been explored in the safe RL literature (Ghavamzadeh et al., 2016; Laroche et al., 2019) and the robust MDP literature (Givan et al., 1997), where algorithms with high-probability performance guarantees are useful in the context of ensuring safety.
Theorem 3 (Uncertainty-aware pessimistic FDPO suboptimality bound).
Consider an uncertainty-aware pessimistic value-based fixed-dataset policy optimization algorithm . Let be any Bellman uncertainty function, be a corresponding value uncertainty function, and be any pessimism hyperparameter. The suboptimality of is bounded with probability at least by
Proof.
See Appendix C.7. ∎
This bound should be contrasted with our result from Theorem 2. With , the family of pessimistic algorithms reduces to the family of naïve algorithms, so the bound is correspondingly identical. We can add pessimism by increasing , and this corresponds to a decrease in the magnitude of the supremum term. When , there is no supremum term at all. In general, the optimal value of lies between the two extremes.
To further understand the power of this approach, it is illustrative to compare it to imitation learning. Consider the case where the dataset contains a small number of expert trajectories but also a large number of interactions from a random policy, i.e. when learning from suboptimal demonstrations (Brown et al., 2019). If the dataset contained only a small amount of expert data, then both an UA pessimistic FDPO algorithm and an imitation learning algorithm would return a high-value policy. However, the injection of sufficiently many random interactions would degrade the performance of imitation learning algorithms, whereas UA pessimistic algorithms would continue to behave similarly to the expert data.
5.2 Proximal Pessimistic Algorithms
The next family of algorithms we study are the proximal pessimistic algorithms, which implement pessimism by penalizing policies that deviate from the empirical policy. The name proximal was chosen to reflect the idea that these algorithms prefer policies which stay “nearby” to the empirical policy. Many FDPO algorithms in the literature, and in particular several recently-proposed deep learning algorithms (Fujimoto et al., 2019; Kumar et al., 2019; Laroche et al., 2019; Jaques et al., 2019; Wu et al., 2019; Liu et al., 2020), resemble members of the family of proximal pessimistic algorithms; see Appendix E. Also, another variant of the proximal pessimistic family, which uses state density instead of state-conditional action density, can be found in Appendix C.9.
Definition 3.
A proximal pessimistic algorithm with pessimism hyperparameter is any algorithm in the family defined by the fixed-point function
Theorem 4 (Proximal pessimistic FDPO suboptimality bound).
Consider any proximal pessimistic value-based fixed-dataset policy optimization algorithm . Let be any state-action-wise decomposable value uncertainty function, and be a pessimism hyperparameter. For any dataset , the suboptimality of is bounded with probability at least by
Proof.
See Appendix C.8. ∎
Once again, we see that as grows, the large supremum term shrinks; similarly, by Lemma 4, when we have , the supremum term is guaranteed to be non-positive.444Initially, it will contain , but this can be removed since it is not dependent on . The primary limitation of the proximal approach is the looseness of the value lower-bound. Intuitively, this algorithm can be understood as performing imitation learning, but permitting minor deviations. Constraining the policy to be near in distribution to the empirical policy can fail to take advantage of highly-visited states which are reached via many trajectories. In fact, in contrast to both the naïve approach and the UA pessimistic approach, in the limit of infinite data this approach is not guaranteed to converge to the optimal policy. Also, note that when , this algorithm is identical to imitation learning.
5.3 The Relationship Between Uncertainty-Aware and Proximal Algorithms
Though these two families may appear on the surface to be quite different, they are in fact closely related. A key insight of our theoretical work is that it reveals the important connection between these two approaches. Concretely: proximal algorithms are uncertainty-aware algorithms which use a trivial value uncertainty function.
To see this, we show how to convert an uncertainty-aware penalty into a proximal penalty. let be any state-action-wise decomposable value uncertainty function. For any dataset , we have
(Lemma 3) | ||||
(Lemma 4) |
We began with the uncertainty penalty. In the first step, we rewrote the uncertainty for into the sum of two terms: the uncertainty for , and the difference in uncertainty between and on various actions. In the second step, we chose our state-action-wise Bellman uncertainty to be , which is a trivial upper bound; we also upper-bound the signed policy difference with the total variation. This results in the proximal penalty.555When constructing the penalty, we can ignore the first term, which does not contain , and so is irrelevant to optimization.
Thus, we see that proximal penalties are equivalent to uncertainty-aware penalties which use a specific, trivial uncertainty function. This result suggests that uncertainty-aware algorithms are strictly better than their proximal counterparts. There is no looseness in this result: for any proximal penalty, we will always be able to find a tighter uncertainty-aware penalty by replacing the trivial uncertainty function with something tighter.
However, currently, proximal algorithms are quite useful in the context of deep learning. This is because the only uncertainty function that can currently be implemented for neural networks is the trivial uncertainty function. Until we discover how to compute uncertainties for neural networks, proximal pessimistic algorithms will remain the only theoretically-motivated family of algorithms.
6 Experiments



We implement algorithms from each family to empirically investigate whether their performance of follows the predictions of our bounds. Below, we summarize the key predictions of our theory.
-
•
Imitation. This algorithm simply learns to copy the empirical policy. It performs well if and only if the data collection policy performs well.
-
•
Naïve. This algorithm performs well only when almost no policies have high value uncertainty. This means that when the data is collected from any mostly-deterministic policy, performance of this algorithm will be poor, since many states will be missing data. Stochastic data collection improves performance. As the size of the dataset grows, this algorithm approaches optimality.
-
•
Uncertainty-aware. This algorithm performs well when there is data on states visited by near-optimal policies. This is the case when a small amount of data has been collected from a near-optimal policy, or a large amount of data has been collected from a worse policy. As the size of the dataset grows, this algorithm to approaches optimality. This approach outperforms all other approaches.
-
•
Proximal. This algorithm roughly mirrors the performance of the imitation approach, but improves upon it. As the size of the dataset grows, this algorithm does not approach optimality, as the penalty persists even when the environment’s dynamics are perfectly captured by the dataset.
Our experimental results qualitatively align with our predictions in both the tabular and deep learning settings, giving evidence that the picture painted by our theoretical analysis truly describes the FDPO setting. See Appendix D for pseudocode of all algorithms; see Appendix F for details on the experimental setup; see Appendix G.3 for additional experimental considerations for deep learning experiments that will be of interest to practicioners. For an open-source implementation, including full details suitable for replication, please refer to the code in the accompanying GitHub repository: github.com/jbuckman/tiopifdpo.
Tabular.
The first tabular experiment, whose results are shown in Figure 1(), compares the performance of the algorithms as the policy used to collect the dataset is interpolated from the uniform random policy to an optimal policy using -greedy. The second experiment, whose results are shown in Figure 1(a), compares the performance of the algorithms as we increase the size of the dataset from 1 sample to 200000 samples. In both experiments, we notice a qualitative difference between the trends of the various algorithms, which aligns with the predictions of our theory.
Neural network.
The results of these experiments can be seen in Figure 2. Similarly to the tabular experiments, we see that the naïve approach performs well when data is fully exploratory, and poorly when data is collected via an optimal policy; the pure imitation approach performs better when the data collection policy is closer to optimal. The pessimistic approach achieves the best of both worlds: it correctly imitates a near-optimal policy, but also learns to improve upon it somewhat when the data is more exploratory. One notable failure case is in Freeway, where the performance of the pessimistic approach barely improves upon the imitation policy, despite the naïve approach performing near-optimally for intermediate values of .
7 Discussion and Conclusion

In this work, we provided a conceptual and mathematical framework for thinking about fixed-dataset policy optimization. Starting from intuitive building blocks of uncertainty and the over-under decomposition, we showed the core issue with naïve approaches, and introduced the pessimism principle as the defining characteristic of solutions. We described two families of pessimistic algorithms, uncertainty-aware and proximal. We see theoretically that both of these approaches have advantages over the naïve approach, and observed these advantages empirically. Comparing these two families of pessimistic algorithms, we see both theoretically and empirically that uncertainty-aware algorithms are strictly better than proximal algorithms, and that proximal algorithms may not yield the optimal policy, even with infinite data.
Future directions.
Our results indicate that research in FDPO should not focus on proximal algorithms. The development of neural uncertainty estimation techniques will enable principled uncertainty-aware deep learning algorithms. As is evidenced by our tabular results, we expect these approaches to yield dramatic performance improvements, rendering the proximal family (Kumar et al., 2019; Fujimoto et al., 2019; Laroche et al., 2019; Kumar et al., 2020) obsolete.
On ad-hoc solutions.
It is undoubtably disappointing to see that proximal algorithms, which are far easier to implement, are fundamentally limited. It is tempting to propose various ad-hoc solutions to mitigate the flaws of proximal pessimistic algorithms in practice. For example, one might consider tuning ; however, doing the tuning requires evaluating each policy in the environment, which involves gaining information by interacting with the environment, which is not permitted by the problem setting. Or, one might consider e.g. an adaptive pessimism hyperparameter which decays with the size of the dataset; however, for such a penalty to be principled, it must be based on an uncertainty function, at which point we may as well just use an uncertainty-aware algorithm.
Stochastic policies.
One surprising property of pessimistic algorithms is that the optimal policy is often stochastic. For the penalty of proximal pessimistic algorithms, it is easy to see that this may be the case for any non-deterministic empirical policy; for UA pessimsitic algorithms, it is dependent on the choice of Bellman uncertainty function, but in general may hold (see Appendix B.2 for the derivation of a Bellman uncertainty function with this property). This observation lends mathematical rigor to the intuition that agents should ‘hedge their bets’ in the face of epistemic uncertainty. This property also means that the simple approach of selecting the argmax action is no longer adequate for policy improvement. In Appendix D.2.2 we discuss a policy improvement procedure that takes into account the proximal penalty to find the stochastic optimal policy.
Implications for RL.
Finally, due to the close connection between the FDPO and RL settings, this work has implications for deep reinforcement learning. Many popular deep RL algorithms utilize a replay buffer to break the correlation between samples in each minibatch (Mnih et al., 2015). However, since these algorithms typically alternate between collecting data and training the network, the replay buffer can also be viewed as a ‘temporarily fixed’ dataset during the training phase. These algorithms are often very sensitive to hyperparemters; in particular, they perform poorly when the number of learning steps per interaction is large (Fedus et al., 2020). This effect can be explained by our analysis: additional steps of learning cause the policy to approach its naïve FDPO fixed-point, which has poor worst-case suboptimality. A pessimistic algorithm with a better fixed-point could therefore allow us to train more per interaction, improving sample efficiency. A potential direction of future work is therefore to incorporate pessimism into deep RL.
Acknowledgements.
The authors thank Alex Slivkins, Romain Laroche, Emmanuel Bengio, Simon Ramstedt, George Tucker, Aviral Kumar, Dave Meger, Ahmed Touati, and Bo Dai for providing valuable feedback, as well as Doina Precup, Jason Eisner, Philip Thomas, Saurabh Kumar, and Marek Petrik for helpful discussions. The authors are grateful for financial support from the CIFAR Canada AI Chair program.
References
- Agarwal et al. (2019) Rishabh Agarwal, Dale Schuurmans, and Mohammad Norouzi. Striving for simplicity in off-policy deep reinforcement learning. arXiv preprint arXiv:1907.04543, 2019.
- Antos et al. (2007) András Antos, Csaba Szepesvári, and Rémi Munos. Value-iteration based fitted policy iteration: learning with a single trajectory. In 2007 IEEE international symposium on approximate dynamic programming and reinforcement learning, pp. 330–337. IEEE, 2007.
- Bellemare et al. (2016) Marc G Bellemare, Georg Ostrovski, Arthur Guez, Philip S Thomas, and Rémi Munos. Increasing the action gap: New operators for reinforcement learning. In Thirtieth AAAI Conference on Artificial Intelligence, 2016.
- Brown et al. (2019) Daniel S. Brown, Wonjoon Goo, Prabhat Nagarajan, and Scott Niekum. Extrapolating beyond suboptimal demonstrations via inverse reinforcement learning from observations. In Proceedings of the International Conference on Machine Learning, 2019.
- Cohen & Hutter (2020) Michael K Cohen and Marcus Hutter. Pessimism about unknown unknowns inspires conservatism. In Conference on Learning Theory, pp. 1344–1373. PMLR, 2020.
- Ernst et al. (2005) Damien Ernst, Pierre Geurts, and Louis Wehenkel. Tree-based batch mode reinforcement learning. Journal of Machine Learning Research, 6:503–556, 2005.
- Fedus et al. (2020) William Fedus, Prajit Ramachandran, Rishabh Agarwal, Yoshua Bengio, Hugo Larochelle, Mark Rowland, and Will Dabney. Revisiting fundamentals of experience replay. arXiv preprint arXiv:2007.06700, 2020.
- Fujimoto et al. (2019) Scott Fujimoto, David Meger, and Doina Precup. Off-policy deep reinforcement learning without exploration. In International Conference on Machine Learning, pp. 2052–2062, 2019.
- Ghavamzadeh et al. (2016) Mohammad Ghavamzadeh, Marek Petrik, and Yinlam Chow. Safe policy improvement by minimizing robust baseline regret. In Advances in Neural Information Processing Systems, pp. 2298–2306, 2016.
- Givan et al. (1997) Robert Givan, Sonia Leach, and Thomas Dean. Bounded parameter markov decision processes. In European Conference on Planning, pp. 234–246. Springer, 1997.
- Goyal & Grand-Clement (2018) Vineet Goyal and Julien Grand-Clement. Robust markov decision process: Beyond rectangularity. arXiv preprint arXiv:1811.00215, 2018.
- Hu et al. (2020) Wei Hu, Lechao Xiao, and Jeffrey Pennington. Provable benefit of orthogonal initialization in optimizing deep linear networks. arXiv preprint arXiv:2001.05992, 2020.
- Hussein et al. (2017) Ahmed Hussein, Mohamed Medhat Gaber, Eyad Elyan, and Chrisina Jayne. Imitation learning: A survey of learning methods. ACM Computing Surveys (CSUR), 50(2):1–35, 2017.
- Iyengar (2005) Garud N Iyengar. Robust dynamic programming. Mathematics of Operations Research, 30(2):257–280, 2005.
- Jaques et al. (2019) Natasha Jaques, Asma Ghandeharioun, Judy Hanwen Shen, Craig Ferguson, Agata Lapedriza, Noah Jones, Shixiang Gu, and Rosalind Picard. Way off-policy batch deep reinforcement learning of implicit human preferences in dialog. arXiv preprint arXiv:1907.00456, 2019.
- Jiang (2019a) Nan Jiang. Note on certainty equivalence, 2019a.
- Jiang (2019b) Nan Jiang. Note on fqi, 2019b.
- Jiang & Huang (2020) Nan Jiang and Jiawei Huang. Minimax confidence interval for off-policy evaluation and policy optimization. arXiv preprint arXiv:2002.02081, 2020.
- Kakade & Langford (2002) Sham Kakade and John Langford. Approximately optimal approximate reinforcement learning. In ICML, volume 2, pp. 267–274, 2002.
- Kidambi et al. (2020) Rahul Kidambi, Aravind Rajeswaran, Praneeth Netrapalli, and Thorsten Joachims. Morel: Model-based offline reinforcement learning. arXiv preprint arXiv:2005.05951, 2020.
- Kumar et al. (2019) Aviral Kumar, Justin Fu, Matthew Soh, George Tucker, and Sergey Levine. Stabilizing off-policy q-learning via bootstrapping error reduction. In Advances in Neural Information Processing Systems, pp. 11784–11794, 2019.
- Kumar et al. (2020) Aviral Kumar, Aurick Zhou, George Tucker, and Sergey Levine. Conservative q-learning for offline reinforcement learning. arXiv preprint arXiv:2006.04779, 2020.
- Lange et al. (2012) Sascha Lange, Thomas Gabel, and Martin Riedmiller. Batch reinforcement learning. Reinforcement learning, pp. 45–73, 2012.
- Laroche et al. (2019) Romain Laroche, Paul Trichelair, and Remi Tachet Des Combes. Safe policy improvement with baseline bootstrapping. In International Conference on Machine Learning, pp. 3652–3661, 2019.
- Lattimore & Szepesvári (2020) Tor Lattimore and Csaba Szepesvári. Bandit algorithms. Cambridge University Press, 2020.
- Levine et al. (2020) Sergey Levine, Aviral Kumar, George Tucker, and Justin Fu. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. arXiv preprint arXiv:2005.01643, 2020.
- Liu et al. (2020) Yao Liu, Adith Swaminathan, Alekh Agarwal, and Emma Brunskill. Provably good batch reinforcement learning without great exploration. arXiv preprint arXiv:2007.08202, 2020.
- Maurer & Pontil (2009) Andreas Maurer and Massimiliano Pontil. Empirical bernstein bounds and sample variance penalization. arXiv preprint arXiv:0907.3740, 2009.
- Mnih et al. (2015) Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare, Alex Graves, Martin Riedmiller, Andreas K Fidjeland, Georg Ostrovski, et al. Human-level control through deep reinforcement learning. Nature, 518(7540):529–533, 2015.
- Munos (2007) Rémi Munos. Performance bounds in l_p-norm for approximate value iteration. SIAM journal on control and optimization, 46(2):541–561, 2007.
- Nadjahi et al. (2019) Kimia Nadjahi, Romain Laroche, and Rémi Tachet des Combes. Safe policy improvement with soft baseline bootstrapping. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 53–68. Springer, 2019.
- Nilim & El Ghaoui (2005) Arnab Nilim and Laurent El Ghaoui. Robust control of markov decision processes with uncertain transition matrices. Operations Research, 53(5):780–798, 2005.
- Puterman (2014) Martin L Puterman. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 2014.
- Riedmiller (2005) Martin Riedmiller. Neural fitted q iteration–first experiences with a data efficient neural reinforcement learning method. In European Conference on Machine Learning, pp. 317–328. Springer, 2005.
- Schulman et al. (2015) John Schulman, Sergey Levine, Pieter Abbeel, Michael Jordan, and Philipp Moritz. Trust region policy optimization. In International conference on machine learning, pp. 1889–1897, 2015.
- Schulman et al. (2017) John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- Simão et al. (2019) Thiago D Simão, Romain Laroche, and Rémi Tachet des Combes. Safe policy improvement with an estimated baseline policy. arXiv preprint arXiv:1909.05236, 2019.
- Smith & Winkler (2006) James E Smith and Robert L Winkler. The optimizer’s curse: Skepticism and postdecision surprise in decision analysis. Management Science, 52(3):311–322, 2006.
- Strehl & Littman (2008) Alexander L Strehl and Michael L Littman. An analysis of model-based interval estimation for markov decision processes. Journal of Computer and System Sciences, 74(8):1309–1331, 2008.
- Sutton & Barto (2018) Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018.
- Swaminathan & Joachims (2015) Adith Swaminathan and Thorsten Joachims. Batch learning from logged bandit feedback through counterfactual risk minimization. The Journal of Machine Learning Research, 16(1):1731–1755, 2015.
- Taleghan et al. (2015) Majid Alkaee Taleghan, Thomas G Dietterich, Mark Crowley, Kim Hall, and H Jo Albers. Pac optimal mdp planning with application to invasive species management. The Journal of Machine Learning Research, 16(1):3877–3903, 2015.
- Thomas et al. (2015a) Philip Thomas, Georgios Theocharous, and Mohammad Ghavamzadeh. High confidence policy improvement. In International Conference on Machine Learning, pp. 2380–2388, 2015a.
- Thomas et al. (2015b) Philip S Thomas, Georgios Theocharous, and Mohammad Ghavamzadeh. High-confidence off-policy evaluation. In Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015b.
- Thomas et al. (2019) Philip S Thomas, Bruno Castro da Silva, Andrew G Barto, Stephen Giguere, Yuriy Brun, and Emma Brunskill. Preventing undesirable behavior of intelligent machines. Science, 366(6468):999–1004, 2019.
- Van Hasselt et al. (2016) Hado Van Hasselt, Arthur Guez, and David Silver. Deep reinforcement learning with double q-learning. In Thirtieth AAAI conference on artificial intelligence, 2016.
- Walsh et al. (2011) Thomas J Walsh, Daniel K Hewlett, and Clayton T Morrison. Blending autonomous exploration and apprenticeship learning. In Advances in Neural Information Processing Systems, pp. 2258–2266, 2011.
- Wang et al. (2020) Ziyu Wang, Alexander Novikov, Konrad Żołna, Jost Tobias Springenberg, Scott Reed, Bobak Shahriari, Noah Siegel, Josh Merel, Caglar Gulcehre, Nicolas Heess, et al. Critic regularized regression. arXiv preprint arXiv:2006.15134, 2020.
- Weissman et al. (2003) Tsachy Weissman, Erik Ordentlich, Gadiel Seroussi, Sergio Verdu, and Marcelo J Weinberger. Inequalities for the l1 deviation of the empirical distribution. Hewlett-Packard Labs, Tech. Rep, 2003.
- Wiesemann et al. (2013) Wolfram Wiesemann, Daniel Kuhn, and Berç Rustem. Robust markov decision processes. Mathematics of Operations Research, 38(1):153–183, 2013.
- Wu et al. (2019) Yifan Wu, George Tucker, and Ofir Nachum. Behavior regularized offline reinforcement learning. arXiv preprint arXiv:1911.11361, 2019.
- Young & Tian (2019) Kenny Young and Tian Tian. Minatar: An atari-inspired testbed for thorough and reproducible reinforcement learning experiments. arXiv preprint arXiv:1903.03176, 2019.
- Yu et al. (2020) Tianhe Yu, Garrett Thomas, Lantao Yu, Stefano Ermon, James Zou, Sergey Levine, Chelsea Finn, and Tengyu Ma. Mopo: Model-based offline policy optimization. arXiv preprint arXiv:2005.13239, 2020.
Appendix A Background
We write vectors using bold lower-case letters, , and matrices using upper-case letters, . To refer to individual cells of a vector or rows of a matrix, we use function notation, . We write the identity matrix as . We use the notation to denote the average value of a function under a distribution , i.e. for any space , distribution , and function , we have .
When applied to vectors or matrices, we use to denote element-wise comparison. Similarly, we use to denote the element-wise absolute value of a vector: . We use to denote the element-wise maximum of and the zero vector. To denote the total variation distance between two probability distributions, we use . When and are conditional probability distributions, we adopt the convention , i.e., the vector of total variation distances conditioned on each .
Markov Decision Processes.
We represent the environment with which we are interacting as a Markov Decision Process (MDP), defined in standard fashion: . and denote the state and action space, which we assume are discrete. We use as the shorthand for the joint state-action space. The reward function maps state-action pairs to distributions over the unit interval, while the transition function maps state-action pairs to distributions over next states. Finally, is the distribution over initial states. We use to denote the expected reward function, , which can also be interpreted as a vector . Similarly, note that can be described as , which can be represented as a stochastic matrix . In order to emphasize that these reward and transition functions correspond to the true environment, we sometimes equivalently denote them as . To denote the vectors of a constant whose sizes are the state and state-action space, we use a single dot to mean state and two dots to mean state-action, e.g., and . A policy defines a distribution over actions, conditioned on a state. We denote the space of all possible policies as . We define an “activity matrix” for each policy, , which encodes the state-conditional state-action distribution of , by letting if , otherwise . Acting in the MDP according to can thus be represented by or . We define a value function as any or whose output is bounded by . Note that this is a slight generalization of the standard definition (Sutton & Barto, 2018) since it accepts a policy as an input. We use the shorthand and to denote the result of applying a value function to a specific policy, which can also be represented as a vector, and . To denote the output of an arbitrary value function on an arbitrary policy, we use unadorned and . The expected return of an MDP , denoted or , is the discounted sum of rewards acquired when interacting with the environment:
Note that . An optimal policy of an MDP, which we denote , is a policy for which the expected return is maximized under the initial state distribution: . The statewise expected returns of an optimal policy can be written as . Of particular interest are value functions whose outputs obey fixed-point relationships, for some . The Bellman consistency equation for a policy , which uniquely identifies the vector of expected returns for , since is the only vector for which holds. Finally, for any state , the probability of being in the state after time steps when following policy is . Furthermore, . We refer to as the discounted visitation of .
Datasets.
We next introduce basic concepts that are helpful for framing the problem of fixed-dataset policy optimization. We define a dataset of transitions , and denote the space of all datasets as . In this work, we specifically consider datasets sampled from a data distribution ; for example, the distribution of state-actions reached by following some stationary policy. We use to denote constructing a dataset of tuples , by first sampling each , and then sampling and i.i.d. from the environment reward function and transition function respectively, i.e. each and .666Note that this is in some sense a simplifying assumption. In practice, datasets will typically be collected using a trajectory from a non-stationary policy, rather than i.i.d. sampling from the stationary distribution of a stationary policy. This greatly complicates the analysis, so we do not consider that setting in this work. We sometimes index using function notation, using to denote the multiset of all such that . We use to denote the vectors of counts, that is, . We sometimes use state-wise versions of these vectors, which we denote with . It is further useful to consider the maximum-likelihood reward and transition functions, computed by averaging all rewards and transitions observed in the dataset for each state-action. To this end, we define empirical reward vector and empirical transition matrix at all state-actions for which . Where with , there is no clear way to define the maximum-likelihood estimates of reward and transition, so we do not specify them. All our results hold no matter how these values are chosen, so long as and is stochastic. The empirical policy of a dataset is defined as except where , where it can similarly be any valid action distribution. The empirical visitation distribution of a dataset is computed in the same way as the visitation distribution, but with replacing , i.e. .
Problem setting.
The primary focus of this work is on the properties of fixed-dataset policy optimization (FDPO) algorithms. These algorithms take the form of a function , which maps from a dataset to a policy.777This formulation hides a dependency on , the start-state distribution of the MDP. In general, can be estimated from the dataset , but this estimation introduces some error that affects the analysis. In this work, we assume for analytical simplicity that is known a priori. Technically, this means that it must be provided as an input to . We hide this dependency for notational clarity. Note that in this work, we consider , so the dataset is a random variable, and therefore is also a random variable. The goal of any FDPO algorithm is to output a policy with minimum suboptimality, i.e. maximum return. Suboptimality is a random variable dependent on , computed by taking the difference between the expected return of an optimal policy and the learned policy under the initial state distribution,
A related concept is the fixed-dataset policy evaluation algorithm, which is any function , which uses a dataset to compute a value function. In this work, we focus our analysis on value-based FDPO algorithms, the subset of FDPO algorithms that utilize FDPE algorithms.888Intuitively, these algorithms use a policy evaluation subroutine to convert a dataset into a value function, and return an optimal policy according to that value function. Importantly, this definition constrains only the objective of the approach, not its actual algorithmic implementation, i.e., it includes algorithms which never actually invoke the FDPE subroutine. For many online model-free reinforcement learning algorithms, including policy iteration, value iteration, and Q-learning, we can construct closely analogous value-based FDPO algorithms; see Appendix D.1. Furthermore, model-based techniques can be interpreted as using a model to implicitly define a value function, and then optimizing that value function; our results also apply to model-based approaches. A value-based FDPO algorithm with FDPE subroutine is any algorithm with the following structure:
We define a fixed-point family of algorithms, sometimes referred to as just a family, in the following way. Any family called family is based on a specific fixed-point identity . We use the notation to denote any FDPE algorithm whose output obeys . Finally, refers to any value-based FDPO algorithm whose subroutine is . We call the set of all algorithms that could implement the family of FDPE algorithms, and the set of all algorithms that could implement as the family of FDPO algorithms.
Appendix B Uncertainty
Epistemic uncertainty measures an agent’s knowledge about the world, and is therefore a core concept in reinforcement learning, both for exploration and exploitation; it plays an important role in this work. Most past analyses in the literature implicitly compute some form of epistemic uncertainty to derive their bounds. An important analytical choice in this work is to cleanly separate the estimation of epistemic uncertainty from the problem of decision making. Our approach is to first define a notion of uncertainty as a function with certain properties, then assume that such a function exists, and provide the remainder of our technical results under such an assumption. We also describe several approaches to computing uncertainty.
Definition 4 (Uncertainty).
A function is a state-action-wise Bellman uncertainty function if for a dataset , it obeys with probability at least for all policies and all values :
A function is a state-wise Bellman uncertainty function if for a dataset and policy , it obeys with probability at least for all policies and all values :
A function is a value uncertainty function if for a dataset and policy , it obeys with probability at least for all policies and all values :
We refer to the quantity returned by an uncertainty function as uncertainty, e.g., value uncertainty refers to the quantity returned by a value uncertainty function.
Given a state-action-wise Bellman uncertainty function , it is easy to verify that is a state-wise Bellman uncertainty function. Similarly, given a state-wise Bellman uncertainty function , it is easy to verify that is a value uncertainty function. Uncertainty functions which can be constructed in this way are called decomposable.
Definition 5 (Decomposablility).
A state-wise Bellman uncertainty function is state-action-wise decomposable if there exists a state-action-wise Bellman uncertainty function such that . A value uncertainty function is state-wise decomposable if there exists a state-wise Bellman uncertainty function such that , and further, it is state-action-wise decomposable if that is itself state-action-wise decomposable.
Our definition of Bellman uncertainty captures the intuitive notion that the uncertainty at each state captures how well an approximate Bellman update matches the true environment. Value uncertainty represents the accumulation of these errors over all future timesteps.
How do these definitions correspond to our intuitions about uncertainty?
An application of the environment’s true Bellman update can be viewed as updating the value of each state to reflect information about its future. However, no algorithm in the FDPO setting can apply such an update, because the true environment dynamics are unknown. We may use the dataset to estimate what such an update would look like, but since the limited information in the dataset may not fully specify the properties of the environment, this update will be slightly wrong. It is intuitive to say that the uncertainty at each state corresponds to how well the approximate update matches the truth. Our definition of Bellman uncertainty captures precisely this notion. Further, value uncertainty represents the accumulation of these errors over all future timesteps.
How can we algorithmically implement uncertainty functions?
In other words, how can we compute a function with the properties required for Definition 4, using only the information in the dataset? All forms of uncertainty are upper-bounds to a quantity, and a tighter upper-bound means that other results down the line which leverage this quantity will be improved. Therefore, it is worth considering this question carefully.
A trivial approach to computing any uncertainty function is simply to return for all and . But although this is technically a valid uncertainty function, it is not very useful, because it does not concentrate with data and does not distinguish between certain and uncertain states. It is very loose and so leads to poor guarantees.
In tabular environments, one way to implement a Bellman uncertainty function is to use a concentration inequality. Depending on which concentration inequality is used, many Bellman uncertainty functions are possible. These approaches lead to Bellman uncertainty which is lower at states with more data, typically in proportion to the square root of the count. To illustrate how to do this, we show in Appendix B.1 how a basic application of Hoeffding’s inequality can be used to derive a state-action-wise Bellman uncertainty. In Appendix B.2, we show an alternative application of Hoeffding’s which results in a state-wise Bellman uncertainty, which is a tighter bound on error. In Appendix B.3, we discuss other techniques which may be useful in tightening further.
When the value function is represented by a neural network, it is not currently known how to implement a Bellman uncertainty function. When an empirical Bellman update is applied to a neural network, the change in value of any given state is impacted by generalization from other states. Therefore, the counts are not meaningful, and concentration inequalities are not applicable. In the neural network literature, many “uncertainty estimation techniques” have been proposed, which capture something analogous to an intuitive notion of uncertainty; however, none are principled enough to be useful in computing Bellman uncertainty.
B.1 State-action-wise Bound
We seek to construct an such that with probability at least .
Firstly, let’s consider the simplest possible bound. is bounded in , so both and must be as well. Thus, their difference is also bounded:
Next, consider that for any , the expression can be equivalently expressed as an expectation of random variables,
each with expected value
Note also that each of these random variables is bounded . Thus, Hoeffding’s inequality tells us that this mean of bounded random variables must be close to their expectation with high probability. By invoking Hoeffding’s at each of the state-actions, and taking a union bound, we see that with probability at least ,
We can left-multiply and rearrange to get:
Finally, we simply intersect this bound with the bound from earlier. Thus, we see that with probability at least ,
B.2 State-wise Bound
This bound is similar to the previous, but uses Hoeffding’s to bound the value at each state all at once, rather than bounding the value at each state-action.
Choose a collection of possible state-local policies . Each state-local policy is a member of the action simplex.
For any and , the expression can be equivalently expressed as a mean of random variables,
each with expected value
Note also that each of these random variables is bounded . Thus, Hoeffding’s inequality tells us that this sum of bounded random variables must be close to its expectation with high probability. By invoking Hoeffding’s at each of the states and local policies, and taking a union bound, we see that with probability at least ,
where the term refers to the elementwise square. Finally, we once again intersect with , yielding that with probability at least ,
Comparing this to the Bellman uncertainty function in Appendix B.1, we see two differences. Firstly, we have replaced a factor of with , typically loosening the bound somewhat (depending on choice of considered local policies). Secondly, has now moved inside of the square root; since the square root is concave, Jensen’s inequality says that
and so this represents a tightening of the bound.
When is the set of deterministic policies, this bound is equivalent to that of Appendix B.1. This can easily be seen by noting that for a deterministic policy, all elements of are either 1 or 0, and so
and also that the size of the set of deterministic policies is exactly .
An important property of this bound is that it shows that stochastic policies can be often be evaluated with lower error than deterministic policies. We prove this by example. Consider an MDP with a single state and two actions , and a dataset with . We can parameterize the policy by a single number by setting . The size of this bound will be proportional to , and setting the derivative equal to zero, we see that the minimum is . (Of course, we would need to increase the size of our local policy set to include this, in order to be able to actually select it, and doing so will increase the overall bound; this example only shows that for a given policy set, the selected policy may in general be stochastic.)
Finding the optimum for larger local policy sets is non-trivial, so we leave a full treatment of algorithms which leverage this bound for future work.
B.3 Other Bounds
There are a few other paths by which this bound can be made tighter still. The above bounds take an extra factor of because we bound the overall return, rather than simply bounding the reward and transition functions. This was done because a bound on the transition function would add a cost of . However, this can be mitigated by intersecting the above confidence interval with a Good-Turing interval, as proposed in Taleghan et al. (2015). Doing so will cause the bound to concentrate much more quickly in MDPs where the transition function is relatively deterministic. We expect this to be the case for most practical MDPs.
Similarly, empirical Bernstein confidence intervals can be used in place of Hoeffding’s, to increase the rate of concentration for low-variance rewards and transitions (Maurer & Pontil, 2009), leading to improved performance in MDPs where these are common.
Finally, we may be able to apply a concentration inequality in a more advanced fashion to compute a value uncertainty function which is not statewise decomposable: we bound some useful notion of value error directly, rather than computing the Bellman uncertainty function and taking the visitation-weighted sum. This would result in an overall tighter bound on value uncertainty by hedging over data across multiple timesteps. However, in doing so, we would sacrifice the monotonic improvement property needed for convergence of algorithms like policy iteration. This idea has a parallel in the robust MDP literature. The bounds in Appendix B.1 can be seen as constructing an sa-rectangular robust MDP, whereas Appendix B.2 is similar to constructing an s-rectangular robust MDP Wiesemann et al. (2013). More recently, approaches have been proposed which go beyond s-rectangular (Goyal & Grand-Clement, 2018), and such approaches likely have natural parallels in implementing value uncertainty functions.
Appendix C Proofs
C.1 Proof of Theorem 1
(valid for any ) | |||||
(using ) | |||||
Since the above holds for all , | |||||
(using ) | |||||
C.2 Tightness of Theorem 1
We show that the bound given in Theorem 1 is tight via a simple example.
Proof.
Let , , , , . Breaking the argmax tie with , we see that the two sides are equal. ∎
C.3 Residual Visitation Lemma
We prove a basic lemma showing that the error of any value function is equal to its one-step Bellman residual, summed over its visitation distribution. Though this result is well-known, it is not clearly stated elsewhere in the literature, so we prove it here for clarity.
Lemma 1.
For any MDP and policy , consider the Bellman fixed-point equation given by, let be defined as the unique value vector such that , and let be any other value vector. We have
(1) | ||||
(2) | ||||
(3) |
Proof.
Thus, we see . An identical proof can be completed starting with , leading to the desired result. ∎
C.4 Naïve FDPE Error Bound
We show how the error of naïve FDPE algorithms is bounded by the value uncertainty of state-actions visited by the policy under evaluation. Next, in Section 4, we use this bound to derive a suboptimality guarantee for naïve FDPO.
Lemma 2 (Naïve FDPE error bound).
Consider any naïve fixed-dataset policy evaluation algorithm . For any policy and dataset , denote . Let be any value uncertainty function. The following component-wise bound holds with probability at least :
Proof.
Notice that the naïve fixed-point function is equivalent to the Bellman fixed-point equation for a specific MDP: the empirical MDP defined by . Thus, invoking Lemma 1, for any values we have
Since is a value vector, this immediately implies that
Since is the solution to the Bellman consistency fixed-point,
Using the definition of a value uncertainty function , we arrive at
completing the proof. ∎
Thus, reducing value uncertainty improves our guarantees on evaluation error. For any fixed policy, value uncertainty can be reduced by reducing the Bellman uncertainty on states visited by that policy. In the tabular setting this means observing more interactions from the state-actions that that policy visits frequently. Conversely, for any fixed dataset, we will have a certain Bellman uncertainty in each state, and policies mostly visit low-Bellman-uncertainty states can be evaluated with lower error.999In the function approximation setting, we do not necessarily need to observe an interaction with a particular state-action to reduce our Bellman uncertainty on it. This is because observing other state-actions may allow us to reduce Bellman uncertainty through generalization. Similarly, the most-certain policy for a fixed dataset may not be the policy for which we have the most data, but rather, the policy which our dataset informs us about the most.
Our bound differs from prior work (Jiang, 2019a; Ghavamzadeh et al., 2016) in that it is significantly more fine-grained. We provide a component-wise bound on error, whereas previous results bound the norm. Furthermore, our bounds are sensitive to the Bellman uncertainty in each individual reward and transition, rather than only relying on the most-uncertain. As a result, our bound does not require all states to have the same number of samples, and is non-vacuous even in the case where some state-actions have no data.
Our bound can also be viewed as an extension of work on approximate dynamic programming. In that setting, the literature contains fine-grained results on the accumulation of local errors (Munos, 2007). However, those results are typically understood as applying to errors induced by approximation via some limited function class. Our bound can be seen as an application of those ideas to the case where errors are induced by limited observations.
C.5 Relative Value Uncertainty
The key to the construction of proximal pessimistic algorithms is the relationship between the value uncertainties of any two policies , for any state-action-wise decomposable value function.
Lemma 3 (Relative value uncertainty).
For any two policies , and any state-action-wise decomposable value uncertainty , we have
Proof.
Firstly, note that since the value uncertainty function is state-action-wise decomposable, we can express it in a Bellman-like form for some state-wise Bellman uncertainty as . Further, is itself state-action-wise decomposable, so it can be written as .
We can use this to derive the following relationship.
Next, we bound the difference between the value uncertainty of and .
This is a geometric series, so . Left-multiplying by we arrive at the desired result. ∎
C.6 Naïve FDPE Relative Error Bound
Lemma 4 (Naïve FDPE relative error bound).
Consider any naïve fixed-dataset policy evaluation algorithm . For any policy and dataset , denote . Then, for any other policy , the following bound holds with probability at least :
Proof.
The goal of this section is to construct an error bound for which we can optimize without needing to compute any uncertainties. To do this, we must replace this quantity with a looser upper bound.
First, consider a state-action-wise decomposable value uncertainty function . We have for some , where for some .
Note that state-action-wise Bellman uncertainty can be trivially bounded as . Also, since , any state-action-wise decomposable value uncertainty can be trivially bounded as .
We now invoke Lemma 3. We then substitute the above bounds into the second term, after ensuring that all coefficients are positive.
The third-to-last step follows from the fact that is stochastic. The final step follows from the fact that the positive and negative components of the state-wise difference between policies must be symmetric, so is precisely equivalent to the state-wise total variation distance, . Thus, we have
Finally, invoking Lemma 2, we arrive at the desired result. ∎
C.7 Suboptimality of Uncertainty-Aware Pessimistic FDPO Algorithms
Let . From the definition of the UA family, we have the fixed-point property , and the standard geometric series rearrangement yields . From here, we see:
We now use this to bound overestimation and underestimation error of by invoking Lemma 2, which holds with probability at least . First, for underestimation, we see:
and thus, . Next, for overestimation, we see:
and thus, . Substituting these bounds into Corollary 1 gives the desired result.
C.8 Suboptimality of Proximal Pessimsitic FDPO Algorithms
Let . From the definition of the proximal family, we have the fixed-point property
and the standard geometric series rearrangement yields
From here, we see:
Next, we define a new family of FDPE algorithms,
We use Lemma 2, which holds with probability at least , to bound the overestimation and underestimation. First, the underestimation:
Next, we analagously bound the overestimation:
We can now invoke Corollary 1 to bound the suboptimality of any value-based FDPO algorithm which uses , which we denote with . Crucially, note that since is not dependent on , it can be removed from the infimum and supremum terms, and cancels. Substituting and rearranging, we see that with probability at least ,
Finally, we see that FDPO algorithms which use as their subroutine will return the same policy as FDPO algorithms which use . First, we once again use the property that is not dependent on . Second, we note that since the total visitation of every policy sums to , we know for all , and thus it is also not dependent on .
Thus, the suboptimality of must be equivalent to that of , leading to the desired result.
C.9 State-wise Proximal Pessimistic Algorithms
In the main body of the work, we focus on proximal pessimistic algorithms which are based on a state-conditional density model, since this approach is much more common in the literature. However, it is also possible to derive proximal pessimistic algorithms which use state-action densities, which have essentially the same properties. In this section we briefly provide the main results.
In this section, we use to indicate the state visitation distribution of any policy .
Definition 6.
A state-action-density proximal pessimistic algorithm with pessimism hyperparameter is any algorithm in the family defined by the fixed-point function
Theorem 5 (State-action-density proximal pessimistic FDPO suboptimality bound).
Consider any state-action-density proximal pessimistic value-based fixed-dataset policy optimization algorithm . Let be any state-action-wise decomposable value uncertainty function, and be a pessimism hyperparameter. For any dataset , the suboptimality of is bounded with probability at least by
Appendix D Algorithms
In the main body of the text, we primarily focus on families of algorithms, rather than specific members. In this section, we provide pseudocode from example algorithms in each family. The majority of these algorithms are simple empirical extensions of well-studied algorithms, so we do not study their properties (e.g. convergence) in detail. The sets of algorithms described here are not intended to be comprehensive, but rather, a few straightforward examples to illustrate key concepts and inspire further research. To simplify presentation, we avoid hyperparameters, e.g. learning rate, wherever possible.
D.1 Naïve
The naïve algorithms presented here are simply standard dynamic programming approaches applied to the empirical MDP construced from the dataset, using in place of . See Puterman (2014) for analysis of convergence, complexity, optimality, etc. of the general dynamic programming approaches, and note that the empirical MDP is simply a particular example of an MDP, so all results apply directly.
D.2 Pessimistic
In this section, we provide pseudocode for concrete implementations of algorithms in the pessimistic families we have discussed. Since the algorithms are not the focus of this work, we do not provide an in-depth analysis of these algorithms. We briefly note that, at a high level, the standard proof technique used to show convergence of policy iteration (Sutton & Barto, 2018) can be applied to all of these approaches. These algorithms utilize a penalized Bellman update, and the penalty is state-wise and independent between states. Thus, performing a local greedy policy improvement in any one state will strictly increase the value of all states. Thus, policy iteration guarantees strict monotonic improvement of the values of all states, and thus eventual convergence to an optimal policy (as measured by the penalized values).
D.2.1 Uncertainty-Aware
In this section, we provide pseudocode for a member of the the UA pessimsitic family of algorithms.
As discussed in the main text, count-based Bellman uncertainty functions, such as those derived from concentration inequalities in Appendix B.1, do not apply to non-tabular environments. The correct way to compute epistemic uncertainty with neural networks is still an open question. Therefore, our pseudocode for a neural implementation of the UA pessimistic approach is in some sense incomplete: a full implementation would need to specify a technique for computing . We hope that future work will identify such a technique.
D.2.2 Proximal
In this section, we provide pseudocode for a member of the proximal pessimsitic family of algorithms.
One important nuance of proximal algorithms is that the optimal policy may not be deterministic, since the penalty term is minimized when the policy matches the empirical policy, which may itself be stochastic. It is not enough to select ; we must instead select , which is a more difficult optimization problem. Fortunately, it has a closed-form solution.
Proposition 1.
Consider any state-action values and empirical policy . Let
. The policy given by
has the property
for all .
Proof.
We provide a brief outline of the proof. Consider any policy in some state . First, assume that exactly two cells of are non-zero, meaning that one term is and the other is (since both distributions sum to 1). If , the change in penalized return is non-positive, so is worse that . To see this, we simply consider all possible mappings between and actions. Actions fall into three categories, corresponding to the three cases of the construction: argmax-actions, empirical-actions, and zero-actions, respectively. It’s easy to check each case, moving probability mass from any category of action to any other, and see that penalized return is always non-positive. Finally, if more than two cells of are non-zero, we can always rewrite it as a sum of positive/negative pairs, and thus the overall change in penalized return is a sum of non-positive terms, and is itself non-positive. ∎
This closed-form can be used anytime the action space is discrete, in both the tabular and neural proximal algorithms. (In the neural algorithm, it replaces the need to optimize for a parameterized policy .)
Appendix E Related Work
Bandits.
Bandits are equivalent to single-state MDPs, and the fixed-dataset setting has been studied in the bandit literature as the logged bandit feedback setting. Swaminathan & Joachims (Swaminathan & Joachims, 2015) describe the Counterfactual Risk Minimization principle, and propose algorithms for maximizing the exploitation-only return. The UA pessimistic approach discussed in this work can be viewed as an extension of these ideas to the many-state MDP setting.
Approximate dynamic programming.
Some work in the approximate dynamic programming (ADP) literature (Munos, 2007) bears resemblance to ours. The setting is very similar; both FDPO and ADP are a modified version of dynamic programming in which errors are introduced, and research on algorithms in these settings studies the emergence and propagation of those errors. The key difference between the two settings lies in the origin of the errors. In ADP, the source of these errors is unspecified. In contrast, in this work, we focus specifically on statistical errors: where they emerge and what their implications are.
Off-policy evaluation and optimization.
The problem of off-policy reinforcement learning is typically posed as the setting where the data is being collected by a fixed stationary policy, but an infinite stream of data is being collected, so statistical issues introduced by finiteness of data can be safely ignored. This is clearly closely related to FDPO, but the focus of our work lies precisely on the statistical issues. However, there are close connections; for example, Jiang and Huang Jiang & Huang (2020) show that in the off-policy function approximation setting, algorithms may obey a version of the pessimism principle in order to select a good approximation.
Exploration.
Since acting pessimistically is the symmetric opposite of acting optimistically, many papers which study exploration leverage math which is nearly identical (though of course used in an entirely different way). For example, the confidence intervals, modified Bellman equation, and dynamic programming approach of Strehl & Littman (2008) closely mirror those used in the UA pessimistic algorithms.
Imitation learning.
Imitation learning (IL) (Hussein et al., 2017) algorithms learn a policy which mimics expert demonstrations. The setting in which these algorithms can be applied is closely related to the FDPO setting, in that IL algorithms map from a dataset of trajectories to a single stationary policy, which is evaluated by its suboptimality. (However, note that IL algorithms can be applied to a somewhat more general class of problems; for example, when no rewards are available.) In the FDPO setting, IL algorithms are limited by the fact that they can never improve on the quality of the empirical policy. In contrast, pessimistic FDPO algorithms will imitate the dataset when other actions are uncertain, but are also sometimes able to deviate and improve.
Safe reinforcement learning.
The sub-field of safe reinforcement learning studies reinforcement learning algorithms with constraints or guarantees that prevent bad behavior from occurring, or reduce it below a certain level (Thomas et al., 2019). To this end, many algorithms utilize variants of pessimism. Our contribution distinguishes itself from this line of work via its focus on the application of pessimism for worst-case guarantees on the more standard objective of expected suboptimality, rather than for guarantees on safety. However, there is a close relationship between our work and the algorithms and theory used for Safe RL. One popular framework is that of robust MDPs Givan et al. (1997); Nilim & El Ghaoui (2005); Weissman et al. (2003); Iyengar (2005), an object which generalizes MDPs to account for specification uncertainty in the transition functions. This framework is closely related to the pessimistic approaches described in this work. In fact, the UA pessimistic approach is precisely equivalent to constructing a robust MDP from data using concentration inequalities, and then solving it for the policy with the optimal performance under an adversarial choice of parameters; this algorithm has been used as a baseline in the literature but not studied in detail Ghavamzadeh et al. (2016); Laroche et al. (2019). Additionally, the sub-problem of Safe RL known as conservative policy improvement focuses on making small changes to a baseline policy which guarantee that the new policy will have higher value Thomas et al. (2015b; a); Ghavamzadeh et al. (2016); Laroche et al. (2019); Simão et al. (2019); Nadjahi et al. (2019), which bears strong resemblance to the proximal pessimistic algorithms discussed in this work.
Apprenticeship learning.
In the apprenticeship-learning setting (Walsh et al., 2011; Cohen & Hutter, 2020), an agent has the choice to take an action of its own selection, or to imitate a (potentially non-optimal) mentor. The process of determining whether or not to imitate a mentor is similar to determining whether or not to imitate the behavior policy used to collect the data. As a result, the underlying principle of pessimism is relevant in both situations. However, the setting of apprenticeship learning is also different from that of FDPO in several ways (availability of a mentor, ability to collect data), and so prior works do not provide a clear analysis of the importance of pessimism in FDPO.
Proximal algorithms.
Several approaches to online deep reinforcement learning, including TRPO and PPO, have been described as “proximal” (Schulman et al., 2015; 2017). These algorithms are derived from CPI (Kakade & Langford, 2002), which also introduces a notion of conservatism. In that body of work, this refers to the notion that, after each policy update, the resulting policy is close to the previous iterate. In contrast, the proximal algorithm described in this work is proximal with respect to a fixed policy (typically, the empirical policy defined by the dataset). The contrast between the importance of pessimism in these two cases can be seen by comparing the settings. CPI-type algorithms use pessimism to ensure that the policy remains good throughout the RL procedure, guaranteeing a monotonically-increasing performance as data is collected and the optimal policy is approached. Whereas, FDPO proximal approaches have no guarantees on the intermediate iterates, but rather, use pessimism to ensure that the final policy selected is as close to optimal as possible.
Deep learning FDPO approaches.
Recently, deep learning FDPO has received significant attention (Agarwal et al., 2019; Fujimoto et al., 2019; Kumar et al., 2019; Laroche et al., 2019; Jaques et al., 2019; Wu et al., 2019; Kidambi et al., 2020; Yu et al., 2020; Wang et al., 2020; Kumar et al., 2020; Liu et al., 2020). At a high level, these works are each primarily focused around proposing and analyzing some specific method. In contrast, the objective of this work is theoretical: we focus on providing a clean mathematical framework through which to understand this setting. We now provide specific details for how our contribution relates to each of these works.
The algorithms introduced in Fujimoto et al. (2019); Laroche et al. (2019); Kumar et al. (2019); Jaques et al. (2019); Liu et al. (2020); Kumar et al. (2020) can all be viewed as variants of the proximal pessimistic approach described in this paper. The implementations vary in a number of ways, but at its core, the primary difference between these algorithms lies in the choice of regularizer: KL, MMD, scalar penalty, or hard constraint. This connection has also been noted in Wu et al. (2019), which provides a unifying algorihtmic framework for these approaches, BRAC, and also performs various empirical ablations. Interestingly, all of these regularizers can be expressed as upper-bounds to , and as such, our proof of the suboptimality of proximal pessimistic algorithms can be used to justify all of these approaches. One aspect of our contribution is therefore providing the theoretical framework justifying BRAC. Furthermore, conceptually, these works differ from our contribution due to their focus on error propagation; in contrast, our results show that poor suboptimality of naïve algorithms is an issue even when there is no function approximation error at all.
Wang et al. (2020) provides an alternative algorithm for pessimistic FDPO, utilizing a policy-gradient based algorithm instead of the value-iteration-style algorithms in other related work. Similarly to Kumar et al. (2019), this approach utilizes a policy constraint to prevent boostrapping from actions which are poorly represented in the training data. However, this difference is purely algorithmic: since the fixed-point found by the optimization is the same, this algorithm can also be justified by our proximal pessimistic suboptimality bounds.
Model-based methods (Yu et al., 2020; Kidambi et al., 2020) have recently been proposed which implement pessimism via planning in a pessimistic model. This procedure can be viewed as learning a model which defines an implicit value function (derived from applying the planner to the model), and from there defines an implicit policy (which is greedy with respect to the implicit value function). The implicit value functions learned by the algorithms in Yu et al. (2020); Kidambi et al. (2020) obey the same fixed-point identity as the solutions to the UA pessimistic approach discussed in our work. Thus, both of these works are implementations of UA pessimistic approaches, and our UA pessimistic suboptimality bound can be viewed as justification for this family of algorithms. However, both of these works rely on ensembles to compute the uncertainty penalty, which is not a theoretically well-motivated technique.
The work of Agarwal et al. (2019) provides empirical evidence supporting the claims in this paper. The authors demonstrate that on realistic tasks like Atari, naïve algorithms are able to get good performance given an extremely large and diverse dataset, but they collapse on smaller or less exploratory datasets. It also highlights the difficulties that emerge when function approximation is introduced. Certain naïve algorithms can be seen to still sometimes collapse when datasets are large and diverse, because of other instabilities associated with function approximation.
Appendix F Experimental Details
F.1 Tabular
The first set of experiments utilize a simple tabular gridworld. The state space is 8x8, and the action space is {Up, Down, Left, Right}. Rewards are Bernoulli-distributed, with the mean reward for each state-action sampled from Beta(3,1); transitions are stochastic, moving in a random direction with probability 0.2; the discount is .99. This environment was selected to be simple and generic. We compare the performance of four approaches: imitation, naïve, uncertainty-aware pessimistic, and proximal pessimistic. The imitation algorithm simply returns the policy which takes actions in proportion to their observed frequencies in the dataset. For the UA pessimistic algorithm, we use the technique described in Appendix B.1 to implement Bellman uncertainty functions. For both pessimistic algorithms, we absorb all constants into the hyperparameter , which we selected to be for both algorithms by a simple manual search. For state-actions with no observations, we select uniformly at random in , transitions uniformly at random, and acts uniformly at random. We report the average of 1000 trials. The shaded region represents a 95% confidence interval.
F.2 Deep Learning
The second setting we evaluate on consists of four environments from the MinAtar suite (Young & Tian, 2019). In order to derive a near-optimal policy on each environment, we run DQN to convergence and save the resulting policy. We report the average of 3 trials. The shaded area represents the range between the maximum and minimum values.
We implement deep learning versions of the above algorithms in the style of Neural Fitted Q-Iteration (Riedmiller, 2005). In this setting, we implement only proximal pessimistic algorithms. To compute the penalty term, we must approximate ; this can be done by training a policy network on the dataset to predict actions via maximum likelihood. Just as in the tabular setting, we absorb all constant coefficients into our pessimism hyperparameter, here setting .
All experiments used identical hyperparameters. Hyperparameter tuning was done on just two experimental setups: Breakout using , and Breakout using . Tuning was very minimal, and done via a small manual search.
Appendix G Additional Content
G.1 An Illustrative Example

Consider the following problem setting. We are given an MDP with a single state and some number of actions, each of which return rewards sampled from a Bernoulli distribution on . (An MDP with a single state is isomorphic to a multi-armed bandit.) Furthermore, for each action, we are given a dataset containing the outcomes of some number of pulls. We are now given the opportunity to take an action, with the goal of maximizing our expected reward. What strategy should we use?
One obvious strategy is to estimate the expected reward for each action by computing its mean reward in the dataset, and then select the arm with the highest empirical reward. However, in certain problem instances, this “naïve strategy” fails. We can illustrate this with a simple example (visualized in Figure 3). Consider an MDP with a single state and 1000 actions. Let the reward distribution of the first action have mean of 0.99, while the reward distributions of all other actions have means of 0.01. Construct a dataset for this bandit by pulling the first arm 10000 times, and the other arms 1 time each.
Although this problem seems easy, the naïve algorithm achieves close to the worst possible performance. It’s clear that in this problem, the best policy selects the first action. The empirical estimate of the mean of the first action will be less than 1 with probability , and there will be at least one other action which has empirical mean of 1 with probability . Thus, the first action is almost never selected.
This issue can be resolved by identifying a fundamental failing of the naïve approach: it ignores epistemic uncertainty around the expected return of each action. In order to guarantee good performance on all problem instances, we need to avoid playing actions that we are uncertain about. One way to do this is to construct a high-probability lower bound on the value of each action (for example using concentration inequalities), and select the action with the highest lower bound. If we consider the upper and lower bounds as defining the set of possible “worlds” that we could be in, acting according to the lower bound of every arm means acting as though we are in the worst possible world. In other words: being pessimistic.
The above example may seem somewhat contrived, due to the enormous size of the action space and skewed data collection procedure. However, we argue that it serves as a good analogy for the more common setting of an MDP with many states and a small number of actions at each state. Roughly speaking, selecting an action in a one-state MDP is analogous to selecting a deterministic policy in a multi-state MDP, and the number of policies is exponentially large in the size of the state space. Additionally, it’s very plausible in practical situations that data is collected according to only a small set of similar policies, e.g. expert demonstrations, leading to skewed data coverage.
G.2 Extreme Pessimsim: Value Lower Bound Algorithms
We begin with a further corollary to Corollary 1. Consider any value-based FDPO algorithm whose value function is guaranteed to always return values which underestimate the expected return with high probability.
Corollary 2 (Value-lower-bound-based FDPO suboptimality bound).
Consider any value-based fixed-dataset policy optimization algorithm , with internal fixed-dataset policy evaluation subroutine , which has the lower-bound guarantee that with probability at least . For any policy , dataset , denote . With probability at least , the suboptimality of is bounded by
Proof.
This result follows directly from Corollary 1, using the fact that . ∎
This bound is identical to the term labeled (a) from Corollary 1. The term labeled (b), which contained a supremum over policies, has vanished, leaving only the term containing an infimum. Where the bound of Corollary 1 demands that some condition hold for all policies, we now only require that there exists any single suitable policy. It is clear that this condition is much easier to satisfy, and thus, this suboptimality bound will typically be much smaller.
A value lower-bound algorithm is in some sense the most extreme example of a pessimistic approach. Any algorithm which penalizes its predictions to decrease overestimation can be described as pessimistic. In such cases, (b) will be decreased, rather than removed entirely. Still, any amount of pessimism reduces dependence on a global condition, decreasing overall suboptimality.
Furthermore, blind pessimism is not helpful. For example, one could trivially construct a value lower-bound algorithm from a naïve algorithm by simply subtracting a large constant from the naïve value estimate of every policy. But this would achieve nothing, as the infimum term would immediately increase by this same amount. To yield a productive change in policy, pessimism must instead vary across states in an intelligent way.
G.3 Practical Considerations
Through the process of running experiments in the deep learning setting, the authors noted that several aspects of the experimental setup, which have not been addressed in previous work, had surprisingly large impacts on the results. In this section, we informally report some of our findings, which future researchers may find useful. Additionally, we hope these effects will be studied more rigorously in future work.
The first consideration is that performance is highly nonmonotonic with respect to training time. In almost all experiments, it was observed that with every target-network update, the performance of the policy would oscillate wildly. Even after performing many Bellman updates, few experiments showed anything resembling convergence. It is therefore important that the total number of steps be selected beforehand, to avoid unintentional cherry-picking of results. Additionally, one common trend was for algorithms to have high performance early on, and then eventually crash. For this reason, it is important that algorithms be run for a long duration, in order to be certain that the conclusions drawn are valid.
The second consideration is the degree to which the inner-loop optimization process succeeds. If throughout training, whenever we update the target network, its error is low, convergence near to the fixed point is guaranteed (Antos et al., 2007). However, computational restrictions force us to make tradeoffs about the degree to which this condition is satisfied. Experimentally, we found this property to be very important: when the error was not properly minimized, performance was negatively impacted, sometimes even leading to divergence. There are three notable algorithmic decisions which we found were required to ensure that the error was adequately minimized.
The size of the network. It is important to ensure that the network is large enough to reasonably fit the values at all points throughout training. Most prior works (Kumar et al., 2019; Fujimoto et al., 2019; Kumar et al., 2020) utilize the same network architecture as the original DQN (Mnih et al., 2015), which is fairly small by modern standards. We found that this size of network was adequate to fit MinAtar environments, but that decreasing the size of the network further led to significant performance degradation. Preliminary experiments indicated that since the full Atari environment is much more complex than MinAtar, a larger network may be required.
The amount of training steps in the inner loop of Neural Fitted Q-Iteration. If the number of steps is too small, error will not be adequately minimized. In our experiments, approximately 250,000 gradient steps per target update were required to consistently minimize error enough to avoid divergence. We note that many prior works (Kumar et al., 2019; Fujimoto et al., 2019; Kumar et al., 2020) do not adjust this hyperparameter; typical results in the literature use fewer than 10,000 gradient steps per update.
The per-update initialization of the network. When changing the target network, we are in essence beginning a new supervised learning problem. Thus, to ensure that we could reliably find a good solution, we found that we needed to fully reinitialize the neural network whenever we updated the target network. The dynamics of training neural networks via gradient descent are still not fully understood, but it is well known that the initialization is of great importance (Hu et al., 2020). It has been observed that certain initializations can seriously impact training: for example, if the final parameters of a classifier trained on randomly-relabeled CIFAR classifier are used as the initialization for the regular CIFAR classification task, the trained network will have worse test-set performance.101010Chelsea Finn, Suraj Nair, Henrik Marklund; personal communication.