This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Fast Slate Policy Optimization: Going Beyond Plackett-Luce

Otmane Sakhi
CREST-ENSAE, Criteo AI Lab [email protected] David Rohde
Criteo AI Lab [email protected] Nicolas Chopin
CREST-ENSAE [email protected]
Abstract

An increasingly important building block of large scale machine learning systems is based on returning slates; an ordered lists of items given a query. Applications of this technology include: search, information retrieval and recommender systems. When the action space is large, decision systems are restricted to a particular structure to complete online queries quickly. This paper addresses the optimization of these large scale decision systems given an arbitrary reward function. We cast this learning problem in a policy optimization framework and propose a new class of policies, born from a novel relaxation of decision functions. This results in a simple, yet efficient learning algorithm that scales to massive action spaces. We compare our method to the commonly adopted Plackett-Luce policy class and demonstrate the effectiveness of our approach on problems with action space sizes in the order of millions.

1 INTRODUCTION

Large scale online decision systems, ranging from search engines to recommender systems, are constantly queried to deliver ordered lists of content given contextual information. As an example, a user who has just seen the film ‘Batman’, might be recommended: Superman, Batman Returns, Bird Man, or a user that is reading a page about ‘technology’ might be interested in other stories about biotechnology, solar power, and large language models. A large scale production system must therefore be able to rapidly respond to a query like ‘Batman’ or ‘technology’ with an ordered list of relevant items.

A proven solution to this problem is to generate the ordered list using an approximate maximum inner product search (MIPS) algorithm (Shrivastava & Li, 2014), which, at the expense of a constraint in the decision rule, provides extremely rapid querying even for massive catalogues. While MIPS based systems are a proven technology at deployment time, the offline optimization of them is not, and the standard algorithm based on policy learning using the Plackett-Luce distribution is infeasibly slow, as the algorithm both iterates slowly and requires very large numbers of iterations to converge. Fortunately, other approaches are possible which both iterate faster and require fewer iterations. In this paper, we propose a simple algorithm that enjoys better convergence properties than competitors.

We denote by x𝒳x\in\mathcal{X} a context; it can be the history of the user, a search query or even a whole web page. The decision system is tasked to deliver, given the context xx, an ordered list of actions AK=[a1,,aK]A_{K}=[a_{1},...,a_{K}] of arbitrary size KK\in\mathbbm{N}^{*}, often referred to as slates. This slate can be an ad banner, a list of recommended content or search results. Our decision system, given contextual information xx, constructs a slate by selecting a subset of actions {a1,,aK}𝒜\{a_{1},...,a_{K}\}\subset\mathcal{A} from a potentially large discrete set 𝒜\mathcal{A} and ordering them. Let P=|𝒜|P=|\mathcal{A}| be the size of the action set. Fixing the slate size KK, we model our system by a decision function d:𝒳𝒮K(𝒜)d:\mathcal{X}\to\mathcal{S}_{K}(\mathcal{A}) that maps contexts xx to the space 𝒮K(𝒜)\mathcal{S}_{K}(\mathcal{A}) of ordered lists of size KK. Each pair of context xx and slate AKA_{K} is associated with a reward function111motivated by business metrics and/or users engagement. r(AK,x)r(A_{K},x) that encodes the relevance of the slate AKA_{K} for xx. We suppose that the contexts are stochastic, and coming from an unknown distribution ν(𝒳)\nu(\mathcal{X}). Our objective is to find decision systems that maximize the expected reward, given by:

𝔼xν(𝒳)[r(AK=d(x),x)].\mathbbm{E}_{x\sim\nu(\mathcal{X})}\left[r(A_{K}=d(x),x)\right].

Assuming that we have access to the reward function, the solution of this optimization problem is given by:

x𝒳,d(x)=argmaxAK𝒮K(𝒜)r(AK,x).\displaystyle\forall x\in\mathcal{X},\quad d(x)=\operatorname*{arg\,max}_{A_{K}\in\mathcal{S}_{K}(\mathcal{A})}r(A_{K},x). (1)

This solution, while optimal, is intractable as it requires, for a given context xx, the search in the enormous space 𝒮K(𝒜)\mathcal{S}_{K}(\mathcal{A}) of size 𝒪(PK)\mathcal{O}(P^{K}) to find the best slate. Instead of maximizing the expected reward over the whole space, we want to restrict ourselves to decision functions of practical use. In this direction, we begin by defining the parametric relevance function fθ:𝒜×𝒳f_{\theta}:\mathcal{A}\times\mathcal{X}\to\mathbbm{R}:

(a,x)𝒜×𝒳,fθ(a,x)=hΞ(x)Tβa,\displaystyle\forall(a,x)\in\mathcal{A}\times\mathcal{X},\quad f_{\theta}(a,x)=h_{\Xi}(x)^{T}\beta_{a},

with the learnable parameter θ=[Ξ,β]\theta=[\Xi,\beta], a parametric transform hΞ:𝒳Lh_{\Xi}:\mathcal{X}\to\mathbbm{R}^{L} that creates a context embedding of size LL, and βa\beta_{a} the embedding of action aa in L\mathbbm{R}^{L}. the embedding dimension LL is usually taken to be much smaller than PP, the size of the action space. We then define our decision function:

x𝒳,dθ(x)=argsortKa𝒜{hΞ(x)Tβa}.\displaystyle\forall x\in\mathcal{X},\quad d_{\theta}(x)=\operatornamewithlimits{argsort^{K}}_{a\in\mathcal{A}}\left\{h_{\Xi}(x)^{T}\beta_{a}\right\}. (2)

with argsortK\operatornamewithlimits{argsort^{K}} the argsort function truncated at the KK-th item. Restricting the space of decision functions to this parametric form reduces the complexity of returning a slate for the context xx from 𝒪(PK)\mathcal{O}(P^{K}) to 𝒪(PlogK)\mathcal{O}(P\log K). This complexity can be further decreased. Equation (2) transforms the querying problem to the MIPS: Maximum Inner Product Search problem (Shrivastava & Li, 2014), for which different algorithms were proposed (Gionis et al., 1999; Malkov & Yashunin, 2020; Guo et al., 2020) to find a solution in a logarithmic time complexity. These algorithms build fixed indexes over the action embeddings β\beta, with particular structures to allow the identification (sometimes approximate) of the KK actions with the largest inner product with the query hΞ(x)h_{\Xi}(x). This allow us to reduce even further the complexity of the argsortK\operatornamewithlimits{argsort^{K}} operator from 𝒪(PlogK)\mathcal{O}(P\log K) to a logarithmic time complexity 𝒪(logP)\mathcal{O}(\log P), making fast decisions possible in problems with massive action spaces 𝒜\mathcal{A}. This leaves us with the problem of finding the optimal decision function within the constraints of this parametric form. This is achieved by solving the following:

θargmaxθ=[Ξ,β]𝔼xν(𝒳)[r(AK=dθ(x),x)].\displaystyle\theta^{*}\in\operatorname*{arg\,max}_{\theta=[\Xi,\beta]}\mathbbm{E}_{x\sim\nu(\mathcal{X})}\left[r(A_{K}=d_{\theta}(x),x)\right].

As we do not have access to ν(𝒳)\nu(\mathcal{X}), we replace the previous objective with its empirical counterpart:

θargmaxθ=[Ξ,β]1Ni=1Nr^(AK=dθ(xi),xi),\displaystyle\theta^{*}\in\operatorname*{arg\,max}_{\theta=[\Xi,\beta]}\frac{1}{N}\sum_{i=1}^{N}\hat{r}\left(A_{K}=d_{\theta}(x_{i}),x_{i}\right), (3)

with {xi}i[N]\{x_{i}\}_{i\in[N]} observed contexts and r^\hat{r} an offline reward estimator; it includes the Direct Method, Inverse Propensity Scoring (Horvitz & Thompson, 1952), Doubly Robust Estimator (Dudík et al., 2014) and many other variants, as presented in (Sakhi et al., 2023b). The optimization problem of Equation (3) is complicated by the fact that the reward can be non-smooth and that our decision function is not differentiable. A way to handle this is by relaxing the optimization objective. Differentiable sorting algorithms (Grover et al., 2019; Prillo & Eisenschlos, 2020) address a similar problem but make strong assumptions about the structure of the reward function, and cannot scale to large action space problems. To be as general as possible, we take another direction and relax the problem into an offline policy learning formulation (Bottou et al., 2013; Swaminathan & Joachims, 2015). We extend our space of parametrized decision functions to a well chosen space of stochastic policies πθ:𝒳𝒫(𝒮𝒦(𝒜))\pi_{\theta}:\mathcal{X}\to\mathcal{P(S_{K}(A))}, that given a context xx, define a probability distribution over the space of slates of size KK. Given a policy πθ\pi_{\theta}, we relax Equation (3), taking an additional expectation under πθ\pi_{\theta} to obtain:

θargmaxθ=[Ξ,β]R^(πθ)\displaystyle\theta^{*}\in\operatorname*{arg\,max}_{\theta=[\Xi,\beta]}\hat{R}(\pi_{\theta}) =1Ni=1N𝔼AKπθ(|xi)[r^(AK,xi)].\displaystyle=\frac{1}{N}\sum_{i=1}^{N}\mathbf{\mathbbm{E}}_{A_{K}\sim\pi_{\theta}(\cdot|x_{i})}\left[\hat{r}(A_{K},x_{i})\right]. (4)

The most common policy class for this type of problem is Plackett-Luce (Plackett, 1975), that generalises the softmax parametrization to slates of size K>1K>1. Under this policy class, computing exact gradients is intractable but we can obtain approximate gradients w.r.t to θ\theta of Equation (4). Common approximations (Williams, 1992) are based on sampling from the policy, are computed in 𝒪(PlogK)\mathcal{O}(P\log K) and suffer from a variance that grows with the slate size KK. In the special case of decomposable rewards over items on the slate (Swaminathan et al., 2017), exploiting this linearity structure (Oosterhuis, 2022) provides gradient estimates with better variance. However, training speed still scales linearly with PP making policy learning infeasible in large action spaces.

In this work, we propose LGP: Latent Gaussian Perturbation, a new policy class based on smoothing the latent space, that is perfectly suitable to optimize decision functions of the form described in (2). As shown in Table 1, our method provides fast sampling, low memory usage, gradient estimates with better computational and statistical properties while being agnostic to the reward structure. When the embeddings are prefixed, this class naturally benefits from approximate MIPS technology, making sampling logarithmic in the action space size and opening the possibility for policy optimization over billion-scale space sizes.

This paper will be structured as follows. In Section 2, we will review the Plackett-Luce policy class and present its limitations. Section 3 will introduce our newly proposed relaxation, motivate its use and propose a learning algorithm. We focus in Section 4 on experiments to validate our findings empirically. Section 5 will cover the related work and we conclude with Section 6.

Arbitrary Reward Low Gradient Variance Time Complexity Space Complexity
PL-PG 𝒪(SP)\mathcal{O}(SP) 𝒪(SP)\mathcal{O}(SP)
PL-Rank 𝒪(SP)\mathcal{O}(SP) 𝒪(SP)\mathcal{O}(SP)
LGP 𝒪(SP)\mathcal{O}(SP) 𝓞(𝑺𝑳)\bm{\mathcal{O}(SL)}
LGP-MIPS 𝓞(𝑺𝐥𝐨𝐠𝑷)\bm{\mathcal{O}(S\log P)} 𝓞(𝑺𝑳)\bm{\mathcal{O}(SL)}
Table 1: High level comparison between the different optimization algorithms for slate decision functions, with PP the size of the action space, LL the size of the embedding space and SS the number of samples used to approximate the gradient. PL-PG is Plackett-Luce trained with the Score Function Gradient (Williams, 1992). PL-Rank is the algorithm proposed in Oosterhuis (2022). LGP is our proposed method and LGP-MIPS is its accelerated variant. Our method works with arbitrary rewards, scales logarithmically with PP and have low memory footprint (LPL\ll P).

2 Plackett-Luce Policies

2.1 A Simple Definition

We relax our objective function and model online decision systems as stochastic policies over the space of slates; ordered lists of actions. There are some natural parametric forms to define a policy on discrete action spaces. If we are dealing with the simple case of K=1K=1 (slate of one action), we can adopt the softmax policy (Swaminathan & Joachims, 2015; Sakhi et al., 2023b) that, conditioned on the context x𝒳x\in\mathcal{X} and for a particular action a𝒜a\in\mathcal{A}, is of the form :

πθ(a|x)=exp{fθ(a,x)}bexp{fθ(b,x)}=exp{fθ(a,x)}Zθ(x).\pi_{\theta}(a|x)=\frac{\exp\{f_{\theta}(a,x)\}}{\sum_{b}\exp\{f_{\theta}(b,x)\}}=\frac{\exp\{f_{\theta}(a,x)\}}{Z_{\theta}(x)}.

This softmax parametrization found great success (Swaminathan & Joachims, 2015; Chen et al., 2019; Sakhi et al., 2023b), and is ubiquitous in applications where the goal is to learn policies or distributions over discrete actions. Once we deal with K>1K>1, one can generalize the previous form giving us the Plackett-Luce policy (Plackett, 1975). For a given xx and a particular slate AK=[a1,,aK]A_{K}=[a_{1},...,a_{K}], we write its probability:

πθ(AK|x)=i=1Kexp{fθ(ai,x)}Zθi1(x)=i=1Kπθ(ai|x,A1:i1),\pi_{\theta}(A_{K}|x)=\prod_{i=1}^{K}\frac{\exp\{f_{\theta}(a_{i},x)\}}{Z^{i-1}_{\theta}(x)}=\prod_{i=1}^{K}\pi_{\theta}(a_{i}|x,A_{1:i-1}), (5)

with Zθ0(x)=Zθ(x)Z^{0}_{\theta}(x)=Z_{\theta}(x) and Zθi(x)=Zθi1(x)exp{fθ(ai,x)}Z^{i}_{\theta}(x)=Z^{i-1}_{\theta}(x)-\exp\{f_{\theta}(a_{i},x)\}.

Computing these probabilities can be done in 𝒪(P)\mathcal{O}(P) which is comparable to the simple case where K=1K=1. The probabilities given by the Plackett-Luce policy are intuitive. Equation (5) can be seen as the probability to generate the slate AK=[a1,,aK]A_{K}=[a_{1},...,a_{K}] by sampling without replacement from a categorical distribution over the discrete space 𝒜\mathcal{A} with action probabilities proportional to exp{fθ(a,x)}\exp\{f_{\theta}(a,x)\}. This sampling procedure can be done in 𝒪(KP)\mathcal{O}(KP), but its sequential nature is a bottleneck for parallelization. Another way to sample from this distribution is to exploit the following expression:

πθ(AK|x)=𝔼γ𝒢P(0,1)[𝟙[AK=argsortKa𝒜{fθ(a,x)+γi}]],\pi_{\theta}(A_{K}|x)=\mathbbm{E}_{\gamma\sim\mathcal{G}^{P}(0,1)}\left[\mathbbm{1}\left[A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{f_{\theta}(a^{\prime},x)+\gamma_{i}\right\}\right]\right],

with γ𝒢P(0,1)\gamma\sim\mathcal{G}^{P}(0,1) a vector of PP independent Gumbel random variables. This is known in the literature as the Gumbel trick (Huijben et al., 2021). This means that sampling from a Plackett-Luce boils down to sampling PP independent Gumbel random variables, which costs 𝒪(P)\mathcal{O}(P), and then computing an argsortK\operatornamewithlimits{argsort^{K}} of the noised f(,x)f(\cdot,x) over the discrete action space. We cannot exploit approximate MIPS for this computation as the noise is added after computing the inner product f(a,x)f(a,x), making this step cost 𝒪(PlogK)\mathcal{O}(P\log K), which makes the total complexity of sampling 𝒪(PlogK)\mathcal{O}(P\log K), slightly better than the first procedure while compatible with parallel acceleration.

2.2 Optimizing The Objective

We want to learn slate policies that can maximize the objective in Equation (4). As our objective is decomposable over contexts, stochastic optimization procedures can be adopted Ruder (2016) making optimization over large datasets possible. For this reason, we can focus on the gradient of the objective for a single context xx. We derive the score function gradient (Williams, 1992):

θR^(πθ|x)\displaystyle\nabla_{\theta}\hat{R}(\pi_{\theta}|x) =𝔼πθ(|x)[r^(AK,x)θlogπθ(AK|x)]\displaystyle=\mathbf{\mathbbm{E}}_{\pi_{\theta}(\cdot|x)}\left[\hat{r}(A_{K},x)\nabla_{\theta}\log\pi_{\theta}(A_{K}|x)\right] (6)
=i=1K𝔼πθ(|x)[r^(AK,x)θlogπθ(ai|x,Ai1)].\displaystyle=\sum_{i=1}^{K}\mathbf{\mathbbm{E}}_{\pi_{\theta}(\cdot|x)}\left[\hat{r}(A_{K},x)\nabla_{\theta}\log\pi_{\theta}(a_{i}|x,A_{i-1})\right]. (7)

This gradient is defined as an expectation under πθ(|x)\pi_{\theta}(\cdot|x) over all possible slates. Computing it exactly requires summing 𝒪(PK)\mathcal{O}(P^{K}) terms which is infeasible. This allows us to approximate the gradient by sampling, which reduces the computation complexity but we will see that this gradient suffers from further problems.

Computational Burden.

The computational complexity of the gradient is crucial for allowing fast learning of slate as it impacts the running time of every gradient step. Even if we avoid computing the gradient exactly, its approximation can still be a bottleneck when dealing with large action spaces for the following reasons: (1) Sampling: Approximating the expectation by sampling slates from πθ(|x)\pi_{\theta}(\cdot|x) can be done in 𝒪(PlogK)\mathcal{O}(P\log K). However, if the action space is large (PP in the order of millions), even a linear complexity on PP can be problematic, massively slowing down our optimization procedure. (2) The Normalizing Constant: Approximating the gradient needs the computation of θlogπθ(AKi|x)\nabla_{\theta}\log\pi_{\theta}(A^{i}_{K}|x) for the sampled slates {AKi}i[S]\{A^{i}_{K}\}_{i\in[S]}. This can slow down the optimization procedure as computing the normalizing constant Zθ(x)Z_{\theta}(x) requires summing over all actions, making the complexity of the operation linear in PP.

We can solve this computational burden by tackling the two previous problems separately. We can get rid of the normalizing constant in the gradient by generalizing the results of Sakhi et al. (2023b). For a single context xx, we can derive a covariance gradient that does not require Zθ(x)Z_{\theta}(x):

θR^(πθ|x)=𝕠𝕧AKπθ(.|x)[r^(AK,x),i=1Kθfθ(ai,x)],\displaystyle\nabla_{\theta}\hat{R}(\pi_{\theta}|x)=\mathbf{\mathbbm{Cov}}_{A_{K}\sim\pi_{\theta}(.|x)}\left[\hat{r}(A_{K},x),\sum_{i=1}^{K}\nabla_{\theta}f_{\theta}(a_{i},x)\right], (8)

with 𝕠𝕧[A,𝑩]=𝔼[(A𝔼[A]).(𝑩𝔼[𝑩])]\mathbf{\mathbbm{Cov}}[A,\bm{B}]=\mathbf{\mathbbm{E}}[(A-\mathbf{\mathbbm{E}}[A]).(\bm{B}-\mathbf{\mathbbm{E}}[\bm{B}])] a covariance between AA a scalar function, and 𝑩\bm{B} a vector. The proof of this new gradient expression is developed in Appendix A.1. One can see that for K=1K=1, we recover the results of Sakhi et al. (2023b). This form of gradient does not involve the computation of a normalizing constant, which solves the second problem, but still requires sampling from the policy πθ(|x)\pi_{\theta}(\cdot|x) to get a good covariance estimation. To lower the time complexity of this step, we can use Monte Carlo techniques such as Importance Sampling/Rejection Sampling (Owen, 2013) with carefully chosen proposals to achieve fast sampling without sacrificing the accuracy of the gradient approximation. We develop a discussion around accelerating Plackett-Luce training in Appendix A.1. While we may have ways to deal with the computation complexity, the Plackett-Luce Policy gradient estimate still suffers from the following problems:

Variance Problems.

Let us focus on the gradient derived in Equation (6). Its exact computation is intractable, and we need to estimate it by sampling from πθ(|x)\pi_{\theta}(\cdot|x). Let us imagine we sample a slate AK=[a1,,aK]A_{K}=[a_{1},...,a_{K}] to estimate the gradient:

Gθ(x)\displaystyle G_{\theta}(x) =r^(AK,x)θlogπθ(AK|x)\displaystyle=\hat{r}(A_{K},x)\nabla_{\theta}\log\pi_{\theta}(A_{K}|x)
=r^(AK,x)i=1Kgθi(x),\displaystyle=\hat{r}(A_{K},x)\sum_{i=1}^{K}g^{i}_{\theta}(x),

with gθi(x)g^{i}_{\theta}(x) set to θlogπθ(ai|x,Ai1)\nabla_{\theta}\log\pi_{\theta}(a_{i}|x,A_{i-1}) to simplify the notation. Gθ(x)G_{\theta}(x) is an unbiased estimator of the gradient θR^(πθ|x)\nabla_{\theta}\hat{R}(\pi_{\theta}|x) and can be used in a stochastic optimization procedure in a principled manner (Ruder, 2016). However, the efficiency of any stochastic gradient descent algorithm depends on the variance of the gradient estimate (Ajalloeian & Stich, 2021), which is defined for vectors XX as:

𝕍[X]=𝔼[X𝔼[X]2]+.\mathbbm{V}[X]=\mathbbm{E}\left[||X-\mathbbm{E}[X]||^{2}\right]\in\mathbbm{R}^{+}.

Gradients with small variances allow practitioners to use bigger step sizes, which reduces the number of iterations as it makes the whole optimization procedure converge faster. Naturally, we would want the variance of our estimator to be small. Unfortunately, the variance of Gθ(x)G_{\theta}(x) grows with the slate size KK. Writing down the variance w.r.t πθ(|x){\pi_{\theta}(\cdot|x)} of Gθ(x)G_{\theta}(x):

𝕍[Gθ(x)]\displaystyle\mathbbm{V}[G_{\theta}(x)] =𝕍[r^(AK,x)θlogπθ(AK|x)]=i=1K𝕍[r^(AK,x)gθi(x)]+2i<j𝕠𝕧[r^(AK,x)gθi(x),r^(AK,x)gθj(x)].\displaystyle=\mathbbm{V}[\hat{r}(A_{K},x)\nabla_{\theta}\log\pi_{\theta}(A_{K}|x)]=\sum_{i=1}^{K}\mathbbm{V}[\hat{r}(A_{K},x)g^{i}_{\theta}(x)]+2\sum_{i<j}\mathbbm{Cov}[\hat{r}(A_{K},x)g^{i}_{\theta}(x),\hat{r}(A_{K},x)g^{j}_{\theta}(x)].

The first term of this variance is a sum over the slate of individual variances, which clearly grows as KK grows. For the covariance terms, we argue that, especially when initializing the parameter θ\theta randomly, the gradients gθi(x)g^{i}_{\theta}(x) and gθj(x)g^{j}_{\theta}(x) will have in expectation different signs making the covariance terms cancel out, leaving the sum of the individual variance terms dominate. This gives a variance that grows in 𝒪(K)\mathcal{O}(K).

Previous work already showed empirically that the score function gradient for the Plackett-Luce distribution has a large variance (Gadetsky et al., 2020; Buchholz et al., 2022), large enough that learning is not possible in difficult scenarios without considering variance reduction methods (Gadetsky et al., 2020). A possible solution is Randomized Quasi-Monte Carlo (Buchholz et al., 2022; L’Ecuyer, 2016), which produces more accurate estimates by covering the sampling space better. Its value is only significant when we sample few slates {AKs}s[S]\{A^{s}_{K}\}_{s\in[S]} to approximate the gradient (Buchholz et al., 2022; L’Ecuyer, 2016). Another direction explores control variates (Gadetsky et al., 2020) as a variance reduction technique. This method requires additional computational costs and a perfect modelling of a differentiable reward proxy to expect variance reduction (Grathwohl et al., 2018).

We want to find a way to both reduce the variance of our method and the computational burden. One of the simplest method to reduce the variance and gain in computation speed is to reduce the number of parameters we want to train (Koch et al., 2021). In our problem of online decision systems, some parameters can be learned independently making policy learning easier (Sakhi et al., 2023b).

2.3 Fixing The Action Embeddings

As we are dealing with large action spaces, we are constrained to the following structure on the relevance function fθf_{\theta} for fast querying:

a𝒜fθ(a,x)=hΞ(x)Tβa,\displaystyle\forall a\in\mathcal{A}\quad f_{\theta}(a,x)=h_{\Xi}(x)^{T}\beta_{a},

with both hΞ(x)h_{\Xi}(x) and βa\beta_{a} living in an embedding space L\mathbbm{R}^{L} with LPL\ll P. The dimension of the matrix β\beta is [L×P][L\times P], which can be very large when PP is large, and dominates Ξ\Xi in terms of number of parameters (Koch et al., 2021; Sakhi et al., 2023b). If we can fix the matrix β\beta, it would benefit our approach both in terms of computational efficiency and variance reduction. Indeed, reducing the number of parameters accelerates learning, makes the problem more identifiable, and reduces drastically the gradient variance as:

𝕍[Gθ(x)]\displaystyle\mathbbm{V}[G_{\theta}(x)] =𝔼[G¯θ(x)2]=𝔼[G¯Ξ(x)2]+𝔼[G¯β(x)2]\displaystyle=\mathbbm{E}[||\overline{G}_{\theta}(x)||^{2}]=\mathbbm{E}[||\overline{G}_{\Xi}(x)||^{2}]+\mathbbm{E}[||\overline{G}_{\beta}(x)||^{2}]
=𝕍[GΞ(x)]+𝕍[Gβ(x)]𝕍[GΞ(x)].\displaystyle=\mathbbm{V}[G_{\Xi}(x)]+\mathbbm{V}[G_{\beta}(x)]\gg\mathbbm{V}[G_{\Xi}(x)].

with G¯θ(x)=Gθ(x)θR^(πθ|x)\overline{G}_{\theta}(x)=G_{\theta}(x)-\nabla_{\theta}\hat{R}(\pi_{\theta}|x) the centred gradient estimate. In many applications (think about information retrieval, recommender systems or ad placement), the action embeddings can be learned from the massive data we have on the actions. In these scenarios, actions boil down to web pages, products we want to recommend or place in an ad. We usually have collaborative filtering signal (Sakhi et al., 2020; Liang et al., 2018) and product descriptions (Vasile et al., 2016) to learn embeddings from. These signals help us obtain good action embeddings β\beta and allow us to fix the action matrix before proceeding to the downstream task we are solving. This approach of fixing β\beta is not new, Koch et al. (2021) fix β\beta to learn a large scale recommender system deployed in production, and Sakhi et al. (2023b) show empirically that learning β\beta actually hurts the performance of softmax policies in large scale scenarios.

We can still learn more about the actions.

Even with β\beta fixed, there are sufficient degrees of freedom to solve our downstream task. If we write down:

hΞ(x)=hΞ(x)Z,\displaystyle h_{\Xi}(x)=h_{\Xi^{\prime}}(x)Z,

with Ξ=[Ξ,Z]\Xi=[\Xi^{\prime},Z], ZZ being a learnable matrix of size [L,L][L^{\prime},L]. the relevance function can be written as:

fθ(a,x)=hΞ(x)T(Zβa),a𝒜.\displaystyle f_{\theta}(a,x)=h_{\Xi^{\prime}}(x)^{T}(Z\beta_{a}),\quad\forall a\in\mathcal{A}.

This means that, even though the matrix β\beta is fixed, we can learn a linear transform of β\beta with the help of ZZ, injecting information from the downstream task and learning a transformed representation of the actions in the embedding space. In the rest of the paper, we will fix the action embeddings β\beta making the parametrization of the relevance function fθf_{\theta} reduce to θ=Ξ\theta=\Xi, giving for any xx:

fθ(a,x)=hθ(x)Tβa,a𝒜.\displaystyle f_{\theta}(a,x)=h_{\theta}(x)^{T}\beta_{a},\quad\forall a\in\mathcal{A}.

3 Latent Gaussian Perturbation

Fixing the embeddings helps to decrease the variance and the number of parameters to optimize substantially, which makes our policy learning routine converge faster. This approach, however, does not deal with the fundamental limit of the Plackett-Luce variance, which grows with the slate size KK. This policy class also needs particular care to accelerate its learning; we should adopt the new gradient formula stated in (8) combined with advanced Monte Carlo techniques to approximate the gradient efficiently, making the implementation of such methods difficult to achieve.

These issues come intrinsically with the adoption of the Plackett-Luce policy, and suggest that we should think differently about how we define policies over slates. Let us investigate this family of policies. For a particular context xx and a slate AKA_{K}, we write down its Gumbel trick (Huijben et al., 2021) expression:

πθ(AK|x)=𝔼γ𝒢P(0,1)[𝟙[AK=argsortKa𝒜{hθ(x)Tβa+γi}]].\displaystyle\pi_{\theta}(A_{K}|x)=\mathbbm{E}_{\gamma\sim\mathcal{G}^{P}(0,1)}\left[\mathbbm{1}\left[A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{h_{\theta}(x)^{T}\beta_{a^{\prime}}+\gamma_{i}\right\}\right]\right].

The Plackett-Luce policy is a smoothed, differentiable relaxation of the following deterministic policy:

bθ(AK|x)\displaystyle b_{\theta}(A_{K}|x) =𝟙[AK=argsortKa𝒜{hθ(x)Tβa}]\displaystyle=\mathbbm{1}\left[A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{h_{\theta}(x)^{T}\beta_{a^{\prime}}\right\}\right]
=𝟙[AK=dθ(x)].\displaystyle=\mathbbm{1}\left[A_{K}=d_{\theta}(x)\right].

bθb_{\theta} is a deterministic policy putting all its mass on the actions chosen by our decision function dθd_{\theta}. Note that, taking an expectation under bθb_{\theta} in Equation (4) recovers Equation (3). It means that introducing noise relaxed Equation (3) to a differentiable objective. This relaxation is achieved by randomly perturbing the scores of the different actions with Gumbel noise (Huijben et al., 2021). As this perturbation is done in the action space level, it induces properties that are not desirable: (1) The gradient of this policy is an expectation under a potentially large action space, accentuating variance problems. (2) The perturbation scales with the size of the action space, as we need PP random draws of Gumbel noises. (3) Sampling from this policy cannot naturally benefit from approximate MIPS algorithms, as discussed previously.

We observe that the majority of these problems emerge from doing this perturbation in the action space level. With this in mind, we introduce the LGP: Latent Gaussian Perturbation policy, that is defined for a context xx and a slate AKA_{K} by:

πθσ(AK|x)\displaystyle\pi^{\sigma}_{\theta}(A_{K}|x) =𝔼ϵ𝒩(0,σ2IL)[𝟙[AK=argsortKa𝒜{(hθ(x)+ϵ)Tβa}]],\displaystyle=\mathbbm{E}_{\epsilon\sim\mathcal{N}(0,\sigma^{2}I_{L})}\left[\mathbbm{1}\left[A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{(h_{\theta}(x)+\epsilon)^{T}\beta_{a^{\prime}}\right\}\right]\right],

with σ>0\sigma>0, a shared standard deviation across dimensions in the latent space L\mathbbm{R}^{L}. The LGP policy defines a smoothed, differentiable relaxation of the deterministic policy bθb_{\theta} by adding Gaussian noise in the latent space L\mathbbm{R}^{L}. This approach can be generalized by perturbing the latent space with any other continuous distribution QQ. The resulting method, called LRP: Latent Random Perturbation is presented in Appendix A.2. Focusing on LGP, this class of policies present desirable properties:

Fast Sampling. For a given xx, sampling from a LGP policy boils down to sampling a Gaussian noise and computing an argsort as:

AKπθσ(|x)AK=argsortKa𝒜{(hθ(x)+σϵ0)Tβa},ϵ0𝒩(0,IL).A_{K}\sim\pi^{\sigma}_{\theta}(\cdot|x)\iff A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{(h_{\theta}(x)+\sigma\epsilon_{0})^{T}\beta_{a^{\prime}}\right\},\epsilon_{0}\sim\mathcal{N}(0,I_{L}).

As the action embeddings β\beta are fixed, and we are performing a perturbation in the latent space, sampling can be done by calling approximate MIPS on the perturbed query hθϵ(x)=hθ(x)+ϵh^{\epsilon}_{\theta}(x)=h_{\theta}(x)+\epsilon, achieving a complexity of 𝒪(logP)\mathcal{O}(\log P) and improving on the sampling complexity 𝒪(PlogK)\mathcal{O}(P\log K) of the Plackett-Luce family.

Well behaved gradient. Similar to the gradient under the Plackett-Luce policy (6), we can derive a score function gradient for LGP policies. Let AK{h}=argsortKa𝒜{hTβa}A_{K}\{h\}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{h^{T}\beta_{a^{\prime}}\right\} the MIPS result for query hh. Let us write the expected reward under πθσ\pi^{\sigma}_{\theta} for a particular context xx:

R^(πθσ|x)\displaystyle\hat{R}(\pi^{\sigma}_{\theta}|x) =𝔼AKπθσ(|x)[r^(AK,x)]\displaystyle=\mathbf{\mathbbm{E}}_{A_{K}\sim\pi^{\sigma}_{\theta}(\cdot|x)}\left[\hat{r}(A_{K},x)\right]
=𝔼ϵ𝒩(0,σ2IL)[r^(AK{hθ(x)+ϵ},x)]\displaystyle=\mathbf{\mathbbm{E}}_{\epsilon\sim\mathcal{N}(0,\sigma^{2}I_{L})}\left[\hat{r}(A_{K}\{h_{\theta}(x)+\epsilon\},x)\right]
=𝔼h𝒩(hθ(x),σ2IL)[r^(AK{h},x)].\displaystyle=\mathbf{\mathbbm{E}}_{h\sim\mathcal{N}(h_{\theta}(x),\sigma^{2}I_{L})}[\hat{r}(A_{K}\{h\},x)].

The last equality allows us to derive the following gradient:

θR^(πθσ|x)\displaystyle\nabla_{\theta}\hat{R}(\pi^{\sigma}_{\theta}|x) =θ𝔼h𝒩(hθ(x),σ2IL)[r^(AK{h},x)]\displaystyle=\nabla_{\theta}\mathbf{\mathbbm{E}}_{h\sim\mathcal{N}(h_{\theta}(x),\sigma^{2}I_{L})}[\hat{r}(A_{K}\{h\},x)]
=𝔼h𝒩(hθ(x),σ2IL)[r^(AK{h},x)θlogqθ(x,h)],\displaystyle=\mathbf{\mathbbm{E}}_{h\sim\mathcal{N}(h_{\theta}(x),\sigma^{2}I_{L})}[\hat{r}(A_{K}\{h\},x)\nabla_{\theta}\log q_{\theta}(x,h)],

with logqθ(x,h)\log q_{\theta}(x,h) the log density of 𝒩(hθ(x),σ2IL)\mathcal{N}(h_{\theta}(x),\sigma^{2}I_{L}) evaluated in hh. We can obtain an unbiased gradient estimate by sampling ϵ0𝒩(0,IL)\epsilon_{0}\sim\mathcal{N}(0,I_{L}), setting h=hθ(x)+σϵ0h=h_{\theta}(x)+\sigma\epsilon_{0} and computing:

Gθσ(x)\displaystyle G^{\sigma}_{\theta}(x) =r^(AK{h},x)θlogqθ(x,h)\displaystyle=\hat{r}(A_{K}\{h\},x)\nabla_{\theta}\log q_{\theta}(x,h)
=1σr^(AK{h},x)θ(ϵ0Thθ(x))\displaystyle=\frac{1}{\sigma}\hat{r}(A_{K}\{h\},x)\nabla_{\theta}(\epsilon_{0}^{T}h_{\theta}(x)) (9)

This gradient expression solves all issues that the Plackett-Luce gradient estimate suffered from:

  • Fast Gradient Estimate. This gradient can be approximated in a sublinear complexity 𝒪(logP)\mathcal{O}(\log P). Building an estimator of the gradient follows these three steps: (1) We sample ϵ0𝒩(0,IL)\epsilon_{0}\sim\mathcal{N}(0,I_{L}), which is done in a complexity 𝒪(L)𝒪(P)\mathcal{O}(L)\ll\mathcal{O}(P). (2) We evaluate the gradient of ϵ0Thθ(x)\epsilon_{0}^{T}h_{\theta}(x) in 𝒪(L)\mathcal{O}(L) with no dependance on PP. (3) We generate the slate AK{h}A_{K}\{h\}. This boils down to computing an argsort that can be accelerated using approximate MIPS technology, giving a complexity of 𝒪(logP)\mathcal{O}(\log P).

  • Better Variance. LGP’s gradient estimate have better statistical properties for two main reasons: (1) The gradient is defined as an expectation under a continuous distribution on the latent space L\mathbbm{R}^{L}, instead of an expectation under a large discrete action space 𝒜\mathcal{A} of size PLP\gg L. This can have an impact on the variance of the gradient estimator as the sampling space is smaller. (2) The approximate gradient defined in Equation (3) does not depend on the slate size KK. This results in a variance that does not grow with KK, which will translate to substantial performance gains on problems with larger slates. However, the expression of this gradient suggests that our attention should be directed towards the standard deviation σ\sigma instead, as the variance of the gradient estimate defined in Equation (3) will scale in 𝒪(1/σ2)\mathcal{O}(1/\sigma^{2}).

Even if σ\sigma can be treated as an additional parameter, we fix it to σ=1/L\sigma=1/L in all our experiments for a fair comparison. The resulting approach will be hyper-parameter free, and will show both statistical and computational benefits. We give a sketch of the resulting optimization procedure in Algorithm 1. This procedure is easy to implement in any automatic differentiation package (Paszke et al., 2019) and is compatible with stochastic first order optimization algorithms (Ruder, 2016). In the next section, we will measure the benefits of the proposed algorithm in different scenarios.

Inputs: D={xi}i=1ND=\{x_{i}\}_{i=1}^{N}, reward estimator r^\hat{r}, the action embeddings β\beta
Parameters: T1T\geq 1, Monte Carlo samples number S1S\geq 1
Initialise: θ=θ0\theta=\theta_{0}, MIPS index of β\beta, σ=1/L\sigma=1/L
for t=0t=0 to T1T-1 do
       sample a context xDx\sim D
       sample SS standard Gaussian noises ϵ1,,ϵS𝒩(0,IL)\epsilon_{1},...,\epsilon_{S}\sim\mathcal{N}(0,I_{L})
       compute for s[S]s\in[S], hs=hθ(x)+ϵsh_{s}=h_{\theta}(x)+\epsilon_{s}
       compute slates AK{hs}A_{K}\{h_{s}\} for s[S]s\in[S] with MIPS
       Estimate the gradient:
gradθ1Sσs=1Sr^(AK{hs},x)θ(ϵsThθ(x))grad_{\theta}\leftarrow\frac{1}{S\sigma}\sum^{S}_{s=1}\hat{r}(A_{K}\{h_{s}\},x)\nabla_{\theta}(\epsilon_{s}^{T}h_{\theta}(x))
Update the policy parameter θ\theta:
       θθαgradθ\theta\leftarrow\theta-\alpha grad_{\theta}
      
end for
return θ\theta
Algorithm 1 Learning with Latent Gaussian Perturbation

4 Experiments

4.1 Experimental Setting

For our experiments, we focus on learning slate decision functions for the particular case of recommendation as collaborative filtering datasets are easily accessible, facilitating the reproducibility of our results. We choose three collaborative filtering datasets with varying action space size, MovieLens25M (Harper & Konstan, 2015), Twitch (Rappaz et al., 2021) and GoodReads (Wan & McAuley, 2018; Wan et al., 2019). We process these datasets to transform them into user-item interactions of shape [U,P][U,P] with UU and PP the number of users and actions. The statistics of these datasets are described in Table 2.

We follow the same procedure as Sakhi et al. (2023b) to build our experimental setup. Given a dataset, we split randomly the user-item interaction session [X,Y][X,Y] into two parts; the observed interactions XX and the hidden interactions YY. the observed part XX represents all the information we know about the user, and will be used by our policy πθ\pi_{\theta} to deliver slates of interest. The hidden part YY is used to define a reward function that will drive the policy to solve a form of session completion task. For a given slate AK=[a1,,aK]A_{K}=[a_{1},...,a_{K}], we define the reward as:

r^(AK,X)=k=1K𝟙[akY]2k1.\hat{r}(A_{K},X)=\sum_{k=1}^{K}\frac{\mathbbm{1}[a_{k}\in Y]}{2^{k-1}}.

Although we can adopt an arbitrary form for the reward function, we want rewards that depend on the whole slate (Aouali et al., 2023b) and that take into account the ordering of the items. We choose a linear reward to be able to compare our method to PL-Rank (Oosterhuis, 2022), which exploits the reward structure to improve the training of Plackett-Luce policies. The objective we want to optimize is the following:

R^(πθ)\displaystyle\hat{R}(\pi_{\theta}) =1Ui=1U𝔼AKπθ(|Xi)[r^(AK,Xi)].\displaystyle=\frac{1}{U}\sum_{i=1}^{U}\mathbf{\mathbbm{E}}_{A_{K}\sim\pi_{\theta}(\cdot|X_{i})}\left[\hat{r}(A_{K},X_{i})\right].

The next step is to parametrize the policy πθ\pi_{\theta}. For large scale problems, and for a given XX, we are restricted to use the following parametrization of the relevance function fθf_{\theta}:

fθ(a,X)=hθ(X)Tβa,a𝒜.\displaystyle f_{\theta}(a,X)=h_{\theta}(X)^{T}\beta_{a},\quad\forall a\in\mathcal{A}.

Given the observed interactions XX, we compute the action embeddings β\beta using an SVD matrix decomposition (Klema & Laub, 1980). This allows us to project the different action into a latent space of lower dimension LPL\ll P, making β\beta of dimension [L,P][L,P]. In all experiments, β\beta will be fixed unless we want to study the impact of training the embeddings. When β\beta is fixed, we create an approximate MIPS index using the HNSW algorithm (Malkov & Yashunin, 2020) with the help of the FAISS library (Johnson et al., 2019). This index will accelerate decision-making online as described in Equation (2) and can also be exploited to speed up the training of LGP policies. With β\beta defined, we still need to parametrize the user embedding function hθh_{\theta}. Given XX, we first define the mean embedding function M:𝒳LM:\mathcal{X}\to\mathbbm{R}^{L}:

M(X)=1|X|aXβa.M(X)=\frac{1}{|X|}\sum_{a\in X}\beta_{a}.

The function MM computes the average of the item embeddings the user interacted with in XX (Koch et al., 2021). hθh_{\theta} follows as:

hθ(X)=M(X)θ\displaystyle h_{\theta}(X)=M(X)\theta (10)

with θ\theta a parameter of dimension [L,L][L,L], much smaller than [L,P][L,P], the dimension of β\beta. All policies in these experiments will use this parametrization. Experiments with deep policies can be found in Appendix A.3.4. The training is conducted on a CPU machine, using the Adam optimizer (Kingma & Ba, 2014), with a batch size of 32. We tune the learning rate on a validation set for all algorithms. We adopt Algorithm 1 to train LGP and its accelerated variant LGP-MIPS. We denote by PL-PG, the algorithm that trains the Plackett-Luce policy trained with the score function gradient; we sample S1S\geq 1 slates {AK1,,AKS}\{A^{1}_{K},...,A^{S}_{K}\} from πθ\pi_{\theta} to derive the gradient estimate for a given XX:

GθS(X)=1Si=1Sr^(AKs,X)θlogπθ(AKs|X).G^{S}_{\theta}(X)=\frac{1}{S}\sum_{i=1}^{S}\hat{r}(A^{s}_{K},X)\nabla_{\theta}\log\pi_{\theta}(A^{s}_{K}|X). (11)

We also compare our results to PL-Rank (Oosterhuis, 2022) that exploits the linearity of the reward to have a better gradient estimate. As we are mostly interested in the performance of the decision system dθd_{\theta}, all the rewards reported in the experiments are computed using:

R^(dθ)\displaystyle\hat{R}(d_{\theta}) =1Ui=1Ur^(dθ(Xi),Xi).\displaystyle=\frac{1}{U}\sum_{i=1}^{U}\hat{r}(d_{\theta}(X_{i}),X_{i}).

In the next section, we study empirically the performance of these approaches by training them with the same time budget on the different datasets. Additional experiments can be found in Appendix A.3 to confirm the robustness of our results and better understand the behaviour of both the Plackett-Luce policy and the newly introduced LGP policy.

#Actions #Users Interactions Density
MovieLens 25M 55K 162K 0.24%
Twitch 750K 580K 0.008%
GoodReads 2.23M 400K 0.01%
Table 2: The statistics of the datasets after preprocessing

4.2 Performance under the same time budget

To measure the performance of our algorithms, we use all three datasets with their statistics described in Table 2. We fix the latent space dimension L=100L=100 and use a slate size of K=5K=5 for these experiments. We split each dataset by users and keep 10% to create a validation set, on which the reward of the decision function is reported. As these algorithms present different iteration speeds, we fix the same time budget for all training methods for a fair comparison. Training with a time budget also simulates a real production environment, where practitioners are bounded by time constraints and scheduled deployments. For all datasets and training routines, we allow a runtime of 60 minutes, and evaluate our policies on the validation set for 10 equally spaced intervals. The results of these experiments are presented in Figure 1 where we plot the evolution of the validation reward on all datasets, for different values of S{1,10,100}S\in\{1,10,100\}; the number of samples used to approximate the gradient.

Our first observation from the graph is that PL-PG cannot compete with other algorithms, even in the simplest scenario of the MovieLens dataset. Its poor performance is mainly due to the high variance of its gradient estimate. This is confirmed by the performance of PL-Rank. Indeed, the PL-Rank algorithm works with the same policy class, has the same iteration cost (scales linearly in PP) and only differs on the quality of the gradient estimate; exploiting the structure of the reward allow us to obtain an estimate with lower variance. These results confirm our first intuition. PL-PG suffer from large variance problems (even in modest sized problems) and is not suitable to solve large scale slate decision problems.

Let us now focus on our newly proposed algorithms; LGP and its accelerated variant LGP-MIPS. We observe that in all scenarios considered, the acceleration brought by the approximate MIPS index benefits our algorithm in terms of performance. For the same time budget, LGP-MIPS obtains better reward than LGP, with the biggest differences observed on datasets with large action spaces; Twitch and GoodReads. LGP-MIPS always gives the best performing decision function, for all datasets and number of Monte Carlo samples SS considered. These results are promising as our algorithm which is agnostic to the form of the reward outperforms PL-Rank that is solely designed to tackle the particular case of linear rewards. This performance is due to LGP-MIPS’s superior sampling complexity combined with an unbiased, low variance gradient estimate. It is also worthy to note that we were unable to run algorithms optimizing Plackett-Luce policies (PL-PG and PL-Rank) on the GoodReads dataset with S=100S=100 due to its massive memory footprint. As sampling is done on the action space, Plackett-Luce-based methods need for each iteration samples of size 𝒪(SP)\mathcal{O}(SP) compared to LGP-based method for which the sampling is done in the latent space requiring 𝒪(SL)\mathcal{O}(SL) memory usage. Being able to increase the number of Monte Carlo Samples SS is desirable, as it helps reduce the variance of the gradient estimates and accelerates further the training.

These results222Code to reproduce the results will be released upon acceptance. demonstrate the utility of our newly proposed method over Plackett-Luce for learning slate decision systems. The LGP policy class combined with accelerated MIPS indices produces unbiased, low variance gradient estimates that are fast to compute, can scale to massive action spaces and exhibit low memory usage, making our algorithm the best candidate for optimizing large scale slate decision systems.

Refer to caption
Figure 1: The performance of slate decision functions obtained after optimizing them by different training algorithms for the same time budget of 60 minutes. Each evaluation on the validation data is done after 6 minutes of training.

4.3 Performance under the same number of iterations

We proposed a new family of policies that enables fast optimization, and naturally benefits from unbiased gradient estimates with a variance that does not depend on the slate size KK. The results reported in Figure 1 show that the newly proposed methods are suitable for optimizing large scale decision systems, under strict time constraints. We provide further experiments and compare the performance of the obtained policies after training them, with different approaches, under the same number of iterations. This gives insight into the behaviour of the variance of the gradient approximations while neglecting the iteration cost. For the same learning rate, gradient approximations with low variance need less optimization iterations to converge. As PL methods tend to iterate slowly, especially in large action space scenarios (Twitch and Goodreads), we fix the number of iterations in all experiments to the iterations made by PL-PG after 60 minutes of optimization. We report the evolution of the training reward of all methods, for different settings, in Figure 2. We identify two evolution patterns from the plots. A first, fast convergence pattern that PL-Rank alongside the LGP family enjoy, and a second, much slower convergence pattern of PL-PG. This difference in training reward evolution conveys the following message; the LGP family benefit from a gradient estimate with a variance similar to PL-Rank, and much lower than the variance of the PL-PG gradient estimate. In addition, not only can the LGP family iterates faster, but it enjoys low variance gradient estimates and does not rely on a linearity assumption on the reward. This further confirms the improvements brought by the LGP family, motivating it as an excellent alternative to Plackett-Luce for optimize large scale decision systems.

Refer to caption
Figure 2: Comparing the performance of slate decision functions obtained by running our different training algorithms for the same iterations. The newly proposed relaxation LGP have a gradient estimate of quality comparable to PL-Rank as they result in policies with a similar training behaviour. LGP methods enjoy low variance gradient estimate without making an assumption on the structure of the reward, contrary to PL-Rank.

5 Related work

Learning from interactions.

Recent advances in learning large scale decision systems adopt the offline Contextual Bandit/Reinforcement Learning framework (Chen et al., 2022; Wang et al., 2022; Ma et al., 2020; Chen et al., 2019) proving itself as a powerful paradigm to align business metrics with the offline optimization problem. Research in this direction either explore the use of known policy learning algorithms (Chen et al., 2022) for learning online decision systems or define better rewards to guide this learning (Wang et al., 2022; Christakopoulou et al., 2022). Chen et al. (2019) showed that large scale recommender systems can benefit from the contextual bandit formulation and introduced a correction to the REINFORCE gradient to encourage softmax policies to recommend multiple items. Their method can be seen as a heuristic applicable when the slate level reward is decomposable into a sum of single item rewards. This assumption is violated in settings where strong interactions exist between the items in the slate. Our method is versatile, does not assume any structure on the reward, and is able to optimize slate decision systems by introducing a new relaxation, smoothing them into a policy that has a better learning behavior than classical Plackett-Luce.

Scalability. The question of scaling offline policy learning to large scale decision systems has received limited attention. It has been shown that offline softmax policy learning can be scaled to production systems (Chen et al., 2019). Recently, Sakhi et al. (2023b) focused on studying the scalability of optimizing policies tailored to one item recommendation, using the covariance gradient estimator combined with importance sampling (Owen, 2013) to exploit approximate MIPS technology in the training phase. Their gradient approximation is provably biased, sacrificing the convergence guarantees provided by stochastic gradient descent (Ruder, 2016). Our method extends the scope of this paper, as it can be applied to slate recommendation. It provides a simpler relaxation than Plackett-Luce, producing a learning algorithm that benefits from approximate MIPS technology, naturally obtaining unbiased gradient estimates with better statistical properties.

Learning to Rank. The learning to rank literature separates algorithms by the output space they operate on, making a clear distinction between pointwise, pairwise and listwise approaches (Liu, 2009). Our method falls in the latter (Xia et al., 2008), as we operate with policies on the slate level. The majority of work in LTR trains decision systems through the optimization of ranking losses (Wang et al., 2018; Oosterhuis, 2022), defined as differentiable surrogates of ranking metrics or by an adapted maximum likelihood estimation (Rendle et al., 2009; Ma et al., 2021). In the same direction, differentiable sorting algorithms (Grover et al., 2019; Prillo & Eisenschlos, 2020) aim at producing a differentiable relaxation to sorting functions that handle reward on the item level These methods also require linear rewards in addition to training an item-item interaction matrix, quadratic on the action space size, making them unsuitable to massive action spaces. We are interested in training reward-driven, large scale slate decision systems, to learn rankings that are more aligned with arbitrary, complex rewards functions.

Smoothing non-differentiable objectives. Our procedure can be interpreted as a stochastic relaxation of non-differentiable decision functions. This relaxation is achieved by introducing a well chosen noise in the latent space. PAC-Bayesian policy learning (London & Sandler, 2019; Sakhi et al., 2023a; Aouali et al., 2023a) and Black-Box optimization algorithms (Bajaj et al., 2021; Rechenberg, 1978; Staines & Barber, 2012) adopt a similar paradigm to optimize non-smooth loss functions. They proceed by injecting noise in the parameter space and derive a gradient with respect to the noise distribution. This parameter level perturbation can suffer from computational issues when the number of parameters increases. Our method is agnostic to the number of parameters. By perturbing the latent space directly, we bypass this problem and simplify the sampling procedure resulting in faster and more efficient training.

6 Conclusion and future work

Countless large scale online decision systems are tasked with delivering slates based on contextual information. We formulate the learning of these systems in the offline contextual bandit setting. This framework relaxes decision systems to stochastic policies, and proceeds at learning them through REINFORCE-like algorithms. Plackett-Luce provides an interpretable distribution over ordered lists of items but its use in a large scale policy learning context is far from being optimal. In this paper, we motivate the Latent Gaussian Perturbation, a new policy class over ordered lists defined as a stochastic, differentiable relaxation of the argsort decision function, induced by perturbing the latent space. The LGP policy provides gradient estimates with better computational and statistical properties. We built an intuition on why this new policy class is better behaved and demonstrated through extensive experiments that not only LGP is faster to train, considerably reducing the computational cost, but it also produces policies with better performance making the use of Plackett-Luce policies in this context obsolete. This work gives practitioners a new way to address large scale slate policy learning and aim at contributing into the adoption of REINFORCE decision systems. The results obtained in this paper suggest that the relaxations used to define policies have a big role in optimization and that we might need to take a step backward and reconsider the massively adopted Softmax/Plackett-Luce parametrization. As we only focused on simple noise distributions, a nice avenue of research will be to study the impact of the choice of these distributions on the learning algorithm produced.

References

  • Ajalloeian & Stich (2021) Ahmad Ajalloeian and Sebastian U. Stich. On the convergence of sgd with biased gradients, 2021.
  • Aouali et al. (2023a) Imad Aouali, Victor-Emmanuel Brunel, David Rohde, and Anna Korba. Exponential smoothing for off-policy learning. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  984–1017. PMLR, 23–29 Jul 2023a. URL https://proceedings.mlr.press/v202/aouali23a.html.
  • Aouali et al. (2023b) Imad Aouali, Achraf Ait Sidi Hammou, Sergey Ivanov, Otmane Sakhi, David Rohde, and Flavian Vasile. Probabilistic Rank and Reward: A Scalable Model for Slate Recommendation. working paper or preprint, January 2023b. URL https://hal.science/hal-03959643.
  • Bajaj et al. (2021) Ishan Bajaj, Akhil Arora, and M. M. Faruque Hasan. Black-Box Optimization: Methods and Applications, pp.  35–65. Springer International Publishing, Cham, 2021. ISBN 978-3-030-66515-9. doi: 10.1007/978-3-030-66515-9_2. URL https://doi.org/10.1007/978-3-030-66515-9_2.
  • Bottou et al. (2013) Léon Bottou, Jonas Peters, Joaquin Quiñonero-Candela, Denis X. Charles, D. Max Chickering, Elon Portugaly, Dipankar Ray, Patrice Simard, and Ed Snelson. Counterfactual reasoning and learning systems: The example of computational advertising. Journal of Machine Learning Research, 14(65):3207–3260, 2013. URL http://jmlr.org/papers/v14/bottou13a.html.
  • Buchholz et al. (2022) Alexander Buchholz, Jan Malte Lichtenberg, Giuseppe Di Benedetto, Yannik Stein, Vito Bellini, and Matteo Ruffini. Low-variance estimation in the plackett-luce model via quasi-monte carlo sampling, 2022. URL https://arxiv.org/abs/2205.06024.
  • Chen et al. (2019) Minmin Chen, Alex Beutel, Paul Covington, Sagar Jain, Francois Belletti, and Ed H. Chi. Top-k off-policy correction for a reinforce recommender system. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, WSDM ’19, pp.  456–464, New York, NY, USA, 2019. Association for Computing Machinery. ISBN 9781450359405. doi: 10.1145/3289600.3290999. URL https://doi.org/10.1145/3289600.3290999.
  • Chen et al. (2022) Minmin Chen, Can Xu, Vince Gatto, Devanshu Jain, Aviral Kumar, and Ed Chi. Off-policy actor-critic for recommender systems. In Proceedings of the 16th ACM Conference on Recommender Systems, RecSys ’22, pp.  338–349, New York, NY, USA, 2022. Association for Computing Machinery. ISBN 9781450392785. doi: 10.1145/3523227.3546758. URL https://doi.org/10.1145/3523227.3546758.
  • Christakopoulou et al. (2022) Konstantina Christakopoulou, Can Xu, Sai Zhang, Sriraj Badam, Trevor Potter, Daniel Li, Hao Wan, Xinyang Yi, Ya Le, Chris Berg, Eric Bencomo Dixon, Ed H. Chi, and Minmin Chen. Reward shaping for user satisfaction in a reinforce recommender, 2022. URL https://arxiv.org/abs/2209.15166.
  • Dudík et al. (2014) Miroslav Dudík, Dumitru Erhan, John Langford, and Lihong Li. Doubly robust policy evaluation and optimization. Statistical Science, 29(4):485–511, 2014.
  • Gadetsky et al. (2020) Artyom Gadetsky, Kirill Struminsky, Christopher Robinson, Novi Quadrianto, and Dmitry Vetrov. Low-variance black-box gradient estimates for the plackett-luce distribution. Proceedings of the AAAI Conference on Artificial Intelligence, 34(06):10126–10135, Apr. 2020. doi: 10.1609/aaai.v34i06.6572. URL https://ojs.aaai.org/index.php/AAAI/article/view/6572.
  • Gionis et al. (1999) Aristides Gionis, Piotr Indyk, and Rajeev Motwani. Similarity search in high dimensions via hashing. In Proceedings of the 25th International Conference on Very Large Data Bases, VLDB ’99, pp.  518–529, San Francisco, CA, USA, 1999. Morgan Kaufmann Publishers Inc. ISBN 1558606157.
  • Grathwohl et al. (2018) Will Grathwohl, Dami Choi, Yuhuai Wu, Geoff Roeder, and David Duvenaud. Backpropagation through the void: Optimizing control variates for black-box gradient estimation. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=SyzKd1bCW.
  • Grover et al. (2019) Aditya Grover, Eric Wang, Aaron Zweig, and Stefano Ermon. Stochastic optimization of sorting networks via continuous relaxations. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=H1eSS3CcKX.
  • Guo et al. (2020) Ruiqi Guo, Philip Sun, Erik Lindgren, Quan Geng, David Simcha, Felix Chern, and Sanjiv Kumar. Accelerating large-scale inference with anisotropic vector quantization. In International Conference on Machine Learning, 2020. URL https://arxiv.org/abs/1908.10396.
  • Harper & Konstan (2015) F. Maxwell Harper and Joseph A. Konstan. The movielens datasets: History and context. ACM Trans. Interact. Intell. Syst., 5(4), dec 2015. ISSN 2160-6455. doi: 10.1145/2827872. URL https://doi.org/10.1145/2827872.
  • Horvitz & Thompson (1952) Daniel G Horvitz and Donovan J Thompson. A generalization of sampling without replacement from a finite universe. Journal of the American statistical Association, 47(260):663–685, 1952.
  • Huijben et al. (2021) Iris AM Huijben, Wouter Kool, Max B Paulus, and Ruud JG van Sloun. A review of the gumbel-max trick and its extensions for discrete stochasticity in machine learning. arXiv preprint arXiv:2110.01515, 2021.
  • Johnson et al. (2019) Jeff Johnson, Matthijs Douze, and Hervé Jégou. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535–547, 2019.
  • Kingma & Ba (2014) Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • Klema & Laub (1980) V. Klema and A. Laub. The singular value decomposition: Its computation and some applications. IEEE Transactions on Automatic Control, 25(2):164–176, 1980. doi: 10.1109/TAC.1980.1102314.
  • Koch et al. (2021) Olivier Koch, Amine Benhalloum, Guillaume Genthial, Denis Kuzin, and Dmitry Parfenchik. Scalable representation learning and retrieval for display advertising. arXiv preprint arXiv:2101.00870, 2021.
  • L’Ecuyer (2016) Pierre L’Ecuyer. Randomized Quasi-Monte Carlo: An Introduction for Practitioners. In 12th International Conference on Monte Carlo and Quasi-Monte Carlo Methods in Scientific Computing (MCQMC 2016), Stanford, United States, August 2016. URL https://hal.inria.fr/hal-01561550.
  • Liang et al. (2018) Dawen Liang, Rahul G. Krishnan, Matthew D. Hoffman, and Tony Jebara. Variational autoencoders for collaborative filtering. In Proceedings of the 2018 World Wide Web Conference, WWW ’18, pp.  689–698, Republic and Canton of Geneva, CHE, 2018. International World Wide Web Conferences Steering Committee. ISBN 9781450356398. doi: 10.1145/3178876.3186150. URL https://doi.org/10.1145/3178876.3186150.
  • Liu (2009) Tie-Yan Liu. Learning to rank for information retrieval. Found. Trends Inf. Retr., 3(3):225–331, mar 2009. ISSN 1554-0669. doi: 10.1561/1500000016. URL https://doi.org/10.1561/1500000016.
  • London & Sandler (2019) Ben London and Ted Sandler. Bayesian counterfactual risk minimization. In Kamalika Chaudhuri and Ruslan Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp.  4125–4133. PMLR, 09–15 Jun 2019. URL https://proceedings.mlr.press/v97/london19a.html.
  • Ma et al. (2020) Jiaqi Ma, Zhe Zhao, Xinyang Yi, Ji Yang, Minmin Chen, Jiaxi Tang, Lichan Hong, and Ed H. Chi. Off-policy learning in two-stage recommender systems. In Proceedings of The Web Conference 2020, WWW ’20, pp.  463–473, New York, NY, USA, 2020. Association for Computing Machinery. ISBN 9781450370233. doi: 10.1145/3366423.3380130. URL https://doi.org/10.1145/3366423.3380130.
  • Ma et al. (2021) Jiaqi Ma, Xinyang Yi, Weijing Tang, Zhe Zhao, Lichan Hong, Ed Chi, and Qiaozhu Mei. Learning-to-rank with partitioned preference: Fast estimation for the plackett-luce model. In Arindam Banerjee and Kenji Fukumizu (eds.), Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, pp.  928–936. PMLR, 13–15 Apr 2021. URL https://proceedings.mlr.press/v130/ma21a.html.
  • Malkov & Yashunin (2020) Yu A. Malkov and D. A. Yashunin. Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs. IEEE Trans. Pattern Anal. Mach. Intell., 42(4):824–836, apr 2020. ISSN 0162-8828. doi: 10.1109/TPAMI.2018.2889473. URL https://doi.org/10.1109/TPAMI.2018.2889473.
  • Oosterhuis (2022) Harrie Oosterhuis. Learning-to-rank at the speed of sampling: Plackett-luce gradient estimation with minimal computational complexity. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’22, pp.  2266–2271, New York, NY, USA, 2022. Association for Computing Machinery. ISBN 9781450387323. doi: 10.1145/3477495.3531842. URL https://doi.org/10.1145/3477495.3531842.
  • Owen (2013) Art B. Owen. Monte Carlo theory, methods and examples. https://artowen.su.domains/mc/, 2013.
  • Paszke et al. (2019) Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pp.  8024–8035. Curran Associates, Inc., 2019. URL http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf.
  • Plackett (1975) R. L. Plackett. The analysis of permutations. Journal of the Royal Statistical Society. Series C (Applied Statistics), 24(2):193–202, 1975. ISSN 00359254, 14679876. URL http://www.jstor.org/stable/2346567.
  • Prillo & Eisenschlos (2020) Sebastian Prillo and Julian Martin Eisenschlos. Softsort: A continuous relaxation for the argsort operator. In Proceedings of the 37th International Conference on Machine Learning, ICML’20. JMLR.org, 2020.
  • Rappaz et al. (2021) Jérémie Rappaz, Julian McAuley, and Karl Aberer. Recommendation on Live-Streaming Platforms: Dynamic Availability and Repeat Consumption, pp.  390–399. Association for Computing Machinery, New York, NY, USA, 2021. ISBN 9781450384582. URL https://doi.org/10.1145/3460231.3474267.
  • Rechenberg (1978) I. Rechenberg. Evolutionsstrategien. In Berthold Schneider and Ulrich Ranft (eds.), Simulationsmethoden in der Medizin und Biologie, pp.  83–114, Berlin, Heidelberg, 1978. Springer Berlin Heidelberg. ISBN 978-3-642-81283-5.
  • Rendle et al. (2009) Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. Bpr: Bayesian personalized ranking from implicit feedback. In Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, UAI ’09, pp.  452–461, Arlington, Virginia, USA, 2009. AUAI Press. ISBN 9780974903958.
  • Ruder (2016) Sebastian Ruder. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747, 2016.
  • Sakhi et al. (2020) Otmane Sakhi, Stephen Bonner, David Rohde, and Flavian Vasile. BLOB: A Probabilistic Model for Recommendation That Combines Organic and Bandit Signals. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &; Data Mining, KDD ’20, pp.  783–793, New York, NY, USA, 2020. Association for Computing Machinery. ISBN 9781450379984. doi: 10.1145/3394486.3403121. URL https://doi.org/10.1145/3394486.3403121.
  • Sakhi et al. (2023a) Otmane Sakhi, Pierre Alquier, and Nicolas Chopin. PAC-Bayesian Offline Contextual Bandits With Guarantees. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp.  29777–29799. PMLR, 23–29 Jul 2023a. URL https://proceedings.mlr.press/v202/sakhi23a.html.
  • Sakhi et al. (2023b) Otmane Sakhi, David Rohde, and Alexandre Gilotte. Fast Offline Policy Optimization for Large Scale Recommendation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8):9686–9694, Jun. 2023b. doi: 10.1609/aaai.v37i8.26158. URL https://ojs.aaai.org/index.php/AAAI/article/view/26158.
  • Shrivastava & Li (2014) Anshumali Shrivastava and Ping Li. Asymmetric lsh (alsh) for sublinear time maximum inner product search (mips). In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger (eds.), Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc., 2014. URL https://proceedings.neurips.cc/paper/2014/file/310ce61c90f3a46e340ee8257bc70e93-Paper.pdf.
  • Staines & Barber (2012) Joe Staines and David Barber. Variational optimization, 2012. URL https://arxiv.org/abs/1212.4507.
  • Swaminathan & Joachims (2015) Adith Swaminathan and Thorsten Joachims. Counterfactual risk minimization: Learning from logged bandit feedback. In Francis Bach and David Blei (eds.), Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pp.  814–823, Lille, France, 07–09 Jul 2015. PMLR. URL https://proceedings.mlr.press/v37/swaminathan15.html.
  • Swaminathan et al. (2017) Adith Swaminathan, Akshay Krishnamurthy, Alekh Agarwal, Miroslav Dudík, John Langford, Damien Jose, and Imed Zitouni. Off-policy evaluation for slate recommendation. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, pp.  3635–3645, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
  • Vasile et al. (2016) Flavian Vasile, Elena Smirnova, and Alexis Conneau. Meta-prod2vec: Product embeddings using side-information for recommendation. In Proceedings of the 10th ACM Conference on Recommender Systems, RecSys ’16, pp.  225–232, New York, NY, USA, 2016. Association for Computing Machinery. ISBN 9781450340359. doi: 10.1145/2959100.2959160. URL https://doi.org/10.1145/2959100.2959160.
  • Wan & McAuley (2018) Mengting Wan and Julian J. McAuley. Item recommendation on monotonic behavior chains. In Sole Pera, Michael D. Ekstrand, Xavier Amatriain, and John O’Donovan (eds.), Proceedings of the 12th ACM Conference on Recommender Systems, RecSys 2018, Vancouver, BC, Canada, October 2-7, 2018, pp.  86–94. ACM, 2018. doi: 10.1145/3240323.3240369. URL https://doi.org/10.1145/3240323.3240369.
  • Wan et al. (2019) Mengting Wan, Rishabh Misra, Ndapa Nakashole, and Julian J. McAuley. Fine-grained spoiler detection from large-scale review corpora. In Anna Korhonen, David R. Traum, and Lluís Màrquez (eds.), Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers, pp.  2605–2610. Association for Computational Linguistics, 2019. doi: 10.18653/v1/p19-1248. URL https://doi.org/10.18653/v1/p19-1248.
  • Wang et al. (2018) Xuanhui Wang, Cheng Li, Nadav Golbandi, Mike Bendersky, and Marc Najork. The lambdaloss framework for ranking metric optimization. In Proceedings of The 27th ACM International Conference on Information and Knowledge Management (CIKM ’18), pp.  1313–1322, 2018.
  • Wang et al. (2022) Yuyan Wang, Mohit Sharma, Can Xu, Sriraj Badam, Qian Sun, Lee Richardson, Lisa Chung, Ed H. Chi, and Minmin Chen. Surrogate for long-term user experience in recommender systems. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD ’22, pp.  4100–4109, New York, NY, USA, 2022. Association for Computing Machinery. ISBN 9781450393850. doi: 10.1145/3534678.3539073. URL https://doi.org/10.1145/3534678.3539073.
  • Williams (1992) Ronald J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8(3):229–256, 1992. doi: 10.1007/BF00992696.
  • Xia et al. (2008) Fen Xia, Tie-Yan Liu, Jue Wang, Wensheng Zhang, and Hang Li. Listwise approach to learning to rank: Theory and algorithm. In Proceedings of the 25th International Conference on Machine Learning, ICML ’08, pp.  1192–1199, New York, NY, USA, 2008. Association for Computing Machinery. ISBN 9781605582054. doi: 10.1145/1390156.1390306. URL https://doi.org/10.1145/1390156.1390306.

Appendix A Appendix

A.1 Accelerating Plackett-Luce training

Input: hΞh_{\Xi}, β\beta, xx, and indexes on β\beta and parameter KK, catalogue size 𝒫\mathcal{P}
Output: aa which is a sample from P(A=a)=exp(hΞ(x)Tβa)aexp(hΞ(x)Tβa)P(A=a)=\frac{\exp(h_{\Xi}(x)^{T}\beta_{a})}{\sum_{a^{\prime}}\exp(h_{\Xi}(x)^{T}\beta_{a^{\prime}})}
α1,,αK=\alpha_{1},...,\alpha_{K}= argsort(hΞ(x)Tβ)1:K(h_{\Xi}(x)^{T}\beta)_{1:K}
𝒵=𝒫exp(hΞ(x)TβαK)\mathcal{Z}^{\prime}=\mathcal{P}\exp(h_{\Xi}(x)^{T}\beta_{\alpha_{K}})
𝒵′′=aKexp(hΞ(x)Tβa)exp(hΞ(x)TβαK)\mathcal{Z}^{\prime\prime}=\sum_{a^{\prime}}^{K}\exp(h_{\Xi}(x)^{T}\beta_{a^{\prime}})-\exp(h_{\Xi}(x)^{T}\beta_{\alpha_{K}})
PK=[exp(hΞ(x)Tβα1)/𝒵′′,,exp(hΞ(x)TβαK)/𝒵′′]P_{K}=[\exp(h_{\Xi}(x)^{T}\beta_{\alpha_{1}})/\mathcal{Z}^{\prime\prime},...,\exp(h_{\Xi}(x)^{T}\beta_{\alpha_{K}})/\mathcal{Z}^{\prime\prime}]
while True do
       dcat([tail,head],[𝒵𝒵+𝒵′′,𝒵′′𝒵+𝒵′′])d\sim{\rm cat}([{\rm tail,head}],[\frac{\mathcal{Z}^{\prime}}{\mathcal{Z}^{\prime}+\mathcal{Z}^{\prime\prime}},\frac{\mathcal{Z}^{\prime\prime}}{\mathcal{Z}^{\prime}+\mathcal{Z}^{\prime\prime}}])
       if d=head then
             rcat(α1,,αK,PK)r\sim{\rm cat}(\alpha_{1},...,\alpha_{K},P_{K})
             return rr
            
       end if
      if d=tail then
             Sample qq uniformly from the set {1,,P}\{1,...,P\}
             Sample uu from a uniform distribution
             if  exp(hΞ(x)Tβq)exp(hΞ(x)TβαK)>u\frac{\exp(h_{\Xi}(x)^{T}\beta_{q})}{\exp(h_{\Xi}(x)^{T}\beta_{\alpha_{K}})}>u  then
                   return qq
             end if
            
       end if
      
end while
Algorithm 2 Categorical Distribution: Rejection sampling using MIPS

As it was discussed in the main paper, we can derive a covariance gradient that does not require the computation of the normalizing constant Zθ(x)Z_{\theta}(x):

θR^(πθ|x)=𝕠𝕧AKπθ(.|x)[r^(AK,x),i=1Kθfθ(ai,x)],\displaystyle\nabla_{\theta}\hat{R}(\pi_{\theta}|x)=\mathbf{\mathbbm{Cov}}_{A_{K}\sim\pi_{\theta}(.|x)}\left[\hat{r}(A_{K},x),\sum_{i=1}^{K}\nabla_{\theta}f_{\theta}(a_{i},x)\right],

with 𝕠𝕧[A,𝑩]=𝔼[(A𝔼[A]).(𝑩𝔼[𝑩])]\mathbf{\mathbbm{Cov}}[A,\bm{B}]=\mathbf{\mathbbm{E}}[(A-\mathbf{\mathbbm{E}}[A]).(\bm{B}-\mathbf{\mathbbm{E}}[\bm{B}])] a covariance between AA a scalar function, and 𝑩\bm{B} a vector. The proof follows:

θR^(πθ|x)\displaystyle\nabla_{\theta}\hat{R}(\pi_{\theta}|x) =𝔼πθ(|x)[r^(AK,x)θlogπθ(AK|x)]\displaystyle=\mathbf{\mathbbm{E}}_{\pi_{\theta}(\cdot|x)}[\hat{r}(A_{K},x)\nabla_{\theta}\log\pi_{\theta}(A_{K}|x)]
=i=1K𝔼πθ(|x)[r^(AK,x)θlogπθ(ai|x,Ai1)]\displaystyle=\sum_{i=1}^{K}\mathbf{\mathbbm{E}}_{\pi_{\theta}(\cdot|x)}[\hat{r}(A_{K},x)\nabla_{\theta}\log\pi_{\theta}(a_{i}|x,A_{i-1})]
=i=1K𝔼πθ(|x)[r^(AK,x)(θfθ(ai,x)θlogZθi1(x))]\displaystyle=\sum_{i=1}^{K}\mathbf{\mathbbm{E}}_{\pi_{\theta}(\cdot|x)}\left[\hat{r}(A_{K},x)\left(\nabla_{\theta}f_{\theta}(a_{i},x)-\nabla_{\theta}\log Z^{i-1}_{\theta}(x)\right)\right]

Using the log trick, we derive the following equality:

θlogZθi1(x)=𝔼πθ(ai|x,Ai1)[θfθ(a,x)].\nabla_{\theta}\log Z^{i-1}_{\theta}(x)=\mathbbm{E}_{\pi_{\theta}(a_{i}|x,A_{i-1})}[\nabla_{\theta}f_{\theta}(a,x)].

This equality is then injected in the gradient formula derived above to obtain:

θR^(πθ|x)=𝕠𝕧AKπθ(.|x)[r^(AK,x),i=1Kθfθ(ai,x)].\nabla_{\theta}\hat{R}(\pi_{\theta}|x)=\mathbf{\mathbbm{Cov}}_{A_{K}\sim\pi_{\theta}(.|x)}\left[\hat{r}(A_{K},x),\sum_{i=1}^{K}\nabla_{\theta}f_{\theta}(a_{i},x)\right].

This concludes the proof. One can see that for K=1K=1, we recover the results of Sakhi et al. (2023b). This form of gradient still requires sampling from the policy πθ(|x)\pi_{\theta}(\cdot|x) to get a good covariance estimation. To lower the time complexity of this step, we can use Monte Carlo techniques such as Importance Sampling/Rejection Sampling (Owen, 2013) with carefully chosen proposals to achieve fast sampling without sacrificing the accuracy of the gradient approximation.

In Sakhi et al. (2023b), a softmax policy learning algorithm was accelerated by approximating the gradients using a self normalized importance sampling algorithm with a proposal distribution that can both exploit the MIPS structure and is a good approximation of the target softmax distribution. This idea can also be used to motivate a rejection sampling algorithm, as a similar proposal can be shown to form an envelope of the target density. It can also be extended from softmax to Plackett-Luce. When the idea of rejection sampling is combined with the MIPS proposal, it results in the rejection sampling algorithm shown in Algorithm 2. While Algorithm 2 can be extended to the slate policy case, enabling the fast evaluation of the covariance gradient estimator, it will still suffer from high variance gradient estimates.

A.2 LRP: Latent Random Perturbation

The method presented in the main paper can be generalized. Instead of focusing on Gaussian distributions, the latent perturbation can be done by any continuous distribution resulting in the more general LRP: Latent Random Perturbation policy. For a context xx and a slate AKA_{K}, its expression is given by:

πθ𝒬(AK|x)\displaystyle\pi^{\mathcal{Q}}_{\theta}(A_{K}|x) =𝔼ϵ𝒬[𝟙[AK=argsortKa𝒜{(hθ(x)+ϵ)Tβa}]],\displaystyle=\mathbbm{E}_{\epsilon\sim\mathcal{Q}}\left[\mathbbm{1}\left[A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{(h_{\theta}(x)+\epsilon)^{T}\beta_{a^{\prime}}\right\}\right]\right],

with 𝒬\mathcal{Q} a continuous distribution on the latent space L\mathbbm{R}^{L}. the LRP policy defines a smoothed, differentiable relaxation of the deterministic policy bθb_{\theta} by adding noise in the latent space L\mathbbm{R}^{L}. This class of policies present desirable properties:

Fast Sampling.

For a given xx, sampling from a LRP policy boils down to sampling from 𝒬\mathcal{Q} and computing an argsort as:

AKπθ𝒬(|x)AK=argsortKa𝒜{(hθ(x)+ϵ)Tβa},ϵ𝒬.A_{K}\sim\pi^{\mathcal{Q}}_{\theta}(\cdot|x)\iff A_{K}=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{(h_{\theta}(x)+\epsilon)^{T}\beta_{a^{\prime}}\right\},\epsilon\sim\mathcal{Q}.

Let us suppose the sampling from 𝒬\mathcal{Q} is easy. As the action embeddings β\beta are fixed, and we are performing a perturbation in the latent space, we can set hθϵ(x)=hθ(x)+ϵh^{\epsilon}_{\theta}(x)=h_{\theta}(x)+\epsilon which makes sampling compatible with approximate MIPS technology making the sampling achievable in 𝒪(logP)\mathcal{O}(\log P), better than the sampling complexity 𝒪(PlogK)\mathcal{O}(P\log K) of the Placett-Luce family.

Well behaved gradient.

Similar to the gradient under the Plackett-Luce policy (6), we can derive a score function gradient for LRP policies. For a given xx, let us write its expected reward under πθ𝒬\pi^{\mathcal{Q}}_{\theta}:

R^(πθ𝒬|x)\displaystyle\hat{R}(\pi^{\mathcal{Q}}_{\theta}|x) =𝔼AKπθ𝒬(|x)[r^(AK,x)]\displaystyle=\mathbf{\mathbbm{E}}_{A_{K}\sim\pi^{\mathcal{Q}}_{\theta}(\cdot|x)}\left[\hat{r}(A_{K},x)\right]
=𝔼ϵ𝒬[r^(AK{hθ(x)+ϵ},x)]\displaystyle=\mathbf{\mathbbm{E}}_{\epsilon\sim\mathcal{Q}}\left[\hat{r}(A_{K}\{h_{\theta}(x)+\epsilon\},x)\right]
=𝔼h𝒬θ(x)[r^(AK{h},x)]\displaystyle=\mathbf{\mathbbm{E}}_{h\sim\mathcal{Q}_{\theta}(x)}[\hat{r}(A_{K}\{h\},x)]
h𝒬θ(x)\displaystyle h\sim\mathcal{Q}_{\theta}(x) h=hθ(x)+ϵ,ϵ𝒬\displaystyle\iff h=h_{\theta}(x)+\epsilon,\quad\epsilon\sim\mathcal{Q}
AK{h}\displaystyle A_{K}\{h\} =argsortKa𝒜{hTβa}.\displaystyle=\operatornamewithlimits{argsort^{K}}_{a^{\prime}\in\mathcal{A}}\left\{h^{T}\beta_{a^{\prime}}\right\}.

𝒬θ(x)\mathcal{Q}_{\theta}(x) is the induced distribution on the user embeddings. 𝒬θ(x)\mathcal{Q}_{\theta}(x) has a tractable density if we choose a classical noise distribution 𝒬\mathcal{Q} (e.g. Gaussian). Working with the induced distribution 𝒬θ(x)\mathcal{Q}_{\theta}(x) transforms the learning parameters θ\theta to parameters of the distribution, allowing us to derive the following gradient:

θR^(πθ𝒬|x)\displaystyle\nabla_{\theta}\hat{R}(\pi^{\mathcal{Q}}_{\theta}|x) =θ𝔼h𝒬θ(x)[r^(AK{h},x)]\displaystyle=\nabla_{\theta}\mathbf{\mathbbm{E}}_{h\sim\mathcal{Q}_{\theta}(x)}[\hat{r}(A_{K}\{h\},x)]
=𝔼h𝒬θ(x)[r^(AK{h},x)θlogqθ(x,h)],\displaystyle=\mathbf{\mathbbm{E}}_{h\sim\mathcal{Q}_{\theta}(x)}[\hat{r}(A_{K}\{h\},x)\nabla_{\theta}\log q_{\theta}(x,h)],

with logqθ(x,h)\log q_{\theta}(x,h) the log density of 𝒬θ(x)\mathcal{Q}_{\theta}(x) evaluated in hh. We can obtain an unbiased gradient estimate by sampling h𝒬θ(x)h\sim\mathcal{Q}_{\theta}(x):

Gθ𝒬(x)=r^(AK{h},x)θlogqθ(x,h).\displaystyle G^{\mathcal{Q}}_{\theta}(x)=\hat{r}(A_{K}\{h\},x)\nabla_{\theta}\log q_{\theta}(x,h). (12)

This gradient expression solves all issues that the Plackett-Luce gradient estimate suffered from:

  • Fast Gradient Estimate. This gradient can be approximated in a sublinear complexity 𝒪(logP)\mathcal{O}(\log P). Building an estimator of the gradient follows these three steps: (1) We sample h𝒬θ(x)h\sim\mathcal{Q}_{\theta}(x), which boils down to adding the noise ϵ𝒬\epsilon\sim\mathcal{Q} to hθ(x)h_{\theta}(x). This can be done in a complexity 𝒪(L)𝒪(P)\mathcal{O}(L)\ll\mathcal{O}(P) if 𝒬\mathcal{Q} is chosen properly. (2) We evaluate the gradient of the log density θlogqθ(x,h)\nabla_{\theta}\log q_{\theta}(x,h) on hh. With a well-chosen 𝒬\mathcal{Q}, this can be done in 𝒪(L)\mathcal{O}(L) as we do not need to normalize over a large discrete action space. (3) We generate the slate AK{h}A_{K}\{h\}. This boils down to computing an argsort which can be accelerated using approximate MIPS technology, giving a complexity of 𝒪(logP)\mathcal{O}(\log P).

  • Better Variance. LRP’s gradient estimate have better statistical properties for two main reasons: (1) The gradient is defined as an expectation under a continuous distribution on the latent space L\mathbbm{R}^{L}, instead of an expectation under a large discrete action space 𝒜\mathcal{A} of size PLP\gg L. This can have an impact on the variance of the gradient estimator as the sampling space is smaller. (2) The approximate gradient defined in Equation (12) does not depend on the slate size KK. This results in a variance that does not grow with KK. This means that we will notice substantial gains when training policies with larger slates.

A.3 Additional Experiments

A.3.1 Robustness of the results

We conduct a further simulation study to be sure that the results obtained in the main paper are robust both to initialisation and to the slate size KK. To this end, we repeat the same experimental setup as in the performance subsection; we use all three datasets, and we fix the latent space dimension L=100L=100. We run the optimization of our algorithms for 1010 minutes and evaluate the obtained policy on the validation split. For each dataset, we define two settings; a setting with a moderate slate size of K=5K=5 and one with a bigger slate size of K=100K=100. The training procedure is repeated for 6 different seeds, and we present the average performance and the standard error for different values of S{1,10,100}S\in\{1,10,100\}; the number of samples used to approximate the gradient. We present results of two different natures:

  • The performance on the validation set, of the obtained policies after training them with different algorithms, under the same time constraint of 10 minutes, is reported in Table 3.

  • The performance on the training set, of the obtained policies, after training them for the same number of iterations. As training with PL algorithms can be very slow, we fix the number of iterations to the iterations done by PL-PG in 10 minutes. The results of these experiments are reported in Table 4.

Let us first focus on the performance obtained after training policies under the same time budget. Looking at Table 3, we can clearly observe that LGP, especially its MIPS accelerated version, always yields the best performing policies no matter the setting adopted. The obtained results are significant, as the standard error is small compared to the difference of performance between the methods. These observations confirm that the LGP family is perfectly suited for industrial applications, as it iterates faster and is robust to different settings. It is noteworthy that the PL family cannot run with big Monte Carlo samples SS, especially in applications with large action spaces, as it suffers from a large memory footprint.

For the sake of completeness, we also report in Table 4, the performance of our decision systems after training them with different approaches under a fixed number of iterations. One can observe that PL-Rank results in the best decision systems, followed by the LGP family. These results suggest that by exploiting the linearity of the reward, the gradient estimates derived by PL-Rank enjoy a better variance than the gradient estimates of the LGP family. The application of PL-Rank however is only restricted to linear rewards, contrary to our newly proposed approach that does not make any assumptions on the structure of the reward. This means that when the reward is complex, LGP methods should be adopted, as they offer a better alternative to the naive, gradient estimate of PL-PG.

Algorithm SS MovieLens Twitch GoodReads
K=5K=5 K=100K=100 K=5K=5 K=100K=100 K=5K=5 K=100K=100
LGP-MIPS S=1S=1 0.834±0.0030.834\pm 0.003 0.599±0.0020.599\pm 0.002 0.438±0.0090.438\pm 0.009 0.343±0.0050.343\pm 0.005 0447±0.0020447\pm 0.002 0.396±0.0070.396\pm 0.007
S=10S=10 0.866±0.0030.866\pm 0.003 0.881±0.0030.881\pm 0.003 0.567±0.0070.567\pm 0.007 0.563±0.0040.563\pm 0.004 0.539±0.0040.539\pm 0.004 0.554±0.0040.554\pm 0.004
S=100S=100 0.865±0.0020.865\pm 0.002 0.867±0.0020.867\pm 0.002 0.607±0.0050.607\pm 0.005 0.609±0.0050.609\pm 0.005 0.580±0.0010.580\pm 0.001 0.592±0.0020.592\pm 0.002
LGP S=1S=1 0.753±0.0050.753\pm 0.005 0.776±0.0040.776\pm 0.004 0.171±0.0040.171\pm 0.004 0.195±0.0040.195\pm 0.004 0.259±0.0130.259\pm 0.013 0.278±0.0130.278\pm 0.013
S=10S=10 0.716±0.0040.716\pm 0.004 0.744±0.0050.744\pm 0.005 0.133±0.0020.133\pm 0.002 0.152±0.0030.152\pm 0.003 0.227±0.0120.227\pm 0.012 0.244±0.0120.244\pm 0.012
S=100S=100 0.700±0.0030.700\pm 0.003 0.721±0.0030.721\pm 0.003 0.097±0.0020.097\pm 0.002 0.111±0.0030.111\pm 0.003 0.192±0.0100.192\pm 0.010 0.204±0.0100.204\pm 0.010
PL-Rank S=1S=1 0.750±0.0040.750\pm 0.004 0.769±0.0050.769\pm 0.005 0.103±0.0030.103\pm 0.003 0.113±0.0030.113\pm 0.003 0.188±0.0090.188\pm 0.009 0.200±0.0100.200\pm 0.010
S=10S=10 0.584±0.0130.584\pm 0.013 0.597±0.0100.597\pm 0.010 0.087±0.0020.087\pm 0.002 0.096±0.0030.096\pm 0.003 0.179±0.0090.179\pm 0.009 0.191±0.0100.191\pm 0.010
S=100S=100 0.342±0.0150.342\pm 0.015 0.353±0.0160.353\pm 0.016 0.084±0.0020.084\pm 0.002 0.093±0.0030.093\pm 0.003 N/A N/A
PL-PG S=1S=1 0.347±0.0190.347\pm 0.019 0.356±0.0200.356\pm 0.020 0.084±0.0020.084\pm 0.002 0.093±0.0020.093\pm 0.002 0.181±0.0090.181\pm 0.009 0.194±0.0090.194\pm 0.009
S=10S=10 0.345±0.0210.345\pm 0.021 0.354±0.0210.354\pm 0.021 0.084±0.0020.084\pm 0.002 0.093±0.0030.093\pm 0.003 0.180±0.0090.180\pm 0.009 0.192±0.0100.192\pm 0.010
S=100S=100 0.323±0.0180.323\pm 0.018 0.332±0.0180.332\pm 0.018 0.084±0.0020.084\pm 0.002 0.093±0.0030.093\pm 0.003 N/A N/A
Table 3: Performance by time budget (10 minutes). Results obtained by the different algorithms while changing the slate size KK and the number of Monte Carlo samples SS. All runs were completed in 10 minutes. We report in this table the average performance and its standard error over 6 different seeds. All statistics are reported on the validation set. The proposed algorithms give robust results in different settings.
Algorithm SS MovieLens Twitch GoodReads
K=5K=5 K=100K=100 K=5K=5 K=100K=100 K=5K=5 K=100K=100
LGP-MIPS S=1S=1 0.645±0.0110.645\pm 0.011 0.507±0.0040.507\pm 0.004 0.098±0.0040.098\pm 0.004 0.107±0.0050.107\pm 0.005 0.192±0.0100.192\pm 0.010 0.205±0.0110.205\pm 0.011
S=10S=10 0.528±0.0120.528\pm 0.012 0.537±0.0110.537\pm 0.011 0.086±0.0030.086\pm 0.003 0.096±0.0040.096\pm 0.004 0.179±0.0070.179\pm 0.007 0.194±0.0070.194\pm 0.007
S=100S=100 0.361±0.0160.361\pm 0.016 0.374±0.0170.374\pm 0.017 0.079±0.0050.079\pm 0.005 0.088±0.0060.088\pm 0.006 0.185±0.0070.185\pm 0.007 0.198±0.0060.198\pm 0.006
LGP S=1S=1 0.623±0.0090.623\pm 0.009 0.646±0.0040.646\pm 0.004 0.094±0.0040.094\pm 0.004 0.108±0.0050.108\pm 0.005 0.192±0.0100.192\pm 0.010 0.207±0.0110.207\pm 0.011
S=10S=10 0.513±0.0120.513\pm 0.012 0.524±0.0100.524\pm 0.010 0.086±0.0030.086\pm 0.003 0.096±0.0040.096\pm 0.004 0.179±0.0070.179\pm 0.007 0.193±0.0070.193\pm 0.007
S=100S=100 0.360±0.0150.360\pm 0.015 0.372±0.0160.372\pm 0.016 0.081±0.0050.081\pm 0.005 0.088±0.0050.088\pm 0.005 0.184±0.0070.184\pm 0.007 0.197±0.0060.197\pm 0.006
PL-Rank S=1S=1 0.750±0.0060.750\pm 0.006 0.779±0.0030.779\pm 0.003 0.102±0.0040.102\pm 0.004 0.113±0.0050.113\pm 0.005 0.188±0.0100.188\pm 0.010 0.203±0.0100.203\pm 0.010
S=10S=10 0.587±0.0110.587\pm 0.011 0.600±0.0090.600\pm 0.009 0.082±0.0030.082\pm 0.003 0.092±0.0030.092\pm 0.003 0.176±0.0070.176\pm 0.007 0.190±0.0070.190\pm 0.007
S=100S=100 0.332±0.0160.332\pm 0.016 0.334±0.0160.334\pm 0.016 0.081±0.0050.081\pm 0.005 0.088±0.0050.088\pm 0.005 N/A N/A
PL-PG S=1S=1 0.342±0.0170.342\pm 0.017 0.356±0.0190.356\pm 0.019 0.084±0.0040.084\pm 0.004 0.096±0.0040.096\pm 0.004 0.181±0.0100.181\pm 0.010 0.196±0.0040.196\pm 0.004
S=10S=10 0.348±0.0180.348\pm 0.018 0.355±0.0200.355\pm 0.020 0.079±0.0020.079\pm 0.002 0.090±0.0030.090\pm 0.003 0.176±0.0070.176\pm 0.007 0.190±0.0070.190\pm 0.007
S=100S=100 0.305±0.0180.305\pm 0.018 0.313±0.0180.313\pm 0.018 0.081±0.0050.081\pm 0.005 0.088±0.0050.088\pm 0.005 N/A N/A
Table 4: Performance by iteration. Results obtained with the different algorithms while changing the slate size KK and the number of Monte Carlo samples SS. The number of iterations is decided by the slowest iterating algorithm (PL-PG) after running for 10 minutes. We report in this table the average performance and its standard error over 6 different seeds. PL-Rank gives the best results, followed by the LGP family.

A.3.2 Effect of fixing β\beta

In this section, we want to validate the intuition we built throughout the paper about the behaviours of both the Plackett-Luce and LGP policy classes. We focus on MovieLens (Harper & Konstan, 2015), a medium scale dataset that allows us to test all our methods, regardless of their potential to scale to harder problems. For all experiments in this section, we fix L=100PL=100\ll P and we study the impact of fixing the action embeddings β\beta. As discussed in Section 3, having the embeddings fixed is a natural solution to improve the learning of these large scale decision systems as it can reduce both the variance of the gradient estimates and the running time of the optimization procedure. To validate this, we focus on the Plackett-Luce policy and define two slightly different parametrizations:

  • Learn θ\theta: this is the parametrization introduced in Equation (10), with β\beta fixed and we only learn the parameter θ\theta.

  • Learn β\beta: We treat β\beta as a parameter after initializing it with the SVD values. Because our user embedding function hθh_{\theta} is linear in θ\theta, we get rid of this parameter as it becomes redundant once β\beta can be optimized. This gives the following parametrization:

(x,a)𝒳×𝒜,fβ(a,x)=M(X)Tβa\forall(x,a)\in\mathcal{X}\times\mathcal{A},\quad f_{\beta}(a,x)=M(X)^{T}\beta_{a}

We train a Plackett-Luce policy with both parametrizations for one epoch, while fixing the slate size K=2K=2 and the Monte Carlo samples to S=1S=1. In this experiment, our objective is not to produce the best policies but to understand how fixing the embeddings can impact the optimization procedure. We report in Figure 3 the evolution of both the gradient estimate variance and the reward on the training data for both parametrizations. We can observe that treating β\beta as a parameter to optimize, even when initialized properly leads to slow learning. The variance of the gradient estimate when learning β\beta is bigger than the variance when learning θ\theta with β\beta fixed, and this will get bigger for problems with larger action spaces as PLP\gg L. We suspect that this is one of the reasons that explain the pace of learning when optimizing β\beta. The same phenomena was observed in Sakhi et al. (2023b). The same experiments demonstrated that training θ\theta alone was twice as fast as training β\beta in this experiment. This suggests that fixing β\beta to a good value is beneficial for training large scale decision systems both in terms of iteration efficiency. We advocate for fixing the action embeddings β\beta when learning large scale MIPS systems.

Refer to caption
Figure 3: Experiments on the MovieLens dataset: We look at the effect of fixing the action embeddings β\beta on the training of Plackett-Luce policies. Training β\beta results in a slow optimization procedure and gradient estimates with bigger variance.

A.3.3 Impact of the slate size KK.

One of the caveats of the Plackett-Luce slate policy is that its gradient estimate has a variance that grows with the slate size KK, reducing its scope of applications to modest slate sizes. the gradient estimate of LGP however does not suffer from this issue, and we want to showcase that with a simple experiment. We derived gradient estimates for LGP-based methods that have a variance that scales in 𝒪(1/σ2)\mathcal{O}(1/\sigma^{2}). Although the standard deviation σ\sigma can be treated as a hyperparameter depending on the task, to allow a fair comparison of the gradient variance of these methods, we set σ\sigma to a particular value coming from the following observation. For a particular action aa:

h𝒩(hθ(x),σ2IL)\displaystyle h\sim\mathcal{N}(h_{\theta}(x),\sigma^{2}I_{L}) hTβa𝒩(hθ(x)Tβa,σ2βa2)\displaystyle\implies h^{T}\beta_{a}\sim\mathcal{N}(h_{\theta}(x)^{T}\beta_{a},\sigma^{2}||\beta_{a}||^{2})
hTβa=hθ(x)Tβa+ϵa\displaystyle\implies h^{T}\beta_{a}=h_{\theta}(x)^{T}\beta_{a}+\epsilon_{a}

with ϵa𝒩(0,(σβa)2)\epsilon_{a}\sim\mathcal{N}(0,(\sigma||\beta_{a}||)^{2}). This can be interpreted as adding a scaled guassian noise to the score of action aa. As we add standardized Gumbel noise γa𝒢(0,1)\gamma_{a}\sim\mathcal{G}(0,1) to the action scores to define the Plackett-Luce policy, we want, for a fair comparison, to have:

a𝒜,σβa1\forall a\in\mathcal{A},\quad\sigma||\beta_{a}||\approx 1

One heuristic to approximately achieve that is to compute the empirical mean of the β\beta norms B=1Pa𝒜βaB=\frac{1}{P}\sum_{a\in\mathcal{A}}||\beta_{a}|| and set σ=1/B\sigma=1/B. This value will be used for this experiment.

We use the MovieLens dataset, and train the LGP policy without exploiting the MIPS index on β\beta, and Plackett-Luce with the parametrization of Equation (10) for 10 epochs with S=1S=1, while varying the slate size K{2,5,10}K\in\{2,5,10\}. Note that all policies are initialized with the same random seed for a fair comparison. We report the evolution of the variance of the gradient estimate alongside the reward on the training data. The results of these experiments are presented in Figure 4.

Focusing on the evolution of the variance, we can see that Plackett-Luce does indeed have a variance that grows with KK contrary to LGP that has a variance of its gradient estimate staying at the same scale no matter the value of KK. We argue that this has a direct impact on the optimization procedure as for the same value of KK, we observe in Figure 1 that LGP-based methods outperform Plackett-Luce learning schemes consistently, making it a good candidate for learning slate policies.

Refer to caption
Figure 4: Impact of the slate size KK on the log variance of the gradient estimate of both PL-PG and LGP on MovieLens. Contrary to LGP, PL-PG has a gradient estimate with a variance that grows with KK.

A.3.4 Experiments with Neural Networks.

For these experiments, we want to explore deep policies and see if we can still empirically validate our findings in this case as well. For this, we adopt the following function to compute the user embedding hθ1,θ2h_{\theta_{1},\theta_{2}}:

hθ1,θ2(X)=sigmoid(M(X)Tθ1)θ2,\displaystyle h_{\theta_{1},\theta_{2}}(X)=\texttt{sigmoid}\left(M(X)^{T}\theta_{1}\right)\theta_{2}, (13)

which boils down to a sigmoid, two layer feed forward neural network with both θ1\theta_{1} and θ2\theta_{2} of size [L,L][L,L]. We run PL-PG, PL-Rank and LGP-MIPS on the three datasets, for the same running time (60 minutes) and cross-validate the learning rate choosing the best value for each algorithm. We aggregate the results on Figure 5. The plot suggests that even for deep policies, LGP-MIPS outperforms Plackett-Luce-based methods on all datasets and for different values of the number of Monte Carlo samples SS. We also observe that training in this case is more unstable, especially for Plackett-Luce-based methods as having more parameters accentuate the variance problems of their gradient estimates. It is noteworthy that, the reward obtained with deep policies in our experiment is less than the one achieved by linear policies. This suggests that deep policies require additional care when training, and we might want to stick to simple policies if we are interested in fast and reliable optimization.

Refer to caption
Figure 5: The performance of slate decision functions, with Neural Network Backbones, obtained by our different training algorithms after running the optimization for 60 minutes.