This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Density Weighting for Multi-Interest Personalized Recommendation

Nikhil Mehta, Anima Singh, Xinyang Yi, Sagar Jain, Lichan Hong and Ed H. Chi
Google
Abstract

Using multiple user representations (MUR) to model user behavior instead of a single user representation (SUR) has been shown to improve personalization in recommendation systems. However, the performance gains observed with MUR can be sensitive to the skewness in the item and/or user interest distribution. When the data distribution is highly skewed, the gains observed by learning multiple representations diminish since the model dominates on head items/interests, leading to poor performance on tail items. Robustness to data sparsity is therefore essential for MUR-based approaches to achieve good performance for recommendations. Yet, research in MUR and data imbalance have largely been done independently. In this paper, we delve deeper into the shortcomings of MUR inferred from imbalanced data distributions. We make several contributions: (1) Using synthetic datasets, we demonstrate the sensitivity of MUR with respect to data imbalance, (2) To improve MUR for tail items, we propose an iterative density weighting scheme (IDW) with user tower calibration to mitigate the effect of training over long-tail distribution on personalization, and (3) Through extensive experiments on three real-world benchmarks, we demonstrate IDW outperforms other alternatives that address data imbalance.

1 Introduction

This work was done when the author was an intern at Google. Correspondence to {nikhilmehta,animasingh}@google.com

Recommender systems are at the core of various online services [49, 16, 34, 38, 2, 40], allowing users to discover items that are of potential interest from a large catalog. Therefore, it is vital for recommendation systems to model user interest behavior effectively. This can be especially challenging for sequential recommender system that takes a sequence of user’s past behavior to predict user’s future interaction, since user’s behavior sequence typically reflects a diverse and dynamic interest profile.

A large scale recommendation system usually consists of two stages, the retrieval stage and the ranking stage. The retrieval stage is critical in narrowing down the candidates items that are relevant to user interests. Existing approaches formulate the retrieval task by assuming a unified low-dimensional representation space for users and items, where a user representation encodes the interest profile of the user and is used to retrieve top candidate items. Recent work [31, 10] has shown that a fixed-length representation can become a bottleneck in modeling user interests for large-scale datasets. Modern approaches [10, 27] circumvent the limitations of a single user representation (SUR) by using multiple user representations (MUR) to model the holistic interest profile from the past behaviors of users. A synthetic experiment in Figure 1 (top-row) illustrates the advantage of using MUR (with 5 representations) over SUR to model the user interests. Specifically, SUR-based model is unable to capture multiple interest regions, whereas MUR can learn multiple representations to enable this, allowing for better personalization.

Refer to caption
Figure 1: The learned item representations using Single (SUR) vs. Multiple User Representations (MUR) on synthetic user-item datasets with different levels of imbalance. The cluster-ids indicate various user interests. Top-row: Low imbalance in user interests. Bottom-row: High imbalance in user interests. The plots depict the learned item representations with SUR/MUR model, and the user representation learned for a random user with two interests.

While MUR can certainly address the desideratum of modeling capacity required for diverse and volatile interests, the overall imbalance in the user interest distribution can hinder the model’s ability to learn representations that reflect different interests of the user. Although there have been recent works on different approaches to learn multiple user representations [31, 10, 27], studying the effect of interest/item imbalance on the ability to capture multiple interests with  MUR is largely unexplored. Without taking into account the imbalance problem, existing methods for MUR implicitly assume that the so-called interests of the users are equally represented in the dataset. Unfortunately, real-world settings rarely offer a balanced distribution of items/interests. In fact, training models over highly skewed long-tail distributions is one of the most notorious problems prevalent when building recommender systems [35]. Figure 1 (bottom-row) highlights the issues observed with imbalanced data. Both SUR and MUR dominate on the head interests, while completely ignoring other interests. Therefore, an MUR-based recommender system can be vulnerable to the imbalanced distribution in the training data. We show that a highly skewed distribution of items/interests causes the gains from using  MUR over  SUR to diminish since the model is unable to capture multiple interests with  MUR.

In this paper, we identify and address the vulnerability of MUR to interest imbalance by making a number of technical innovations. We begin by analyzing MUR on synthetic datasets generated for the sequential recommendation task. The generative mechanism, despite its simplicity, can be used to control several characteristics that are inspired from the real-world datasets, including sparsity in items and user interests, user interest volatility, and interest exploration. The main motivation behind using a synthetic dataset is to formulate a framework to analyze the effect of interest and item imbalance on the retrieval performance. Using synthetic experiments, we derive two important insights. First, we empirically show that the relative improvement observed by using MUR over SUR decreases, especially for tails items, as we increase the skewness in the interest distribution. Second, we show, both quantitatively and qualitatively, that a model trained using MUR leads to a better clustering structure for item representations compared to a model trained with a SUR. However, as we increase the skewness in the data, the quality of clustering structure decreases due to underrepresented tail items in the training dataset.

Our findings from the analysis on synthetic data motivate our key contribution to address the vulnerability of MUR to skewness and improve its robustness to data imbalance. We propose a novel density-based weighting approach, referred to as Iterative Density Weighting (IDW), designed to reduce the sensitivity of MUR to imbalanced training distribution. IDW leverages the item representation space to mitigate the effect of training the recommender model on a highly skewed item/interest distribution. In our synthetic experiments, we show that IDW improves the robustness of MUR to distribution skewness by consistently outperforming the baseline  MUR for overall and tail performance. Moreover, IDW leads to an improved clustering structure of the item representations even when the model is trained on a highly skewed interest distribution. On applying the proposed weighting method on real-world datasets, we observe improved performance relative to several other baselines.

To summarize, the main contributions of this work are as follows:

  • A comprehensive analysis on synthetic data, describing the sensitivity of model performance and clustering structure to data skewness for SUR and MUR. Our analysis depicts the advantage of MUR over SUR from the standpoint of the structure of the learned item representations, which could be of independent interest to the community when comparing MUR and SUR models.

  • We propose a novel density-based weighting mechanism that leverage improved item representation structure to make MUR robust to imbalanced interest distributions. Combination of density-based weighting with user model calibration results in improved tail item recommendations along with better clustering structure of the learned item representations.

  • Extensive experimental results on three real-world benchmarks show that IDW consistently improves tail item recommendations significantly over baseline MUR, while also achieving solid improvements for the overall performance in both hit ratio (HR@K) and ranking (NDCG@K) metric.

2 Related Work

Multi-Interest User Representations. Extracting multiple interest user representations for recommendation systems was first proposed in [47], where a user is modeled by a set of latent vectors, and the recommendation score of each item is defined as the maximum dot product of the item representation and each latent vector. This was extended in the Embedding-Mixture Factorization (EM-F) and the Projection-Mixture Factorization (PM-F) [27] where the item recommendation score is a weighted combination of recommendation scores obtained using each user interest representation. Recently, there has been work on sequential recommender systems where multiple interest representations are used to model a user’s item history. In [27], Mixture LSTM (M-LSTM) was proposed that combines the LSTM network with the PM-F model to extract multiple user representations in a sequential setting. ComiRec [30] and MIND [10] are recent methods that propose multi-interest extractor layers based on the capsule routing mechanism. Although there is a vast amount of work that has focused on model architectures for MUR, the robustness of  MUR to imbalanced user interest distributions is largely unexplored. We analyze the effect of skewness in user interests on  MUR and propose a novel density-based weighting scheme for improving the recommendation performance of tail items with MUR. While this paper focuses on a transformer encoder-based architecture to model  MUR (details in Section 3.1), our work nicely complements prior advances made in modeling  MUR since the proposed density-based weighting can be used with other multi-interest frameworks [30, 10, 27] proposed for recommendation systems.

Long-tail Learning. In most real-world datasets, the data distribution is highly skewed and often follows a long-tail distribution. Learning models on long-tail distribution has been studied across many areas including computer vision [12, 33], natural language processing [13], and recommendation systems [35]. In recommendation datasets, the long-tail distribution is especially apparent due to the large scale of items and highly skewed user interactions. The recommender systems are heavily influenced [14, 5] by the long-tail distribution, leading to poor performance for the tail items.

There are several approaches that can allow learning from imbalanced data. Recent methods [46, 36, 50] that perform head-to-tail knowledge transfer have gained attention in computer vision. Knowledge transfer has also been extended in recommendation systems [51]. Approaches to tackle cold-start recommendation [53, 32, 15, 29] can also improve tail item prediction. Aforementioned approaches rely on additional side information (item and user content features), which can be difficult to acquire, especially when there are heterogeneous items [52] that do not share a common feature space. In this work, we only focus on approaches that do not depend on additional side information.

Frequency-based regularizers [49, 37, 4] have been used to reduce the sampling bias when training models on long-tail distribution. Some methods modify the loss objective [33, 12, 5, 18] using sample weighting for each class/item in the dataset. However, these methods use the item frequencies in the training dataset to determine the weights, which can lead to sub-optimal performance for sequential recommendation systems as item frequencies alone lack the contextual information contained in user’s past history. Unlike the above methods, our proposed density weighting uses the learned item representations that encapsulate the contextual information from the user’s history to determine and mitigate the effective imbalance. Closest to our method is the recent work [44, 48] used in regression models, which uses density estimation in the target space to mitigate imbalance. However, in contrast to the regression datasets that have fixed labels, in recommendation system, the labels for the user tower (i.e.i.e. item representations) are also learned using a separate model, making it a more challenging task. We make several contributions to enable density weighting for recommendations.

3 Methods

Problem Statement. Consider a set of users 𝒰{\mathcal{U}} and items {\mathcal{I}}, where u𝒰u\in{\mathcal{U}} denotes a user and ii\in{\mathcal{I}} denotes an item. We will use the notation |𝒰||{\mathcal{U}}| and |||{\mathcal{I}}| to denote the total number of users and items respectively. Given the user’s historical behaviour in the form of a sequence of item engagements xu={i1(u),,in(u)}x_{u}=\{i^{(u)}_{1},\dots,i^{(u)}_{n}\}, our recommendation task is to predict the next item that the user is likely to interact with at the (n+1)th(n+1)^{th} step. For ease of notation, we will denote the training dataset as 𝒟={xu,yu}u=1|𝒰|{\mathcal{D}}=\{x_{u},y_{u}\}_{u=1}^{|{\mathcal{U}}|}, where yuy_{u} refers to the in+1(u)i^{(u)}_{n+1} item and is considered the label for the user uu.

3.1 Background

Neural Deep Retrieval Model. Motivated by the success of deep learning in computer vision and natural language processing, applying deep neural networks (DNNs) for recommender systems is prevalent. Deep representation learning provides a powerful framework to encode the users and the items in a low-dimensional representation space, which can be used for retrieval in recommender systems. We apply the two-tower DNN model that aims to learn a common low-dimensional representation space for users and items. Figure 2 provides an illustration of a typical two-tower model architecture used for next item prediction, where the left and right tower encode the user item history and the candidate items respectively. A user tower consists of a sequence encoder (e.g.e.g. Transformer, LSTM, etcetc.) and a representation extraction layer that reduces the outputs of the sequence encoder to a single user representation (e.g.e.g. by weighted aggregation or by picking the hidden output of the encoder from the latest time step). The item tower is a multi-layer perceptron (MLP) model that learns the representation of item candidates given a one-hot encoding of the item. The towers can also easily incorporate side-information of users and items, if available. We denote the user tower as fθf_{\theta} and the item tower as gϕg_{\phi}, where θ\theta and ϕ\phi are the set of trainable parameters of the two respective architectures. The two-tower model uses the batch softmax optimization [4, 11] for training. The item probabilities P(y|xu;θ,ϕ)P\left(y|x_{u};\theta,\phi\right) are calculated over a random batch:

P(y|xu;θ,ϕ)\displaystyle P\left(y|x_{u};\theta,\phi\right) =exp(s(fθ(xu),gϕ(y)))yexp(s(fθ(xu),gϕ(y)))\displaystyle=\frac{exp\left(s\left(f_{\theta}(x_{u}),g_{\phi}(y)\right)\right)}{\sum_{y^{\prime}\in{\mathcal{B}}}exp\left(s\left(f_{\theta}(x_{u}),g_{\phi}(y^{\prime})\right)\right)} (1)

where {\mathcal{B}} is the set of item labels in the sampled batch and s(a,b)s(a,b) is the affinity score defined as the standard dot product. Naively using the batch softmax probabilities leads to a biased-estimate of P(y|xu;θ,ϕ)P\left(y|x_{u};\theta,\phi\right), where popular items are overly penalized as negatives due to high probability of being included in the batch. To mitigate this bias, LogQ correction [21, 49] is used with the sampled softmax loss, where the logit scores are corrected as follows:

sc(fθ(xu),gϕ(y))=s(fθ(xu),gϕ(y))log(p(y)).\displaystyle s^{c}\left(f_{\theta}(x_{u}),g_{\phi}(y)\right)=s\left(f_{\theta}(x_{u}),g_{\phi}(y)\right)-\log(p(y)). (2)

Here p(y) denotes the sampling probability of item yy in a random batch. Given the softmax probabilities, the recommendation task can be formalized as a multi-class classification problem with yuy_{u} as the label for the input xux_{u}, and the negative log-likelihood ((θ,ϕ){\mathcal{L}}(\theta,\phi)) of the training dataset can be defined as:

(θ,ϕ)=1|𝒰|{xu,yu}𝒟logP(yu|xu;θ,ϕ).\displaystyle{\mathcal{L}}(\theta,\phi)=\frac{1}{|{\mathcal{U}}|}\sum_{\{x_{u},y_{u}\}\in{\mathcal{D}}}-\log{P\left(y_{u}|x_{u};\theta,\phi\right)}. (3)

The models are trained to minimize (θ,ϕ){\mathcal{L}}(\theta,\phi) with respect to the parameters θ\theta and ϕ\phi. Thus, the affinity score for the user representation and the corresponding label item (also referred to as the “positive item”) is increased, while the affinity for the user representation and other item labels in the mini-batch (referred to as the “negative items”) is reduced. The optimization procedure implicitly pulls items that correspond to the same user interest close to each other in the representation space, while items belonging to different interests are pushed away.

Refer to caption
Figure 2: A typical two-tower DNN model for next item prediction. (Left) User tower that outputs a single user representation. (Right) Item tower.
Refer to caption
(a) Sampled interests.
Refer to caption
(b) Sampled items.
Refer to caption
(c) Item distribution.
Refer to caption
(d) % HR@5 for [C3, C4, C5].
Refer to caption
(e) % HR@5 for [C1, C2].
Refer to caption
(f) % HR@5 for all clusters.
Refer to caption
(g) Clustering Quality.
Refer to caption
(h) SUR Synthetic
Refer to caption
(i) MUR Synthetic
Figure 3: Analysis on the synthetic data having 50 items, 5 user interests (clusters) and varying degree of skewness in the interest distribution.

Module for Multi-Interest Extraction. We extend the standard user tower described in Figure 2 to model MUR using a multi-interest extraction layer. The goal of such a layer is to output multiple user representations, i.e.i.e. fθ(xu)={zu(1),,zu(M)}f_{\theta}(x_{u})=\left\{z_{u}^{(1)},\dots,z_{u}^{(M)}\right\}, where MM denotes the number of user representations extracted. While {zu(m)}m=1M\{z_{u}^{(m)}\}_{m=1}^{M} could be learned separately using MM separate user towers, such an approach will be inefficient in both computation and the number of parameters. Humeau et al. [20] recently proposed an efficient and scalable approach that uses global attention to extract multiple context embeddings for the natural language sentence matching tasks. We adapt this idea of global attention for a recommender system to extract multiple representations for a given user history. Let the output of the sequence encoder be {hu(1),,hu(n)}\{h^{(1)}_{u},\dots,h^{(n)}_{u}\}, then to obtain multi-interest user representations, we learn mm global query vectors {q1,,qM}\{q_{1},\dots,q_{M}\}, where each qmq_{m} extracts the user representation zmz_{m} using attention:

zu(m)=i=1nαm(i)hu(i)whereαm(i)=exp(qmThu(i))i=1nexp(qmThu(i))\displaystyle z_{u}^{(m)}=\sum_{i=1}^{n}\alpha^{(i)}_{m}h^{(i)}_{u}\quad\text{where}\quad\alpha^{(i)}_{m}=\frac{exp\left(q_{m}^{T}h_{u}^{(i)}\right)}{\sum_{i=1}^{n}exp\left(q_{m}^{T}h_{u}^{(i)}\right)}

Given multiple user representations, we compute the affinity score for each candidate yy as follows:

s(fθ(xu),gϕ(y))\displaystyle s\left(f_{\theta}(x_{u}),g_{\phi}(y)\right) =(zuTgϕ(y)),where\displaystyle={\left({z_{u}}^{T}g_{\phi}(y)\right)},\enspace\text{where} (4)
zu=mw(m)zu(m),\displaystyle z_{u}=\sum_{m}w^{(m)}{z^{(m)}_{u}},\quad w(m)=exp(zu(m)Tgϕ(y))j=1mexp(zu(j)Tgϕ(y))\displaystyle w^{(m)}=\frac{exp\left({z^{(m)}_{u}}^{T}g_{\phi}(y)\right)}{\sum_{j=1}^{m}exp\left({z^{(j)}_{u}}^{T}g_{\phi}(y)\right)}

As the above multi-interest extraction approach uses global query parameters in attention, we refer this representation extraction layer as parametric attention. Note that there are alternate methods for multi-interest extraction layers [10, 27, 31], but in this work we only consider parametric attention to extract MUR due to its simplicity and scalability. More importantly, the focus of this work is on studying the effect of long-tail distribution on MUR in the two-tower DNN architecture [11], which has become a popular framework for recommendations in the recent years.

3.2 Synthetic Data Analysis

In this section, we discuss our analysis on the synthetic data, which forms the basis of our density weighting scheme. The synthetic data for sequential recommendation is generated using a finite state hidden Markov model [3, 22], where we consider the hidden state clusters as the so-called user interests. To define the Markovian process for the sequence, we assume a user-specific interest transition matrix which is a function of the user interests, distribution skewness, and user interest volatility i.e., probability of switching between different user interests. A detailed description of the data generating process is provided in Appendix A. Figure  3(a)-3(c) show examples of synthetically generated data distributions. The synthetic data models long-tailed distributions for both interest and item space to capture the power-law behavior of items encountered in real recommendation systems.

While the Markovian assumption may not reflect complex real-world scenarios, we find that our observations are consistent in real-world datasets as described later in Section 4. Our key motivation to use synthetic data is to study the effect of distribution skewness on modeling capabilities of MUR. Moreover, item clusters in the synthetic data can be treated as ground-truth interests for the users, allowing us to quantitatively (using   mean-silhouette (MS) score [41]) and qualitatively (See Figure 1) evaluate the structure of the learned item representations for different models.

Retrieval performance. To study the sensitivity with respect to an imbalanced distribution, we evaluate the performance of the models trained with varying degree of data skewness. Specifically, we vary the imbalance in the user interest distribution and plot the recommendation performance (Hit Ratio, expressed as %) across different splits: tail/infrequent clusters (Figure 3(d)), head/frequent clusters (Figure 3(e)), and all clusters (Figure 3(f)). As seen in Figure 3(d), increasing the skewness causes the performance of MUR-based model to decrease for tail interests, which also results in decreasing average performance across all the interest clusters (Figure 3(f)). In fact, when the user interests are highly skewed, the performance of MUR-based model becomes comparable to the performance obtained using the SUR-based model. This is due to the under-representation of items in the tail interest clusters, causing the model to dominate on the head interest items. On the other hand, using our proposed density weighting with MUR, referred to as MUR-IDW (described in Section 3.3), makes the model robust to imbalance as shown in Figure 3(d) and Figure 3(f).

Clustering quality. We inspect the structure of the representation space learned using different models. A good representation space would have localized clusters representing user interests, such that the items belonging to a similar interest have item representations close to each other. First, we do a qualitative analysis of the clustering structure learned using MUR and SUR. Figure 3(h)3(i) shows that MUR leads to a better clustering structure of the representation space compared to the representations learned with SUR on the synthetic data.

Next, we measure the clustering quality of the item representations using the mean-silhouette (MS[41] score. The MS score measures the average tightness and separation of the clusters. For a point pp belonging to the ground truth cluster cc, we define the silhouette as sp=bpapmax(ap,bp)s_{p}=\frac{b_{p}-a_{p}}{max(a_{p},b_{p})}, where apa_{p} is the mean distance between pp and other points in the cluster cc, and bpb_{p} is the mean distance between pp and the points in the closest cluster. Specifically, we can calculate apa_{p} and bpb_{p} as:

ap=1|c|1qc,pqdist(p,q);bp=minc~c1|c~|qc~dist(p,q)\displaystyle a_{p}=\frac{1}{|c|-1}\sum_{q\in c,p\neq q}dist(p,q);\quad b_{p}=\min_{\tilde{c}\neq c}\frac{1}{|\tilde{c}|}\sum_{q\in\tilde{c}}dist(p,q)

The MS score is computed by taking the average over all points, i.ei.e MS:=𝔼p[sp]MS:=\mathbb{E}_{p}\left[s_{p}\right]. The MS score takes the value [1,1]\in[-1,1], where a higher MS depicts distinct and separated clusters, and a lower MS depicts overlapping clusters. Figure 3(g) shows the MS score for varying degrees of skewness in the synthetic dataset. While the clustering structure obtained with MUR is consistently better than with SUR , the clustering quality in the representation space deteriorates with increasing degree of imbalance.

3.3 Iterative Density Weighting

In this section, we describe our novel density-weighted loss for the two-tower recommendation system, which advocates leveraging the structure of the item representation space to learn and mitigate the effective imbalance of the items. This proposal aims to leverage the item representation space enabled by MUR to enhance the overall recommendation performance.

To determine the effective item imbalance, we use item distribution smoothing that convolves a symmetric kernel with the empirical distribution of the items. This results in a kernel-smoothed version of the item distribution that accounts for the overlap in information of nearby items in the representation space. We then use this kernel-smoothed item distribution to learn a balanced loss that is uniformly weighted in the item representation space.

We first consider the simple frequency-balanced [18] version of the loss function described in (3):

BAL(θ,ϕ)=1|𝒰|{xu,yu}𝒟1p(y)logP(yu|xu;θ,ϕ), where\displaystyle{\mathcal{L}}_{BAL}(\theta,\phi)=\frac{-1}{|{\mathcal{U}}|}\sum_{\{x_{u},y_{u}\}\in{\mathcal{D}}}\frac{1}{p(y)}\log{P\left(y_{u}|x_{u};\theta,\phi\right)},\text{ where} (5)

p(y):=ny/|𝒰|p(y):=n_{y}/|{\mathcal{U}}| corresponds to the empirical distribution of the items and nyn_{y} denotes the item frequency. Minimizing the balanced loss in  (5) aims to learn a uniform loss function across all the items in the training dataset. However, the imbalance in p(y)p(y) may not reflect the effective imbalance of the items in the representation space, and naively optimizing  (5) can lead to overfitting of the model to tail items that have low p(y)p(y). As we show in Section 4, such a frequency-based balanced loss can lead to poor overall recommendation performance.

In contrast, our density-weighted loss uses the kernel-smoothed version of the empirical item distribution:

p~(y):=k(y,y)p(y)𝑑y, where\displaystyle\tilde{p}(y):=\int k(y,y^{\prime})\,{p(y^{\prime})}dy^{\prime},\text{ where} (6)

k(y,y)k(y,y^{\prime}) is the symmetric Gaussian kernel that characterizes the distance of items in the representation space. Therefore, p~(y)\tilde{p}(y) leverages the structure of the representation space in determining the effective imbalance of items. In particular, p~(y)\tilde{p}(y) captures the effective weight of the region defined by the neighborhood of item yy in the representation space, thereby reducing the model’s sensitivity to p(y)p(y) through kernel smoothing. Note that since p~(y)\tilde{p}(y) is determined using the item representations learned in the item tower, in practice we use p~(y|ϕ)\tilde{p}(y|\phi). As both the kernel-smoothed density p~(y|ϕ)\tilde{p}(y|\phi) and the loss used to train the parameters of the two-tower recommender model {θ,ϕ\{\theta,\phi} are inter-dependent, we propose an iterative two-phase procedure (Algorithm 1) to train the two-tower architecture. Given the model parameters {θ(t),ϕ(t)}\{\theta^{(t)},\phi^{(t)}\} and the density p~(y|ϕ(t))\tilde{p}(y|\phi^{(t)}) at iteration tt,

  • In the Sleep Phase, the item representations from the item tower are used to update the density through kernel-smoothing.

  • In the Wake Phase, the two-tower model is trained with the balanced loss given by the item density from the previous iteration.

Sleep Phase: Update Item Weights. To learn a balanced loss function with respect to p~(y|ϕ(t))\tilde{p}(y|\phi^{(t)}), we introduce item-specific weights wy(t)w_{y}^{(t)}, such that the initial wy(0)=ny/|𝒰|w_{y}^{(0)}=n_{y}/|{\mathcal{U}}|. The sleep phase updates the item weights based on the item representations defined by the current state of the item tower. In particular, the density update is defined as:

p~(y|ϕ(t))\displaystyle\tilde{p}(y|\phi^{(t)}) =1k=1||wk(t)𝒦(yvk), where\displaystyle=\frac{1}{\ell}\sum_{k=1}^{|{\mathcal{I}}|}w_{k}^{(t)}{\mathcal{K}}\left(\frac{||y-v_{k}||}{\ell}\right),\text{ where} (7)

vk=gϕ(t)(ik)v_{k}=g_{\phi^{(t)}}(i_{k}) refers to the representation of the kthk^{th} item, 𝒦{\mathcal{K}} is the Gaussian kernel function with the bandwidth parameter \ell (the bandwidth \ell is selected based on Scott’s Rule [43], which is a standard bandwidth selection method). In practice, we found using relative normalized density [44] as a better estimate:

p(y|ϕ(t))\displaystyle p^{\prime}(y|\phi^{(t)}) =p~(y|ϕ(t))minyp~(y|ϕ(t))maxyp~(y|ϕ(t))minyp~(y|ϕ(t))\displaystyle=\frac{\tilde{p}(y|\phi^{(t)})-\min_{y}\tilde{p}(y|\phi^{(t)})}{\max_{y}\tilde{p}(y|\phi^{(t)})-\min_{y}\tilde{p}(y|\phi^{(t)})} (8)

The relative item density of each item is \in [0,1][0,1] and the item weights are updated with hk(t):=1p(y=k|ϕ(t))h^{(t)}_{k}:=1-p^{\prime}(y=k|\phi^{(t)}) resulting in a negative correlation with the relative item density. To avoid a drastic change in the loss function, the weights are updated gradually using a momentum parameter mm:

wk(t+1)=mwk(t)+(1m)(hk(t)/khk(t))\displaystyle w_{k}^{(t+1)}=m\cdot w_{k}^{(t)}+(1-m)\cdot{\left(h^{(t)}_{k}/\sum_{{k^{\prime}}}h^{(t)}_{k^{\prime}}\right)} (9)

where we use the subscript kk to denote the kthk^{th} item. Since the updates are negatively correlated with the item density, they allow the model to progressively upweight/downweight the regions with low/high item density respectively. The complete procedure followed in the sleep phase is summarized in Algorithm 2.

Wake Phase: Train Two-Tower Model. Given wy(t+1)w^{(t+1)}_{y}, the two-tower model is trained by minimizing the following weighted loss:

~BAL(θ,ϕ)\displaystyle\tilde{{\mathcal{L}}}_{BAL}(\theta,\phi) =1|𝒰|{x,y}𝒟wy(t+1)logP(y|x;θ(t),ϕ(t)).\displaystyle=\frac{-1}{|{\mathcal{U}}|}\sum_{\{x,y\}\in{\mathcal{D}}}w^{(t+1)}_{y}\cdot\log{P\left(y|x;\theta^{(t)},\phi^{(t)}\right)}. (10)
{θ(t+1),ϕ(t+1)}\displaystyle\{\theta^{(t+1)},\phi^{(t+1)}\} =argminθ,ϕ~BAL(θ,ϕ)\displaystyle=\operatorname*{arg\,min}_{\theta,\phi}\tilde{{\mathcal{L}}}_{BAL}(\theta,\phi) (11)
Algorithm 1 Iterative Density Weighting
1:Input: Training Dataset 𝒟{\mathcal{D}}, Weight momentum mm.
2:Initialize: wk(0)nk/|𝒰|w_{k}^{(0)}\leftarrow n_{k}/|{\mathcal{U}}| \triangleright Normalized Item Frequency.
3:Initialize: θ(0),Φ(0)\theta^{(0)},\Phi^{(0)} \triangleright Trained with unweighted loss (3).
4:for t0,,MAXTt\in 0,\dots,MAX_{T} do
5:     Sleep: w(t+1)UpdateWeights(m,w(t),ϕ(t))w^{(t+1)}\leftarrow\text{UpdateWeights}(m,w^{(t)},\phi^{(t)})
6:     if w(t+1)w(t)<η||w^{(t+1)}-w^{(t)}||<\eta then
7:         break
8:     end if
9:     Wake: {θ(t+1),ϕ(t+1)}\{\theta^{(t+1)},\phi^{(t+1)}\}\leftarrow TrainTowers(w(t+1)w^{(t+1)}).
10:end for
11:ϕϕ(t)\phi^{*}\leftarrow\phi^{(t)}
12:θ\theta^{*}\leftarrow UserTowerCalibration(θ(t),ϕ\theta^{(t)},\phi^{*})
Algorithm 2 Sleep Phase: Update Item Weights.
1:function UpdateWeights(mm, w(t)w^{(t)}, ϕ(t)\phi^{(t)})
2:     vkv_{k}\leftarrow gϕ(t)(ik)g_{\phi^{(t)}}\left(i_{k}\right). \triangleright Item tower forward pass.
3:     Obtain p(y|ϕ(t))p(y|\phi^{(t)}) using Equation (7).
4:     Obtain p(y|ϕ(t))p^{\prime}(y|\phi^{(t)}) using Equation (8).
5:     hk(t)1p(y=k|ϕ(t))h^{(t)}_{k}\leftarrow 1-p^{\prime}(y=k|\phi^{(t)}).
6:     Obtain wk(t+1)w_{k}^{(t+1)} using equation (9).
7:     return w(t+1)w^{(t+1)}
8:end function
Refer to caption
Refer to caption
Figure 4: The evolution of item embedding space and item distribution over the representation space over several iterations of IDW. The improvement in the clustering structure of the item embeddings is illustrated in the first row and also captured by the MS score, increasing from 0.240.24 at iteration 1 to 0.830.83 in iteration 7.

Stopping Criterion. To detect convergence in IDW, we track w(t+1)w(t)||w^{(t+1)}-w^{(t)}||. As the model is trained, the loss contributions becomes uniform in the item representation space. We visualize the evolution of the item representation space for the synthetic data in Figure 4. We observed that making the loss uniform in the representation space, progressively improves the clustering structure of the item representations. The improvement in the clustering structure is also reflected in the mean-silhouette (MS) score in Figure 4 (Top).

User Tower Calibration. While the uniform loss structure improves the clustering structure of the items, the retrieval performance over the head clusters can drop compared to when the model is trained using the unweighted likelihood. Therefore, to mitigate the performance drop for head clusters, once we improve the item clustering structure using IDW, we freeze the item tower and train only the user tower with the unweighted loss. We study the importance of user tower calibration in section 4.6.3.

4 Experiments

Table 1: The recommendation performance on three real-world benchmark datasets. We use blue color for the best baseline. We use bold and underline to denote the best metric. (Higher is better).
Method # User Representations (M) MovieLens 1M Kindle Store Clothing, Shoes and Jewelry
HR@20 NDCG@20 HR@20 NDCG@20 HR@20 NDCG@20
SUR [20] 1 80.62 ±\pm 0.18 47.24 ±\pm 0.14 63.33 ±\pm 0.38 30.87 ±\pm 0.56 32.43 ±\pm 0.66 13.44 ±\pm 0.54
MUR [20] 5 80.82 ±\pm 0.20 47.72 ±\pm 0.23 64.66 ±\pm 0.45 31.16 ±\pm 0.38 33.92 ±\pm 0.94 14.90 ±\pm 0.52
Frequency Balance [18] 5 75.75 ±\pm 0.32 40.36 ±\pm 0.24 55.96 ±\pm 0.77 23.49 ±\pm 0.33 22.38 ±\pm 0.52 7.44 ±\pm 0.31
Focal Balance [33] 5 80.88 ±\pm 0.22 45.48 ±\pm 0.37 62.17 ±\pm 0.94 28.56 ±\pm 0.83 32.39 ±\pm 1.16 13.19 ±\pm 0.61
Qmargin [9] 5 80.35 ±\pm 0.37 45.43 ±\pm 0.46 63.75 ±\pm 0.71 29.87 ±\pm 0.50 34.71 ±\pm 1.16 14.13 ±\pm 0.52
Item Norm  [25] 5 81.36 ±\pm 0.42 47.19 ±\pm 0.47 65.20 ±\pm 0.49 32.19 ±\pm 0.24 34.57 ±\pm 0.67 14.59 ±\pm 0.71
Item Norm (Post-hoc) [23] 5 80.20 ±\pm 0.53 45.87 ±\pm 0.67 64.05 ±\pm 0.18 30.68 ±\pm 0.22 34.32 ±\pm 0.77 13.83 ±\pm 0.52
SUR-IDW [Ours] 1 81.85 ±\pm 0.40 48.95 ±\pm 0.51 64.62 ±\pm 0.28 31.67 ±\pm 0.26 34.88 ±\pm 0.38 15.45 ±\pm 0.16
MUR-IDW [Ours] 5 82.65 ±\pm 0.03 49.67 ±\pm 0.14 65.24 ±\pm 0.25 32.25 ±\pm 0.32 37.34 ±\pm 0.63 16.33 ±\pm 0.18
MUR vs. SUR - +0.25% +1.02% +2.10% +0.94% +4.59% +10.86%
MUR-IDW [Ours] vs. SUR - +2.52% +5.14% +3.02% +4.47% +15.14% +21.50%
Refer to caption
Refer to caption
Refer to caption
Figure 5: NDCG@20 on the overall, head and tail item sets for three real-world benchmarks. The relative increase/decrease in the performance is calculated relative to the MUR (BASE) performance on the respective item set.

4.1 Datasets

We evaluate the proposed method on three public benchmarks that are used for recommender models. The statistics of the datasets are summarized in Table 2. In all the three datasets, the item distribution follows a highly skewed long-tail distribution (See the item distributions in Appendix C.). For each benchmark dataset, we preprocess the data to construct item sequences using the history of items that users have interacted with. We consider a maximum sequence length of 30, where padding is added if the item sequences are smaller than 30, and the item sequences larger than 30 are truncated and treated as multiple sequences.

Table 2: Dataset Statistics.
Dataset # Users # Items Avg Seq Length
MovieLens-1M [17] 6,040 3,706 165
Kindle Store [39] 139,769 98,769 15
Clothing Shoes & Jewelry [39] 1,219,594 376,438 9

4.2 Evaluation Metrics

We employ top-k Hit Ratio (HR@k) and NDCG@K with k=20k=20 to evaluate the recommendation performance of all the models. Additional results with k=5k=5 are included in the Appendix B. As the goal of IDW is to improve the performance of the tail items without hurting the overall performance across all the items, we further compare the metrics on the tail and head item sets separately. Assuming the Pareto Principle [7], the 20% most frequent items are considered head items, and the remaining items as the tail items. Following the standard evaluation protocol [24, 42], we use the leave-one-out strategy for evaluation. For each item sequence, the last item is used for testing, the item before the last is used for validation, and the rest is used for training. As the item set is typically large, we use randomly sampled negative items [24, 19, 26, 42] at evaluation. Specifically, we pair the ground-truth item with 99 randomly sampled negative items that the user has not previous interacted with to compute the evaluation metrics.

4.3 Baselines

To demonstrate the effectiveness of the proposed IDW on long-tail item distribution, we compare IDW with other commonly used weighting methods that deal with imbalance in recommender models. All the baselines use the two-tower model as the backbone, which has proven to be scalable because of efficient inference through inner products. For extracting multiple user representations to capture diverse user interests, we use parametric attention as described in Section 3.1. We list all the baselines below. We note that methods that rely on knowledge transfer of user/item features [6, 51] or augmentation methods (over/under sampling [8]) to mitigate imbalance can be easily combined with all the following baselines. Our proposed density-weighting and the baselines considered do not depend on user/item features.

  • SUR: We use a single user representation as the output of the user tower by using M=1M=1 in the parametric attention [20] layer.

  • MUR: The multiple user representations are extracted using the parametric attention [20] layer with M=5M=5 output representations. We do an ablation to study the effect of MM in Section 4.6.1.

  • Frequency Balance: We use the frequency-based balancing [18] as described in Equation 5. Note that this is a static weighting scheme, where the item weights do not change during training.

  • Focal Balance: A recent work by Lin et al. [33] modifies the loss function to train the model on an imbalanced dataset by down-weighting the loss of well-classified examples. This is a dynamic weighting scheme, where the item weights depend on the loss of each input during training.

  • Qmargin: Cao et al. [9] proposed to add a per-class (per-item) margin δy\delta_{y} into the cross-entropy loss, where δyp(y)1/4\delta_{y}\propto p(y)^{-1/4}.

  • Item Norm: [25] upweights the score of rare items by normalizing the item representations. We also experiment with a post-hoc version of normalized item, where a standard two-tower model is trained and the item representation normalization is only carried out during the inference. [23].

4.4 Hyper-parameter Settings

We set the embedding dimension as 16 across all the models for the three benchmarks. We use the two-tower architecture as the backbone for all the models, where the user tower is a self-attention encoder [45] with two multi-head attention layers and four attention heads per layer. The item tower is an MLP with two fully connected layers each with 16 output nodes. For IDW, the momentum update of item weights (Equation 9) and η\eta are determined using grid search in the range of {0.7,0.8,0.9,0.99}\{0.7,0.8,0.9,0.99\} and {1.0,2.0,3.0,4.0}104\{1.0,2.0,3.0,4.0\}*10^{-4}. We use Adam to optimize all the models with a learning rate of 0.01. For all models, we used early stopping based on the validation dataset. The models are implemented with TensorFlow Recommenders (TFRS) [28].

4.5 Recommendation Performance

4.5.1 Overall Performance

We compare our proposed method with several representative baselines proposed for learning on imbalanced/skewed datasets in Table 1. For a fair comparison, all the methods use the same two-tower architecture described in Section 3.1. We first observe that the baseline MUR has marginal gains compared to the vanilla SUR in all the benchmarks. This is likely because of highly skewed long-tail item distributions in the real world datasets (See item distribution in Figure 9). This observation is consistent with our findings from synthetic datasets having high skewness (Figure  3(f)). Whereas, with IDW we see significant gains in performance for all real-world benchmarks we considered. In fact, the relative gains with IDW-based MUR ranged from 1.5x - 10x compared to the gains observed with the naive MUR. Our results also clearly show that IDW outperforms other imbalanced baselines that are typically used to address data imbalance. Finally, MUR coupled with IDW yield the best overall performance, highlighting that IDW is able to successfully leverage the improved clustering structure of MUR over SUR to yield further improvements.

4.5.2 Head vs. Tail Performance

We conduct a performance analysis on the head and tail items separately to get insights where the overall performance gains are coming from. We compare our method with the top three baseline methods from Table 1; namely, MUR, Qmargin, and Item Norm. The NDCG@20 metric for all the three benchmarks is shown in Figure 5111 We omit HR@20 metric for this analysis as it follows the same pattern as NDCG@20.. Compared to the naive MUR, we observe that IDW-based MUR consistently improves the performance on the tail items, while having better or comparable performance across the head items, leading to better overall performance. Whereas, both Qmargin and Item Norm focus primarily on the tail items, while sacrificing significant performance for the head items. This makes IDW a much more appealing method compared to other methods proposed for training on imbalanced datasets.

4.6 Ablation Study

4.6.1 Effect of Number of User Representations

We analyze the effect of the number of user representations (MM) on the model performance for several representative methods. As shown in figure 6, IDW-based models not only perform better than other counterparts for the same MM, they are the only ones that show gains with increasing MM. In contrast, the baseline methods do not observe any trend with respect to MM. This shows that MUR and IDW nicely complement each other.

Refer to caption
Refer to caption
Figure 6: The effect of number of user representations (M) on all the methods for the MovieLens-1M dataset.
Refer to caption
(a) Ablation with varying η\eta.
Refer to caption
(b) Ablation with varying mm.
Figure 7: The effect of η\eta and momentum (m) on IDW performance for the MovieLens-1M dataset.
Refer to caption
Refer to caption
Figure 8: The effect of user tower calibration for all datasets.

4.6.2 Sensitivity to IDW Hyperparameters

We also analyze the effect of the two hyperparameters introduced with IDW: η\eta and momentum (mm). The hyperparameter η\eta is used to define the stopping criterion in IDW (See Algorithm 1). Figure 7(a) shows the performance for η[0.0001,0.0007]\eta\in[0.0001,0.0007]. The model performance can be sensitive to η\eta; a value too low increases the number of iterations causing over-fitting, whereas a value too high reduces the number of iterations causing under-fitting. We found η=0.0004\eta=0.0004 to be the best in this ablation. We also perform an ablation for mm, which determines the update in item weights (Equation 9). We found that setting a low value for mm makes the model unstable due to an abrupt change in the loss function. In our ablation, m=0.9m=0.9 yields the best results, followed by m=0.99m=0.99 and m=1.0m=1.0. The setting with m=1m=1 doesn’t use density-based weights and is equivalent to the baseline MUR with our proposed user tower calibration.

4.6.3 Importance of User Tower Calibration

In section 4, we proposed calibrating the user tower on the unweighted training distribution with the item tower fixed, once the model is trained with IDW. To evaluate the effectiveness of calibrating the user tower, we do an ablation study comparing IDW with and without user tower calibration (IDW w/o user calib). Figure 8 shows that the proposed calibration of user tower boosts the overall recommendation performance, making it an essential component of IDW.

5 Conclusions

In conclusion, this paper makes contributions in two key aspects. Firstly, to the best of our knowledge, this is the first study to investigate the impact of skewed item distribution on learning MUR, emphasizing the importance of loss balancing to fully realize the benefits of MUR. This finding alone could hold substantial interest for the user-modeling community. Secondly, the paper introduces a novel density-weighting framework that utilizes item representations to balance the loss, resulting in improved performance with MUR. Experimental results demonstrate that MUR with IDW yields superior overall performance on both public benchmarks and synthetic datasets, highlighting that IDW is successfully able to use improved learned item representation to improve overall recommendation performance. The presented work also sets the stage for further exploration in the field, opening avenues for the development of alternative balancing techniques and more robust MUR extraction layers. We hope that the findings in this work contribute to advancing the field of user modeling in recommendation systems and pave the way for future research endeavors in this domain.

Appendix

Appendix A Synthetic Data Generation

We generate synthetic data containing a sequence of item clicks for each user. The generated item sequences are used for training on the next item prediction task. The parameters for generating the dataset are summarized in Table 3. The data generation process is specified as follows:

  1. 1.

    Assign clusters to users: We assign each user a set of clusters (YuY_{u}) that depict user’s interests. The cluster assignment is done by sampling from a multinomial distribution given by weights {π1,,πC}\{\pi_{1},\dots,\pi_{C}\}. The sampling weights are chosen to follow a power-law, which results in skewness in the user interest distribution as shown in Figure 3(a).

  2. 2.

    Generate cluster sequence: Given a user’s interests, we assume a finite state Hidden Markov model to define the cluster transitions over time. The resulting sequence reflects a user’s interest over time. The cluster transition is given by:

    Pij(u)={αif i=j and iYuγ|Yu1|if ij and {i,j}Yu1αγC|Yu|if jYu and iYuϵif i=j and iYu1ϵ|Yu|if jYu and iYu0if ij and {i,j}YuP^{(u)}_{ij}=\begin{cases}\alpha&\text{if }i=j\text{ and }i\in Y_{u}\\ \frac{\gamma}{|Y_{u}-1|}&\text{if }i\neq j\text{ and }\{i,j\}\in Y_{u}\\[4.0pt] \frac{1-\alpha-\gamma}{C-|Y_{u}|}&\text{if }j\notin Y_{u}\text{ and }i\in Y_{u}\\ \epsilon&\text{if }i=j\text{ and }i\notin Y_{u}\\ \frac{1-\epsilon}{|Y_{u}|}&\text{if }j\in Y_{u}\text{ and }i\notin Y_{u}\\ 0&\text{if }i\neq j\text{ and }\{i,j\}\notin Y_{u}\end{cases} (12)

    where Pij(u)P^{(u)}_{ij} depicts the probability of transition from cluster ii to cluster jj for a user uu. In (12), γ\gamma depicts the user’s interest volatility: a high γ\gamma would lead to frequent switching between different clusters that the user is interested in.

  3. 3.

    Generate item sequence: Finally, we generate the item sequence using the cluster sequence generated above. We assume that the items from each cluster are drawn based on a multinomial distribution with {p1,,pK}\{p_{1},\dots,p_{K}\}, where KK is the number of items. For all clusters, we assume K=||/CK=|{\mathcal{I}}|/C. Similar to cluster assignment, we induce a power-law distribution for {p1,,pK}\{p_{1},\dots,p_{K}\} as shown in Figure 3(b).

Table 3: Parameters used to generate the synthetic sequential item history.
Param Description
CC The number of total clusters or interests.
TT The sequence length.
|𝒰||{\mathcal{U}}| The number of total users.
|||{\mathcal{I}}| The number of total items.
YuY_{u} The set of clusters that a user uu is interested in.
α\alpha Probability of staying in the current cluster cYuc\in Y_{u}.
γ\gamma User interest volatility: The total probability of a user switching to a new cluster in YuY_{u}.
ϵ\epsilon Probability that a user stays in a “non-interesting” cluster cYuc\notin Y_{u}.

Note that the ground-truth cluster correspondences are not available to the recommendation model during training on the synthetic dataset, i.e.i.e. the model only has access to the marginal item distribution (Figure 3(c)).

Parameters for generating synthetic data: For the experiments presented in Section 3.2, we use a set of 50 items with 10 items per interest (cluster). The degree of skewness in Figure 3 is set by changing the imbalance in the cluster sampling distribution governed by πc\pi_{c}: the power law exponent is increased from 0.0 to 2.0 to make the interest distribution skewed. The conditional item sampling distribution pkp_{k} also uses power law weights, however, for all the experiments the item sampling distribution uses 0.5 as the power law exponent. For all experiments, we set the interest volatility parameters γ=0.3\gamma=0.3 and α=0.6\alpha=0.6. We set ϵ=0.1\epsilon=0.1 to induce exploration in the user engagements.

Table 4: The recommendation performance on three real-world benchmark datasets with K=5.
Method MovieLens 1M Kindle Store Clothing, Shoes …
HR@5 NDCG@5 HR@5 NDCG@5 HR@5 NDCG@5
SUR 54.69 ±\pm 0.20 39.72 ±\pm 0.14 32.98 ±\pm 0.85 22.13 ±\pm 0.66 11.87 ±\pm 0.67 7.76 ±\pm 0.55
MUR 55.21 ±\pm 0.29 40.25 ±\pm 0.24 33.44 ±\pm 0.58 22.35 ±\pm 0.44 13.13 ±\pm 0.63 9.15 ±\pm 0.43
Frequency Bal. 46.07 ±\pm 0.34 31.72 ±\pm 0.25 23.10 ±\pm 0.31 14.46 ±\pm 0.22 5.34 ±\pm 0.31 3.16 ±\pm 0.19
Focal Bal. 53.27 ±\pm 0.48 37.42 ±\pm 0.43 30.93 ±\pm 1.21 19.65 ±\pm 0.90 11.49 ±\pm 0.73 7.42 ±\pm 0.49
Qmargin 52.80 ±\pm 0.63 37.40 ±\pm 0.53 32.03 ±\pm 0.64 20.82 ±\pm 0.47 12.32 ±\pm 0.53 7.94 ±\pm 0.42
Item Norm 54.93 ±\pm 0.67 39.48 ±\pm 0.54 35.02 ±\pm 0.44 23.55 ±\pm 0.37 12.68 ±\pm 0.79 8.55 ±\pm 0.75
Post-hoc 53.09 ±\pm 0.92 37.98 ±\pm 0.77 32.38 ±\pm 0.37 21.66 ±\pm 0.32 11.94 ±\pm 0.61 7.64 ±\pm 0.50
SUR-IDW 56.88 ±\pm 0.69 41.57 ±\pm 0.59 34.29 ±\pm 0.31 23.01 ±\pm 0.28 13.79 ±\pm 0.18 9.62 ±\pm 0.10
MUR-IDW 57.79 ±\pm 0.16 42.42 ±\pm 0.15 35.07 ±\pm 0.51 23.62 ±\pm 0.39 14.67 ±\pm 0.16 10.43 ±\pm 0.20

Appendix B Recommendation Performance measured as HR@5 and NDCG@5

In addition to the main results showing the recommendation performance measured as HR@20 and NDCG@20 in Table 1, we also include metrics HR@5 and NDCG@5 for overall item sets in Table 4. Once again, we observe that IDW performs better than the other baselines we considered.

Appendix C Item distribution in the datasets

In this section, we visualize the empirical distribution of the three benchmarks that we considered in our experiments in Figure 9. In the y-axis we plot the normalized frequency of items and x-axis is the item ids arranged based on their frequency.

Refer to caption
Figure 9: The item distribution in the three benchmarks. (Left) Kindle Store. (Middle) Clothing, Shoes and Jewelry. (Right) MovieLens-1M.

References

  • [1]
  • Abel et al. [2011] Fabian Abel, Qi Gao, Geert-Jan Houben, and Ke Tao. 2011. Analyzing User Modeling on Twitter for Personalized News Recommendations. In User Modeling, Adaption and Personalization, Joseph A. Konstan, Ricardo Conejo, José L. Marzo, and Nuria Oliver (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 1–12.
  • Baum and Petrie [1966] Leonard E. Baum and Ted Petrie. 1966. Statistical Inference for Probabilistic Functions of Finite State Markov Chains. The Annals of Mathematical Statistics 37, 6 (1966), 1554 – 1563. https://doi.org/10.1214/aoms/1177699147
  • Bengio and Senecal [2008] Yoshua Bengio and Jean-Sébastien Senecal. 2008. Adaptive Importance Sampling to Accelerate Training of a Neural Probabilistic Language Model. IEEE Transactions on Neural Networks 19 (2008), 713–722.
  • Beutel et al. [2017a] Alex Beutel, Ed H. Chi, Zhiyuan Cheng, Hubert Pham, and John Anderson. 2017a. Beyond Globally Optimal: Focused Learning for Improved Recommendations. In Proceedings of the 26th International Conference on World Wide Web (WWW ’17). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 203–212. https://doi.org/10.1145/3038912.3052713
  • Beutel et al. [2017b] Alex Beutel, Ed H. Chi, Zhiyuan Cheng, Hubert Pham, and John Anderson. 2017b. Beyond Globally Optimal: Focused Learning for Improved Recommendations. In Proceedings of the 26th International Conference on World Wide Web, WWW 2017, Perth, Australia, April 3-7, 2017.
  • Box and Meyer [1986] George E. P. Box and R. Daniel Meyer. 1986. An Analysis for Unreplicated Fractional Factorials. Technometrics 28, 1 (1986), 11–18. http://www.jstor.org/stable/1269599
  • Brownlee [2020] Jason Brownlee. 2020. Imbalanced classification with Python: better metrics, balance skewed classes, cost-sensitive learning. Machine Learning Mastery.
  • Cao et al. [2019] Kaidi Cao, Colin Wei, Adrien Gaidon, Nikos Arechiga, and Tengyu Ma. 2019. Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss. Curran Associates Inc., Red Hook, NY, USA.
  • Cen et al. [2020] Yukuo Cen, Jianwei Zhang, Xu Zou, Chang Zhou, Hongxia Yang, and Jie Tang. 2020. Controllable Multi-Interest Framework for Recommendation. Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (2020).
  • Covington et al. [2016] Paul Covington, Jay Adams, and Emre Sargin. 2016. Deep Neural Networks for YouTube Recommendations. In Proceedings of the 10th ACM Conference on Recommender Systems (RecSys ’16). Association for Computing Machinery, New York, NY, USA, 191–198. https://doi.org/10.1145/2959100.2959190
  • Cui et al. [2019] Yin Cui, Menglin Jia, Tsung-Yi Lin, Yang Song, and Serge Belongie. 2019. Class-balanced loss based on effective number of samples. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 9268–9277.
  • Czarnowska et al. [2019] Paula Czarnowska, Sebastian Ruder, Edouard Grave, Ryan Cotterell, and Ann Copestake. 2019. Don’t Forget the Long Tail! A Comprehensive Analysis of Morphological Generalization in Bilingual Lexicon Induction. arXiv preprint arXiv:1909.02855 (2019).
  • Domingues et al. [2012] Marcos Aurélio Domingues, Fabien Gouyon, Alípio Mário Jorge, José Paulo Leal, João Vinagre, Luís Lemos, and Mohamed Sordo. 2012. Combining Usage and Content in an Online Music Recommendation System for Music in the Long-Tail. In Proceedings of the 21st International Conference on World Wide Web (WWW ’12 Companion). Association for Computing Machinery, New York, NY, USA, 925–930. https://doi.org/10.1145/2187980.2188224
  • Dong et al. [2020] Manqing Dong, Feng Yuan, Lina Yao, Xiwei Xu, and Liming Zhu. 2020. MAMO: Memory-Augmented Meta-Optimization for Cold-Start Recommendation. Association for Computing Machinery, New York, NY, USA, 688–697. https://doi.org/10.1145/3394486.3403113
  • Hallinan and Striphas [2016] Blake Hallinan and Ted Striphas. 2016. Recommended for you: The Netflix Prize and the production of algorithmic culture. New Media & Society 18, 1 (2016), 117–137. https://doi.org/10.1177/1461444814538646
  • Harper and Konstan [2015] F. Maxwell Harper and Joseph A. Konstan. 2015. The MovieLens Datasets: History and Context. ACM Trans. Interact. Intell. Syst. 5, 4, Article 19 (Dec. 2015), 19 pages. https://doi.org/10.1145/2827872
  • He and Garcia [2009] Haibo He and Edwardo A. Garcia. 2009. Learning from Imbalanced Data. IEEE Transactions on Knowledge and Data Engineering 21, 9 (2009), 1263–1284. https://doi.org/10.1109/TKDE.2008.239
  • He et al. [2017] Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. 2017. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web (WWW ’17). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 173–182. https://doi.org/10.1145/3038912.3052569
  • Humeau et al. [2020] Samuel Humeau, Kurt Shuster, Marie-Anne Lachaux, and Jason Weston. 2020. Poly-encoders: Architectures and Pre-training Strategies for Fast and Accurate Multi-sentence Scoring. In International Conference on Learning Representations. https://openreview.net/forum?id=SkxgnnNFvH
  • Jean et al. [2014] Sébastien Jean, Kyunghyun Cho, Roland Memisevic, and Yoshua Bengio. 2014. On using very large target vocabulary for neural machine translation. arXiv preprint arXiv:1412.2007 (2014).
  • Juang and Rabiner [1991] B. H. Juang and L. R. Rabiner. 1991. Hidden Markov Models for Speech Recognition. Technometrics 33, 3 (1991), 251–272. http://www.jstor.org/stable/1268779
  • Kang et al. [2020] Bingyi Kang, Saining Xie, Marcus Rohrbach, Zhicheng Yan, Albert Gordo, Jiashi Feng, and Yannis Kalantidis. 2020. Decoupling Representation and Classifier for Long-Tailed Recognition. In International Conference on Learning Representations. https://openreview.net/forum?id=r1gRTCVFvB
  • Kang and McAuley [2018] Wang-Cheng Kang and Julian McAuley. 2018. Self-Attentive Sequential Recommendation. 197–206. https://doi.org/10.1109/ICDM.2018.00035
  • Kim and Kim [2020] Byungju Kim and Junmo Kim. 2020. Adjusting Decision Boundary for Class Imbalanced Learning. IEEE Access 8 (2020), 81674–81685. https://doi.org/10.1109/ACCESS.2020.2991231
  • Koren [2008] Yehuda Koren. 2008. Factorization Meets the Neighborhood: A Multifaceted Collaborative Filtering Model. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’08). Association for Computing Machinery, New York, NY, USA, 426–434. https://doi.org/10.1145/1401890.1401944
  • Kula [2017] Maciej Kula. 2017. Mixture-of-tastes Models for Representing Users with Diverse Interests. ArXiv abs/1711.08379 (2017).
  • Kula et al. [2020] Maciej Kula, Tiansheng Yao, Maheswaran Sathiamoorthy, Yihua Chen, Xinyang Yi, Sequoia Eyzaguirre, Lichan Hong, and Ed H. Chi. 2020. TensorFlow Recommenders. https://github.com/tensorflow/recommenders.
  • Lee et al. [2019] Hoyeop Lee, Jinbae Im, Seongwon Jang, Hyunsouk Cho, and Sehee Chung. 2019. MeLU: Meta-Learned User Preference Estimator for Cold-Start Recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’19). Association for Computing Machinery, New York, NY, USA, 1073–1082. https://doi.org/10.1145/3292500.3330859
  • Li et al. [2019a] Chao Li, Zhiyuan Liu, Mengmeng Wu, Yuchi Xu, Pipei Huang, Huan Zhao, Guoliang Kang, Qiwei Chen, Wei Li, and Dik Lun Lee. 2019a. Multi-Interest Network with Dynamic Routing for Recommendation at Tmall. arXiv:cs.IR/1904.08030
  • Li et al. [2019b] Chao Li, Zhiyuan Liu, Mengmeng Wu, Yuchi Xu, Huan Zhao, Pipei Huang, Guoliang Kang, Qiwei Chen, Wei Li, and Dik Lun Lee. 2019b. Multi-Interest Network with Dynamic Routing for Recommendation at Tmall. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management (CIKM ’19). Association for Computing Machinery, New York, NY, USA, 2615–2623. https://doi.org/10.1145/3357384.3357814
  • Liang et al. [2020] Tingting Liang, Congying Xia, Yuyu Yin, and Philip S. Yu. 2020. Joint Training Capsule Network for Cold Start Recommendation. arXiv:cs.IR/2005.11467
  • Lin et al. [2018] Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollár. 2018. Focal Loss for Dense Object Detection. IEEE Transactions on Pattern Analysis and Machine Intelligence 42, 2 (2018), 318–327. https://doi.org/10.1109/TPAMI.2018.2858826
  • Linden et al. [2003] G. Linden, B. Smith, and J. York. 2003. Amazon.com recommendations: item-to-item collaborative filtering. IEEE Internet Computing 7, 1 (2003), 76–80. https://doi.org/10.1109/MIC.2003.1167344
  • Liu and Zheng [2020] Siyi Liu and Yujia Zheng. 2020. Long-Tail Session-Based Recommendation. In Fourteenth ACM Conference on Recommender Systems (RecSys ’20). Association for Computing Machinery, New York, NY, USA, 509–514. https://doi.org/10.1145/3383313.3412222
  • Liu et al. [2019] Ziwei Liu, Zhongqi Miao, Xiaohang Zhan, Jiayun Wang, Boqing Gong, and Stella X. Yu. 2019. Large-Scale Long-Tailed Recognition in an Open World. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
  • Menon et al. [2021] Aditya Krishna Menon, Sadeep Jayasumana, Ankit Singh Rawat, Himanshu Jain, Andreas Veit, and Sanjiv Kumar. 2021. Long-tail learning via logit adjustment. In International Conference on Learning Representations. https://openreview.net/forum?id=37nvvqkCo5
  • Messica and Rokach [2018] Asi Messica and Lior Rokach. 2018. Personal Price Aware Multi-Seller Recommender System: Evidence from eBay. Knowledge-Based Systems 150 (02 2018). https://doi.org/10.1016/j.knosys.2018.02.026
  • Ni et al. [2019] Jianmo Ni, Jiacheng Li, and Julian McAuley. 2019. Justifying Recommendations using Distantly-Labeled Reviews and Fine-Grained Aspects. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 188–197. https://doi.org/10.18653/v1/D19-1018
  • Pal et al. [2020] Aditya Pal, Chantat Eksombatchai, Yitong Zhou, Bo Zhao, Charles Rosenberg, and Jure Leskovec. 2020. PinnerSage. Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (Jul 2020). https://doi.org/10.1145/3394486.3403280
  • Rousseeuw [1987] Peter J. Rousseeuw. 1987. Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20 (1987), 53–65. https://doi.org/10.1016/0377-0427(87)90125-7
  • Ruiyang Ren [2020] Yaliang Li Wayne Xin Zhao Hui Wang Bolin Ding Ji-Rong Wen Ruiyang Ren, Zhaoyang Liu. 2020. Sequential Recommendation with Self-Attentive Multi-Adversarial Network. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, China, July 25-30, 2020.
  • Scott [2012] David W. Scott. 2012. Multivariate Density Estimation and Visualization.
  • Steininger et al. [2021] Michael Steininger, Konstantin Kobs, Padraig Davidson, Anna Krause, and Andreas Hotho. 2021. Density-based weighting for imbalanced regression. Machine Learning 110, 8 (01 Aug 2021), 2187–2211. https://doi.org/10.1007/s10994-021-06023-5
  • Vaswani et al. [2017] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In Advances in Neural Information Processing Systems, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), Vol. 30. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf
  • Wang et al. [2017] Yu-Xiong Wang, Deva Ramanan, and Martial Hebert. 2017. Learning to Model the Tail. In Advances in Neural Information Processing Systems, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), Vol. 30. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2017/file/147ebe637038ca50a1265abac8dea181-Paper.pdf
  • Weston et al. [2013] Jason Weston, Ron J. Weiss, and Hector Yee. 2013. Nonlinear latent factorization by embedding multiple user interests. Proceedings of the 7th ACM conference on Recommender systems (2013).
  • Yang et al. [2021] Yuzhe Yang, Kaiwen Zha, Ying-Cong Chen, Hao Wang, and Dina Katabi. 2021. Delving into Deep Imbalanced Regression. In ICML. 11842–11851. http://proceedings.mlr.press/v139/yang21m.html
  • Yi et al. [2019] Xinyang Yi, Ji Yang, Lichan Hong, Derek Zhiyuan Cheng, Lukasz Heldt, Aditee Ajit Kumthekar, Zhe Zhao, Li Wei, and Ed Chi (Eds.). 2019. Sampling-Bias-Corrected Neural Modeling for Large Corpus Item Recommendations.
  • Yin Cui [2018] Chen Sun Andrew Howard Serge Belongie Yin Cui, Yang Song. 2018. Large Scale Fine-Grained Categorization and Domain-Specific Transfer Learning. In CVPR.
  • Zhang et al. [2021] Yin Zhang, Derek Zhiyuan Cheng, Tiansheng Yao, Xinyang Yi, Lichan Hong, and Ed H. Chi. 2021. A Model of Two Tales: Dual Transfer Learning Framework for Improved Long-tail Item Recommendation. arXiv:cs.IR/2010.15982
  • Zhu et al. [2021] Feng Zhu, Yan Wang, Chaochao Chen, Jun Zhou, Longfei Li, and Guanfeng Liu. 2021. Cross-Domain Recommendation: Challenges, Progress, and Prospects. arXiv:cs.IR/2103.01696
  • Zhu et al. [2018] Yu Zhu, Jinhao Lin, Shibi He, Beidou Wang, Ziyu Guan, Haifeng Liu, and Deng Cai. 2018. Addressing the Item Cold-start Problem by Attribute-driven Active Learning. arXiv:cs.IR/1805.09023
DNN
Deep Neural Network
DNNs
Deep Neural Networks
MS
mean-silhouette
IDW
Iterative Density Weighting
KDE
Kernel Density Estimation
MUR
multiple user representations
SUR
single user representation