This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Can One Embedding Fit All? A Multi-Interest Learning Paradigm Towards Improving User Interest Diversity Fairness

Yuying Zhao [email protected] Vanderbilt UniversityNashvilleTNUSA Minghua Xu [email protected] Visa ResearchFoster CityCAUSA Huiyuan Chen [email protected] Visa ResearchFoster CityCAUSA Yuzhong Chen [email protected] Visa ResearchFoster CityCAUSA Yiwei Cai [email protected] Visa ResearchFoster CityCAUSA Rashidul Islam [email protected] Visa ResearchFoster CityCAUSA Yu Wang [email protected] Vanderbilt UniversityNashvilleTNUSA  and  Tyler Derr [email protected] Vanderbilt UniversityNashvilleTNUSA
(2024)
Abstract.

Recommender systems (RSs) have gained widespread applications across various domains owing to the superior ability to capture users’ interests. However, the complexity and nuanced nature of users’ interests, which span a wide range of diversity, pose a significant challenge in delivering fair recommendations. In practice, user preferences vary significantly; some users show a clear preference toward certain item categories, while others have a broad interest in diverse ones. Even though it is expected that all users should receive high-quality recommendations, the effectiveness of RSs in catering to this disparate interest diversity remains under-explored.

In this work, we investigate whether users with varied levels of interest diversity are treated fairly. Our empirical experiments reveal an inherent disparity: users with broader interests often receive lower-quality recommendations. To mitigate this, we propose a multi-interest framework that uses multiple (virtual) interest embeddings rather than single ones to represent users. Specifically, the framework consists of stacked multi-interest representation layers, which include an interest embedding generator that derives virtual interests from shared parameters, and a center embedding aggregator that facilitates multi-hop aggregation. Experiments demonstrate the effectiveness of the framework in achieving better trade-off between fairness and utility across various datasets and backbones.

Fairness; Diversity; Multi-Interest Recommendations
journalyear: 2024copyright: acmlicensedconference: Proceedings of the ACM Web Conference 2024; May 13–17, 2024; Singapore, Singaporebooktitle: Proceedings of the ACM Web Conference 2024 (WWW ’24), May 13–17, 2024, Singapore, Singaporedoi: 10.1145/3589334.3645662isbn: 979-8-4007-0171-9/24/05ccs: Information systems Recommender systemsccs: Social and professional topics User characteristics

1. Introduction

Recommender systems (RSs) have been widely applied in different domains, such as news recommendation (Liu et al., 2010), friend recommendation (Fan et al., 2019), etc. While a plethora of RSs have been proposed (He et al., 2020; Wang et al., 2023b; Chen et al., 2021, 2022; Rendle et al., 2012), the main focus is on maximizing the overall utility, typically measured by metrics like Recall, F1, and NDCG (Aggarwal, 2016). These metrics offer a comprehensive view on the accuracy of recommendations and the system’s ability in capturing user interests. However, solely relying on these utility-based metrics can cause issues: (1) it hides biases across distinct user groups, posing fairness concerns; and (2) it overshadows underlying performance bottlenecks, impeding potential utility enhancements. In light of these issues, recent studies have adopted a group-centric lens for recommendations (Li et al., 2021; Wu et al., 2021, 2022). Investigations have been conducted on user groups defined by explicit attributes (i.e., sensitive features) (Wang et al., 2022b), such as gender (Wu et al., 2021), race (Zhu et al., 2018), as well as implicit features (i.e., extracted from interactions) such as number of interactions and amount of purchases (Li et al., 2021; Wu et al., 2022). These studies highlight group-specific biases and advocate for solutions that ensure fairness. Given the rich existing literature focused on explicit sensitive attributes, our study dives into the implicit features and specifically focuses on a novel perspective termed user interest diversity. We investigate the following research question:

Refer to caption
Figure 1. Why diverse interests matter? Real-world RS examples (A) Food recommendation (B) Dating recommendation222Note that the illustration does not represent authors’ perspective on binary genders..

Are users of varied interest diversity treated fairly in RSs?

Firstly, imbalanced user satisfaction could undermine the overall utility of the platform and even result in dissatisfied users leaving (i.e., increased user defection) (Li et al., 2021; Wu et al., 2022). For example, in the context of food recommendation in Fig. 2(A), some users prefer a limited number of cuisines while others have more flexible tastes. Satisfying all users is a primary goal. Secondly, if the platform fails to equitably accommodate these diverse preferences, it not only raises issues of user satisfaction but also poses significant ethical concerns. Online dating recommendation in Fig. 2(B) serves as a pertinent example. Users exhibit a spectrum of sexual orientations, including homosexuality, bisexuality, heterosexuality, and more. While homosexual and heterosexual users have more specific preferences related to gender interests, bisexual users might exhibit a broader range of interests. Ensuring a fair system for users with varied interest diversity is a core requirement for ethical consideration (Zhao et al., 2024).

To explore the fairness of existing models towards users exhibiting various levels of interest diversity, we conduct a preliminary experiment with detailed analysis in Sec. 2. In particular, we contemplate two scenarios: one where item category information (e.g., movie genres) is available, and another where it is not. We then define two interest diversity metrics. Following this, we categorize users into groups based on interest diversity and compare the utility metrics of the recommendations they receive. The results reveal a pattern that users with higher interest diversity tend to receive lower recommendation performance. This observation remains consistent across multiple datasets, models, definitions of interest diversity, and group partitions. Our experiments indicates that the unfairness among user groups with varied interest diversity (i.e., user interest diversity unfairness) indeed exists. To alleviate such unfairness without compromising the overall utility performance, it’s necessary to enhance the recommendations for users with high interest diversity, as this is the system’s performance bottleneck. We explore the cause of performance disparity among user groups, and our conclusion aligns with prior work (Zhang et al., 2022; Cen et al., 2020), which suggests that a single embedding is insufficient to capture users’ interests.

To this end, we propose a multi-interest framework to improve user interest diversity fairness, that can be integrated into existing RS models. In our multi-interest framework, each user is composed of a center embedding representing users’ main characteristic and multiple virtual embeddings, reflecting users’ interests derived from their interacted items. We develop multi-interest representation layers to learn better user embeddings, especially for users with high interest diversity. Each layer includes an interest embedding generator that derives virtual interest embeddings from globally shared interest parameters, and a center embedding aggregator that facilitates multi-hop aggregation. As such, the designed mechanism can automatically assign different interest numbers that are generally consistent with the interest diversity in an implicit manner. Experimental results validate the effectiveness of our framework in achieving a better trade-off between fairness and utility performance. Our main contributions are summarized as follows:

  • Consistent Disparity Identification: We identify the unfair treatment among users with varied interest diversity, where users with broader interests tend to receive lower-quality recommendation. This pattern has been empirically verified to be consistent across datasets, models, diversity metrics, and group partitions.

  • Multi-interest Framework Design: We delve into the potential reason causing the disparity from the embedding space where we observe the insufficiency of using single embedding to represent users and items due to their complex multi-faceted interactions. This motivates us to propose a multi-interest framework which is both model-agnostic and parameter-efficient.

  • Better Fairness-Utility Tradeoff: Our proposed multi-interest framework outperforms the backbone models and fairness baselines by achieving the optimal balance between fairness and utility. Also, it offers superior and more balanced embedding alignment, along with more diverse recommendations.

2. User Interest Diversity Unfairness

Table 1. Notations.
Notations Descriptions
u\mathcal{I}_{u} User uu’s interactions u=[i1,i2,,idu]\mathcal{I}_{u}=[i_{1},i_{2},...,i_{d_{u}}]
dud_{u} Number of interactions
𝒞u\mathcal{C}_{u} Category set of u\mathcal{I}_{u}
NucN_{u}^{c} Number of user uu’s interaction in category cc
Dcat/Demb\text{D}_{\text{cat}}/\text{D}_{\text{emb}} User interest diversity via item category/embedding
ϕ(,)\phi(\cdot,\cdot) Similarity function
𝐞u\mathbf{e}_{u}/𝐞i\mathbf{e}_{i} User/Item embeddings
𝐞~u\tilde{\mathbf{e}}_{u}/𝐞~i\tilde{\mathbf{e}}_{i} Normalized User/Item embeddings
𝐀\mathbf{A}/𝐃\mathbf{D} Adjacency/Degree matrix
K/kK/k Number of interests/kk-th interest
NN Number of users and items
dd Embedding dimension
y^ui\hat{y}_{ui} Relevance score between user uu and item ii
𝒩vi\mathcal{N}_{v_{i}} The neighborhood set of node viv_{i}
𝐄Cl\mathbf{E}_{C}^{l} Center embeddings at layer ll
𝐄Vl\mathbf{E}_{V}^{l} Virtual interest embeddings at layer ll
𝐰kl\mathbf{w}_{k}^{l} Global interest parameter of kk-th interest at layer ll
/\downarrow/\uparrow The lower/higher the better
Refer to caption
Figure 2. Group recommendation performance (Recall \uparrow): the pattern that users with more diverse interests generally receive lower recommendation quality is consistent across various datasets, models, diversity metrics, and group partitions. A larger group ID indicates a higher level of user interest diversity.

In this section, we investigate how existing RSs treat users with varied levels of interest diversity. First, we formally define interest diversity, concatering two scenarios where item category is available or not. Then, we categorize users into groups with varied levels of interest diversity. Ultimately, we demonstrate the performance across different groups using two representative recommendation models: LightGCN (He et al., 2020) and CAGCN (Wang et al., 2023b). The disparate group performance reveals the existence of user interest diversity unfairness. Notations used in the paper are summarized in Table 1.

2.1. Interest Diversity Definition

User interest diversity aims to measure the dissimilarity of the items interacted with each user in the training data (i.e., users’ historical interactions). Based on whether category information is available, we define interest diversity based on item category or item embedding.

Definition 2.1.

Interest Diversity via Item Category. Given user uu’s historical interaction u=[i1,i2,,idu]\mathcal{I}_{u}=[i_{1},i_{2},...,i_{d_{u}}] where dud_{u} is the number of interactions and 𝒞u\mathcal{C}_{u} is the set of categories of items user uu has interacted with, NucN_{u}^{c} denotes the number of items from user uu’s interaction belonging to category cc, we define user uu’s interest diversity Dcat(u)\text{D}_{\text{cat}}(u) following Simpson’s Index of Diversity (Simpson, 1949):

(1) Dcat(u)=1c𝒞uNuc(Nuc1)|u|(|u|1).\text{D}_{\text{cat}}(u)=1-\frac{\sum_{c\in\mathcal{C}_{u}}N_{u}^{c}(N_{u}^{c}-1)}{|\mathcal{I}_{u}|(|\mathcal{I}_{u}|-1)}.
Definition 2.2.

Interest Diversity via Item Embedding. Given the pretrained item embeddings, user uu’s interest diversity Demb(u)\text{D}_{\text{emb}}(u) is as follows:

(2) Demb(u)=1𝔼(i,i)u×uϕ(𝐞i,𝐞i),\text{D}_{\text{emb}}(u)=1-\mathbb{E}_{(i,i^{\prime})\in\mathcal{I}_{u}\times\mathcal{I}_{u}}\phi(\mathbf{e}_{i},\mathbf{e}_{i^{\prime}}),

where ϕ(𝐞i,𝐞i)=𝐞i𝐞i𝐞i𝐞i\phi(\mathbf{e}_{i},\mathbf{e}_{i^{\prime}})=\frac{\mathbf{e}_{i}\cdot\mathbf{e}_{i^{\prime}}}{{\|{\mathbf{e}_{i}}\|}{\|{\mathbf{e}_{i^{\prime}}}\|}} is the cosine similarity between the embeddings of two items i,ii,i^{\prime}.

Essentially, Dcat(u)\text{D}_{\text{cat}}(u) measures the probability that two randomly sampled items are from different categories, and Demb(u)\text{D}_{\text{emb}}(u) measures the dissimilarity between the interacted items in their embedding space. For both scenarios, a larger value indicates a higher level of interest diversity. Unless specified, we use Dcat\text{D}_{\text{cat}} for default.

2.2. Group Partition

Given users’ interest diversity, we group users with k-means clustering (MacQueen et al., 1967). The number of clusters is determined using the commonly-used elbow method (Thorndike, 1953). The assignment of clusters subsequently defines the group partition, with a higher group ID indicating a higher diversity of interests. It’s worth noting that there are alternative methods to group users, e.g., dividing users into equal sized groups based on number of users, or range of user interest diversity. Unless specified otherwise, we primarily rely on k-means clustering in the experiments.

2.3. Preliminary Results

Given the exceptional performance of utilizing graphs in RSs, we select two graph-based models for evaluation: LightGCN (He et al., 2020) and CAGCN (Wang et al., 2023b). The former is a widely recognized and frequently used model. The latter is a newer development and improves the overall utility by reducing the emphasis of neighbors not adhering to the main interest which is closely related to our topic.

We evaluate them on four datasets including ml-1m, epinion, embmetics, and anime, the details of which will be described in Sec. 5.1.1. The preliminary results across different scenarios are illustrated in Fig. 2. Specifically, Fig. 2(A) is the group utility performance (Recall) where groups are divided based on k-means clustering with Dcat\text{D}_{\text{cat}} as the diversity metric. The curves suggest a trend that as interest diversity increases, the group utility performance generally decreases. This pattern is observable across multiple datasets and models. We also explore another diversity definition Demb\text{D}_{\text{emb}} in Fig. 2(B) which shows a similar trend. Additionally, we obtain results based on different group partitions including the equal user number and equal user interest diversity range in Fig. 2(C). The results show a consistent trend across various datasets, models, diversity metrics, and group partitions that users with diverse interests generally receive a lower recommendation quality. This indicates the existence of user interest diversity unfairness, which jeopardizes the user experience for user with diverse interests.

3. Source of Unfairness and Motivation of Multi-Interest

To mitigate user interest diversity unfairness identified in Sec. 2, we dive into the source from the alignment and misalignment between user and item embeddings. Our empirical findings indicate a trend in alignment that correlates with the observed performance disparities: user group with diverse interests has poor performance as well as poor alignment. We hypothesize that the suboptimal alignment arises from the inadequacy of using single embedding to align user’s diverse interests (illustrated in Fig. 4).

Since the core component in majority RSs is to learn high-quality user and item embeddings, we investigate the root cause of user interest diversity unfairness from the embedding space. Prior research has underscored the correlation between embedding alignment (i.e., the capacity to bring users and their associated items closer in the embedding space) and utility performance (Wang et al., 2022a; Wang and Isola, 2020). A superior alignment typically correlates with a better performance. The alignment definition is as follows:

(3) Alignment=𝔼(u,i)ppos𝐞~u𝐞~i2,\displaystyle{\text{Alignment}}=\mathbb{E}_{(u,i)\sim p_{\text{pos}}}\left\|\tilde{\mathbf{e}}_{u}-\tilde{\mathbf{e}}_{i}\right\|^{2},

where 𝐞~u\tilde{\mathbf{e}}_{u} and 𝐞~i\tilde{\mathbf{e}}_{i} are the l2l_{2} normalized user and item embeddings from historical interacted pairs. It measures the Euclidean distance in the unit hypersphere and a lower Alignment score (aka. shorter distance) corresponds to better utility performance. To uncover the potential reason for unfair recommendation performance across different user groups, we measure the average Alignment in each group. Results on ml-1m in Fig. 3 (results for other datasets are included in Appendix A) show that (1) CAGCN exhibits superior alignment compared to LightGCN, a consistency mirrored in the performance illustrated in Fig. 2; (2) Users displaying a broader spectrum of interests tend to have larger Alignment scores in the embedding space. This suggests that the current recommendation models are not effective in aligning users and items, particularly when users have a wide array of interests.

Fig.4(A) depicts the alignment challenge for user with high interest diversity. When the user is represented by a single embedding, to achieve an optimal alignment with every interacted item, the learned single embedding falls in-between the interacted items. This results in a poor alignment with the real interests. Such insufficiency of using single embedding to align interacted items, that are from diverse interests, motivates us to use multiple embeddings to represent different user interests (Zhang et al., 2022; Li et al., 2019; Cen et al., 2020). As shown in Fig. 4(B), the user has multiple embeddings. For items belonging to diverse interests, the embeddings can be automatically obtained and they have a better alignment with the corresponding interacted items in the embedding space. Comparing the scenarios of single-interest and multi-interest, we find that owing to a better alignment, the recommended items in Fig. 4(B) are more accurate than Fig. 4(A). This underscores the potential of the multi-interest approach. We also conduct experiments in Appendix B to investigate whether the unfairness is due to interest shift, which is not the major cause.

Refer to caption
Figure 3. Group-level embedding alignment (\downarrow) of ml-1m dataset based on LightGCN and CAGCN.

4. The Multi-Interest Framework

To mitigate unfairness, we propose a multi-interest framework where each user is represented by multiple (virtual) interest embeddings. Based on the proposed framework, we improve the alignment for users with high interest diversity, thereby improving their recommendation performance and alleviating the performance bias. Next, we give an overview of the framework in Sec. 4.1, elaborate the component details in Sec. 4.2 and optimization in Sec. 4.3.

4.1. Model Architecture

Fig. 5 shows the multi-interest framework where each user/item has different types of embeddings, including (1) center embeddings 𝐄ClN×d\mathbf{E}_{C}^{l}\in\mathbb{R}^{N\times d} representing users/items main characteristic/features where NN is the number of node (including users and items) and dd is the dimension; (2) interest (virtual) embeddings 𝐄VlN×K×d\mathbf{E}_{V}^{l}\in\mathbb{R}^{N\times K\times d} which relate to specific interests where KK is the number of interests (for simplicity, we denote 𝐄V\mathbf{E}_{V} as virtual embeddings hereafter). Among these embeddings, center embeddings are learnable parameters while the virtual embeddings are calculated based on center embeddings via attentions. This mechanism avoids introducing a large number of learnable parameters by sharing the global interest 𝐰kl\mathbf{w}_{k}^{l} in the attention mechanism. We represent the kk-th virtual embedding of node vuv_{u} as 𝐄VL[vu,k]\mathbf{E}_{V}^{L}[v_{u},k] and the user center embedding as 𝐄CL[vu]\mathbf{E}_{C}^{L}[v_{u}]. Similar notations apply to the item side. We illustrate the motivation of using multiple embeddings for items in Appendix C.

With these notations, the framework is as follows: (1) Given the user-item bipartite graph, user and item embeddings are obtained through the multi-interest representation layers (details in Sec. 4.2);(2) After obtaining the embeddings, the relevance score y^ui\hat{y}_{ui} for user, item pair (vu,vi)(v_{u},v_{i}) is calculated based on the last layer representations where L is the number of hops:

(4) y^ui=maxk=1K𝐄VL[vu,k]𝐄CL[vi]+maxk=1K𝐄VL[vi,k]𝐄CL[vu];\hat{y}_{ui}=\max_{k=1}^{K}\mathbf{E}_{V}^{L}[v_{u},k]^{\top}\mathbf{E}_{C}^{L}[v_{i}]+\max_{k=1}^{K}\mathbf{E}_{V}^{L}[v_{i},k]^{\top}\mathbf{E}_{C}^{L}[v_{u}];

(3) These predicted relevance scores are optimized via Bayesian Personalized Ranking Loss (BPR) loss (Rendle et al., 2012) BPR\mathcal{L}_{\text{BPR}}.

Refer to caption
Figure 4. Multi-interest motivation: single embedding is insufficient to capture users’ diverse interests.
Refer to caption
Figure 5. Multi-interest framework (interest number equals two): rather than a single embedding, each user/item is represented by multiple embeddings (i.e., center and virtual). Center embeddings and global interest embeddings are learnable parameters while the interest (virtual) embeddings are calulated without assigning extra parameters.

Note that the relevance score in Eq.(4) is different from the calculation in previous recommendation models (Wang et al., 2023b; He et al., 2020) or multi-interest-based session recommendation (Zhang et al., 2022; Cen et al., 2020). In previous works, because user and item only have single embeddings, the dot product between the user and the item embedding (i.e., 𝐞u𝐞i{\mathbf{e}_{u}}^{\top}\mathbf{e}_{i}) denotes their relevance score. In multi-interest based session recommendation, only items have learnable parameters and users/sessions are calculated based on items (KK embeddings with 𝐞uk\mathbf{e}_{u}^{k} denoting the kk-th interest) and therefore maxk=1K𝐞uk𝐞i\max_{k=1}^{K}{\mathbf{e}_{u}^{k}}^{\top}\mathbf{e}_{i} is sufficient to update the item embeddings. However, similar to LightGCN, we have both user and item embeddings to learn. Simply optimizing the traditional multi-interest relevance score that is commonly used in session-based representation cannot utilize user embeddings, indicating it is not suitable in our case. Therefore, we use the symmetric scores shown in Eq.(4) where both user and item embeddings are optimized.

4.2. Multi-Interest Representation Layer

Next, we introduce the details of multi-interest representation layer, which is at the core of the architecture and designed to learn, calculate and aggregate multiple embeddings. The model is composed of stacked layers to deliver the final user and item embeddings.

4.2.1. Interest embedding generation:

Virtual embeddings of ll-th layer for node vv and kk-th interest (i.e., 𝐄Vl[v,k]\mathbf{E}_{V}^{l}[v,k]) is calculated in Eq.(5) as the weighted average of the center embeddings of neighbors. The weight is calculated in Eq.(6) based on Softmax attention mechanism where TT is the temperature to control the Softmax smoothness. The input logits to Softmax function are cosine distances between virtual embeddings and the global interest 𝐰kl\mathbf{w}_{k}^{l}. Intuitively, if an item is related to the kk-th interest, the attention will be higher and lead to larger contribution to the aggregates from this item. Therefore, 𝐄Vl[v,k]\mathbf{E}_{V}^{l}[v,k] captures information related to kk-th interest.

(5) 𝐄Vl[v,k]=vn𝒩vak,vnl𝐄Cl[vn]\mathbf{E}_{V}^{l}[v,k]=\sum_{v_{n}\in\mathcal{N}_{v}}a_{k,v_{n}}^{l}\mathbf{E}_{C}^{l}[v_{n}]
(6) ak,vnl=exp(ϕ(𝐄Cl[vn],𝐰kl)/T)iexp(ϕ(𝐄Cl[vn],𝐰il)/T)a_{k,v_{n}}^{l}=\frac{\text{exp}(\phi(\mathbf{E}_{C}^{l}[v_{n}],\mathbf{w}_{k}^{l})/T)}{\sum_{i}\text{exp}(\phi(\mathbf{E}_{C}^{l}[v_{n}],\mathbf{w}_{i}^{l})/T)}

4.2.2. Center embedding aggregator:

We adopt the similar approach as LightGCN (He et al., 2020) to update embeddings based on topology of the graph. Different from LightGCN, we use virtual embeddings to update the center embedding as in Eq.(7). Since virtual embeddings have extra dimension in interest, these embeddings need to be transformed to the same dimension as center embedding before the aggregation. We use an argmax operator to select the interest id of the “matching slide” called mid. The embeddings of mid index has the highest dot product similarity with the node’s center embedding. Such operator has been commonly used in multi-interest literature (Cen et al., 2020; Zhang et al., 2022) and has been verified to have faster convergence and better performance compared with other ways to use multi-interests (Li et al., 2019). For each node vv whose center embedding is 𝐄Cl[v]\mathbf{E}_{C}^{l}[v], the id of the matching slide for one neighbor node vn𝒩vv_{n}\in\mathcal{N}_{v} is:

mid(v,vn,l)=argmaxk=1K(𝐄Vl[vn,k]𝐄Cl[v])\text{mid}(v,v_{n},l)=\text{argmax}_{k=1}^{K}({\mathbf{E}_{V}^{l}[v_{n},k]}^{\top}\mathbf{E}_{C}^{l}[v])

Given the “matching slide,” the aggregation process is as follows:

(7) 𝐄Cl+1[v]=vn𝒩v1dvdvn𝐄Vl[vn,mid(v,vn,l)]\mathbf{E}_{C}^{l+1}[v]=\sum_{v_{n}\in\mathcal{N}_{v}}{\frac{1}{\sqrt{d_{v}d_{v_{n}}}}\mathbf{E}_{V}^{l}[v_{n},\text{mid}(v,v_{n},l)]}

4.3. Optimization

We utilize the BPR loss (Rendle et al., 2012) (BPR\mathcal{L}_{\text{BPR}}) to train our multi-interest RS.

BPR=(u,i,j)𝒟logσ(y^uiy^uj)+λΘΘ2,\mathcal{L}_{\text{BPR}}=-\sum_{(u,i,j)\in\mathcal{D}}\log\sigma(\hat{y}_{ui}-\hat{y}_{uj})+\lambda_{\Theta}\|\Theta\|^{2},

where 𝒟={(u,i,j)|u𝒰iu+ju}\mathcal{D}=\{(u,i,j)|u\in\mathcal{U}\land i\in\mathcal{I}_{u}^{+}\land j\in\mathcal{I}_{u}^{-}\} is the training dataset and 𝒰\mathcal{U} is the total user set, u+\mathcal{I}_{u}^{+}/u\mathcal{I}_{u}^{-} are the item sets that user uu has/hasn’t interacted with. σ()\sigma(\cdot) is Sigmoid function. Θ\Theta denotes the model parameter with λΘ\lambda_{\Theta} controlling the L2L_{2} norm regulation to prevent over-fitting. y^ui\hat{y}_{ui} is the predicted preference/relevance score computed based on Eq.(4).

5. Experiments

In this section, we evaluate the performance of our multi-interest framework on on real-world datasets and compare the utility and fairness performance with various representative methods. Through experiments, we aim to answer the following research questions:

  • RQ1: Does our proposed multi-interest framework achieve a better utility-fairness trade-off than the baseline methods?

  • RQ2: Is the multi-interest framework able to learn higher-quality embeddings with better alignment?

  • RQ3: Can the proposed framework learn to match the number of interest embeddings with the diversity of historical interactions?

  • RQ4: Can the multi-interest framework provide extra benefits beyond accuracy and fairness, e.g., recommendation diversity?

  • RQ5: How do the hyperparameters affect the performance?

Table 2. Dataset statistics.
Dataset # Edges # Users # Items # Category
ml-1m 223305 5645 2357 18
epinion 163320 11875 11164 26
cosmetics 930275 53238 28310 400
anime 901328 40112 4514 76
Table 3. Fairness and utility performance for various models denoted by Backbone-Method (Each model is composed of the recommendation backbone and fairness method). The best is highlighted in bold and the runner-up is underlined.
Backbone Method ml-1m epinion cosmetics anime Avg Rank
Recall\uparrow Unfairness\downarrow Recall\uparrow Unfairness\downarrow Recall\uparrow Unfairness\downarrow Recall\uparrow Unfairness\downarrow
LightGCN Vanilla 0.3087 0.0376/0.1018 0.0904 0.0320/0.0378 0.2116 0.1260/0.1942 0.4015 0.0384/0.098 2.08
DRO 0.3143 0.0409/0.1047 0.0926 0.0377/0.0.451 0.2104 0.1296/0.2013 0.4029 0.0453/0.1102 2.92
ARL 0.2973 0.0376/0.1058 0.0850 0.0316/0.0381 0.1941 0.1199/0.1730 0.3844 0.0407/0.1110 2.83
Multi 0.3116 0.0385/0.0856 0.0901 0.0364/0.0222 0.2405 0.1193/0.1494 0.4239 0.0430/0.1136 1.92
CAGCN Vanilla 0.3141 0.0429/0.1054 0.0948 0.0380/0.0373 0.2286 0.1332/0.1903 0.4044 0.0415/0.1096 2.83
DRO 0.3173 0.0401/0.0946 0.0927 0.0383/0.0375 0.2294 0.1350/0.1891 0.4024 0.0414/0.1012 2.58
ARL 0.3024 0.0367/0.1121 0.0912 0.0354/0.0380 0.2167 0.1257/0.1734 0.3884 0.0413/0.0985 2.67
Multi 0.3107 0.0411/0.0921 0.0922 0.0378/0.0212 0.2548 0.1297/0.1368 0.4237 0.0417/0.0904 1.92

5.1. Experimental Setup

5.1.1. Datasets.

We evaluate the proposed multi-interest framework on four datasets including ml-1m, epinion, cosmetics, and anime333Datasets are available at: ml-1m, epinion, cosmetics, anime. We pre-process data by (1) filtering edges by maintaining the highest rating score so that the remaining edges show strong preferences; and (2) applying k-core filtering iteratively to remove users with interaction number smaller than 55. After that, we randomly split the dataset into train/validation/test based on 60%/20%/20%60\%/20\%/20\% proportions. The statistics of the pre-processed datasets are summarized in Table 2.

5.1.2. Baselines.

To verify whether our framework can achieve a better trade-off between fairness and utility, and further generalize to different backbones, we compare the performance of two representative recommendation backbones (LightGCN (He et al., 2020) and CAGCN (Wang et al., 2023b)) before/after equipping our proposed multi-interest framework. For a fair comparison, we also apply other fair baselines to the backbones including DRO (Hashimoto et al., 2018) and ARL (Lahoti et al., 2020). Note that all these methods are group-agnostic which means that the group partition is unavailable during training. More descriptions about compared methods and implementation details are in Appendix D444Our code and datasets are available at: Code.. The number of parameters are reported in Appendix E.

5.1.3. Metrics.

For utility performance, we adopt Recall@20 and NDCG@20. For fairness performance, we use the standard deviation of the utility performance across user groups. The deviation measures the performance gap among groups, and a larger score signifies lower fairness. Based on group partitions via interest diversity metrics Dcat\text{D}_{\text{cat}} and Demb\text{D}_{\text{emb}}, we report two corresponding (un)fairness scores. This setting can evaluate whether the group-agnostic models are effective for different group partitions.

5.2. Performance Comparison (RQ1)

We present the utility and fairness scores in Table 3 (The results based on another utility metric NDCG is included in Appendix G). Since the standard deviations for all methods across various seeds are negligible compared with the main performance, we leave them out. From the result, we draw several observations:

  • The multi-interest framework has the best fairness-utility trade-off in general. Our proposed method achieves the best and runner up performance in most of the times when compared with other methods. Upon calculating the average rank for each method, ours emerges as the leader in both backbones. While the current rank of 1.921.92 indicates some room for enhancement towards the optimal rank of 11, it underscores the efficacy and potential of the multi-interest framework in balancing fairness and utility.

  • The multi-interest framework works better with large dataset. In cosmetics dataset, which has the highest count of items and categories, our method consistently delivers enhanced performance in both fairness and utility. Given the diversity of items and categories, learning varied interests becomes more essential, amplifying the advantages. We dive into more details about the impact of items and categories in Appendix H.

  • The multi-interest framework is more stable across backbones compared with other fairness baselines (i.e., DRO and ARL). DRO and ARL rank higher than the base model CAGCN, however, their rank drops when integrated into LightGCN. This underscores the complexity of maintaining an optimal balance across different models. Furthermore, such distinct performance variations of DRO and ARL across different backbones can be attributed to their inherent design. These methods were specifically designed to enhance the performance for instances with suboptimal recommendations. While Fig. 2 demonstrates that the user group with diverse interests has the poorest average performance and is expected to gain the most, other factors, such as the percentage of under-performing users in each group, play a role. If other groups have a higher proportion of users with poor recommendations, they might obtain greater benefits, thereby increasing the unfairness. Therefore, we can observe in some cases (e.g., DRO in ml-1m with LightGCN backbone) that the utility improves and the fairness drops. Such percentage in each group can vary across models, resulting in high instability of DRO and ARL due to their heavy reliance on the performance distribution. This suggests that DRO and ARL are not universally effective in the current context. In contrast, the multi-interest framework relies on the underlying interests rather than performance, which is more closely related to the current setting and more stable.

5.3. Representation Quality (RQ2)

Multiple embeddings are expected to learn a better embedding distribution compared with single embeddings (Fig. 4), especially for the embedding alignment between user and interacted items. To evaluate this, we calculate the average alignment based on the backbones and their multi-version. Table 4 shows that multi-interest improves the alignment consistently. This suggests that the framework effectively brings users and their interacted items closer in the embedding space. However, an intriguing observation arises when examining performance metrics. While the improved alignment in CAGCN leads to superior utility performance compared to LightGCN in Table 3, the enhanced alignment in the multi-version does not always result in better utility performance relative to the backbones. This inconsistency may arise from the trade-off between alignment and uniformity (Wang et al., 2022a). Specifically, while alignment improves, it could lead to reduced uniformity in the multi-version due to more user embeddings, which offsets the anticipated enhancements. The nuanced interplay between alignment and uniformity, and strategies to effectively balance them, present intriguing avenues for future exploration in multi-interest scenario.

Table 4. Embedding alignment (Results with improved alignment compared with backbone are highlighted in bold).
Method ml-1m epinion cosmetics anime
LightGCN 0.8774 0.5951 0.7937 1.0165
Multi-LightGCN 0.5007 0.4111 0.5396 0.7514
CAGCN 0.7512 0.5429 0.7069 0.9118
Multi-CAGCN 0.4315 0.2973 0.4694 0.7176

Beyond evaluating overall alignment, we delve into embedding alignment at the group level. In Fig. 6, there’s a discernible trend when comparing the backbone to its multi-version: alignment appears more evenly distributed across different groups. Since alignment is closely related to the utility performance, it contributes to a fair recommendation across groups, which follows our expectation.

Refer to caption
Figure 6. Group-level embedding alignment of ml-1m dataset based on two backbones.

5.4. Interest Matching (RQ3)

For each user, our multi-interest framework initially assigns the same number of interests (i.e., KK). Given the underlying assumption that users exhibit varied levels of interest diversity, can the model autonomously adjust the number of interests even if it begins with an equal allocation? To answer this question, we obtain the set of interests that matches the recommended items (i.e., for each item, the matched interest is the specific interest that has the maximum relevance score) and calculate the average matched interest number for each group. Results in Fig. 7 show that for the first three groups, users with more diverse interests have been assigned a larger interest number, indicating that our model has the ability to distinguish different interest diversity and can automatically cater to user preferences to some extend. However, the trend for the last two groups is not consistent, which leave us a future direction to explicitly assign interest number based on user interest diversity in addition to the current implicit way.

Table 5. Diversity measured by Dcat\text{D}_{\text{cat}} and Demb\text{D}_{\text{emb}} (Results with improved diversity compared with backbone are in bold).
Diversity Method ml-1m epinion cosmetics anime
Dcat\text{D}_{\text{cat}} LightGCN 0.3852 0.5477 0.6849 0.3193
Multi-LightGCN 0.3768 0.5454 0.6110 0.3300
CAGCN 0.3786 0.5382 0.6611 0.3206
Multi-CAGCN 0.4182 0.6667 0.7639 0.3573
Demb\text{D}_{\text{emb}} LightGCN 0.3189 0.2871 0.4271 0.5134
Multi-LightGCN 0.3206 0.3338 0.3781 0.4557
CAGCN 0.3934 0.3292 0.3833 0.4259
Multi-CAGCN 0.5229 0.3987 0.4919 0.4009
Refer to caption
Figure 7. Average interest number for each group on ml-1m.

5.5. Recommendation Diversity (RQ4)

We measure the diversity of the recommended item sets. The results are presented in Table 5 based on two diversity metrics: Dcat\text{D}_{\text{cat}} in Eq.(1) and Demb\text{D}_{\text{emb}} in Eq.(2). First, the cosmetics dataset, which has the highest number of categories among the datasets, consistently exhibits the greatest diversity in comparison to the other datasets. Second, CAGCN has a slightly reduced Dcat\text{D}_{\text{cat}} than LightGCN. This is attributed to CAGCN’s mechanism: it assigns higher pre-computed topological-based weights to neighbors that are more densely connected to the center node (i.e., nodes that are topologically more similar). While certain nodes gain emphasis, others get overshadowed. This reduces the likelihood of recommendations based on less-similar users, resulting in the drop in diversity. Third, multi-CAGCN has a consistent diversity enhancement (in both Dcat\text{D}_{\text{cat}} and Demb\text{D}_{\text{emb}}) compared with the backbone (with enhancements in 7/87/8 cases). We hypothesize that CAGCN learns more accurate user interests and incorporating higher-quality embeddings amplifies the advantages of our multi-interest framework.

5.6. Sensitivity Analysis (RQ5)

There are two hyperparameters in the model: the number of interests and the number of hops. From Fig. 8, we draw the following observations. A larger interest number could contribute to the utility performance but not necessary maintain a higher performance. This could be due to the increasing learning difficulty and over-fitting risk. Our multi-model prefers a smaller hop since (1) the multi-interest representation layer in Sec. 4.2 aggregates neighborhood information, serving as an implicit hop; (2) more layers would result in a higher level of smoothness which hides the diversity.

Refer to caption
Figure 8. Sensitivity analysis on Multi-LightGCN.

6. Related Works

6.1. Fairness in Recommender Systems

The majority RS development is concentrated predominantly on utility performance enhancement. However, emergent concerns regarding the equitable treatment of diverse user groups have motivated the advent of fairness-aware recommender systems (Li et al., 2022; Zhao et al., 2023; Wang et al., 2023a). Researchers have divided users into groups and investigated the group-level unfairness based on various criteria which can be summarized into two primary categories (Zhao et al., 2023): (1) explicit features, which involve sensitive features such as gender (Deldjoo et al., 2021; Wu et al., 2021), race (Gorantla et al., 2021; Zhu et al., 2018) and age (Farnadi et al., 2018); (2) implicit features, which are extracted from interactions such as the number of interactions (i.e., degree) and the amount of purchases (Li et al., 2021; Wu et al., 2022; Fu et al., 2020; Rahmani et al., 2022). While the explicit features are vital to fairness discourse, they are often inaccessible due to privacy policies or users’ reluctance to share such information. Consequently, our research focuses on implicit features given the profusion of user interactions in recommendation scenarios. Despite the significance of all previously mentioned features, our study explores a novel perspective within the realm of implicit features called user interest diversity considering its close relationship with the RS goal and its high relevance to real-world applications. Additionally, while most works adopt group information during training (Wang et al., 2022a; Li et al., 2022), recent works have also explored group-agnostic directions with the assumption that group partitions are not available during the optimization (Lahoti et al., 2020; Hashimoto et al., 2018). In this work, we follow this setting considering there are various ways to divide users into groups. Our goal is to develop a model that upholds fairness across diverse group divisions rather than catering to specific partition.

6.2. Multi-Interest Recommender Systems

The main idea of multi-interest solutions is that single embedding is insufficient to represent node’s features, hence necessitating the deployment of multiple embeddings. This idea has been extensively employed in session-based recommendations - depending on how the interests are obtained, the solutions fall into attention-based and category-based methods. The attention-based methods extract interests from the interactions into interest embeddings based on the attention mechanism. MIND (Li et al., 2019) initializes the effort to extract interests based on dynamic capsule routing. After that, ComiRec (Cen et al., 2020) leverages self-attention to learn multiple interests. While ComiRec (Cen et al., 2020) considers the item-to-interest relationship, Re4 (Zhang et al., 2022) models interest-to-item relationship by adding regularizations. The cluster-based methods perform clustering on the interacted items and obtain representative embedding per cluster to depict interests. PinnerSage (Pal et al., 2020) clusters interacted items with Ward hierarchical clustering method (Ward Jr, 1963), and utilizes the embedding of the center item, which minimizes distance sum to other items within the cluster, to depict user’s interests. MIP (Shi et al., 2022) assigns each interest as representation of the latest interacted item in each cluster. Additionally, MIP learns weight to represent preference over each interest and integrates it into the relevance score.

Beyond their application within RSs, multi-interest idea has also been applied in other representation learning tasks. For instance, the multi-interest-based random walk (Park et al., 2020; Yan et al., 2021) assigns each node a target embedding along with multiple context embeddings. Similarly, in the multi-interest-based Graph Neural Network (GNN)(Choudhary et al., 2022; Yan et al., 2023a), each node is characterized by several embeddings and an additional membership embedding that signifies the association with each interest. The principle of node partitioning(Epasto and Perozzi, 2019) has also been adapted to accommodate multi-interest strategies, where a node is divided into several virtual nodes based on neighborhood structure, whose embeddings represent the original node (Yan et al., 2023b).

In contrast, we delve into multi-interest in direct recommendation, emphasizing the importance of learning both user and item embeddings. Notably, in contrast to numerous studies (Epasto and Perozzi, 2019; Choudhary et al., 2022; Park et al., 2020; Yan et al., 2022) that increase parameter size for user representation, we employ shared global interest parameters for all users. This approach allows us to compute virtual interests in a parameter-efficient manner.

7. Conclusion

In this study, we examine whether users with varied levels of interest diversity are treated similarly/fairly in recommendation systems. Initial findings reveal a consistent disparity among user groups across different models, datasets, diversity metrics, and group partitions. This indicates the existence of User Interest Diversity Unfairness. Specifically, users with a broader range of interests often receive lower-quality recommendations, which has a negative impact on the user fairness and overall utility. Delving into the embedding space, we notice a trend linking group embedding alignment and utility performance. This suggests that a single embedding may not adequately represent diverse interests. To address this, we introduce a multi-interest framework where users are characterized by multiple (virtual) interest embeddings. Evaluation on two representative recommendation system backbones demonstrates that our approach better balances fairness and utility. Additionally, the learned embeddings have higher-quality and more balanced alignment in the embedding space. The proposed framework also provides more diverse recommendations. In future research, we aim to enhance the interest generation component. Currently this component is based on Softmax attention, other attentions or generative methods can be used to derive interest embeddings. For instance, we can incorporate text information and leverage large language models (LLM) for interest extraction/generation (Christakopoulou et al., 2023; Jin et al., 2024). The trade-off between alignment and uniformity within the realm of multi-interest also merits investigation.

Acknowledgements

This research is supported by Visa Research and the National Science Foundation (NSF) under grant number IIS2239881. The authors appreciate the anonymous reviewers for dedicating their time and efforts during the review process and offering insightful and constructive feedback.

References

  • (1)
  • Aggarwal (2016) Charu C Aggarwal. 2016. Evaluating recommender systems. Recommender Systems: The Textbook (2016), 225–254.
  • Cen et al. (2020) Yukuo Cen, Jianwei Zhang, Xu Zou, Chang Zhou, Hongxia Yang, and Jie Tang. 2020. Controllable multi-interest framework for recommendation. In Proceedings of KDD. 2942–2951.
  • Chen et al. (2021) Huiyuan Chen, Lan Wang, Yusan Lin, Chin-Chia Michael Yeh, Fei Wang, and Hao Yang. 2021. Structured graph convolutional networks with stochastic masks for recommender systems. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval.
  • Chen et al. (2022) Huiyuan Chen, Chin-Chia Michael Yeh, Fei Wang, and Hao Yang. 2022. Graph neural transport networks with non-local attentions for recommender systems. In Proceedings of the ACM Web Conference 2022.
  • Choudhary et al. (2022) Gautam Choudhary, Iftikhar Ahamath Burhanuddin, Eunyee Koh, Fan Du, and Ryan A Rossi. 2022. PersonaSAGE: A Multi-Persona Graph Neural Network. arXiv preprint arXiv:2212.13709 (2022).
  • Christakopoulou et al. (2023) Konstantina Christakopoulou, Alberto Lalama, Cj Adams, Iris Qu, Yifat Amir, Samer Chucri, Pierce Vollucci, Fabio Soldo, Dina Bseiso, Sarah Scodel, et al. 2023. Large Language Models for User Interest Journeys. arXiv preprint arXiv:2305.15498 (2023).
  • Deldjoo et al. (2021) Yashar Deldjoo, Vito Walter Anelli, Hamed Zamani, Alejandro Bellogin, and Tommaso Di Noia. 2021. A flexible framework for evaluating user and item fairness in recommender systems. User Modeling and User-Adapted Interaction (2021), 1–55.
  • Epasto and Perozzi (2019) Alessandro Epasto and Bryan Perozzi. 2019. Is a single embedding enough? learning node representations that capture multiple social contexts. In The world wide web conference. 394–404.
  • Fan et al. (2019) Wenqi Fan, Yao Ma, Qing Li, Yuan He, Eric Zhao, Jiliang Tang, and Dawei Yin. 2019. Graph neural networks for social recommendation. In The world wide web conference. 417–426.
  • Farnadi et al. (2018) Golnoosh Farnadi, Pigi Kouki, Spencer K Thompson, Sriram Srinivasan, and Lise Getoor. 2018. A fairness-aware hybrid recommender system. arXiv preprint arXiv:1809.09030 (2018).
  • Fu et al. (2020) Zuohui Fu, Yikun Xian, Ruoyuan Gao, Jieyu Zhao, Qiaoying Huang, Yingqiang Ge, Shuyuan Xu, Shijie Geng, Chirag Shah, Yongfeng Zhang, et al. 2020. Fairness-aware explainable recommendation over knowledge graphs. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. 69–78.
  • Gorantla et al. (2021) Sruthi Gorantla, Amit Deshpande, and Anand Louis. 2021. On the problem of underranking in group-fair ranking. In International Conference on Machine Learning. PMLR, 3777–3787.
  • Hashimoto et al. (2018) Tatsunori Hashimoto, Megha Srivastava, Hongseok Namkoong, and Percy Liang. 2018. Fairness without demographics in repeated loss minimization. In International Conference on Machine Learning. PMLR, 1929–1938.
  • He et al. (2020) Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yongdong Zhang, and Meng Wang. 2020. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval. 639–648.
  • Jin et al. (2024) Hongye Jin, Xiaotian Han, Jingfeng Yang, Zhimeng Jiang, Zirui Liu, Chia-Yuan Chang, Huiyuan Chen, and Xia Hu. 2024. LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning. arXiv preprint arXiv:2401.01325 (2024).
  • Lahoti et al. (2020) Preethi Lahoti, Alex Beutel, Jilin Chen, Kang Lee, Flavien Prost, Nithum Thain, Xuezhi Wang, and Ed Chi. 2020. Fairness without demographics through adversarially reweighted learning. Advances in neural information processing systems 33 (2020), 728–740.
  • Li et al. (2019) Chao Li, Zhiyuan Liu, Mengmeng Wu, Yuchi Xu, Huan Zhao, Pipei Huang, Guoliang Kang, Qiwei Chen, Wei Li, and Dik Lun Lee. 2019. Multi-interest network with dynamic routing for recommendation at Tmall. In Proceedings of the 28th ACM international conference on information and knowledge management. 2615–2623.
  • Li et al. (2021) Yunqi Li, Hanxiong Chen, Zuohui Fu, Yingqiang Ge, and Yongfeng Zhang. 2021. User-oriented fairness in recommendation. In Proceedings of the Web Conference 2021. 624–632.
  • Li et al. (2022) Yunqi Li, Hanxiong Chen, Shuyuan Xu, Yingqiang Ge, Juntao Tan, Shuchang Liu, and Yongfeng Zhang. 2022. Fairness in recommendation: A survey. arXiv preprint arXiv:2205.13619 (2022).
  • Liu et al. (2010) Jiahui Liu, Peter Dolan, and Elin Rønby Pedersen. 2010. Personalized news recommendation based on click behavior. In Proceedings of the 15th international conference on Intelligent user interfaces. 31–40.
  • MacQueen et al. (1967) James MacQueen et al. 1967. Some methods for classification and analysis of multivariate observations. In Proceedings of the fifth Berkeley symposium on mathematical statistics and probability, Vol. 1. Oakland, CA, USA, 281–297.
  • Pal et al. (2020) Aditya Pal, Chantat Eksombatchai, Yitong Zhou, Bo Zhao, Charles Rosenberg, and Jure Leskovec. 2020. Pinnersage: Multi-modal user embedding framework for recommendations at pinterest. In SIGKDD. 2311–2320.
  • Park et al. (2020) Chanyoung Park, Carl Yang, Qi Zhu, Donghyun Kim, Hwanjo Yu, and Jiawei Han. 2020. Unsupervised differentiable multi-aspect network embedding. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 1435–1445.
  • Rahmani et al. (2022) Hossein A Rahmani, Yashar Deldjoo, Ali Tourani, and Mohammadmehdi Naghiaei. 2022. The unfairness of active users and popularity bias in point-of-interest recommendation. In International Workshop on Algorithmic Bias in Search and Recommendation. Springer, 56–68.
  • Rendle et al. (2012) Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2012. BPR: Bayesian personalized ranking from implicit feedback. arXiv preprint arXiv:1205.2618 (2012).
  • Shi et al. (2022) Hui Shi, Yupeng Gu, Yitong Zhou, Bo Zhao, Sicun Gao, and Jishen Zhao. 2022. Every Preference Changes Differently: Neural Multi-Interest Preference Model with Temporal Dynamics for Recommendation. arXiv (2022).
  • Simpson (1949) Edward H Simpson. 1949. Measurement of diversity. nature 163, 4148 (1949), 688–688.
  • Thorndike (1953) Robert L Thorndike. 1953. Who belongs in the family? Psychometrika 18, 4 (1953), 267–276.
  • Wang et al. (2022a) Chenyang Wang, Yuanqing Yu, Weizhi Ma, Min Zhang, Chong Chen, Yiqun Liu, and Shaoping Ma. 2022a. Towards representation alignment and uniformity in collaborative filtering. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 1816–1825.
  • Wang and Isola (2020) Tongzhou Wang and Phillip Isola. 2020. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In International Conference on Machine Learning. PMLR, 9929–9939.
  • Wang et al. (2023a) Yifan Wang, Weizhi Ma, Min Zhang, Yiqun Liu, and Shaoping Ma. 2023a. A survey on the fairness of recommender systems. ACM Transactions on Information Systems 41, 3 (2023), 1–43.
  • Wang et al. (2022b) Yu Wang, Yuying Zhao, Yushun Dong, Huiyuan Chen, Jundong Li, and Tyler Derr. 2022b. Improving fairness in graph neural networks via mitigating sensitive attribute leakage. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.
  • Wang et al. (2023b) Yu Wang, Yuying Zhao, Yi Zhang, and Tyler Derr. 2023b. Collaboration-Aware Graph Convolutional Network for Recommender Systems. In Proceedings of the ACM Web Conference 2023. 91–101.
  • Ward Jr (1963) Joe H Ward Jr. 1963. Hierarchical grouping to optimize an objective function. Journal of the American statistical association 58, 301 (1963), 236–244.
  • Wu et al. (2022) Chuhan Wu, Fangzhao Wu, Tao Qi, and Yongfeng Huang. 2022. Are Big Recommendation Models Fair to Cold Users? arXiv preprint arXiv:2202.13607 (2022).
  • Wu et al. (2021) Chuhan Wu, Fangzhao Wu, Xiting Wang, Yongfeng Huang, and Xing Xie. 2021. Fairness-aware news recommendation with decomposed adversarial learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 4462–4469.
  • Yan et al. (2023a) Yuchen Yan, Yuzhong Chen, Huiyuan Chen, Minghua Xu, Mahashweta Das, Hao Yang, and Hanghang Tong. 2023a. From Trainable Negative Depth to Edge Heterophily in Graphs. In Thirty-seventh Conference on Neural Information Processing Systems.
  • Yan et al. (2023b) Yuchen Yan, Baoyu Jing, Lihui Liu, Ruijie Wang, Jinning Li, Tarek Abdelzaher, and Hanghang Tong. 2023b. Reconciling Competing Sampling Strategies of Network Embedding. In Thirty-seventh Conference on Neural Information Processing Systems.
  • Yan et al. (2021) Yuchen Yan, Si Zhang, and Hanghang Tong. 2021. Bright: A bridging algorithm for network alignment. In Proceedings of the Web Conference 2021.
  • Yan et al. (2022) Yuchen Yan, Qinghai Zhou, Jinning Li, Tarek Abdelzaher, and Hanghang Tong. 2022. Dissecting cross-layer dependency inference on multi-layered inter-dependent networks. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management.
  • Zhang et al. (2022) Shengyu Zhang, Lingxiao Yang, Dong Yao, Yujie Lu, Fuli Feng, Zhou Zhao, Tat-Seng Chua, and Fei Wu. 2022. Re4: Learning to Re-contrast, Re-attend, Re-construct for Multi-interest Recommendation. In WebConf. 2216–2226.
  • Zhao et al. (2023) Yuying Zhao, Yu Wang, Yunchao Liu, Xueqi Cheng, Charu Aggarwal, and Tyler Derr. 2023. Fairness and Diversity in Recommender Systems: A Survey. arXiv preprint arXiv:2307.04644 (2023).
  • Zhao et al. (2024) Yuying Zhao, Yu Wang, Yi Zhang, Pamela Wisniewski, Charu Aggarwal, and Tyler Derr. 2024. Leveraging Opposite Gender Interaction Ratio as a Path towards Fairness in Online Dating Recommendations Based on User Sexual Orientation. In Proceedings of the AAAI Conference on Artificial Intelligence.
  • Zhu et al. (2018) Ziwei Zhu, Xia Hu, and James Caverlee. 2018. Fairness-aware tensor-based recommendation. In Proceedings of the 27th ACM international conference on information and knowledge management. 1153–1162.
Table 6. Group interest diversity in various datasets.
Interest Diversity G1G_{1} G2G_{2} G3G_{3} G4G_{4} G5G_{5}
ml-1m-train 0.08±\pm0.00 0.32±\pm0.00 0.51±\pm0.00 0.65±\pm0.00 0.83±\pm0.00
ml-1m-test 0.18±\pm 0.01 0.39±\pm0.01 0.50±\pm0.00 0.57±\pm0.01 0.55±\pm0.10
ml-1m-train+test 0.11±\pm0.00 0.33±\pm0.00 0.51±\pm0.00 0.63±\pm0.00 0.76±\pm0.00
epinion-train 0.05±\pm0.00 0.31±\pm0.00 0.52±\pm0.00 0.72±\pm0.00 0.87±\pm0.00
epinion-test 0.21±\pm0.01 0.44±\pm0.02 0.57±\pm0.00 0.72±\pm0.01 0.82±\pm0.00
epinion-train+test 0.09±\pm0.00 0.34±\pm0.00 0.54±\pm0.00 0.72±\pm0.00 0.86±\pm0.00
cosmetics-train 0.04±\pm0.00 0.31±\pm0.00 0.53±\pm0.00 0.72±\pm0.00 0.90±\pm0.00
cosmetics-test 0.16±\pm0.00 0.42±\pm0.01 0.58±\pm0.01 0.73±\pm0.00 0.87±\pm0.00
cosmetics-train+test 0.07±\pm0.00 0.33±\pm0.00 0.54±\pm0.00 0.72±\pm0.00 0.89±\pm0.00
anime-train 0.12±\pm0.00 0.32±\pm0.00 0.48±\pm0.00 0.66±\pm0.00 0.82±\pm0.01
anime-test 0.27±\pm0.01 0.35±\pm0.00 0.42±\pm0.00 0.45±\pm0.00 0.45±\pm0.02
anime-train+test 0.16±\pm0.00 0.32±\pm0.00 0.46±\pm0.00 0.60±\pm0.00 0.73±\pm0.01

Appendix A Group-level Embedding Alignment

In this section, we present the group embedding alignment on LightGCN and CAGCN on the other three datasets in Fig. 9. Generally, the trend is similar to the one in Fig. 3. When interest diversity increases, the embedding alignment shows a poor performance. The trend when using Demb{}_{\text{emb}} as diversity metric is more consistent.

Refer to caption
Figure 9. Group-level embedding alignment on epinion, cosmetics, and anime datasets based on LightGCN and CAGCN.

Appendix B Interest Shift

Table 7. Correlation between train and test dataset.
Correlation metric (train / test) ml-1m epinion cosmetics anime
Pearson Correlation 0.9376 0.9981 0.9986 0.9492
Spearman Correlation Coefficient 0.9000 1.000 1.000 0.9000
Kendall Correlation Coefficient 0.8000 1.000 1.000 0.8000
Table 8. Correlation between train and train+test dataset.
Correlation metric (train / train+test) ml-1m epinion cosmetics anime
Pearson Correlation 0.9990 0.9999 0.9999 0.9997
Spearman Correlation Coefficient 1.000 1.000 1.000 1.000
Kendall Correlation Coefficient 1.000 1.000 1.000 1.000
Table 9. The number of parameters in backbones and the fairness approaches.
Backbone Method ml-1m epinion cosmetics anime dataset(#user\#\text{user}, #item\#\text{item})
LightGCN/CAGCN Vanilla 256064 737248 2609536 1428032 (#user+#item)d(\#\text{user}+\#\text{item})\cdot d
DRO +0 +0 +0 +0 +0
ARL +33 +33 +33 +33 +(d+1)
Multi +1536 +1536 +1536 +1536 +(#hops#interestd)+(\#\text{hops}\cdot\#\text{interest}\cdot d)

To investigate weather the unfairness is due to the disparity in interest diversity or interest shift. We conducted additional experiments to measure the correlation among interest diversity in the train and test datasets. We adopt two settings where we (1) compare the interest diversity between train dataset and test dataset; (2) compare the interest diversity between train dataset and the dataset including both train and test datasets. Specifically, we calculate the interest diversity for each user in the train and test (or train+test) datasets respectively, obtain an average score for each group, and then use different correlation metrics to obtain the correlation scores.

We reported the mean and standard deviation for group interest diversity for all datasets in Table 6, which shows that interest diversity in train and test (or train+test) datasets at a group level has a similar pattern. We also reported their correlation under various correlation metrics in Tables 7 and 8. The high correlations between the train and test datasets indicate that at the group level, there are no major interest diversity shifts. The correlation between train and train+test is even higher since including the train dataset makes the difference in interactions smaller. We would like to note however these high correlations are likely related to the dataset splitting for obtaining train/val/test datasets, which was randomly split following usual convention. In the future, we hope to explore the combination of multi-interest, continual learning, and interest diversity shifts with temporal splits, as we believe this will require dedicated methods and more detailed initial analysis to eventually tackle the potential interest diversity shifts of some users.

Appendix C Motivation of item side

The virtual embeddings of items function similarly to those of users. Different users have different interests. Accordingly, they interact with items based on different interests. The virtual embeddings of items capture the aspects related to those interests. For instance, if a user purchased a painting of a basketball player due to an interest in sports, then the aspect related to sports rather than art aspect will be extracted from this item (i.e., virtual item embedding) for the aggregation. If we only use a single embedding for the item, it then becomes challenging to distinguish/extract these different aspects and irrelevant/noisy aspect information will be aggregated. Therefore, the virtual embeddings of items facilitate learning user virtual interest embeddings. Additionally, such virtual embeddings can be further utilized for more fine-grained recommendations (i.e., interest-aware recommendations where explicit interest is given rather than the current implicit interest encoded in user embeddings), which we leave as future work.

This is also a design choice based on the observation that most RS have a symmetric architecture (e.g., the symmetric proprieties of the user-item bipartite graph) where users and items are treated in the same way during the computation. The symmetric style has the following benefits (1) it has a unified logic for processing users and items, which is easier for implementation and understanding; (2) it facilitates the matrix computation since user and item embeddings can be stacked together, which would be efficient.

Appendix D Baseline descriptions and implementation details

The descriptions for the compared methods are as follows:

  • LightGCN (He et al., 2020) is a GNN-based method that aggregates high-order neighborhood information and simplifies traditional GCN by removing the linear transformation and nonlinear activation.

  • CAGCN (Wang et al., 2023b) is a fusion model of LightGCN (He et al., 2020) and Collaboration-Aware Graph Convolutional Network (CAGCN) (Wang et al., 2023b). It analyzes how message-passing captures collaborative filtering (CF) effect and pre-computes a topological metric, Common Interacted Ratio (CIR), for collaboration-aware propagation.

  • DRO (Hashimoto et al., 2018) is a group-agnostic optimization approach that aims to improve the performance of the worst-case instances via distributionally robust learning.

  • ARL (Lahoti et al., 2020) is a group-agnostic optimization approach that leverages an adversary module to automatically adjust the weight in the training loss so that instances with higher loss will be assigned higher weights.

  • Multi is the multi-interest framework proposed in this paper. It learns multiple interest embeddings to represent each user to mitigate the performance gap among user groups.

For all methods, we use Adam optimizer for training and set the learning rate to 0.001, batch size to 2048, L2 coefficient to 0.001, and embedding dimension to 32. We early stop the training process when the best validation score remains unchanged for 25 epochs. Trend coefficient in CAGCN is set to 1.0. Temperature in the Softmax function is set to 2.0. The model hyperparameters are selected based on the best recall value during validation. For each model, we tune the number of hops within {1,2,3}\{1,2,3\}. Additionally, for DRO-based model, we tune the hyperparameter η\eta within {0.0,0.2,0.4,0.6,0.8,1.0}\{0.0,0.2,0.4,0.6,0.8,1.0\}. For our model, we tune interest number within {2,4,8,16}\{2,4,8,16\}. We run the experiments three times and report the average results. The best hyperparameters for each model are reported in Appendix F. When applied on CAGCN backbone, the aggregation weights in Eq.(7) is substituted with the pre-computed topological-based weights introduced in work (Wang et al., 2023b).

Appendix E Number of Parameters

We report the parameter numbers in Table 9 where the + means extra parameters compared with the backbone and dd is the embedding dimension. The number of parameters in LightGCN and CAGCN backbones are the same which equals to (#user+#item)d(\#\text{user}+\#\text{item})\cdot d where dd is the embedding dimension (d=32d=32 in our paper). DRO does not introduce new parameters so its number equals the one of backbone. ARL has an adversarial module in which we adopt the same one-layer architecture as in the original paper. Its parameters include the transformation from dd to 1. The number equals d+1d+1 where 1 is the bias. For our model, the extra parameters are the global embeddings 𝐰\mathbf{w} in Eq.(6) whose shape is (#hops,#interest,d)(\#\text{hops},\#\text{interest},d). Its size equals (#hops#interestd)(\#\text{hops}\cdot\#\text{interest}\cdot d). As claimed in the paper, we tune the number of hops within {1,2,3}\{1,2,3\} and interest number within {2,4,8,16}\{2,4,8,16\}. We report the maximum number among all hyperparameters (i.e., 3 hops and 16 interests). For all models, the number of extra parameters compared with the backbones is negligible.

As discussed in related works, traditional multi-interest frameworks that directly learn multiple embeddings will introduce a large number of additional parameters. The small number of extra parameters in our model is owing to the design of virtual embedding. We proposed the globally shared parameter 𝐰\mathbf{w} and then computed the virtual embeddings based on the original embedding and shared embedding, largely reducing the parameter size.

Appendix F Best Hyperparameters

For each model, we tune the number of hops within {1,2,3}\{1,2,3\}. Additionally, for DRO-based model, we tune the hyperparameter η\eta within {0.0,0.2,0.4,0.6,0.8,1.0}\{0.0,0.2,0.4,0.6,0.8,1.0\}. For our model, we tune interest number within {2,4,8,16}\{2,4,8,16\}.

Best hyperparameters for each model on three seeds are as follows (the order of datasets is ml-1m, epinion, cosmetics, and anime):

(1) LightGCN as backbone:

  • LightGCN (number of hops): [3,3,2], [3,3,3], [2,2,2], [3,3,3].

  • DRO (number of hops): [2,2,2], [3,3,3], [2,2,2], [2,2,2], [2,2,2];
    (η\eta): [0.6,0.6,0.6], [0.6,0.6,0.6], [0.0,0.0,0.0], [0.6,0.6,0.6].

  • ARL (number of hops): [2,2,2], [2,3,3], [3,2,2], [2,3,2].

  • Multi (number of hops): [2,1,1], [2,2,2], [2,2,2], [1,1,1];
    (number of interests): [16,8,4], [8,4,8], [4,16,16], [2,4,2].

(2) CAGCN as backbone:

  • CAGCN (number of hops): [3,3,3], [3,3,3], [3,3,3], [1,1,1].

  • DRO (number of hops): [3,3,3], [3,3,3], [3,3,3], [1,2,1];
    (η\eta): [0.6,0.6,0.6], [0.0,0.0,0.4], [0.0,0.0,0.0], [0.6,0.6,0.6].

  • ARL (number of hops): [3,3,2], [3,3,3], [3,3,3], [2,2,2].

  • Multi (number of hops): [1,1,1], [2,2,2], [2,2,2], [1,1,1];
    (number of interests): [16,8,4], [16,8,2], [4,16,8], [8,2,2].

Table 10. Performance (NDCG) on LightGCN backbone (The best is highlighted in bold and the runner-up is underlined).
Method ml-1m epinion cosmetics anime
NDCG\uparrow Unfairness\downarrow NDCG\uparrow Unfairness\downarrow NDCG\uparrow Unfairness\downarrow NDCG\uparrow Unfairness\downarrow Avg Rank\downarrow
LightGCN 0.2335 0.0133/0.0307 0.0462 0.0166/0.0153 0.1154 0.0643/0.1021 0.2594 0.0154/0.0513 2.33
DRO-LightGCN 0.2368 0.0139/0.0292 0.0473 0.0192/0.0188 0.1158 0.0670/0.1076 0.2553 0.0141/0.0514 2.75
ARL-LightGCN 0.2258 0.0138/0.0324 0.0438 0.0167/0.0178 0.1063 0.0607/0.0897 0.2466 0.0156/0.0483 3.00
Multi-LightGCN 0.2363 0.0137/0.0499 0.0464 0.0184/0.0093 0.1373 0.0592/0.0756 0.2852 0.0176/0.0471 1.92
Table 11. Performance (NDCG) on CAGCN backbone (The best is highlighted in bold and the runner-up is underlined).
Method ml-1m epinion cosmetics anime
NDCG\uparrow Unfairness\downarrow NDCG\uparrow Unfairness\downarrow NDCG\uparrow Unfairness\downarrow NDCG\uparrow Unfairness\downarrow Avg Rank\downarrow
CAGCN 0.2382 0.0138/0.0249 0.0492 0.0196/0.0143 0.1288 0.0692/0.1027 0.2619 0.0164/0.0521 2.33
DRO-CAGCN 0.2418 0.0144/0.0246 0.0483 0.0197/0.0148 0.1296 0.0708/0.1025 0.2613 0.0161/0.0554 2.67
ARL-CAGCN 0.2295 0.0126/0.0318 0.0471 0.0183/0.0180 0.1218 0.0640/0.0899 0.2591 0.0185/0.0556 3.00
Multi-CAGCN 0.2352 0.0132/0.0592 0.0480 0.0192/0.0107 0.1472 0.0663/0.0704 0.2875 0.0168/0.0446 2.00

Appendix G Fairness and Utility Trade-off

In this section, we report another utility performance NDCG and its corresponding fairness metric (i.e., the standard deviation of group NDCG performance) in Table 10 and Table 11 based on two backbones. Note that the models are the same as in Table 3 which are selected based on the best recall value. Similar to Table 3, our proposed multi-interest framework has the highest rank among all compared methods, indicating its effectiveness in balancing fairness and utility performance.

Appendix H Impact of Item and Category Number

To dive into the impact of item number and category number on fairness performance, we conducted additional experiments. The main idea is to sample datasets of different sizes (e.g., in terms of item number, category number) from the original large datasets, run the framework, and observe fairness performance to see the impact. For the convenience of sampling, we chose the largest dataset cosmetics to allow a wide range of sizes. We conducted two settings (1) Datasets with different numbers of items and categories (2) Datasets with different numbers of items and a similar number of categories. The results indicate that category number has a large impact on fairness and a dataset with a larger category number tends to have a larger unfairness.

H.1. Datasets with different numbers of items and categories

We randomly sampled a fixed number of items (5000/10000/15000) from the original item sets and formed a new interaction list by keeping (user, item) pairs where item is in the sampled set. After that, we performed the same preprocessing steps and trained our multi-framework based on LightGCN as described in the paper. Note that the item numbers will be smaller than the initial sample number due to the preprocessing steps. The dataset statistics and fairness results are reported where the dataset name is denoted as cosmetics-sample number. Table 12 shows that according to the sampling method, a larger item number would lead to a larger dataset with more edges, users, items, and categories. As the dataset becomes larger in all metrics, unfairness also increases.

Table 12. Fairness on sampled data based on item number.
Dataset #Edges #Users #Items #Category Fairness
cosmetics-5000 24574 4306 2134 254 0.0662
cosmetics-10000 82269 12354 5999 335 0.0784
cosmetics-15000 145752 19795 10089 366 0.0929

H.2. Datasets with different numbers of items and a similar number of categories

In this setting, we aim to control the number of categories to have a better understanding of the impact. Therefore, we designed a new sampling method where we kept a fixed proportion (0.5/0.7/0.9) of items in each category so that the category number would roughly remain the same. Similarly, we also do the preprocessing steps. The datasets are denoted as cosmetics-proportion. The results in Table 13 show that the category number of each dataset is the same under our control. The increased sample proportions result in a larger edge, user, and item number. However, the fairness performance is similar for these datasets.

Table 13. Fairness on sampled data based on proportion.
dataset #Edges #Users #Items #Category Fairness
cosmetics-0.5 257494 28942 13082 400 0.1160
cosmetics-0.7 386516 39082 18999 400 0.1084
cosmetics-0.9 519466 48499 25051 400 0.1178

Based on the results in these two settings, we hypothesize that the fairness performance is largely likely to be impacted by the category number.