This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\useunder

\ul

Knowledge Graph Context-Enhanced Diversified Recommendation

Xiaolong Liu [email protected] University of Illinois ChicagoChicagoUSA Liangwei Yang [email protected] University of Illinois ChicagoChicagoUSA Zhiwei Liu Salesforce AI ResearchPalo AltoUSA [email protected] Mingdai Yang [email protected] Chen Wang [email protected] University of Illinois ChicagoChicagoUSA Hao Peng School of Cyber Science and Technology, Beihang University,Beijing, China
Yunnan Key Laboratory of Artificial Intelligence, Kunming University of Science and Technology,Kunming, China
[email protected]
 and  Philip S. Yu University of Illinois ChicagoChicagoUSA [email protected]
(2024)
Abstract.

The field of Recommender Systems (RecSys) has been extensively studied to enhance accuracy by leveraging users’ historical interactions. Nonetheless, this persistent pursuit of accuracy frequently engenders diminished diversity, culminating in the well-recognized ”echo chamber” phenomenon. Diversified RecSys has emerged as a countermeasure, placing diversity on par with accuracy and garnering noteworthy attention from academic circles and industry practitioners. This research explores the diversified RecSys within the intricate context of knowledge graphs (KG). These KGs act as repositories of interconnected information concerning entities and items, offering a propitious avenue to amplify recommendation diversity through the incorporation of insightful contextual information. Our contributions include introducing an innovative metric, Entity Coverage, and Relation Coverage, which effectively quantifies diversity within the KG domain. Additionally, we introduce the Diversified Embedding Learning (DEL) module, meticulously designed to formulate user representations that possess an innate awareness of diversity. In tandem with this, we introduce a novel technique named Conditional Alignment and Uniformity (CAU). It adeptly encodes KG item embeddings while preserving contextual integrity. Collectively, our contributions signify a substantial stride towards augmenting the panorama of recommendation diversity within the KG-informed RecSys paradigms. We release the code at https://github.com/Xiaolong-Liu-bdsc/KG-diverse.

Recommender System, Knowledge Graph, Graph Neural Network
journalyear: 2024copyright: acmlicensedconference: Proceedings of the 17th ACM International Conference on Web Search and Data Mining; March 4–8, 2024; Merida, Mexicobooktitle: Proceedings of the 17th ACM International Conference on Web Search and Data Mining (WSDM ’24), March 4–8, 2024, Merida, Mexicoprice: 15.00doi: 10.1145/3616855.3635803isbn: 979-8-4007-0371-3/24/03ccs: Information systems Personalizationccs: Information systems Recommender systemsccs: Information systems Information retrieval diversity

1. Introduction

Recommender system (RecSys) has emerged as a vital solution for mitigating information overload in the contemporary landscape of extensive data (Mayer-Schönberger and Cukier, 2013). RecSys acquires the underlying preference by gaining insights from historical user-item interactions (Liu et al., 2023; Yang et al., 2023b). Subsequently, it offers personalized suggestions from a pool of potential items. Its pervasiveness is evident across multiple facets of our quotidian existence, encompassing domains such as news feeds (Wu et al., 2019), cinematic suggestions (Gomez-Uribe and Hunt, 2015), and e-commerce recommendation (Schafer et al., 2001).

Predominantly driven by the imperative of accuracy, commercial entities (Cheng et al., 2016; Gomez-Uribe and Hunt, 2015; Ying et al., 2018; Liu et al., 2022; Li et al., 2023; Wang et al., 2023b, 2022c, a) have endeavored to construct intricate algorithms. These algorithms predict the item that most likely engages each user, predicated upon their historical interactions. The mechanisms (Zhou et al., 2010) striving for a balance between accuracy and diversity have been developed to address this issue. This deficiency in diversity results in the echo chamber or filter bubble phenomenon (Ge et al., 2020), where users are repeatedly exposed to content aligning with their preferences, limiting their exposure to diverse options.

The paradigm of Diversified Recommender Systems (Diversified RecSys) has emerged to counteract the aforementioned constraints by actively prioritizing diversity during the recommendation process (Zheng et al., 2021; Yang et al., 2023c). These systems incite users to embark upon a broader exploration of choices. Consequently, users are enabled to encounter novel items and potentially discover items of interest that might have otherwise eluded their attention. Diversified RecSys has garnered escalating scholarly and industrial attention (Zheng et al., 2021; Yang et al., 2023c; Cheng et al., 2017; Chen et al., 2018a). Given the inherent conundrum between diversity and accuracy (Zhou et al., 2010), these systems are geared towards optimizing diversity while keeping the compromise on accuracy to a minimum, thus yielding an improved trade-off. Current methods gauge diversity through the lens of item categorization attributes, such as category coverage (Zheng et al., 2021; Yang et al., 2023c; Chen et al., 2018a). This approach primarily operates at a coarse-grained categorical level, thereby exhibiting limitations in comprehensively assessing diversity. It can not adequately discriminate between recommending items from within the same category. For example, recommending both (”iPhone 14” and ”Galaxy S23”) and (”iPhone 14” and ”iPhone 14 Pro”) yields identical diversity outcomes due to their shared electronic category, which falls short of a comprehensive measure for diversification.

To cope with this, we delve into the diversified RecSys under the knowledge graphs (KG) framework. Extensive literature has been published on recommender systems that incorporate KG (Wang et al., 2019; Yang et al., 2022; Wang et al., 2021a, 2023c). To diversify RecSys, a primary focal point of this investigation revolves around the nuanced integration of knowledge graphs. This entails leveraging the intricate details within the knowledge graph. To illustrate this, consider the previous example. In this context, a recommendation such as (”iPhone 14,” ”Galaxy S23”) acquires heightened diversification by enlisting entities encompassing not only the specific products but also broader contextual elements, including nations (”United States,” ”Korea”), manufacturers (”Apple,” ”Samsung”), and operating systems (”iOS,” ”Android”), all drawn from the encompassing knowledge graph. As a result, users are endowed with an augmented sphere of exposure to a diverse spectrum of entities, thereby enriching their engagement through the recommended selections.

Effective enhancement of recommendation diversity through incorporating KG information presents several pivotal challenges that merit careful consideration. (1) Actively prioritizing diversity during the recommendation process with KG: The intrinsic nature of KG structures poses a distinctive predicament where KG entities do not directly encapsulate user attributes. This inherent disconnect consequently engenders a deficiency in accurately representing user embeddings derived from KG information. Consequently, proficiently depicting and diversifying user embeddings using KG data constitutes a significant challenge. (2) Striking a Balance between Accuracy and Diversity: Higher accuracy often comes at the cost of reduced diversity. For all diversified RecSys, the inherent challenge lies in attaining elevated diversity without compromising recommendation accuracy. (3) Inadequate Characterization of Item KG Context Similarity: Present methodologies (Wang et al., 2019, 2021a, 2022a) exhibit limitations in the robust encoding of item KG context similarity. This deficiency impairs the capacity to comprehensively discern item similarity from the vantage point of the KG. The apt assessment of item similarity within the KG milieu bears particular significance when striving to impart diversity to recommendations.

Refer to caption
Figure 1. Diversified RecSys with knowledge graph.

In this study, we introduce a novel framework, denoted as KG-Diverse, aimed at addressing the aforementioned challenges. The cornerstone of our approach involves the formulation of two comprehensive metrics for gauging recommendation diversity within Knowledge Graphs (KGs): (1) Entity Coverage (EC) and (2) Relation Coverage (RC). These metrics assess the extent to which recommended items encapsulate a wide array of entities and relations within the KG context. To illustrate the practical implications of our proposed metrics, consider the scenario depicted in Figure 1, wherein a user’s viewing history encompasses the film The Avengers. A conventional RecSys would potentially propose Captain America 3 owing to the shared lead actors Chris Evans and Robert Downey Jr. with The Avengers, thus highlighting a thematic relevance. However, the introduction of diversified recommendations such as the film Dolittle, notable for its comedic genre and featuring actor Robert Downey Jr., or the movie Before We Go, notable for actor Chris Evans assuming both acting and directorial roles, serves to introduce elements of entity and relation diversification, respectively. Such nuanced recommendations contribute to a more holistic and enriching user experience. Consequently, we proceed to present an innovative module coined the Diversified Embedding Learning (DEL) module. This module is thoughtfully devised to engender personalized user representations imbued with an awareness of diversity, thereby fostering the augmentation of recommendation diversity without compromising recommendation accuracy. The underpinning of effective KG embedding learning is fundamentally pivotal for diversification. In response, we proffer a novel strategy labeled ”conditional alignment and uniformity.” This strategy is meticulously designed to uphold the integrity of KG embeddings and preserve the intrinsic similarity between items that share common entities. This dual-pronged approach ensures both the quality of KG embeddings and the coherence of item similarity within the KG context. Our contributions are summarized as follows:

  • To the best of our knowledge, this paper firstly introduces the novel measurement of recommendation diversity in KG through the use of Entity Coverage and Relation Coverage metrics.

  • We propose a simple yet effective Diversified Embedding Learning module to generate diversity-aware representations for users. Additionally, we design a novel technique, conditional alignment and uniformity, to effectively encode item embeddings in KG.

  • We evaluate the recommendation performance of our method on three public datasets. The extensive experimentation validates the effectiveness of our proposed model, showing the significance of each module in augmenting diversity while incurring only a negligible decrease in the accuracy of recommendations.

2. Methodology

The overall architecture of the proposed KG-Diverse is displayed in Figure 2, including knowledge graph propagation, diversified embedding learning and conditional alignment.

2.1. PROBLEM FORMULATION

We introduces the two data structures: user-item interaction graph, knowledge graph and the formulated problem.

User-Item interactions graph. We consider a collection of users denoted as 𝒰\mathcal{U}, and a set of items represented by \mathcal{I}, where mm and nn correspond to the number of users and items, respectively. We establish a user-item bipartite graph 𝒢={𝒱,}\mathcal{G}=\{\mathcal{V},\mathcal{E}\}, where 𝒱={𝒰,}\mathcal{V}=\{\mathcal{U},\mathcal{I}\} denotes the node set, and \mathcal{E} represents the edge set. An edge (u,i)(u,i)\in\mathcal{E} signifies that the user uu purchased the item ii before.

Knowledge graph. Knowledge graph is defined as 𝒢K={(h,r,t)}\mathcal{G}_{K}=\{(h,r,t)\}, where each triplet represents a relation rr\in\mathcal{R} from the head entity hh\in\mathcal{E} to the tail entity tt\in\mathcal{E}. Here, \mathcal{E} and \mathcal{R} represent the sets of entities and relations within KG, respectively. Notably, the entity set \mathcal{E} comprises both items (\mathcal{I}) and non-item entities (\\mathcal{E}\textbackslash\mathcal{I}).

Task Description. Given the 𝒢\mathcal{G} and 𝒢K\mathcal{G}_{K}, the goal of knowledge-aware recommendation is to recommend top kk items to each user. Moreover, the diversified recommendation task in KG encourages more entities (or relations) covered by recommended items.

Refer to caption
Figure 2. The framework of KG-Diverse. The leftmost figure is the collaborative knowledge graph.

2.2. Item entity-aware representation via Knowledge Graph propagation

KG is a multi-relational graph comprising entities and relations, denoted by a set of triplets, i.e., (h,r,t)(h,r,t). To facilitate understanding, we denote Ni={(r,v)|(i,r,v)𝒢K}N_{i}=\{(r,v)|(i,r,v)\in\mathcal{G}_{K}\} to represent the corresponding relations and entities connected to the item ii. A single entity may participate in multiple KG triplets, and it possesses the ability to adopt the linked entities as its attributes, thus revealing content similarity among items. Take an example in Figure 1, the star Chris Evans involves in multiple KG triplets, i.e., (actor, The Avengers), (actor, Before We Go), and (director, Before We Go). Then, we perform an aggregation technique to integrate the semantic information from NiN_{i} to generate the representation of item ii:

(1) ei(l)=f({(ei(l1),er,ev(l1))|(r,v)Ni}),\displaystyle\textbf{e}_{i}^{(l)}=f(\{(\textbf{e}_{i}^{(l-1)},\textbf{e}_{r},\textbf{e}_{v}^{(l-1)})|(r,v)\in N_{i}\}),

where ei(l1)\textbf{e}_{i}^{(l-1)} and ev(l1)\textbf{e}_{v}^{(l-1)} represent the embedding of item ii and entity vv after l1l-1 layers aggregation respectively, and er\textbf{e}_{r} is the embedding of relation rr. ei(l)\textbf{e}_{i}^{(l)} is the aggregated representation obtained by f()f(\cdot) which is the aggregation function that integrates information from ego-network NiN_{i} of item ii. Regarding the function f()f(\cdot), prior research efforts (Wang et al., 2019) predominantly focus on incorporating the relation rr solely into the weight factors during propagation. Nevertheless, it is vital to acknowledge that the combination of connected relation and entity (r,v)(r,v) holds distinct contextual implications. For instance, consider the star Chris Evans, who is linked twice to the film Before We Go in KG triplets, each time in a different role (director and actor). To address this contextual difference effectively, it becomes imperative to integrate relational contexts into the aggregator f()f(\cdot). It not only facilitates discerning the diverse effects of relational nuances but also augments the overall diversity of relations within the model. Therefore, we model the relation into the aggregation function f()f(\cdot) following (Wang et al., 2021a):

(2) ei(l)=1|Ni|(r,v)Nierev(l1),\displaystyle\textbf{e}_{i}^{(l)}=\frac{1}{|N_{i}|}\sum_{(r,v)\in N_{i}}\textbf{e}_{r}\odot\textbf{e}_{v}^{(l-1)},

where \odot is the element-wise product. The relational message erev(l1)\textbf{e}_{r}\odot\textbf{e}_{v}^{(l-1)} is propagated by modeling the relation rr as either a projection or rotation operator (Sun et al., 2019). This design allows the relational message to effectively unveil distinct meanings inherent in the triplets.

2.3. Diversified Embedding Learning

After obtaining item embedding representation ei(l)\textbf{e}_{i}^{(l)} from ll-th layer of knowledge graph, we leverage it to generate diversified user embedding. We denote 𝒩u={i|(u,i)}\mathcal{N}_{u}=\{i|(u,i)\in\mathcal{E}\} to represent the uu’s interacted items. Then, we formulate the temporary user representation by applying mean pooling on the interacted items:

(3) eutemp(l)=1|𝒩u|i𝒩uei(l).\displaystyle\textbf{e}^{(l)}_{u_{\text{temp}}}=\frac{1}{|\mathcal{N}_{u}|}\sum_{i\in\mathcal{N}_{u}}\textbf{e}_{i}^{(l)}.

In this way, the user representation is expressed by the representation of adopted items. The motivation of conventional RecSys revolves around aligning the representations of users and their purchased items. For example, in the left part of Figure 2, the user uu interacted with i1i_{1}, i2i_{2}, and i3i_{3}. A common approach involves driving the vector representation of user uu to approximate the vectors of items i1i_{1} and i2i_{2}, as these two items display high similarity due to their overlapping connections on the KG. However, it localizes the representation of uu in the surrounding area of i1i_{1} and i2i_{2} with their connected entities in KG (i.e., v1v_{1}, v2v_{2}, and v3v_{3}), which makes it hard to be exposed to the diverse items and entities around i3i_{3} (i.e., i4i_{4} and v5v_{5}). Hence, we devise the user diversified embedding learning layer with the intent of liberating the user representation from localization constraints and fostering diversity. The temporary user representation eutemp(l)\textbf{e}^{(l)}_{u_{\text{temp}}}, where temp is short for temporary, is utilized to measure the dissimilarities with ei(l)\textbf{e}_{i}^{(l)} in the ll-th layer:

(4) di=eutemp(l)ei(l)2,i𝒩u,\displaystyle d_{i}=\|\textbf{e}^{(l)}_{u_{\text{temp}}}-\textbf{e}_{i}^{(l)}\|_{2},\indent i\in\mathcal{N}_{u},
(5) ai=exp(di)j𝒩uexp(dj).\displaystyle a_{i}=\frac{\text{exp}(d_{i})}{\sum_{j\in\mathcal{N}_{u}}\text{exp}(d_{j})}.

Here, we utilize the Euclidean distance to compute dissimilarity, as it serves as a reliable measure to gauge the level of diversity. A higher distance value implies a greater potential for diversity in the resulting outcomes. The abundance of similar items may confine the user representation, resulting in a resemblance to these item representations. To counteract this, we mitigate the issue while preserving the intrinsic preference by generating diversity-aware embedding for the user through the following step:

(6) eu(l)=i𝒩uaiei(l).\displaystyle\textbf{e}_{u}^{(l)}=\sum_{i\in\mathcal{N}_{u}}a_{i}\textbf{e}_{i}^{(l)}.

Consequently, the user representation is emancipated from the constraints of localization, which often result from an abundance of similar items. For instance, eu\textbf{e}_{u} will be compelled to distance itself from ei\textbf{e}_{i} and e2\textbf{e}_{2}, while drawing nearer to e3\textbf{e}_{3}, thereby facilitating access to novel entities and items (e.g., i4i_{4} and v5v_{5}). Moreover, stacking multiple layers has been empirically demonstrated to effectively exploit high-order connectivity (Wang et al., 2019, 2021a). Therefore, the eu(l)\textbf{e}_{u}^{(l)} can capture diverse information from ll-hop neighbors of linked items, thereby enhancing exposure to diverse entities explicitly. We further sum up the embeddings obtained at each layer to form the final user and item representation, respectively:

(7) eu=eu(0)++eu(L),ei=ei(0)++ei(L),\displaystyle\textbf{e}_{u}=\textbf{e}_{u}^{(0)}+...+\textbf{e}_{u}^{(L)},\indent\textbf{e}_{i}=\textbf{e}_{i}^{(0)}+...+\textbf{e}_{i}^{(L)},

where LL is the number of knowledge graph propagation layers.

2.4. Conditional Alignment and Uniformity

To effectively encode KG in RecSys, we design a novel conditional alignment on KG entity embedding to preserve the semantic information between two similar items. Consider item i1i_{1} and item i2i_{2}, both sharing the entities v1v_{1} and v2v_{2}. Rather than a straightforward alignment (Wang et al., 2022b; Yang et al., 2023a), the item representations ei1\textbf{e}_{i_{1}} and ei2\textbf{e}_{i_{2}} should be aligned based on the shared information of entity representations ev1\textbf{e}_{v_{1}} and ev2\textbf{e}_{v_{2}}. We define conditional embedding as follows:

(8) ei1|voverlap=ei1evoverlap,ei2|v=ei2evoverlap,\displaystyle\textbf{e}_{i_{1}|v_{overlap}}=\textbf{e}_{i_{1}}\odot\textbf{e}_{v_{\text{overlap}}},\indent\textbf{e}_{i_{2}|v}=\textbf{e}_{i_{2}}\odot\textbf{e}_{v_{\text{overlap}}},

where Oi1,i2O_{i_{1},i_{2}} is the set of entities shared by item i1i_{1} and item i2i_{2} in KG and evoverlap=1|Oi1,i2|vOi1,i2ev\textbf{e}_{v_{\text{overlap}}}=\frac{1}{|O_{i_{1},i_{2}}|}\sum_{v\in O_{i_{1},i_{2}}}\textbf{e}_{v} is the average embedding of the overlapping entities between item i1i_{1} and item i2i_{2}. After modeling the overlapping entities as the projection or rotation on two items, we then perform the alignment loss:

(9) align=𝔼ei1|vei2|v2,\displaystyle\mathcal{L}_{\text{align}}=\mathbb{E}\|\textbf{e}_{i_{1}|v}-\textbf{e}_{i_{2}|v}\|^{2},

This design plays an essential role in pulling two similar items closer according to their connectivity on KG. Furthermore, the uniformity loss (Yu et al., 2022) is also applied to scatter the embedding uniformly:

(10) uniform=log𝔼i,ie2eiei2.\mathcal{L}_{\text{uniform}}=\log\mathbb{E}_{i,i^{{}^{\prime}}\in\mathcal{I}}\textbf{e}^{-2\|\textbf{e}_{i}-\textbf{e}_{i^{{}^{\prime}}}\|^{2}}.

2.5. Model Prediction

After performing LL propagation layers in KG, we further employ convolutional operation to capture collaborative signals from user-item interaction. Due to the effectiveness and simple architecture of LightGCN (He et al., 2020), where feature transformation and activation function are removed, we adopt its Light Graph Convolution layer (LGC) to encode the collaborative signals from user-item interactions:

(11) 𝐞u(k)=i𝒩u1|𝒩u||𝒩i|𝐞i(k1),𝐞i(k)=u𝒩i1|𝒩i||𝒩u|𝐞u(k1),\begin{gathered}\mathbf{e}_{u}^{(k)}=\sum_{i\in\mathcal{N}_{u}}\frac{1}{\sqrt{\left|\mathcal{N}_{u}\right|}\sqrt{\left|\mathcal{N}_{i}\right|}}\mathbf{e}_{i}^{(k-1)},\\ \mathbf{e}_{i}^{(k)}=\sum_{u\in\mathcal{N}_{i}}\frac{1}{\sqrt{\left|\mathcal{N}_{i}\right|}\sqrt{\left|\mathcal{N}_{u}\right|}}\mathbf{e}_{u}^{(k-1)},\end{gathered}

where 𝐞u(k)\mathbf{e}_{u}^{(k)} and 𝐞i(k)\mathbf{e}_{i}^{(k)} are the embedding of user uu and item ii at the kk-th LGC layer, respectively. Meanwhile, 𝐞u(0)\mathbf{e}_{u}^{(0)} and 𝐞i(0)\mathbf{e}_{i}^{(0)} are the embedding obtained by Eq. 7. Following KK layers of propagation, we integrate the embeddings acquired at each layer into the ultimate embeddings for both users and items:

(12) 𝐞u=1Kk=0K𝐞u(k),𝐞i=1Kk=0K𝐞i(k).\displaystyle\begin{aligned} \mathbf{e}_{u}=\frac{1}{K}\sum_{k=0}^{K}\mathbf{e}_{u}^{(k)},\mathbf{e}_{i}=\frac{1}{K}\sum_{k=0}^{K}\mathbf{e}_{i}^{(k)}.\end{aligned}

The embeddings produced by different LGC layers stem from distinct receptive fields, which also facilitates diversity from high-order neighbors. We choose the BPR loss (Rendle et al., 2012) for the optimization:

(13) CF=(u,i,j)𝒪logσ(𝐞uT𝐞i𝐞uT𝐞j),\displaystyle\mathcal{L}_{CF}=-\sum_{(u,i,j)\in\mathcal{O}}\text{log}\sigma(\mathbf{e}_{u}^{\text{T}}\mathbf{e}_{i}-\mathbf{e}_{u}^{\text{T}}\mathbf{e}_{j}),

where 𝒪={(u,i,j)|(u,i),(u,j)}\mathcal{O}=\{(u,i,j)|(u,i)\in\mathcal{E},(u,j)\in\mathcal{E}^{-}\}. \mathcal{E}^{-} is the set of unobserved interactions, and the jj is sampled from items that user uu has not interacted. Finally, the overall loss function is defined as:

(14) =CF+λ1align+λ2uniform+λ3Θ22,\mathcal{L}=\mathcal{L}_{\text{CF}}+\lambda_{1}\mathcal{L}_{\text{align}}+\lambda_{2}\mathcal{L}_{\text{uniform}}+\lambda_{3}{\left\|\Theta\right\|}_{2}^{2},

where λ1\lambda_{1} and λ2\lambda_{2} decide the weight of alignment and uniformity loss, respectively, and λ3\lambda_{3} denotes the parameter regularizing factor.

3. Experiments

This section aims to answer the 3 research questions (RQ).

  • RQ1: How does KG-Diverse perform compared to other state-of-the-art recommendation methods?

  • RQ2: Does every designed module play a role in KG-Diverse?

  • RQ3: What are the impacts of hyper-parameters: KG propagation layers LL, alignment weight λ1\lambda_{1}, and uniformity weight λ2\lambda_{2}?

3.1. Experimental Settings

3.1.1. Datasets

To evaluate the effectiveness of our KG-Diverse against other baseline models, we perform experiments on three benchmark datasets: Amazon-book 111http://jmcauley.ucsd.edu/data/amazon/, Last-fm 222https://grouplens.org/datasets/hetrec-2011, and Movielens 333https://grouplens.org/datasets/movielens/, which are widely used in previous works (Wang et al., 2019, 2021b, 2023c; He and McAuley, 2016). To ensure the quality of user-item interactions, the 10-core setting (preserve the users who have at least ten interactions) was adopted on all datasets as previous research. We partitioned the historical interactions of each user into three sets: training (80%), validation (10%), and test (10%). Table 1 presents detailed statistics for these datasets.

Table 1. Statistics of the Datasets.
Dataset Amazon-Book Last.FM Movielens
#Users 70,679 1,872 37,385
#Items 24,915 3,846 6,182
#Interactions 846,434 42,346 539,300
#Entities 113,487 9,366 24,536
#Relations 39 60 20
#Triplets 2,557,746 15,518 237,155

3.1.2. Baselines

We compare KG-Diverse with several representative methods as baselines:

(1) General Recommender Systems

  • MF (Koren et al., 2009) factorizes the user-item interaction matrix into user/item latent vectors via the BPR loss.

  • LightGCN (He et al., 2020) is the state-of-the-art GNN recommender system. It is a simplified GCN (Kipf and Welling, 2017) by removing the transformation function and non-linear activation.

  • DirectAU (Wang et al., 2022b) optimizes alignment and uniformity as the objective function to enhance collaborative filtering methods.

(2) Knowledge Graph for Recommendation

  • KGAT (Wang et al., 2019) utilizes an attention mechanism to discriminate the entity importance in KG during recursive propagation.

  • KGIN (Wang et al., 2021a) uncovers the users’ intents via the attentive combination of relations in KG and utilizes relational path-aware aggregation to refine user/item representations.

(3) Diversified Recommender Systems

  • DGCN (Zheng et al., 2021) is the GNN-based method for diversified recommendation with rebalanced neighbor and adversarial learning.

  • DGRec (Yang et al., 2023c) is the current state-of-the-art diversified RecSys model based on GNN. It designs the submodular function to select a diversified neighbors to enhance diversity.

3.1.3. Evaluation Metrics

In our experiments, we evaluate accuracy and diversity, respectively.

Accuracy. We use Recall@k (R@k) and NDCG@k (N@k) (Krichene and Rendle, 2020) with k ranges in {20, 40} to evaluate the performance of the top-k recommendation. The average metrics of all test users are reported.

Diversity. Some studies (Gan et al., 2020; Wu et al., 2022) employ the ILAD metric to quantify diversity based on dissimilarities among recommended items, yet its effectiveness heavily relies on the quality of learned embedding. For assessing the diversity of recommendations on KG, we establish two metrics on KG: (1) Entity Coverage (EC) and (2) Relation Coverage (RC). Entity/Relation Coverage refers to the number of entities/relations the recommended itemsets cover. We report top-20 and top-40 retrieval results to accord with accuracy.

3.2. Performance Comparison (RQ1)

Table 2. Performance Comparison. The best and runner-ups are marked in bold and underlined separately.
Datasets Methods Accuracy Diversity
R@20 R@40 N@20 N@40 EC@20 EC@40 RC@20 RC@40
Amazon- Book MF 0.0999 0.1431 0.0524 0.0634 63.1383 97.2867 20.2529 21.5524
LightGCN 0.1090 0.1647 0.0558 0.0699 64.1135 95.1836 19.9618 21.8544
DirectAU 0.1225 0.1716 0.0650 0.0776 48.0056 73.0553 18.0014 20.0823
KGAT 0.1245 0.1846 0.0620 0.0773 52.4172 79.2991 18.4723 20.5078
KGIN 0.1281 0.1817 0.0673 0.0809 50.2582 75.2091 18.3580 20.2204
DGCN 0.0805 0.1285 0.0407 0.0528 56.1095 84.3621 19.6098 21.5630
DGRec 0.0920 0.1414 0.0458 0.0584 65.2395 97.5764 20.1816 21.9992
KG-Diverse 0.1071 0.1621 0.0547 0.0687 67.8613 101.1123 20.5630 22.5242
Last.FM MF 0.3292 0.4395 0.1802 0.2056 68.2359 122.5158 11.5179 14.3109
LightGCN 0.3685 0.4719 0.2016 0.2257 67.0864 117.2707 10.6190 13.4332
DirectAU 0.3379 0.4349 0.1780 0.2005 60.4386 108.7641 8.8190 11.8342
KGAT 0.3334 0.4373 0.1818 0.2057 66.2712 118.7859 11.1397 14.0098
KGIN 0.3415 0.4537 0.1635 0.1895 66.1098 117.9679 11.8299 14.5141
DGCN 0.2430 0.3474 0.1242 0.1482 59.8146 114.2739 10.7978 13.8157
DGRec 0.2602 0.3618 0.1294 0.1529 60.5130 111.8712 11.1902 14.1957
KG-Diverse 0.3539 0.4716 0.1785 0.2059 87.2005 154.7543 13.9125 16.6293
Movielens MF 0.4065 0.5297 0.2187 0.2490 454.3015 805.9360 17.1941 17.3594
LightGCN 0.4157 0.5462 0.2236 0.2557 452.7703 803.3682 17.1809 17.3426
DirectAU 0.4077 0.5256 0.2257 0.2547 441.1399 790.9293 17.2679 17.4271
KGAT 0.4097 0.5340 0.2228 0.2533 444.9880 792.5169 17.2219 17.3757
KGIN 0.4374 0.5669 0.2386 0.2705 447.2437 797.7831 17.1909 17.3531
DGCN 0.3625 0.4941 0.1874 0.2198 444.5471 799.3097 17.2358 17.4053
DGRec 0.3676 0.4981 0.1917 0.2238 445.0291 798.6214 17.2310 17.3980
KG-Diverse 0.4132 0.5425 0.2241 0.2559 458.8848 815.3123 17.1993 17.3710

3.2.1. Parameter Setting

We implement KG-Diverse and all baselines on Pytorch with Adam (Kingma and Ba, 2015) as the optimizer. We train each model until there is no performance improvement on the validation set within 10 epochs. Following the setting of KGAT (Wang et al., 2019), we fix the embedding size to 64 for all models. Grid search is performed to tune the hyper-parameters for each method. The model parameters are initialized by Xavier initializer. For KGAT, the pre-trained MF embeddings of users and items are used for initialization.

Specifically, the learning rate and weight decay are tuned in {0.1, 0.05, 0.01, 0.005, 0.001} and {1e-4, 1e-5, 1e-6, 1e-7} respectively. The number of KG layers and LightGCN layers are searched with the range of {1, 2, 3, 4, 5, 6}. Finally, we tune the coefficients of alignment λ1\lambda_{1} and uniformity λ2\lambda_{2} among {0.1, 0.2, …, 1.0} with increment of 0.1.

We report the overall performance of all baselines on the three datasets in Table 2. We have the following observations:

  • KG-Diverse outperforms all the other baselines on diversity metric: Entity Coverage (EC) and Relation Coverage (RC) except the RC on Movielens dataset. However, all the performances on RC of Movielens dataset are almost the same, making models indistinguishable with each other regards RC.

  • Despite achieving the highest Entity Coverage, KG-Diverse shows comparable performance on Recall and NDCG metrics, particularly attaining the second-best results on Last.FM and exhibiting similarity to the second-best LightGCN on MovieLens dataset. This observation indicates that KG-Diverse effectively enhances diversity while incurring a minimal cost to accuracy, thereby striking a well-balanced accuracy-diversity trade-off.

  • DGRec is the state-of-the-art diversified RecSys that models diversity at the coarse-grained category information. Our model, by consistently outperforming DGRec across three datasets in all measures, especially on Last.FM, demonstrates the efficacy of modeling diversity at the fine-grained level of KG entities. It enables accurate recommendations and facilitates the provision of diverse and relevant information to a significant extent.

  • In contrast to general accuracy-based methods (MF, LightGCN, and DirectAU), KG-Diverse exhibits superior enhancements in diversity metrics while maintaining comparable performance in accuracy metrics. This observation serves as a testament to the model’s effective capability in encoding KG.

To be more specific, Figure 3 illustrates the accuracy-diversity trade-off comparison of all models on Last-FM datasets, where Recall@40 and EC@40 measure accuracy and diversity, respectively. It is evident from the plot that KG-Diverse occupies the most favorable position in the upper-right quadrant, demonstrating that KG-Diverse achieves the optimal accuracy-diversity trade-off. Our model KG-Diverse outperforms all other baselines in both accuracy and diversity metrics except LightGCN. When compared to LightGCN, KG-Diverse exhibits a substantial increase in diversity while making only a minor sacrifice in accuracy.

3.3. Ablation Study (RQ2)

Refer to caption
Figure 3. Accuracy-Diversity trade-off comparison. The upper-right model is better.

We explore the effects of each module in KG-Diverse on the Amazon-book dataset, including (1) KG propagation layer for item representation learning, (2) Relation encoding into item representation in Eq. 2, (3) Diversified Embedding Learning (DEL) module in generating user diversity-aware representations, and (4) Conditional Alignment and Uniformity (CAU) that regularizes KG embedding learning. The results are shown in Table 3:

  • After removing the KG propagation layer, the knowledge within entities and relations will not be incorporated to enrich the item representations, and there is a single layer to generate user diversified representation without KG. Even though there are 2.7% and 5.5 % improvements on R@20 and N@20, respectively, the performance on EC@20 decreased by 7.1%. It shows the effectiveness of learning entity-aware item representation via Eq. 2.

  • Upon removing relation encoding, the item representation in ll-th layer is obtained by averaging the embeddings of its connected entities in the previous layer. The results indicate significant decreases in all evaluation metrics, particularly a 5.9% drop in RC@20. It indicates the crucial role of relations in enriching item representations. The presence of relations in the encoding process allows the model to capture and leverage the semantic connections between items and entities, contributing to more accurate and diverse recommendations.

  • In the ll-th KG propagation layer, the user diversity-aware embedding eu(l)\textbf{e}_{u}^{(l)} in Eq. 6 is replaced by eutemp(l)\textbf{e}^{(l)}_{u_{\text{temp}}} in Eq. 3 if DEL is removed. The declines in all four metrics: R@20 decreased by 4.3%, N@20 decreased by 7.0%, EC@20 decreased by 3.4%, and RC@20 decreased by 2.1%, underscores our model’s advantage in enabling user representations to transcend localization constraints.

  • Without conditional alignment and uniformity, there are substantial drops in R@20, N@20, and EC@20 metrics. This underscores the crucial role played by the regularization on KG embedding learning, as it serves as the foundational prerequisite for KG application. We find that the similarity and distinctions among items in the KG significantly influence the overall framework.

Table 3. Ablation Study.
Method R@20 N@20 EC@20 RC@20
KG-Diverse 0.1071 0.0547 67.8613 20.5630
w/o KG propagation layer 0.1101 0.0577 63.3891 20.1389
w/o Relation encoding 0.0902 0.0452 66.4480 19.4177
w/o DEL 0.1027 0.0511 65.6006 20.3592
w/o CAU 0.0929 0.0456 64.7545 20.2670

3.4. Parameter Sensitivity (RQ3)

Additional exploratory experiments are conducted to comprehensively understand the KG-Diverse structure and investigate the underlying reasons for its effectiveness. We explore the effects of four essential hyper-parameters in KG-Diverse on Last.FM dataset, including the number of KG propagation layers LL, the coefficient λ1\lambda_{1} on alignment loss, and the weight decay λ2\lambda_{2} on uniformity loss.

3.4.1. Number of KG propagation layers LL

We vary the number of KG propagation layers in the range of {1, 2, 3, 4, 5, 6} to analyze its influence in our model. The two line plots reveal a clear trend: as the depth of KG propagation layers increases, the entity coverage steadily improves, reaching its peak at a depth of 5, and subsequently exhibits a decline. This trend signifies that the stacking of multiple KG layers aids item representations in capturing high-order connectivity among entities, thereby encoding a diverse range of entities within the item representations, as depicted in Equation 2. In contrast, the accuracy metric (Recall) exhibits a consistent downward trend as the depth LL increases. It may be attributed to the over-smooth problem (Liu et al., 2020), where the item embeddings become overly similar, negatively impacting the accuracy.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 4. Parameter sensitivity of LL, λ1\lambda_{1}, and λ2\lambda_{2}

3.4.2. Coefficient λ1\lambda_{1} on alignment loss

. The introduction of this coefficient aims to govern the weight of aligning item embeddings in the Knowledge Graph (KG). By adjusting the value of λ1\lambda_{1}, KG-Diverse becomes more attentive to aligning similar item representations, especially when two items share common entities in the KG. As demonstrated in Figure 4, the substantial enhancements observed in both Recall and Entity Coverage metrics serve as tangible evidence of the notable efficacy of conditional alignment in effectively encoding KG embeddings. As similar items are successfully aligned, KG-Diverse could easily retrieve the items that cover different and diverse entities.

3.4.3. Weight decay λ2\lambda_{2} on uniformity

. The incorporation of uniformity in KG-Diverse serves the purpose of dispersing the item representations learned from the KG, preventing them from becoming identical and promoting diversity among the representations. Our observations indicate that both accuracy and diversity exhibit an upward trend, achieving their optimal performance when the value of λ2\lambda_{2} is approximately 0.4. Nonetheless, as λ2\lambda_{2} increases further, the results for Recall and Entity Coverage witness a rapid decline. This analysis leads us to the conclusion that a larger weight assigned to uniformity would promote more uniform embedding dispersal and potentially deteriorate KG alignment. However, a lighter emphasis on uniformity can still yield beneficial effects on the encoding of KG embeddings to some extent.

3.5. Case Study

In this case study, we randomly selecte a user, u52279u_{52279}, from the Amazon-book dataset to compare the entity and relation coverages of two recommendation models: KGIN and KG-Diverse. The visualization of their performance comparison is illustrated in Figure 5. From both models, we retrieved the top-5 recommendations for u52279u_{52279}, with i4079,i4075,i1153,i3613i_{4079},i_{4075},i_{1153},i_{3613} being common items recommended by both models. To simplify the analysis, we focus on two distinct recommended items: i6583i_{6583} from KGIN and i1041i_{1041} from KG-Diverse. We have the following observations:

  • The item i6583i_{6583} with the name The Affair covers only a few entities and relations, namely NThe AffairN_{\text{The Affair}} = ((Genre, Noval), (Author, Lee Child)). Remarkably, the item i1041i_{1041} recommended by our model, KG-Diverse, also includes these two pairs of entities and relations. It exemplifies the strong relevance and effectiveness of KG-Diverse in providing relevant recommendations.

  • In addition to the strong relevance, the item i1041i_{1041} titled Echo Burning covers a significantly broader range of entities and relations in the KG compared to i6583i_{6583}. The entities and relations covered by i1041i_{1041} include (Genre, Suspense), (Genre, Fiction), (Subjects, Adventure), (Subjects, Texas). It underscores the substantial capacity of KG-Diverse in enhancing the diversity of recommendations. By recommending items like Echo Burning, which encompass a wide variety of entities and relations, KG-Diverse successfully introduces diversity and enriches the user experience by offering a more varied and engaging selection of recommendations.

Refer to caption
(a) KGIN
Refer to caption
(b) KG-Diverse
Figure 5. Real Example from Amazon-Book.

4. Related Works

4.1. Knowledge Graph for Recommendation

CKE (Zhang et al., 2016) model enriches the item representation by incorporating (1) structural embedding learned from items’ structural knowledge via TransR, (2) textual and visual representations extracted by applying two different auto-encoders on textual and visual knowledge separately. DKN (Wang et al., 2018), a news recommendation model, fuses semantic-level and knowledge-level information to news embedding by a multi-channel and word-entity-aligned knowledge-aware convolutional neural network (KCNN) (Kim, 2014). The user’s vector is aggregated from the attention-weighted sum of news he/she clicked. KGIN (Wang et al., 2021a) models a finer-grained level of each intent by a combination of KG relations and designs a relational path-aware aggregation scheme that integrates information from high-order connectivities. KGAT (Wang et al., 2019) is designed to update embeddings by recursive propagation to capture high-order connectivities and compute the attention weights of a node’s neighbors which aggregate information from high-order relations during propagation. KGCL (Yang et al., 2022) defines knowledge graph structure consistency score to identify items less sensitive to structure variation and tolerate noisy entities in KG. MetaKRec (Wang et al., 2022a) extracts KG triples into direct edges between items via prior knowledge and achieves improved performance with graph neural network.

While the mentioned works primarily concentrate on enhancing the accuracy of recommendations by leveraging KG, our work distinguishes itself by focusing on elevating the diversity of recommendations at a fine-grained level while ensuring accuracy.

4.2. Diversified Recommendation

Diversified recommendation technique is used to provide users with a diverse set of recommendations rather than a narrow set of highly similar items. The concept of diversified recommendation was first proposed by Ziegler et al. (Ziegler et al., 2005). They performed a greedy algorithm to retrieve the diversified top-K items. DUM (Ashkan et al., 2015) designs the submodular function to select items by a greedy method and balance the utility of items and diversity in the re-ranking procedure. Determinantal point process (DPP) (Chen et al., 2018b), a post-processing method, retrieves the set of diverse items depending on the largest determinant of the item similarity matrix with MAP inference. DGCN (Zheng et al., 2021) stands as the initial diversified recommendation approach based on GNNs. It adopts the inverse category frequency to select node neighbors for diverse aggregation and employs category-boosted negative sampling and adversarial learning to enhance item diversity within the embedding space. DGRec (Yang et al., 2023c) is the current state-of-the-art model that produces varied recommendations through an enhanced embedding generation process. The proposed modules (i.e., submodular neighbor selection, layer attention, and loss reweighting) collectively aim to boost diversity without compromising accuracy. DivKG (Gan et al., 2020) learns knowledge graph embeddings via TransH (Wang et al., 2014) and designs a DPP kernel matrix to ensure the trade-off between accuracy and diversity. EMDKG (Gan et al., 2021) encodes the diversity into item representations by an Item Diversity Learning module to reflect the semantic diversity from KG.

The previous works study diversity as the covered categories, which is measured at the coarse-grained level. This paper first investigates the fine-grained diversity with KG context information.

5. Conclusion

In this paper, we design a straightforward and effective framework named KG-Diverse to enhance the diversity of recommendations with the rich information from KG. It first adopts the KG propagation method to incorporate semantic information into item representations. Then, we propose a diversified embedding learning module to generate diversity-aware embedding for users. Finally, we design Conditional Alignment and Uniformity to effectively encode the KG to preserve the similarity between two items with shared entities. Furthermore, we introduce Entity Coverage (EC) and Relation Coverage (RC) to measure the diversity in KGs and evaluate the final performance. Extensive experiments on three public datasets demonstrate the effectiveness of KG-Diverse in balancing the accuracy and diversity trade-off.

Acknowledge

Hao Peng is supported by National Key R&D Program of China through grant 2022YFB3104700, NSFC through grants 62322202, U21B2027, 62002007, 61972186 and 62266028, Beijing Natural Science Foundation through grant 4222030, S&T Program of Hebei through grant 21340301D, Yunnan Provincial Major Science and Technology Special Plan Projects through grants 202302AD080003, 202202AD080003 and 202303AP140008, General Projects of Basic Research in Yunnan Province through grants 202301AS070047, 202301AT070471, and the Fundamental Research Funds for the Central Universities. This work is supported in part by NSF under grant III-2106758.

References

  • (1)
  • Ashkan et al. (2015) Azin Ashkan, Branislav Kveton, Shlomo Berkovsky, and Zheng Wen. 2015. Optimal Greedy Diversity for Recommendation. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, IJCAI 2015, Buenos Aires, Argentina, July 25-31, 2015, Qiang Yang and Michael J. Wooldridge (Eds.). AAAI Press, 1742–1748.
  • Chen et al. (2018a) Laming Chen, Guoxin Zhang, and Eric Zhou. 2018a. Fast greedy map inference for determinantal point process to improve recommendation diversity. Advances in Neural Information Processing Systems 31 (2018).
  • Chen et al. (2018b) Laming Chen, Guoxin Zhang, and Eric Zhou. 2018b. Fast Greedy MAP Inference for Determinantal Point Process to Improve Recommendation Diversity. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, December 3-8, 2018, Montréal, Canada, Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicolò Cesa-Bianchi, and Roman Garnett (Eds.). 5627–5638.
  • Cheng et al. (2016) Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, et al. 2016. Wide & deep learning for recommender systems. In Proceedings of the 1st workshop on deep learning for recommender systems. 7–10.
  • Cheng et al. (2017) Peizhe Cheng, Shuaiqiang Wang, Jun Ma, Jiankai Sun, and Hui Xiong. 2017. Learning to recommend accurate and diverse items. In Proceedings of the 26th international conference on World Wide Web. 183–192.
  • Gan et al. (2020) Lu Gan, Diana Nurbakova, Léa Laporte, and Sylvie Calabretto. 2020. Enhancing Recommendation Diversity using Determinantal Point Processes on Knowledge Graphs. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, China, July 25-30, 2020, Jimmy X. Huang, Yi Chang, Xueqi Cheng, Jaap Kamps, Vanessa Murdock, Ji-Rong Wen, and Yiqun Liu (Eds.). ACM, 2001–2004.
  • Gan et al. (2021) Lu Gan, Diana Nurbakova, Léa Laporte, and Sylvie Calabretto. 2021. EMDKG: Improving Accuracy-Diversity Trade-Off in Recommendation with EM-based Model and Knowledge Graph Embedding. In WI-IAT ’21: IEEE/WIC/ACM International Conference on Web Intelligence, Melbourne VIC Australia, December 14 - 17, 2021, Jing He, Rainer Unland, Eugene Santos Jr., Xiaohui Tao, Hemant Purohit, Willem-Jan van den Heuvel, John Yearwood, and Jie Cao (Eds.). ACM, 17–24.
  • Ge et al. (2020) Yingqiang Ge, Shuya Zhao, Honglu Zhou, Changhua Pei, Fei Sun, Wenwu Ou, and Yongfeng Zhang. 2020. Understanding echo chambers in e-commerce recommender systems. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval. 2261–2270.
  • Gomez-Uribe and Hunt (2015) Carlos A Gomez-Uribe and Neil Hunt. 2015. The netflix recommender system: Algorithms, business value, and innovation. ACM Transactions on Management Information Systems (TMIS) 6, 4 (2015), 1–19.
  • He and McAuley (2016) Ruining He and Julian McAuley. 2016. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In proceedings of the 25th international conference on world wide web. 507–517.
  • He et al. (2020) Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yong-Dong Zhang, and Meng Wang. 2020. LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, China, July 25-30, 2020, Jimmy X. Huang, Yi Chang, Xueqi Cheng, Jaap Kamps, Vanessa Murdock, Ji-Rong Wen, and Yiqun Liu (Eds.). ACM, 639–648.
  • Kim (2014) Yoon Kim. 2014. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, October 25-29, 2014, Doha, Qatar, A meeting of SIGDAT, a Special Interest Group of the ACL, Alessandro Moschitti, Bo Pang, and Walter Daelemans (Eds.). ACL, 1746–1751.
  • Kingma and Ba (2015) Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, Yoshua Bengio and Yann LeCun (Eds.).
  • Kipf and Welling (2017) Thomas N. Kipf and Max Welling. 2017. Semi-Supervised Classification with Graph Convolutional Networks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net.
  • Koren et al. (2009) Yehuda Koren, Robert M. Bell, and Chris Volinsky. 2009. Matrix Factorization Techniques for Recommender Systems. Computer 42, 8 (2009), 30–37.
  • Krichene and Rendle (2020) Walid Krichene and Steffen Rendle. 2020. On Sampled Metrics for Item Recommendation. In KDD ’20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, August 23-27, 2020, Rajesh Gupta, Yan Liu, Jiliang Tang, and B. Aditya Prakash (Eds.). ACM, 1748–1757.
  • Li et al. (2023) Xiaohan Li, Yuqing Liu, Zheng Liu, and Philip S. Yu. 2023. Time-aware Hyperbolic Graph Attention Network for Session-based Recommendation. CoRR abs/2301.03780 (2023).
  • Liu et al. (2020) Meng Liu, Hongyang Gao, and Shuiwang Ji. 2020. Towards Deeper Graph Neural Networks. In KDD ’20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, August 23-27, 2020, Rajesh Gupta, Yan Liu, Jiliang Tang, and B. Aditya Prakash (Eds.). ACM, 338–348.
  • Liu et al. (2023) Xiaolong Liu, Liangwei Yang, Zhiwei Liu, Xiaohan Li, Mingdai Yang, Chen Wang, and Philip S Yu. 2023. Group-Aware Interest Disentangled Dual-Training for Personalized Recommendation. arXiv preprint arXiv:2311.09577 (2023).
  • Liu et al. (2022) Zhuoran Liu, Leqi Zou, Xuan Zou, Caihua Wang, Biao Zhang, Da Tang, Bolin Zhu, Yijie Zhu, Peng Wu, Ke Wang, et al. 2022. Monolith: real time recommendation system with collisionless embedding table. arXiv preprint arXiv:2209.07663 (2022).
  • Mayer-Schönberger and Cukier (2013) Viktor Mayer-Schönberger and Kenneth Cukier. 2013. Big data: A revolution that will transform how we live, work, and think. Houghton Mifflin Harcourt.
  • Rendle et al. (2012) Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. 2012. BPR: Bayesian personalized ranking from implicit feedback. arXiv preprint arXiv:1205.2618 (2012).
  • Schafer et al. (2001) J Ben Schafer, Joseph A Konstan, and John Riedl. 2001. E-commerce recommendation applications. Data mining and knowledge discovery 5 (2001), 115–153.
  • Sun et al. (2019) Zhiqing Sun, Zhi-Hong Deng, Jian-Yun Nie, and Jian Tang. 2019. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net.
  • Wang et al. (2022b) Chenyang Wang, Yuanqing Yu, Weizhi Ma, Min Zhang, Chong Chen, Yiqun Liu, and Shaoping Ma. 2022b. Towards Representation Alignment and Uniformity in Collaborative Filtering. In KDD ’22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, August 14 - 18, 2022, Aidong Zhang and Huzefa Rangwala (Eds.). ACM, 1816–1825.
  • Wang et al. (2023c) Hao Wang, Yao Xu, Cheng Yang, Chuan Shi, Xin Li, Ning Guo, and Zhiyuan Liu. 2023c. Knowledge-Adaptive Contrastive Learning for Recommendation. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, WSDM 2023, Singapore, 27 February 2023 - 3 March 2023, Tat-Seng Chua, Hady W. Lauw, Luo Si, Evimaria Terzi, and Panayiotis Tsaparas (Eds.). ACM, 535–543.
  • Wang et al. (2018) Hongwei Wang, Fuzheng Zhang, Xing Xie, and Minyi Guo. 2018. DKN: Deep Knowledge-Aware Network for News Recommendation. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, WWW 2018, Lyon, France, April 23-27, 2018, Pierre-Antoine Champin, Fabien Gandon, Mounia Lalmas, and Panagiotis G. Ipeirotis (Eds.). ACM, 1835–1844.
  • Wang et al. (2022a) Shen Wang, Liangwei Yang, Jibing Gong, Shaojie Zheng, Shuying Du, Zhiwei Liu, and S Yu Philip. 2022a. MetaKRec: Collaborative Meta-Knowledge Enhanced Recommender System. In 2022 IEEE International Conference on Big Data (Big Data). IEEE, 665–674.
  • Wang et al. (2019) Xiang Wang, Xiangnan He, Yixin Cao, Meng Liu, and Tat-Seng Chua. 2019. KGAT: Knowledge Graph Attention Network for Recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, August 4-8, 2019, Ankur Teredesai, Vipin Kumar, Ying Li, Rómer Rosales, Evimaria Terzi, and George Karypis (Eds.). ACM, 950–958.
  • Wang et al. (2021a) Xiang Wang, Tinglin Huang, Dingxian Wang, Yancheng Yuan, Zhenguang Liu, Xiangnan He, and Tat-Seng Chua. 2021a. Learning Intents behind Interactions with Knowledge Graph for Recommendation. In WWW ’21: The Web Conference 2021, Virtual Event / Ljubljana, Slovenia, April 19-23, 2021, Jure Leskovec, Marko Grobelnik, Marc Najork, Jie Tang, and Leila Zia (Eds.). ACM / IW3C2, 878–887.
  • Wang et al. (2021b) Yu Wang, Zhiwei Liu, Ziwei Fan, Lichao Sun, and Philip S. Yu. 2021b. DSKReG: Differentiable Sampling on Knowledge Graph for Recommendation with Relational GNN. In CIKM ’21: The 30th ACM International Conference on Information and Knowledge Management, Virtual Event, Queensland, Australia, November 1 - 5, 2021, Gianluca Demartini, Guido Zuccon, J. Shane Culpepper, Zi Huang, and Hanghang Tong (Eds.). ACM, 3513–3517.
  • Wang et al. (2023a) Yu Wang, Zhiwei Liu, Liangwei Yang, and Philip S Yu. 2023a. Conditional Denoising Diffusion for Sequential Recommendation. arXiv preprint arXiv:2304.11433 (2023).
  • Wang et al. (2023b) Yu Wang, Zhengyang Wang, Hengrui Zhang, Qingyu Yin, Xianfeng Tang, Yinghan Wang, Danqing Zhang, Limeng Cui, Monica Cheng, Bing Yin, et al. 2023b. Exploiting intent evolution in e-commercial query recommendation. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 5162–5173.
  • Wang et al. (2022c) Yu Wang, Hengrui Zhang, Zhiwei Liu, Liangwei Yang, and Philip S Yu. 2022c. Contrastvae: Contrastive variational autoencoder for sequential recommendation. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management. 2056–2066.
  • Wang et al. (2014) Zhen Wang, Jianwen Zhang, Jianlin Feng, and Zheng Chen. 2014. Knowledge Graph Embedding by Translating on Hyperplanes. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, July 27 -31, 2014, Québec City, Québec, Canada, Carla E. Brodley and Peter Stone (Eds.). AAAI Press, 1112–1119.
  • Wu et al. (2019) Chuhan Wu, Fangzhao Wu, Mingxiao An, Jianqiang Huang, Yongfeng Huang, and Xing Xie. 2019. NPA: neural news recommendation with personalized attention. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining. 2576–2584.
  • Wu et al. (2022) Haolun Wu, Yansen Zhang, Chen Ma, Fuyuan Lyu, Fernando Diaz, and Xue Liu. 2022. A Survey of Diversification Techniques in Search and Recommendation. CoRR abs/2212.14464 (2022).
  • Yang et al. (2023a) Liangwei Yang, Zhiwei Liu, Chen Wang, Mingdai Yang, Xiaolong Liu, Jing Ma, and Philip S Yu. 2023a. Graph-based Alignment and Uniformity for Recommendation. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management. 4395–4399.
  • Yang et al. (2023c) Liangwei Yang, Shengjie Wang, Yunzhe Tao, Jiankai Sun, Xiaolong Liu, Philip S. Yu, and Taiqing Wang. 2023c. DGRec: Graph Neural Network for Recommendation with Diversified Embedding Generation. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, WSDM 2023, Singapore, 27 February 2023 - 3 March 2023, Tat-Seng Chua, Hady W. Lauw, Luo Si, Evimaria Terzi, and Panayiotis Tsaparas (Eds.). ACM, 661–669.
  • Yang et al. (2023b) Mingdai Yang, Zhiwei Liu, Liangwei Yang, Xiaolong Liu, Chen Wang, Hao Peng, and Philip S Yu. 2023b. Unified Pretraining for Recommendation via Task Hypergraphs. arXiv preprint arXiv:2310.13286 (2023).
  • Yang et al. (2022) Yuhao Yang, Chao Huang, Lianghao Xia, and Chenliang Li. 2022. Knowledge Graph Contrastive Learning for Recommendation. In SIGIR ’22: The 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022, Enrique Amigó, Pablo Castells, Julio Gonzalo, Ben Carterette, J. Shane Culpepper, and Gabriella Kazai (Eds.). ACM, 1434–1443.
  • Ying et al. (2018) Rex Ying, Ruining He, Kaifeng Chen, Pong Eksombatchai, William L Hamilton, and Jure Leskovec. 2018. Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. 974–983.
  • Yu et al. (2022) Junliang Yu, Hongzhi Yin, Xin Xia, Tong Chen, Lizhen Cui, and Quoc Viet Hung Nguyen. 2022. Are Graph Augmentations Necessary?: Simple Graph Contrastive Learning for Recommendation. In SIGIR ’22: The 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022, Enrique Amigó, Pablo Castells, Julio Gonzalo, Ben Carterette, J. Shane Culpepper, and Gabriella Kazai (Eds.). ACM, 1294–1303.
  • Zhang et al. (2016) Fuzheng Zhang, Nicholas Jing Yuan, Defu Lian, Xing Xie, and Wei-Ying Ma. 2016. Collaborative Knowledge Base Embedding for Recommender Systems. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, Balaji Krishnapuram, Mohak Shah, Alexander J. Smola, Charu C. Aggarwal, Dou Shen, and Rajeev Rastogi (Eds.). ACM, 353–362.
  • Zheng et al. (2021) Yu Zheng, Chen Gao, Liang Chen, Depeng Jin, and Yong Li. 2021. DGCN: Diversified Recommendation with Graph Convolutional Networks. In WWW ’21: The Web Conference 2021, Virtual Event / Ljubljana, Slovenia, April 19-23, 2021, Jure Leskovec, Marko Grobelnik, Marc Najork, Jie Tang, and Leila Zia (Eds.). ACM / IW3C2, 401–412.
  • Zhou et al. (2010) Tao Zhou, Zoltán Kuscsik, Jian-Guo Liu, Matúš Medo, Joseph Rushton Wakeling, and Yi-Cheng Zhang. 2010. Solving the apparent diversity-accuracy dilemma of recommender systems. Proceedings of the National Academy of Sciences 107, 10 (2010), 4511–4515.
  • Ziegler et al. (2005) Cai-Nicolas Ziegler, Sean M. McNee, Joseph A. Konstan, and Georg Lausen. 2005. Improving recommendation lists through topic diversification. In Proceedings of the 14th international conference on World Wide Web, WWW 2005, Chiba, Japan, May 10-14, 2005, Allan Ellis and Tatsuya Hagino (Eds.). ACM, 22–32.