This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Beyond Clicks: Modeling Multi-Relational Item Graph for Session-Based Target Behavior Prediction

Wen Wang [email protected] School of Computer Science and Technology,
East China Normal University
Wei Zhang 0000-0001-6763-8146 [email protected] School of Computer Science and Technology,
East China Normal University
Shukai Liu Qi Liu Tencent [email protected] [email protected] Bo Zhang Tencent [email protected] Leyu Lin Tencent [email protected]  and  Hongyuan Zha Georgia Institute of Technology [email protected]
(2020)
Abstract.

Session-based target behavior prediction aims to predict the next item to be interacted with specific behavior types (e.g., clicking). Although existing methods for session-based behavior prediction leverage powerful representation learning approaches to encode items’ sequential relevance in a low-dimensional space, they suffer from several limitations. Firstly, they focus on only utilizing the same type of user behavior for prediction, but ignore the potential of taking other behavior data as auxiliary information. This is particularly crucial when the target behavior is sparse but important (e.g., buying or sharing an item). Secondly, item-to-item relations are modeled separately and locally in one behavior sequence, and they lack a principled way to globally encode these relations more effectively. To overcome these limitations, we propose a novel Multi-relational Graph Neural Network model for Session-based target behavior Prediction, namely MGNN-SPred for short. Specifically, we build a Multi-Relational Item Graph (MRIG) based on all behavior sequences from all sessions, involving target and auxiliary behavior types. Based on MRIG, MGNN-SPred learns global item-to-item relations and further obtains user preferences w.r.t. current target and auxiliary behavior sequences, respectively. In the end, MGNN-SPred leverages a gating mechanism to adaptively fuse user representations for predicting next item interacted with target behavior. The extensive experiments on two real-world datasets demonstrate the superiority of MGNN-SPred by comparing with state-of-the-art session-based prediction methods, validating the benefits of leveraging auxiliary behavior and learning item-to-item relations over MRIG.

Sequential recommendation, graph neural networks, user behavior modeling
journalyear: 2020copyright: iw3c2w3conference: Proceedings of The Web Conference 2020; April 20–24, 2020; Taipei, Taiwanbooktitle: Proceedings of The Web Conference 2020 (WWW ’20), April 20–24, 2020, Taipei, Taiwandoi: 10.1145/3366423.3380077isbn: 978-1-4503-7023-3/20/04ccs: Information systems Personalizationccs: Computing methodologies Neural networks

1. Introduction

Unlike conventional recommendation algorithms which get accustomed to modeling each user-item interaction separately (Koren et al., 2009), recent sequential recommendation approaches meet more realistic requirements for its ability of modeling user dynamic interest. Session-based target behavior prediction (Hidasi et al., 2016) is the one of the main studied problem in this regard, aiming to predict the next item to be interacted with a user under a specific type of behavior (e.g., clicking an item). Based on the predictions, information providers can effectively deliver items to appropriate users and at the same time, and users can quickly find the items what they actually want. Note that we use session-based prediction and session-based recommendation interchangeably throughout this paper.

Early studies for this problem assume that the appearance of the next item depends only on its previous item (Rendle et al., 2010; Zhang and Wang, 2015) in the same sequence. With such a strong assumption, they could only model the last item in each sequence and ignore other information from the sequence. To relieve this assumption, various methods adopt sequential models for session-based recommendation system to learn behavior sequences. Recurrent Neural Networks (RNN) (Hochreiter and Schmidhuber, 1997) is commonly leveraged to obtain promising performance. The relevant methods could roughly be attributed into two categories: single-session based recommendation models (Hidasi et al., 2016; Xu et al., 2019) and multi-session based recommendation models (Quadrana et al., 2017; You et al., 2019). As the latter category requires the user ID of each behavior sequence should be known in advance to link multiple sequences of the same user together, it is not so universal than the first category due to privacy issues and user scalability problem (e.g., a billion of active users each day in WeChat). As such, we study session-based target behavior prediction from the perspective of single-session based modeling.

In the domain of single-session based behavior prediction, some studies (Liu et al., 2018; Ren et al., 2019; Sun et al., 2019) adopt attention mechanism (Bahdanau et al., 2015; Vaswani et al., 2017) and outperform the pioneering RNN based methods (Hidasi et al., 2016). Recent advances in graph neural networks (GNN) (Defferrard et al., 2016; Hamilton et al., 2017) further boost the performance of session-based behavior prediction by modeling each session-based behavior sequence as a graph to achieve the state-of-the-art performance (Wu et al., 2019; Xu et al., 2019). However, existing studies in this regard still suffer from several limitations. Firstly, they focus on only using the same type of user behavior as input for the next item prediction, but ignore the potential of leveraging other type of behavior as auxiliary information. This is particularly crucial when the target behavior is sparse but important (e.g., buying or sharing an item). Secondly, item-to-item relations are modeled separately and locally, since both RNN based and GNN based recommendation models only utilize one behavior sequence each time. It is intuitive that abundant item-to-item relations are hidden in various behavior sequences. For example, if many other users who have bought item B after buying item A, the relation between item A and B is especially vital if a target user just bought item A.

To overcome these limitations, we propose a novel Multi-relational Graph Neural Network model for Session-based target behavior Prediction, namely MGNN-SPred for short. The target behavior we focused on is the aforementioned sparse behavior beyond the dense click behavior. MGNN-SPred jointly considers target behavior and auxiliary behavior sequences and explores global item-to-item relations for accurate prediction. Specifically, for the purpose of considering the global item-to-item relations, we build a Multi-Relational Item Graph (MRIG) based on the past behavior sequences of all sessions. There might exist multiple relations between two graph nodes, denoting target and auxiliary behavior types. Based on MRIG, MGNN-SPred encodes global item-to-item relations into node representations and further obtains local representations for current target and auxiliary behavior sequences, respectively. In the end, MGNN-SPred leverages a gating mechanism to adaptively fuse the representations from target behavior sequence and auxiliary behavior sequence to produce current user interest representation.

The main contributions of this work is summarized as follows:

1. We address the two limitations of existing methods by breaking the restriction of only using one type of behavior sequence in session-based recommendation and exploring another type of behavior as auxiliary information. We further construct the multi-relational item graph for learning global item-to-item relations.

2. To effectively model MRIG w.r.t. target and auxiliary behavior sequences, we develop the novel graph model MGNN-SPred which learns global item-to-item relations through graph neural network and integrates representations of target and auxiliary of current sequences by the gating mechanism.

3. We carry out extensive experiments and demonstrate MGNN-SPred achieves the best performance among strong competitors, showing the benefits of overcoming the two limitations. As a byproduct, we release the source code111https://github.com/Autumn945/MGNN-SPred of our model for relevant studies.

2. Related Work

Session-Based Behavior Prediction. In the literature, the pioneering study (Hidasi et al., 2016) in the direction of single-session based recommendation first adopts a recurrent neural network based approach with past interacted items as the input of different time steps for session-based recommendation. Following that, (Tan et al., 2016) improves the model with data augmentation and the consideration of temporal user behavior shift. In addition to using RNN, (Li et al., 2017) also adopts attention mechanism to capture a user’s sequential behavior and its main purpose in a current session. Similarly, (Liu et al., 2018) proposes a novel attention mechanism to capture both the users’ long-term interests in general and their short-term attention. More recently, with the flourish Graph Neural Networks (GNN) methodologies, (Wu et al., 2019) first separates each session sequence into different graphs and uses graph neural networks to capture complex item transitions in a specific graph. Afterwards, each session is represented as the combination of the global preference and current interests of this session using an attention network. (Xu et al., 2019) is similar to (Wu et al., 2019), which uses a multi-layered self-attention network as an alternative to capture long-range dependencies between items within a session. As discussed in the introduction, these existing relevant methods suffer from two limitations which motivate the proposal of our model in this paper.

Multi-Behavior Modeling. Multi-behavior modeling for recommender system aims to leverage other types of user behavior to boost the recommendation performance on the target behavior. A few studies have already investigated this scenario from different perspectives.  (Krohn-Grimberghe et al., 2012) considers to leverage users’ social interactions as auxiliary behavior for target behavior prediction by collective matrix factorization (CMF) techniques. In a similar fashion, (Zhao et al., 2015) builds multiple matrices from user different behaviors which cover user resharing behavior, user commenting behavior, user posting behavior, etc. CMF is adopted to learn shared user representation for recommendation as well.  (Loni et al., 2016) proposes multi-feedback Bayesian personalized ranking (BPR), an extension of the classical Bayesian personalized ranking approach and tailored for different user behaviors. It differentiates different preference levels between different user behaviors in the sampling stage for ranking.  (Ding et al., 2018) also considers the assignment different preference levels of various user behaviors. Instead of BPR, it incorporates this useful information into element-wise alternating least squares learner. More recently, a neural network approach is proposed by (Gao et al., 2019) to learn representations for user-item interactions with different behaviors. Multi-task learning is conducted to predict multi-behaviors with respect to a certain item in a cascading way. Our work fundamentally differs from the above studies since all of them assume the independence of different user-item interactions while our study is more realistic by considering to model user behaviors in a sequential setting.

Graph Neural Networks. Graph neural networks are the methods used to generate representation of graph structured data, such as social network and knowledge graph. (Perozzi et al., 2014) extends Word2vec (Mikolov et al., 2013) by proposing a model, DeepWalk, to learn node representations based on sequences sampled from graphs. LINE(Tang et al., 2015) encodes first-order and second-order proximity of nodes into a low-dimensional space. Recently, a surge of methods related on graph convolutional networks (GCN) have been raised. (Bruna et al., 2014) presents a method with a graph-based analogue of convolutional architectures, which is the original version of GCN. Later, a number of improvements, extensions, and approximations of these spectral convolutions be proposed (Kipf and Welling, 2017; Duvenaud et al., 2015; Hamilton et al., 2017; Monti et al., 2017). These approaches outperform other methods based on random walks (e.g., DeepWalk and node2vec). With the success in mind, an amount of GCN based methods are widely applied in various domains such as recommendation systems (Monti et al., 2017). But most GCN based methods require that all nodes in the graph are present in each propagation step of GNN. Different from GCN, GraphSAGE (Hamilton et al., 2017) can train GNN with a minibatch setting. Inspired by this, we design our GNN to learn from the constructed multi-relational item graph for session-based behavior prediction.

Refer to caption\Description

[Graph of model architecture]Graph of model architecture containing input MRIG and output item ranking list.

Figure 1. The architecture of our model. We use a toy MRIG and two current behavior sequences as input. The number of recommended items is set to 2.

3. Technologies

3.1. Problem Definition

For a session ss in the session set SS, let Ps=[p1s,p2s,p3s,,p|Ps|s]P^{s}=[p^{s}_{1},p^{s}_{2},p^{s}_{3},...,p^{s}_{|P^{s}|}] denote the target behavior sequence and Qs=[q1s,q2s,q3s,,q|Qs|s]Q^{s}=[q^{s}_{1},q^{s}_{2},q^{s}_{3},...,q^{s}_{|Q^{s}|}] represent the auxiliary behavior sequence. Moreover, we construct a Multi-Relational Item Graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) based on all behavior sequences from all sessions, where 𝒱\mathcal{V} is the set of nodes in the graph containing all available items and \mathcal{E} is the edge sets involving multiple types of directed edges. Each edge is a triple consisting of the head item, the tail item, and the type of this edge. For instance, if we construct the graph based on behaviors of sharing and clicking, then an edge (a,b,share)(a,b,\textnormal{share})\in\mathcal{E} means that a user shared item aa and subsequently shared item bb, and an edge (a,b,click)(a,b,\textnormal{click})\in\mathcal{E} means that a user clicked item bb after clicking item aa. Given the above notations, we formulate the problem as follows:

Problem 1 (session-based target behavior prediction).

Given a session sSs\in S and its target and auxiliary behavior sequences PsP^{s} and QsQ^{s}, along with MRIG 𝒢\mathcal{G}, the target of this problem is to learn a model that can generate KK items which are most likely to be interacted with the user of the session in the next.

Algorithm 1 Multi-relational item graph construction
0:  Session set SS, both target and auxiliary behavior sequences PsP^{s} and QsQ^{s}, sS{\forall}s\in S
0:  MRIG 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E})
1:  𝒱\mathcal{V}\leftarrow\varnothing, \mathcal{E}\leftarrow\varnothing
2:  for s𝒮s\in\mathcal{S} do
3:     𝒱𝒱{Ps[1]}\mathcal{V}\leftarrow\mathcal{V}\cup\{P^{s}[1]\}
4:     for i=2i=2 to |Ps||P^{s}| do
5:        𝒱𝒱{Ps[i]}\mathcal{V}\leftarrow\mathcal{V}\cup\{P^{s}[i]\}, {(Ps[i1],Ps[i],target)}\mathcal{E}\leftarrow\mathcal{E}\cup\{(P^{s}[i-1],P^{s}[i],\textnormal{target})\}
6:     end for
7:     𝒱𝒱{Qs[1]}\mathcal{V}\leftarrow\mathcal{V}\cup\{Q^{s}[1]\}
8:     for i=2i=2 to |Qs||Q^{s}| do
9:        𝒱𝒱{Qs[i]}\mathcal{V}\leftarrow\mathcal{V}\cup\{Q^{s}[i]\}, {(Qs[i1],Qs[i],auxiliary)}\mathcal{E}\leftarrow\mathcal{E}\cup\{(Q^{s}[i-1],Q^{s}[i],\textnormal{auxiliary})\}
10:     end for
11:  end for

3.2. Overview

The overall architecture of the proposed MGNN-SPred is depicted in Figure 1. The input to MGNN-SPred contains a Multi-Relational Item Graph (MRIG) and the two types of behavior sequences. SR-MRIG first learns item correlations from MRIG by graph neural networks and encode them into item representations. Afterwards, a user’s two behavior sequences are regarded as two sub-graphs in the MRIG where the items in each sub-graph are connected with a virtual node (“T” or “A” in Figure 1), respectively. Subsequently, SR-MRIG aggregates the nodes of each sub-graph to the corresponding virtual node, thus getting the representation of each behavior sequence. Finally, to fuse the two behavior representations and obtain user preference representations, a gating mechanism is adopted to adaptively decide the importance of different behaviors and perform weighted summation over them. For the purpose of recommendation, SR-MRIG calculates each item’s score by user and item representations via a bi-linear product and use the scores to rank them for recommendation.

3.3. Graph Construction

There are abundant relationships between items lying in users’ historical behaviors. If a user buys item aa, and subsequently buys item bb in the same session, it indicates that item aa and item bb probably have some dependency, but does not reflect similarity too much since a user less likely buys two very similar items within a short duration. In comparison, if a user clicks item aa, and subsequently clicks item bb, it indicates that item aa and item bb are probably with large similarity. This is intuitive because a user usually browses a number of similar items, and picks the most suitable one to buy.

We construct the multi-relational item graph by taking all items as nodes and each type of behavior corresponds as one directed edge, denoting different relationships between items. The process of constructing MRIG is shown in Algorithm 1. The both target and auxiliary behavior sequences from all sessions PsP^{s} and QsQ^{s} (s𝒮{\forall}s\in\mathcal{S}) are provided as input. The algorithm browses all behavior sequences, collects all items in the sequences as the nodes of the graph, and constructs edges between two consequent items in the same sequence with their behavior types as the edge types. After constructing the graph with target and auxiliary behaviors, there are two types of directional edges in the graph.

3.4. Item Representation Learning

For each node v𝒱v\in\mathcal{V}, we use 𝐞¯v|𝒱|\bar{\mathbf{e}}_{v}\in\mathbb{R}^{|\mathcal{V}|} denotes its one-hot representation. Before we feed the one-hot representations of nodes into GNN, we first convert each of them into a low-dimensional dense vector 𝐞vd\mathbf{e}_{v}\in\mathbb{R}^{d} by a learnable embedding matrix 𝐄|𝒱|×d\mathbf{E}\in\mathbb{R}^{|\mathcal{V}|\times d}: 𝐞v=𝐄𝐞¯v\mathbf{e}_{v}=\mathbf{E}^{\top}\bar{\mathbf{e}}_{v}.

After collecting the vectors 𝐞v(v𝒱)\mathbf{e}_{v}({\forall}v\in\mathcal{V}), we feed them with MRIG 𝒢\mathcal{G} into GNN to generate global representations of nodes 𝐠v\mathbf{g}_{v}. The representations are expected to encode multiple item-to-item relations. We take node vv as an example for illustration. First of all, we collect neighbors of node vv. Each node in the graph has four types of neighboring node sets. According to the type and direction, we name the four sets as “target-forward”, “target-backward”, “auxiliary-forward”, and “auxiliary-backward”. Take the type of “target” as an example, we obtain neighbor groups corresponding to forward and backward directions as below:

(1) 𝒩t+(v)={v|(v,v,target)},𝒩t(v)={v|(v,v,target)}.\mathcal{N}_{\textnormal{t}+}(v)=\{v^{\prime}|(v^{\prime},v,\textnormal{target})\in\mathcal{E}\},~{}~{}\mathcal{N}_{\textnormal{t}-}(v)=\{v^{\prime}|(v,v^{\prime},\textnormal{target})\in\mathcal{E}\}.

For the type of “auxiliary”, its neighbor groups, i.e., 𝒩a+(v)\mathcal{N}_{\textnormal{a}+}(v) and 𝒩a(v)\mathcal{N}_{\textnormal{a}-}(v), are acquired by the same way.

At each step of representation propagation in GNN, we first aggregate each group of neighbors by mean-pooling to obtain the representation of this group, defined as below:

(2) 𝐡t+,vk=v𝒩t+(v)𝐡vk1|𝒩t+(v)|.\mathbf{h}^{k}_{\textnormal{t}+,v}=\frac{\sum_{v^{\prime}\in\mathcal{N}_{\textnormal{t}+}(v)}\mathbf{h}^{k-1}_{v^{\prime}}}{|\mathcal{N}_{\textnormal{t}+}(v)|}.

The representations of the three remaining groups are calculated in a similar fashion. Consequently, for the propose of joint considering different relations between items, we combine these four representations of different neighbor groups by sum-pooling:

(3) 𝐡¯vk=𝐡t+,vk+𝐡t,vk+𝐡a+,vk+𝐡a,vk.\mathbf{\bar{h}}^{k}_{v}=\mathbf{h}^{k}_{\textnormal{t}+,v}+\mathbf{h}^{k}_{\textnormal{t}-,v}+\mathbf{h}^{k}_{\textnormal{a}+,v}+\mathbf{h}^{k}_{\textnormal{a}-,v}.

Finally, we update the representation of the center node vv by:

(4) 𝐡vk=𝐡vk1+𝐡¯vk.\mathbf{h}^{k}_{v}=\mathbf{h}^{k-1}_{v}+\mathbf{\bar{h}}^{k}_{v}.

After performing KK iterations, we take the node representation of the last step as the representation of the corresponding item: 𝐠v=𝐡vK\mathbf{g}_{v}=\mathbf{h}^{K}_{v}. In practice, we implement the GNN in a minibatch setting which is inspired by (Hamilton et al., 2017) to ensure scalability.

3.5. Sequence Representation Learning

We have tried different ways to compute the representation of the virtual node for the target and auxiliary behavior sequences, including using attention mechanism to assign different importance weights to the nodes and performing sub-graph propagating for several times. Empirically, we have found that simple mean-pooling could already achieve comparable performance while retaining low complexity. We denote the summarized representations of target behavior sequence PP and auxiliary behavior sequence QQ as 𝐩\mathbf{p} and 𝐪\mathbf{q}, respectively, which are given as:

(5) 𝐩=i=1|P|𝐠pi|P|,𝐪=i=1|Q|𝐠qi|Q|.\mathbf{p}=\frac{\sum_{i=1}^{|P|}\mathbf{g}_{p_{i}}}{|P|},~{}~{}\mathbf{q}=\frac{\sum_{i=1}^{|Q|}\mathbf{g}_{q_{i}}}{|Q|}.

We argue that the two different types of behavior sequence representations might contribute differently when building an integrated representation. This is because the auxiliary behavior is not exactly the same with the target behavior to be predicted, and different users might have different concentration on different behaviors. For instances, some users might browse the item pages frequently and click various items arbitrarily, and another users might only click the items they want to buy. It is self-evident that the contributions of auxiliary behavior sequence for the next item prediction are different in these situations. We define the following gating mechanism to calculate the relative importance weight α\alpha:

(6) α=σ(𝐖g[𝐩;𝐪]),\alpha=\sigma(\mathbf{W}_{g}[\mathbf{p};\mathbf{q}]),

where [𝐩;𝐪][\mathbf{p};\mathbf{q}] denotes the concatenation of the two representations, σ\sigma is the sigmoid function, and 𝐖gR1×2d\mathbf{W}_{g}\in R^{1\times 2d} is a trainable parameter of our model. Finally, we obtain the user preference representation 𝐨\mathbf{o} for the current session by the weighted summation of 𝐩\mathbf{p} and 𝐪\mathbf{q}:

(7) 𝐨=α𝐩+(1α)𝐪.\mathbf{o}=\alpha\cdot\mathbf{p}+(1-\alpha)\cdot\mathbf{q}.

3.6. Model Prediction and Training

We further calculate the recommendation score svs_{v} of each item v𝒱v\in\mathcal{V} using the item embedding 𝐞v\mathbf{e}_{v}. A bi-linear matching scheme is employed by:

(8) sv=𝐨𝐖𝐞v,s_{v}=\mathbf{o}^{\top}\mathbf{W}\mathbf{e}_{v},

where 𝐖Rd×d\mathbf{W}\in R^{d\times d} is a trainable parameter matrix of our model.

To learn the parameters of our model, we apply a softmax function to normalize the scores 𝐬R|𝒱|\mathbf{s}\in R^{|\mathcal{V}|} over all items to get the probability distribution 𝐲^\hat{\mathbf{y}}:

(9) 𝐲^=softmax(𝐬).\hat{\mathbf{y}}=\textnormal{softmax}(\mathbf{s}).

Backpropagation for neural networks is adopted to optimize the model by minimizing the cross-entropy loss of the predicted probability distribution 𝐲^\hat{\mathbf{y}} w.r.t. the ground truth. The loss function is defined as follows:

(10) RS=i=1|𝒱|yilog(y^i),\mathcal{L}_{RS}=-\sum_{i=1}^{|\mathcal{V}|}y_{i}\log(\hat{y}_{i})\,,

where (y1,,y|𝒱|)(y_{1},\cdots,y_{|\mathcal{V}|}) denotes the one-hot representation of the ground truth. Note that RS\mathcal{L}_{RS} is easily extended to a minibatch loss.

4. Experiment

Table 1. Basic statistics of the datasets.
Data WeChat Yoochoose
#items 56,561 52,740
#sessions 100,000 9,249,729
Time duration 2019/09/17~23 2014/04/01~09/30
#edge of target 217,774 225,879
#edge of auxiliary 1,546,220 3,277,411
Average length of target 9.76 3.31
Average length of auxiliary 33.49 8.56
#training data 167,931 163,005
#validation data 12,333 12,985
#test data 24,667 25,971

4.1. Experimental Setup

4.1.1. Dataset

We evaluate our model on two real-world datasets named WeChat and Yoochoose. The Yoochoose dataset is obtained from the RecSys Challenge 2015. The user behavior sequences in the dataset are already segmented into sessions and all the users are anonymized. The WeChat dataset is collected from Top Stories (看一看) of WeChat, where we choose videos are regarded as items. We randomly select one hundred thousand active users and collect their behavior records for a duration of one week. Since the duration is relatively short, we retain an entire behavior sequence of each user by taking the sequence as a single session. In this paper, we treat the behavior of purchase in Yoochoose and the behavior of sharing in WeChat as the target behavior and regard the behavior of clicking in both datasets as the auxiliary behavior.

Given a session with the target behavior sequence P=[p1,p2,,p|P|]P=[p_{1},p_{2},...,\\ p_{|P|}] and the auxiliary behavior sequence Q=[q1,q2,,q|Q|]Q=[q_{1},q_{2},...,q_{|Q|}], we adopt a similar way to construct training example as (Li et al., 2017; Wu et al., 2019). That is, we treat each item pip_{i}, (i2i\geq 2) as the label and use [p1,p2,pi1][p_{1},p_{2},...p_{i-1}] as input of target behavior. The treatment for the auxiliary behavior is a little different, because a user is very likely to click an item before buying or sharing it. To avoid the auxiliary input already sees the labels, we only keep the clicked items before the target item that is also bought or shared by the user. We set a maximum length LL for both types of sequences and only keep the last LL items longer than the maximum length. Considering the fact that two datasets have different average sequence length (see details in Table 1), we set the maximum length LL to 10 for WeChat and 3 for Yoochoose. We discuss the impact of different maximum length in Section 4.4.3.

We split the datasets in a chronological order for evaluation, consistent with real situations. We take the first 6/7 of datasets as the training data, and use 1/3 of the remaining data as the validation data to determine optimal hyper-parameter settings. MRIG used throughout the experiments are constructed only based on training data. The basic statistics of two datasets are summarized in Table 1.

4.1.2. Baselines

We compare the proposed model with several strong competitors, including state-of-the-art graph neural network based model for session-based recommendation.

  • POP. It just recommends the top-n frequent items in the training set regardless of behaviors in current sessions.

  • Item-KNN (Sarwar et al., 2001). It recommends items most similar to the previously interacted items belonging to the same sessions.

  • GRU4Rec (Hidasi et al., 2016). GRU4Rec is the pioneering RNN-based deep sequential model for session-based recommendation.

  • NARM (Li et al., 2017). It employs attention mechanism to capture different importance of each item according to their hidden states obtained by RNN. A weighted integration of different item representations is performed to obtain final representation.

  • STAMP (Liu et al., 2018). This model learns users’ general interest from the long-term memory of session context and current interest from the short-term memory of their last behaviors.

  • SR-GNN (Wu et al., 2019) and GC-SAN (Xu et al., 2019). Both of the graph-based models only use a current session to construct graph for applying GNN to learn item representations. The difference is SR-GNN represents each session by a traditional attention network while GC-SAN is based on a multi-layered self-attention mechanism.

  • R-DAN. Reasoning-DAN (R-DAN) (Nam et al., 2017) is used to model both behavior sequences simultaneously.

  • CoAtt. Co-Attention (CoAtt) (Lu et al., 2016) with alternative calculation for interactive attention is adopted for comparison.

  • HetGNN. Heterogeneous graph neural network (Zhang et al., 2019) is applied for recommendation, with two edge types and one node type.

It is worth noting only target behavior is considered by the above baselines originally developed for session-based recommendation, i.e., GRU4Rec, NARM, STAMP, SR-GNN, and GC-SAN. To make the comparison more fairable, we revise these methods through the following manner. We use their original forms to model the target behavior sequence and auxiliary behavior sequence respectively, And afterwards, we utilize the proposed gating mechanism to fuse the two types of representations as ours. In addition, we also compare our model with the baselines in the situation of only considering target behavior (see Table 3 for details).

Table 2. Evaluation results of all methods.
Methods WeChat Yoochoose
H@100 M@100 N@100 H@100 M@100 N@100
POP 13.565 1.1247 3.2621 6.095 0.2529 1.2231
Item-KNN 15.770 1.1624 3.7222 15.286 1.9415 4.4040
GRU4Rec 18.831 1.3956 4.3966 19.114 2.5292 5.5830
NARM 19.131 1.4034 4.4416 18.775 2.5819 5.5813
STAMP 17.757 1.3083 4.1078 20.361 2.3487 5.6879
SR-GNN 18.940 1.3827 4.3967 21.262 2.6892 6.1232
GC-SAN 19.034 1.2090 4.2490 19.718 2.5218 5.6861
HetGNN 20.290 1.4171 4.6504 24.031 2.9546 6.8732
R-DAN 18.952 1.3879 4.3980 15.956 2.3107 4.8608
CoAtt 17.700 1.1931 4.0137 20.080 2.5742 5.8206
Ours 21.271 1.4797 4.8529 28.632 3.6564 8.2722
Table 3. Results of not using auxiliary behavior sequences.
Methods WeChat Yoochoose
H@100 M@100 N@100 H@100 M@100 N@100
GRU4Rec (w/o a) 16.889 1.2346 3.9128 14.817 1.6032 4.0012
NARM (w/o a) 17.773 1.3123 4.1298 14.443 1.5540 3.8900
SR-GNN (w/o a) 18.093 1.2621 4.1368 15.302 1.5782 4.0852
Ours (w/o a) 19.252 1.3933 4.4473 21.089 2.3798 5.8221

4.1.3. Implementation Details

We implement our proposed model based on Tensorflow. The dimension of item embedding is set to 64. Adam with default parameter setting is adopted to optimize the model, with the mini-batch size of 64. GNN is ensured to run in a minibatch setting and the depth KK is set to 2. We terminate the learning process with an early stopping strategy. We test different forms of attention computation formulas for the baselines based on attention mechanism and report their best results. The hyper-parameters of baselines are turned on validation datasets as well.

4.2. Model Comparison

We consider the top-100 ranked predictions as recommended items. Following (Wu et al., 2019; Xu et al., 2019), we adopt HR@100 (H@100), MRR@100 (M@100), and NDCG@100 (N@100) to evaluate the recommendation performance of all models after obtaining their recommendation lists. Table 2 shows the performance comparison between our model and the adopted baselines. (1) The first part of the table corresponds to the simple baselines. We observe their results are significantly worse than other methods. (2) The second part involves standard sequential based methods for session-based recommendation. We observe that their results keep at the same level, except for STAMP on WeChat. It shows that: 1) taking session-based recommendation as a sequential modeling task can improve performance; 2) although NARM and STAMP are more advanced approaches which use attention mechanism to combine hidden representations of different time steps, they do not show advantages on the sparse behavior prediction problem we studied (not the same as previous studies focusing on click prediction). (3) The third part is GNN based models. SR-GNN and GC-SAN seem to be better than the sequential methods, and HetGNN further boost the performance. (4) The second-to-last part involves approaches of learning two sequences in other research domains. Their best results are worse than the best performance of the above recommendation methods, which suggests that considering the interaction of items in two sequences might have no benefit for the studied problem. Finally, we can see that our method outperforms all the other methods, demonstrating the superiority of our model for session-based recommendation.

4.3. Impact of Auxiliary Behavior Sequence

We choose several representative methods in Table 3 to test whether considering the auxiliary behavior sequence indeed boosts the performance of session-based recommendation. The methods with “(w/o a)” mean removing the auxiliary behavior sequence from their full version. Firstly, we observe that our proposed model still consistently achieves better performance in this situation. Moreover, by comparing each method in Table 2 with its “(w/o a)” version, we can find every method beats the one of “(w/o a)” with significant margins. Based on the above illustrations, we demonstrate that considering the auxiliary behavior sequence is indeed meaningful.

Table 4. Ablation study of MGNN-SPredl.
Methods WeChat Yoochoose
H@100 M@100 N@100 H@100 M@100 N@100
Ours (w/o ae) 20.923 1.4665 4.7945 25.463 2.7678 6.8907
Ours (w/o asg) 19.742 1.3949 4.5167 22.517 2.6025 6.2631
Ours (w/o g) 20.363 1.3707 4.6154 27.577 3.3531 7.7896
Ours 21.271 1.4797 4.8529 28.632 3.6564 8.2722

4.4. Model Analysis

4.4.1. Ablation Study

We conduct ablation studies of our model, using “w/o ae” to denote removing the edges related to the auxiliary behavior, using “w/o asg” to denote that not modeling the sub-graph of the auxiliary behavior sequence in getting user preference representation, and using “w/o g” to indicate merging the two representations of the target and auxiliary behavior sequences by simple summation instead of the gating mechanism. Table 4 shows the corresponding results. We observe that the incorporation of the auxiliary edge into the built graph is beneficial for the problem by seeing “w/o ae”. The integration of the auxiliary behavior with target behavior sequence have a notable contribution by seeing “w/o asg”. Besides, we find that the performance becomes worse if we do not use the gating mechanism to merge the two representations of the target and auxiliary behavior sequences by investigating “w/o g”. Through the above comparison, we conclude the main components in our model are effective.

Refer to caption
(a) WeChat
Refer to caption
(b) Yoochoose
Figure 2. Results of our model with different depths of GNN.
\Description

[Histogram of results]Histogram of results corresponding to different versions of our model with with different number of layers in GNN.

4.4.2. Impact of Depth of GNN

We test different depth settings (from 0 to 3) about graph representation propagation. The depth setting with value 0 means the our model does not use GNN and could not learn any information from MRIG. Figure 2(b) shows the corresponding results. We can see that the performance of depth 0 is without doubt much worse than the results with depths from 1 to 3. This comparison clarifies the significance of considering MRIG for our model. Moreover, the performance becomes significantly better when the depth grows from 1 to 2, showing modeling high-order relation between items through GNN is indispensable. When the number of graph representation propagation is larger than 3, the representations of nodes might become less distinguishable, which is not ideal for further improving the performance.

Refer to caption
(a) WeChat
Refer to caption
(b) Yoochoose
Figure 3. Results for different maximum lengths.
\Description

[Line charts]Line charts of different results corresponding to different sequence lengths.

4.4.3. Impact of Sequence Length

We visualize the performance variation with the change of the maximum behavior sequence length LL in Figure 3(b), where we set LL in the range from 1 to 20. As expected, with larger maximum sequence length at the beginning, the performance of both our model and SR-GNN grows to be better. After reaching the peaks, the results slightly become worse, and finally the variation trends turn to be stable. Overall, our model outperforms SR-GNN consistently. Besides, we find the lengths with the best performance are not the same in the two datasets. This is due to the fact the average length of Yoochoose is much smaller than that of WeChat, as shown in Table 1.

5. Conclusion

In this paper, we study session-based target behavior prediction. Two limitations of existing relevant models are addressed: using only target behavior for next item prediction and lacking a principled way to encode global item-to-item relations. To alleviate the issues, MGNN-SPred is proposed, with the major novelties of building and modeling of the multi-relational item graph. In addition, a gating mechanism is adopted to adaptively fuse target behavior sequences and auxiliary behavior sequences into the user preference representations for the next item prediction. Comprehensive experiments on two real-world datasets demonstrate MGNN-SPred achieves the best performance and its design is rational.

References

  • (1)
  • Bahdanau et al. (2015) Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural Machine Translation by Jointly Learning to Align and Translate. ICLR (2015).
  • Bruna et al. (2014) Joan Bruna, Wojciech Zaremba, Arthur Szlam, and Yann LeCun. 2014. Spectral networks and locally connected networks on graphs. ICLR (2014).
  • Defferrard et al. (2016) Michaël Defferrard, Xavier Bresson, and Pierre Vandergheynst. 2016. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In NIPS. 3837–3845.
  • Ding et al. (2018) Jingtao Ding, Guanghui Yu, Xiangnan He, Yuhan Quan, Yong Li, Tat-Seng Chua, Depeng Jin, and Jiajie Yu. 2018. Improving Implicit Recommender Systems with View Data. In IJCAI. 3343–3349.
  • Duvenaud et al. (2015) David K Duvenaud, Dougal Maclaurin, Jorge Iparraguirre, Rafael Bombarell, Timothy Hirzel, Alán Aspuru-Guzik, and Ryan P Adams. 2015. Convolutional networks on graphs for learning molecular fingerprints. In NIPS. 2224–2232.
  • Gao et al. (2019) Chen Gao, Xiangnan He, Dahua Gan, Xiangning Chen, Fuli Feng, Yong Li, Tat-Seng Chua, and Depeng Jin. 2019. Neural Multi-task Recommendation from Multi-behavior Data. In ICDE. 1554–1557.
  • Hamilton et al. (2017) Will Hamilton, Zhitao Ying, and Jure Leskovec. 2017. Inductive representation learning on large graphs. In NIPS. 1024–1034.
  • Hidasi et al. (2016) Balázs Hidasi, Alexandros Karatzoglou, Linas Baltrunas, and Domonkos Tikk. 2016. Session-based recommendations with recurrent neural networks. ICLR (2016).
  • Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735–1780.
  • Kipf and Welling (2017) Thomas N Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. ICLR (2017).
  • Koren et al. (2009) Yehuda Koren, Robert M. Bell, and Chris Volinsky. 2009. Matrix Factorization Techniques for Recommender Systems. IEEE Computer 42, 8 (2009), 30–37.
  • Krohn-Grimberghe et al. (2012) Artus Krohn-Grimberghe, Lucas Drumond, Christoph Freudenthaler, and Lars Schmidt-Thieme. 2012. Multi-relational matrix factorization using bayesian personalized ranking for social network data. In WSDM. 173–182.
  • Li et al. (2017) Jing Li, Pengjie Ren, Zhumin Chen, Zhaochun Ren, Tao Lian, and Jun Ma. 2017. Neural attentive session-based recommendation. In CIKM. 1419–1428.
  • Liu et al. (2018) Qiao Liu, Yifu Zeng, Refuoe Mokhosi, and Haibin Zhang. 2018. STAMP: short-term attention/memory priority model for session-based recommendation. In KDD. 1831–1839.
  • Loni et al. (2016) Babak Loni, Roberto Pagano, Martha Larson, and Alan Hanjalic. 2016. Bayesian Personalized Ranking with Multi-Channel User Feedback. In RecSys. 361–364.
  • Lu et al. (2016) Jiasen Lu, Jianwei Yang, Dhruv Batra, and Devi Parikh. 2016. Hierarchical question-image co-attention for visual question answering. In NIPS. 289–297.
  • Mikolov et al. (2013) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. In NIPS. 3111–3119.
  • Monti et al. (2017) Federico Monti, Michael Bronstein, and Xavier Bresson. 2017. Geometric matrix completion with recurrent multi-graph neural networks. In NIPS. 3697–3707.
  • Nam et al. (2017) Hyeonseob Nam, Jung-Woo Ha, and Jeonghee Kim. 2017. Dual attention networks for multimodal reasoning and matching. In CVPR. 299–307.
  • Perozzi et al. (2014) Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. 2014. Deepwalk: Online learning of social representations. In KDD. 701–710.
  • Quadrana et al. (2017) Massimo Quadrana, Alexandros Karatzoglou, Balázs Hidasi, and Paolo Cremonesi. 2017. Personalizing session-based recommendations with hierarchical recurrent neural networks. In RecSys. 130–137.
  • Ren et al. (2019) Pengjie Ren, Zhumin Chen, Jing Li, Zhaochun Ren, Jun Ma, and Maarten de Rijke. 2019. RepeatNet: A Repeat Aware Neural Recommendation Machine for Session-Based Recommendation. In AAAI. 4806–4813.
  • Rendle et al. (2010) Steffen Rendle, Christoph Freudenthaler, and Lars Schmidt-Thieme. 2010. Factorizing personalized markov chains for next-basket recommendation. In WWW. 811–820.
  • Sarwar et al. (2001) Badrul Munir Sarwar, George Karypis, Joseph A Konstan, John Riedl, et al. 2001. Item-based collaborative filtering recommendation algorithms. Www 1 (2001), 285–295.
  • Sun et al. (2019) Fei Sun, Jun Liu, Jian Wu, Changhua Pei, Xiao Lin, Wenwu Ou, and Peng Jiang. 2019. BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer. In CIKM.
  • Tan et al. (2016) Yong Kiam Tan, Xinxing Xu, and Yong Liu. 2016. Improved recurrent neural networks for session-based recommendations. In Proceedings of the 1st Workshop on Deep Learning for Recommender Systems. ACM, 17–22.
  • Tang et al. (2015) Jian Tang, Meng Qu, Mingzhe Wang, Ming Zhang, Jun Yan, and Qiaozhu Mei. 2015. Line: Large-scale information network embedding. In WWW. 1067–1077.
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In NIPS. 6000–6010.
  • Wu et al. (2019) Shu Wu, Yuyuan Tang, Yanqiao Zhu, Liang Wang, Xing Xie, and Tieniu Tan. 2019. Session-based recommendation with graph neural networks. In AAAI. 346–353.
  • Xu et al. (2019) Chengfeng Xu, Pengpeng Zhao, Yanchi Liu, Victor S Sheng, Jiajie Xu, Fuzhen Zhuang, Junhua Fang, and Xiaofang Zhou. 2019. Graph Contextualized Self-Attention Network for Session-based Recommendation. In IJCAI. 3940–3946.
  • You et al. (2019) Jiaxuan You, Yichen Wang, Aditya Pal, Pong Eksombatchai, Chuck Rosenberg, and Jure Leskovec. 2019. Hierarchical Temporal Convolutional Networks for Dynamic Recommender Systems. In WWW. 2236–2246.
  • Zhang et al. (2019) Chuxu Zhang, Dongjin Song, Chao Huang, Ananthram Swami, and Nitesh V. Chawla. 2019. Heterogeneous Graph Neural Network. In KDD. 793–803.
  • Zhang and Wang (2015) Wei Zhang and Jianyong Wang. 2015. Location and Time Aware Social Collaborative Retrieval for New Successive Point-of-Interest Recommendation. In CIKM. 1221–1230.
  • Zhao et al. (2015) Zhe Zhao, Zhiyuan Cheng, Lichan Hong, and Ed Huai-hsin Chi. 2015. Improving User Topic Interest Profiles by Behavior Factorization. In WWW. 1406–1416.