This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Graph Neural Ordinary Differential Equations-based method for Collaborative Filtering

Ke Xu University of Illinois Chicago
Chicago, USA
[email protected]
   Yuanjie Zhu University of Illinois Chicago
Chicago, USA
[email protected]
   Weizhi Zhang University of Illinois Chicago
Chicago, USA
[email protected]
   Philip S. Yu University of Illinois Chicago
Chicago, USA
[email protected]
Abstract

Graph Convolution Networks (GCNs) are widely considered state-of-the-art for collaborative filtering. Although several GCN-based methods have been proposed and achieved state-of-the-art performance in various tasks, they can be computationally expensive and time-consuming to train if too many layers are created. However, since the linear GCN model can be interpreted as a differential equation, it is possible to transfer it to an ODE problem. This inspired us to address the computational limitations of GCN-based models by designing a simple and efficient NODE-based model that can skip some GCN layers to reach the final state, thus avoiding the need to create many layers. In this work, we propose a Graph Neural Ordinary Differential Equation-based method for Collaborative Filtering (GODE-CF). This method estimates the final embedding by utilizing the information captured by one or two GCN layers. To validate our approach, we conducted experiments on multiple datasets. The results demonstrate that our model outperforms competitive baselines, including GCN-based models and other state-of-the-art CF methods. Notably, our proposed GODE-CF model has several advantages over traditional GCN-based models. It is simple, efficient, and has a fast training time, making it a practical choice for real-world situations.

Index Terms:
recommendation, neural ordinary differential equations, graph neural ordinary differential equations, collaborative filtering, graph convolutional networks

I Introduction

In recent years, machine learning tasks on graph-structured data have become increasingly popular, and Graph Convolutional Networks (GCNs) [1, 2, 3, 4] have emerged as a powerful tool for these tasks. GCNs are effective in various tasks, particularly in recommendation systems, due to their simplicity and effectiveness. The core idea of GNNs is to update each node embedding iteratively using multiple graph propagation layers. This process aggregates node embeddings from both their neighbors and themselves. Essentially, GNNs learn the embeddings of nodes by leveraging the graph’s structure. In practice, it has been observed that most tasks require only two or three layers [5] to achieve good results. Using a larger number of layers in GCNs may result in inferior performance, as the model becomes increasingly complex and its parameters more difficult to optimize [1, 2, 6].

Collaborative Filtering(CF)[7, 8, 9, 10, 11, 12, 13, 14] is a popular approach used in recommender systems to predict users’ interests based on their past behavior or preferences. The goal is to learn user and product embeddings and compute dot-products for recommendations. However, this approach has long been a research problem in the field of recommender systems. And Graph Convolutional Networks (GCNs) can effectively capture the relationships between users and items in a graph, enhancing the performance of CF. This approach represents users and products as nodes in a graph, with edges between them representing interactions or relationships. A GCN will learn node embeddings by iteratively aggregating information from neighboring nodes in the graph. GCNs have shown promising results in improving the performance of traditional CF methods and are increasingly used in various recommendation systems. It is particularly useful in scenarios where user-item interactions are complex and sparse, as they can leverage graph structure to capture higher-order relationships between nodes.

Several studies[15, 16] have found that linear GCN architectures with layer combination outperform non-linear GCN [11] architectures for collaborative filtering (CF). Additionally, a linear GCN can be easily interpreted as differential equations, this concept[17, 18] has led to a Neural Ordinary Differential Equations(NODEs)-based CF method called LT-OCF[19]. This method demonstrates the suitability of NODEs-based approaches for CF. The main idea behind LT-OCF is to develop a purely continuous version of the GCN layer. The architecture can be viewed as a continuous version of LightGCN[15] with a learnable number of layers.

Recent study[17] on neural ordinary differential equations(NODEs) suggests that discrete layers with residual connections may not be the only perspective for building neural networks. Instead, this line of research proposes that discrete layers with residual connections can be understood as a discretization of a continuous ordinary differential equation(ODE). This approach is more memory-efficient since NODEs can significantly reduce the required number of parameters when building neural networks and models[20], and can also model the dynamics of hidden states more smoothly. The implications of this approach are significant because it opens up new possibilities for constructing neural networks that are more flexible and efficient.

The concept behind neural ordinary differential equations(NODEs) are to learn implicit differential equations from data. NODEs calculate the output h(t1)=h(t0)+t0t1f(h(t),θ)𝑑th(t_{1})=h(t_{0})+\int^{t_{1}}_{t_{0}}f(h(t),\theta)dt, where f(h(t),θ)f(h(t),\theta) is a neural network. θ\theta is trained from data. And t0t_{0} is the staring time, t1t_{1} is ending time. The resulting h(t1)h(t_{1}) values provide an approximation of the solution to the differential equation. There are various ODE solvers that can solve integral problems, and they can also generalize different types of neural network architectures[17]. For instance, the explicit Euler method for soloving ODE problems can be viewed as a residual connection[17, 19]. Meanwhile, the fourth-order Runge-Kutta(RK4) ODE solver can be seen as a dense convolutional network[21].

Because of the conceptual connections between Graph Neural Networks (GNNs)[22, 1] and NODEs, which have led to the development of a continuous version of GNNs known as Continuous Graph Neural Networks (CGNNs)[23]. CGNNs extend existing discrete GNNs to the continuous case by defining the evolution of node embeddings using ordinary differential equations (ODEs). A CGNN calculates final node embeddings h(t1)=h(0)+0t1(AI)h(t)𝑑th(t_{1})=h(0)+\int_{0}^{t_{1}}(A-I)h(t)dt, where AA is the adjacency matrix. The node embeddings of h(t1)h(t_{1}) at step t1t_{1} can be viewed as all the information propagated up to t1t_{1} steps with the node embeddings h(t)h(t). In this way t1t_{1} can be viewed as the number of layers for GNNs. The method we mentioned earlier, LT-OCF, utilizes a similar idea to CGNNs.

Different than CGNNs, Graph Neural Ordinary Differential Equations(GODEs)[24] do not develop a continuous message-passing layer. Instead, they use the NODE framework and parameterize the derivative function directly using multiple GCN layers. Since GCNs perform message passing to derive final states, it is similar to solving an initial value problem for an ODE. The final states of node embeddings h(t1)h(t_{1}) can be calculated using h(t1)=h(t0)+t0t1G(h(t))𝑑th(t_{1})=h(t_{0})+\int^{t_{1}}_{t_{0}}G(h(t))dt, where G(h(t))G(h(t)) is a multiple GCN layer with the node embeddings at state tt, h(t)h(t). GODEs retains the concept of the number of layers nn and views the time step t1t_{1} as another hyperparameter. To put it simply, GODEs seek to estimate the end node embeddings by utilizing data captured by just 1 or 2 GCN layers.

Inspired by the concept of Graph Neural Ordinary Differential Equations (GODEs), we propose a new method called Graph Neural Ordinary Differential Equations-based Collaborative Filtering (GODE-CF). GODE-CF is designed to estimate the final embeddings by utilizing an ODE. Unlike many other GCN methods for Collaborative Filtering (CF), GODE-CF does not utilize layer combinations. Instead, we will use the output of the ODE as the final embeddings. Our proposed method has several advantages over other linear GCN models. It has a faster training time without sacrificing accuracy, and it achieves state-of-the-art accuracy in CF tasks.

Figure 1 shows the architecture of GODE-CF and a comparison between LightGCN and LT-OCF. To evaluate the performance of GODE-CF, we use four public review datasets and compare them with state-of-the-art methods, such as LT-OCF[19], LightGCN[15] and UltraGCN[25]. The results show that our method consistently outperforms these methods in all datasets. Furthermore, we demonstrate the effectiveness of our method by showing that it can be trained faster than most GCN-based methods. Additionally, we investigate the impact of the choice of solver for ODEs on different datasets. Our contributions can be summarized as follows:

  • We propose a novel architecture for CF, Graph Neural Ordinary Differential Equations-based Collaborative Filtering (GODE-CF), which incorporates both GCN and GODE components..

  • We demonstrate that our method outperforms state-of-the-art methods in all datasets.

  • We show that our method can be trained faster than most GCN-based methods.

  • We investigate the impact of the choice of solver for ODEs on different datasets.

Refer to caption

Figure 1: The architectures of LT-OCF, LightGCN and GODE-CF. Left figure: a LT-OCF with three ODE blocks; Right figure: a LightGCN with three layers, and \bigoplus is the layer combination; Middle figure: GODE-CF with two layers.

II PRELIMINARIES

II-A Graph Convolution Networks(GCN)

Graph convolution network(GCN) is variant of Graph Neural Networks(GNNs). The basic idea of GCNs is to learn embeddings for nodes by smoothing features over the graph[1, 2, 3, 4]. For each layer, it aggregates the features of neighbors as the new embeddings of the target node. The propagation rule can be written as:

Hn+1=σ(AGG(W,A,Hn))H_{n+1}=\sigma(AGG(W,A,H_{n})) (1)

Where AGGAGG is aggregation function, HnH_{n} is node embedding for the nnth layer, Hn+1H_{n+1} is node embedding for n+1n+1th layer, WW is the transformation matrix, and AA is adjacency matrix of the graph. However, there are studies[1] have shown that the GCN can only learn graph moments of a specific power of the Graph. To overcome this limitation, many GCN models[1, 11, 15] concatenate the node embeddings of each layer, which allows the model to capture information from different powers of AA and improves its ability to model complex graph structures.

II-B Neural Ordinary Differential Equations(NODEs)

Neural Ordinary Differential Equations(NODEs)[17] is a method that aims to model continuous dynamics on hidden state of neural networks. This is achieved by characterizing dynamics through an ordinary differential equation (ODE), which is parameterized by a neural network. The goal of this approach is to learn implicit differential equations from data. One of the key advantages of Neural ODEs is that they provide a flexible framework for modeling complex systems. By using ODEs, it is possible to model systems that exhibit continuous behavior. Additionally, by using neural networks to parameterize the ODEs, it is possible to capture complex patterns in the data that may be difficult to model using discrete methods. Neural ordinary differential equations calculate h(t1)=h(t0)+t0t1f(h(t),θ)𝑑th(t_{1})=h(t_{0})+\int_{t_{0}}^{t_{1}}f(h(t),\theta)dt, where ff is a neural network parameterized by θ\theta that approximates dh(t)dt\frac{dh(t)}{dt}, to derive h(t1)h(t_{1}) from h(t0)h(t_{0}). θ\theta is trained from data. The variable t0t_{0} and t1t_{1} are staring time and ending time, commonly set t0=0t_{0}=0 and then t1t_{1} can be viewed as the number of layers in a neural networks, whereas it can be any non-negative real number in NODEs. In order to solve ODE problem, there are different ODE solvers that can be used to solve the ODE problem, such as Euler, RK2, RK4, and they can also be viewed as various neural network architectures. For example, the general form of the residual connection can be written as h(t+1)=h(t)+f(h(t);θ)h(t+1)=h(t)+f(h(t);\theta), which is same as the explicit Euler method for solving ODE problems. In general, the choice of ODE solver depends on the specific problem being solved. For instance, the explicit Euler method is simple and easy to implement, but it can be unstable. On the other hand, the fourth-order Runge-Kutta(RK4) method is more stable and accurate, but it can be computationally expensive[24, 19].

III RELATED WORK

III-A Continuous Graph Neural Networks(CGNNs)

Deep learning has traditionally been dominated by discrete models. However, there are some studies[26, 17] have proposed treating neural networks as equipped with a continuum of layers, rather than a discrete set of layers. This approach enables a reformulation of the forward and backward pass as the solution to the initial value problem of an ODE. Such approaches allow direct modeling of ODEs and can guide discovery of novel general purpose deep learning models. This concept also build some connections between GCNs and traditional dynamical systems. Continuous Graph Neural Networks (CGNNs)[23] are an architecture that models continuous dynamics on node embeddings. First, it employ a neural encoder to project each node into a latent space based on its features. Afterwards, the encoded nodes are treated as the initial value H(0)H(0), and an ODE is designed to define the continuous dynamics on node embeddings. This allows CGCNs to capture complex patterns of interactions between nodes over time. Finally, the node embeddings obtained at the ending time t1t_{1} can be used for downstream applications such as node classification. Motivated by diffusion-based methods on graphs, an effective way for characterizing the propagation process is to use the following propagation equations:

Hn+1=AHn+H0H_{n+1}=AH_{n}+H_{0} (2)

where HnH_{n} is the embeddings for all the nodes at stage nn, and H0H_{0} is the initial embeddings. Intuitively, each node at stage n+1n+1 learns the node information from its neighbors through AHnAH_{n} and remembers its original node features through H0H_{0}. This way is to learn the graph structure without forgetting the original node features. And then, the formula can be derived as follows:

Hn=(i=0nAi)H0=(AI)1(AnI)H0H_{n}=(\sum_{i=0}^{n}A^{i})H_{0}=(A-I)^{-1}(A^{n}-I)H_{0} (3)

where the embeddings HnH_{n} at stage nn incorporates all the information propagated up to nn steps with the initial embeddings H0H_{0}. With the idea of NODEs, we can replace the discrete step nn with a continuous variable tt, and use an ODE to characterize such a continuous propagation dynamic. This allows us to naturally move from the discrete propagation process to the continuous case. Specifically, the embeddings HtH_{t} at time tt incorporates all the information propagated up to time tt with the initial embeddings H0H_{0}. We can view this as a Riemann sum of an integral from time 0 to time tt, which leads us to the following proposition. To derive the equation using Taylor expansion, we have:

dH(t)dt=(AI)H(t)+H(0)\frac{dH(t)}{dt}=(A-I)H(t)+H(0) (4)

And,

H(t1)=H(0)+0t1(AI)H(t)𝑑tH(t_{1})=H(0)+\int^{t_{1}}_{0}(A-I)H(t)dt (5)

where H(0)H(0) represents the initial embeddings of all nodes, H(t1)H(t_{1}) represents the node embeddings at time t1t_{1} and AA represents the adjacency matrix of the graph. Typically, t0t_{0} is set to 0 and t1t_{1} can be treated as the number of layers. Therefore, H(t1)H(t_{1}) can be interpreted as the node embeddings after t1t_{1} convolutions.

III-B Graph Neural Ordinary Differential Equations (GODEs)

Unlike Continuous Graph Neural Networks (CGNNs), which develop a continuous message-passing layer, Graph Neural Ordinary Differential Equations (GODEs)[24] parametrize the derivative function using 2 or 3 GCN layers directly. Based on graph spectral theory, the residual version of GCN layers[1, 27] are in the following form:

Hn+1=σ(AGG(W,A,Hn))H_{n+1}=\sigma(AGG(W,A,H_{n})) (6)

where AA is the graph Laplacian and σ\sigma is a nonlinear activation function, WW is the transformation matrix, and Hn+1H_{n+1} is the node embedding at the (n+1)(n+1)th layer. We denote the graph convolution operator as GCG_{C}. Where GC(H(n+1))=σ(AGG(W,A,Hn))G_{C}(H(n+1))=\sigma(AGG(W,A,H_{n})). The nn-layer convolution can be written as follows:

H(n)=GCN(n,H(0))=GCnGCn1GC1H(0)H(n)=GCN(n,H(0))=G_{C}^{n}\circ G_{C}^{n-1}\circ...\circ G_{C}^{1}H(0) (7)

Where GCG_{C} is the graph convolution layer, nn is number of layers. And the GODEs to calculate the final node embeddings is:

H(t1)=H(0)+t0t1GCN(n,H(t))𝑑tH(t_{1})=H(0)+\int^{t_{1}}_{t_{0}}GCN(n,H(t))dt (8)

where H(t)H(t) represents the node embeddings at time tt, H(0)H(0) represents the initial embeddings, GCN(n,H(t))GCN(n,H(t)) represents the nn graph convolution layer with input node embeddings H(t)H(t). H(t1)H(t_{1}) can be treat as the final node embeddings that estimate by the ODE, with the input of H(t)H(t).

III-C ODE Solver

There are various ODE solvers can solve the integral problem, and it can generalize various neural network architectures[17]. For instance, the explicit Euler method can be written as following step:

H(t+1)=H(t)+hf(H(t),t;θ)H(t+1)=H(t)+hf(H(t),t;\theta) (9)

where H(t)H(t) is node embeddings at state tt. To estimate the solution of the differential equation, we can integrate the states up to H(t1)H(t_{1}) using this formula, starting from a given initial value of H(t0)H(t_{0}). The resulting H(t)H(t) provide an approximation of the solution to the differential equation. Note that this equation is identical to a residual connection when h=1h=1. Other ODE solvers like RK4 it is also can be viewed as dense convolution networks[17]. Other than Euler and RK4, there are many other solvers exist, such as dopri5, dopri8, fixed adams, midpoint and so on. For GODEs, higher-order ODE solvers are generally more effective, as long as the graph is dense enough to benefit from the additional computation [24]. In this study, we will only test two solvers: explicit Euler and RK4.

III-D Collaborative Filtering(CF)

Collaborative filtering(CF)[7, 8, 9, 10, 11, 12, 13, 14] is a technique that can filter out items that a user might like on the basis of reactions by similar users. A typical CF paradigm involves learning user and item node embeddings, from historical interaction data, and give a top-kk recommendation based on the pairwise similarity of the user and item.

Let E0uRN×DE_{0}^{u}\in R^{N\times D} and E0iRM×DE_{0}^{i}\in R^{M\times D} be the initial user and item embeddings, respectively. Since user-item interactions can be represented as bipartite graphs, it is popular to adopt GCNs for CF. One of the early used GCN-based CF methods is NGCF[9], which utilizes non-linear activations and transformation matrices to map the user embedding space to the item embedding space. At each layer, user and item embeddings are combined by concatenation. However, non-linear activations and transformation matrices are not necessary for CF, as the user-item graph is sparse and non-linear GCNs are prone to overfitting. Moreover, empirical evidence suggests that transforming between user and item embedding spaces does not improve CF accuracy significantly. A simpler and more effective GCN-based CF method, called LightGCN[15], has been proposed. LightGCN removes the activation function and transformation matrix, and achieves state-of-the-art accuracy on many datasets. This suggests that linear GCNs with layer combination perform best among various design choices. The definition of the linear graph convolutional layer is as follows:

Eku=AkEk1i,Eki=AkEk1u\begin{split}E_{k}^{u}&=A^{k}E_{k-1}^{i},\\ E_{k}^{i}&=A^{k}E_{k-1}^{u}\end{split} (10)

Here, AA represents the normalized adjacency matrix, and EkE_{k} represents the user/item embedding matrix at the kk-th layer. The initial embeddings, which are denoted as E0uE^{u}_{0} and E0pE^{p}_{0}. The final embeddings EfuE^{u}_{f} and EfpE^{p}_{f} are calculated using layer combination, expressed as follows:

Efu=l=0KwlElu,Efi=l=0KwlEli\begin{split}E^{u}_{f}&=\sum^{K}_{l=0}w_{l}E_{l}^{u},\\ E^{i}_{f}&=\sum^{K}_{l=0}w_{l}E_{l}^{i}\end{split} (11)

where KK is the number of layers, wlw_{l} is weight for llth layer, and EfuE^{u}_{f} and EfiE^{i}_{f} are the final node embeddings. All CF methods learn the initial embeddings of users and items. The model prediction is defined as the inner product of user and item final embeddings:

yu,i=EfuTEfiy_{u,i}=E_{f}^{u^{T}}E_{f}^{i} (12)

where EfuE_{f}^{u} is the final users embeddings and EfiE_{f}^{i} is the final items embeddings. Typically, the following Bayesian personalized ranking (BPR)[10] loss is used to train the initial embedding vectors in the field of collaborative filtering.

III-E Learnable-Time ODE-based Collaborative Filtering(LT-OCF)

The aggregation of messages in LightGCN, a linear GCN model, can be viewed as the integration of an ordinary differential equation (ODE). Building on this concept, LT-OCF[19] was developed as an ODE-based method. The complete model can be viewed as a continuous version of LightGCN with a learnable number of layers. Figure 1 offers a comparison of the two methods. The message passing of LT-OCF for single ODE can be written as follows:

u(K)=u(0)+0K(AI)i(t)𝑑t,i(K)=i(0)+0K(AI)u(t)𝑑t\begin{split}u(K)&=u(0)+\int^{K}_{0}(A-I)i(t)dt,\\ i(K)&=i(0)+\int^{K}_{0}(A-I)u(t)dt\end{split} (13)

where, u(t)RN×Du(t)\in R^{N\times D} refers to user embeddings, and i(t)RM×Di(t)\in R^{M\times D} refers to item embeddings at state tt. KK can be treated as the number of layers. u(K)u(K) and i(K)i(K) are the user embeddings and item embeddings at time KK, respectively. For the whole model, the time interval [0,K][0,K] is split into TT parts. For instance, if T=2T=2, the model will have 22 ODE blocks. The first ODE block at time [0,t1][0,t_{1}] and the second ODE block at time [t1,K][t_{1},K]. The final embedding will be a combination of the output of these two blocks, the initial embeddings, and the embeddings at time KK. The final embeddings are calculated as follows:

Efu=w0u(0)+l=1Twlu(tl)+wKu(K),Efi=w0i(0)+l=1Twli(tl)+wKi(K)\begin{split}E^{u}_{f}&=w_{0}u(0)+\sum^{T}_{l=1}w_{l}u(t_{l})+w_{K}u(K),\\ E^{i}_{f}&=w_{0}i(0)+\sum^{T}_{l=1}w_{l}i(t_{l})+w_{K}i(K)\end{split} (14)

where u(0)u(0) and i(0)i(0) are initial user embeddings and initial item embeddings respectively, wiw_{i} is weight for each ODE block, and KK is the ending time or number of layers. EfuE^{u}_{f} and EfiE^{i}_{f} are final user embeddings and final item embeddings, respectively.

IV Proposed Method

IV-A Motivation

As we previously mentioned, the linear GCN model can be interpreted as an ODE problem [19, 17]. This means that we can view the GCN process as a continuous flow of information across the graph, where the state of each node changes continuously over time, governed by a differential equation. This concept has led to the development of Graph Neural Ordinary Differential Equations (GODEs), which extend the GCN framework to a continuous-time setting. With GODEs, we can estimate the final node embedding with one or two GCN layers. Specifically, by solving the ODEs governing the GCN process, we can obtain a continuous-time dynamical system that describes the evolution of the node embeddings over time. This can be viewed as a generalization of the static GCN model, where we can capture more complex temporal dynamics in the graph. In addition, we can also use the GODE framework to incorporate other types of information, such as node attributes or edge features. Inspired by the success of GODE in modeling continuous-time dynamics, we believe that this framework is also suitable for the collaborative filtering (CF) task. In CF, the goal is to estimate user and item embeddings. By leveraging GODEs, we can estimate these embeddings more accurately.

IV-B Method

We propose a new method based on the idea of Graph Neural ODE, as shown in Fig 1. Instead of creating a continuous message passing layer, we parametrize the derivative function using multiple layers of GCN directly. In other words, we use the information captured by one or two GCN layers to estimate the final state of the embedding by solving an ODE problem without creating all of the layers. Unlike the concept of CGNN[18], we treat t1t_{1} as hyperparameters, and our model doesn’t have any layer combinations, since integration can be viewed as the summation of all layers from time 0 to t1t_{1}. The initial embeddings are directly fed into the ODEs, and the output of the ODEs becomes the final embedding. The overall formula can be written as:

Efu=Eu(t1)=Eu(0)+0t1W(AnI)Ei(t)𝑑tEfi=Ei(t1)=Ep(0)+0t1W(AnI)Eu(t)𝑑t\begin{split}E^{u}_{f}=E^{u}(t_{1})&=E^{u}(0)+\int^{t_{1}}_{0}W(A^{n}-I)E^{i}(t)dt\\ E^{i}_{f}=E^{i}(t_{1})&=E^{p}(0)+\int^{t_{1}}_{0}W(A^{n}-I)E^{u}(t)dt\end{split} (15)

In this model, AA is the adjacency matrix, E0uE^{u}_{0} and E0iE^{i}_{0} are the initial user and item embeddings, respectively. WW is a trainable weight and nn is the number of layers. The user and item embeddings at time tt are denoted as Eu(t)E^{u}(t) and Ei(t)E^{i}(t), respectively. The final user embeddings and item embeddings are denoted as EfuE^{u}_{f} and EfiE^{i}_{f}, respectively. Unlike other GCN-based models [11, 15], where the final embeddings are a combination of all the layers. Our method aims to estimate the final embeddings using the information that captured by several GCN layers through an ODE function. Same as other methods, we use BPR[10] loss to train the embeddings and use ODE solvers, such as explicit Euler and RK4, to solve the ODE problem.

When it comes to solving ODEs, there are many options available, each with their own strengths and weaknesses. In fact, different ODE solvers may produce vastly different results even when applied to the same dataset. As a result, we believe that it is crucial to carefully consider the choice of ODE solver based on the characteristics of the dataset being used. By doing so, we can ensure that we are using an appropriate solver that will provide accurate and reliable results.

And the model prediction is defined as the inner product of the final embeddings of the user and item:

yu,i=EfuTEfiy_{u,i}={E^{u}_{f}}^{T}E^{i}_{f} (16)

where yu,iy_{u,i} is the rating matrix for all users.

IV-C Connection with CGNN and LT-OCF

Suppose the following setting in our method:i) the number of layers nn is 11, ii) We use euler’s method to solve the ODE problem, iii) we set WW to 11. Then our model can be written as follows:

Eu(t1)=E0u+0t1(AI)Ei(t)𝑑tEi(t1)=E0i+0t1(AI)Eu(t)𝑑t\begin{split}E^{u}(t_{1})&=E^{u}_{0}+\int^{t_{1}}_{0}(A-I)E^{i}(t)dt\\ E^{i}(t_{1})&=E^{i}_{0}+\int^{t_{1}}_{0}(A-I)E^{u}(t)dt\end{split} (17)

which is same as the Continuous Graph Neural Networks(CGCN), or a LT-OCF with single ODE block (K=1K=1).

V Experiment

TABLE I: The statistics of the datasets
Datasets # of interactions for training # of interactions for validation # of interactions for testing Sparsity
Office 43448 4905 4905 0.44867%
Health 269137 38609 38609 0.0484%
Cell Phone 138681 27879 27879 0.0668%
Beauty 153776 22363 22363 0.07335%

V-A Experimental Environments

In this section, we introduce our experimental environments and results. All experiments were conducted in the following hardware environments: 12th Gen Intel(R) Core(TM) i9-12900F, with one GeForce RTX 3080 GPU. We utilize the torchdiffeq PyTorch package to solve and backpropagate through the ODE solver.

V-B Datasets and Baselines

Datasets. We use the public Amazon Reviews dataset[28] with four benchmark categories, including: Beauty, Health, Cell Phones, Office Product. The details of the datasets are summarized in Table I. We follow the 5-core setting as existing works on users and the same transformation[11, 15, 29, 30] of treating the existence of reviews as positives. We sort each user’s interactions chronologically and adopt the leave-one-out setting, with the last interacted item for testing and the second last interaction for validation.

Eveluation. We use the standard ranking evaluation metrics, including Recall@N and NDCG@N, to evaluate the averaged ranking performance over all users. These metrics are widely used in existing works, such as [11, 15, 29, 19]. Recall@N measures the correctness of ranking the ground-truth items in the top-N list, regardless of the ranking position. It is a widely used metric for evaluating the effectiveness of top-N recommendation algorithms. NDCG@N extends the idea of Recall@N by giving higher weights to higher ranking positions. It measures the quality of a ranking algorithm based on the graded relevance of the items in the ranked list. We only report the N=20 due to limited space.

Baselines. In total, we compare GODE-CF with various types of the state-of-the-art models:

  • NGCF[11] is a GCN-based CF method. It uses feature transformation and non-linear activations

  • LightGCN[15] are linear GCN-based CF methods.

  • UltraGCN[25] is an ultra-simplified formulation of GCN which skips explicit message passing and directly approximate the limit of infinite message passing layers.

  • layerGCN[31] is a layer-refined GCN model, that refines layer representations during information propagation and node updating of GCN.

  • GTN[29] a graph trend filtering networks framework to capture the adaptive reliability of the interactions between users and items. Extensive experiments on four real-world datasets demonstrate the.

  • LT-OCF[19] is a NODE-based method that learn the optimal architecture of the model for CF.

V-C Hyper-Parameters Grid Search

In this section we introduce the grid search of the hyperparameters. The learning rate is 0.0010.001, and the embedding size is 128128 for all models and the Normal distribution of N(0,1)N(0,1) is used to set initial embeddings.

  • For LT-OCF, the regularization coefficient λ\lambda in all method is in {1e4,1e3,1e2}\{1e-4,1e-3,1e-2\}. The number of elements KK is in {2,3,4}\{2,3,4\}. The number of learnable intermediate time points T is in {1,2,3}\{1,2,3\}, we consider the following solvers: the explicit Euler, and RK4.

  • For NGCF, we search the node and message dropouts from {0.1,0.3,0.5}\{0.1,0.3,0.5\} and number of layers from {1,2,3,4,5}\{1,2,3,4,5\}.

  • For GTN, we search parameters β\beta and γ\gamma from {0.001,0.005,0.01,0.05,0.1}\{0.001,0.005,0.01,0.05,0.1\}.

  • For UltraGCN, we search their weights from {1,1e3,1e5,1e7}\{1,1e-3,1e-5,1e-7\} and negative weights from {10,50,100}\{10,50,100\}.

  • For layerGCN, we search number of layers from {1,2,3,4,5}\{1,2,3,4,5\}

  • For LightGCN, we search the node and message dropouts from {0.1,0.3,0.5}\{0.1,0.3,0.5\} and number of layers from {1,2,3,4,5}\{1,2,3,4,5\}

  • For GODE-CF, we search number of layers from {1,2,3}\{1,2,3\}, and tt from {0.7,0.75,0.8,0.85,0.9,0.95,1}\{0.7,0.75,0.8,0.85,0.9,0.95,1\}

V-D Experimental Results

In Table II, we summarize the overall accuracy in terms of Recall and NDCG.

  • Among GCN-based methods, GODE-CF consistently achieves the best performance across all four datasets. In particular, on Health, GODE-CF significantly outperforms the strongest GCN-based baseline, GTN, by 11% on Recall@20. For other baselines, LightGCN exhibits the best performance on Beauty, while GTN demonstrates the best performance on Health compared to other GCN-based baselines.

  • Among all baselines, the ODE-based model LT-OCF shows state-of-the-art accuracy for Beauty. However, our method, GODE-CF, achieves the best accuracy in all cases. The Euler solver yields the best results for all four datasets compared to RK4. We believe that the choice of solver may affect performance depending on the dataset.

We compare the training curve of LightGCN and GODE-CF for Beauty and Health. In general, our method provides a faster training speed in terms of the number of epochs than LightGCN.

Refer to caption

Figure 2: Training curve for Beauty and Health
TABLE II: Overall Performance
Dataset Beauty Health Cell Phone Office
Method Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20
NDCG 0.070787 0.029954 0.030641 0.012258 0.043868 0.016914 0.050968 0.022137
layerGCN 0.076197 0.031435 0.024528 0.010094 0.039671 0.015078 0.047095 0.021048
UltraGCN 0.056611 0.026183 0.031702 0.013475 0.036049 0.015575 0.046891 0.023363
GTN 0.071457 0.030585 0.032868 0.013517 0.045590 0.017802 0.043629 0.020263
LightGCN 0.077762 0.032996 0.030304 0.012023 0.044299 0.016609 0.056473 0.026355
LT-OCF(Euler) 0.078791 0.033095 0.030226 0.011975 0.046415 0.017393 0.056269 0.025960
LT-OCF(RK4) 0.078254 0.033194 0.030330 0.011930 0.045626 0.017161 0.055250 0.025634
GODE-CF(RK4) 0.078701 0.033480 0.033619 0.013565 0.048531 0.018325 0.055861 0.025178
GODE-CF(Euler) 0.080758 0.034068 0.033878 0.013340 0.050791 0.019095 0.056677 0.027020

V-E Efficiency Comparison

In this section, we provide empirical evidence demonstrating the superiority of GODE-CF’s training efficiency compared to other baselines. We believe that it is essential to prove this convincingly, so we compare the total training time and epochs required to achieve the best performance among the models. Our experiment includes validation time in the training time to provide a comprehensive overview of the training process. To ensure a fair comparison, we use a uniform code framework that we implemented ourselves for all models. We train all models with a fixed number of 1000 epochs to eliminate the effect of varying epoch numbers. Additionally, we conduct our experiments on all four datasets using the same machine:12th Gen Intel(R) Core(TM) i9-12900F, with one GeForce RTX 3080 GPU for all compared models.

Table III shows the details of the results. The explicit Euler providing the best performance for our model for all four datasets. Since this solver is simple and fast, which is a contributing factor to GODE-CF’s short training time. By comparing the total training time and epochs required to achieve the best performance, we demonstrate that GODE-CF consistently outperforms other baselines.

TABLE III: Efficiency comparison
Training Time
Dataset Beauty Health Cell Phone Office
NDCG 39m 90m 43m 7m
UltraGCN 54m 93m 52m 14m
GTN 73m 207m 67m 8m
LightGCN 40m 84m 42m 6m
LT-OCF(Euler) 48m 115m 48m 8m
LT-OCF(RK4) 98m 263m 94m 13m
GODE-CF(RK4) 57m 146m 56m 9m
GODE-CF(Euler) 35m 80m 36m 7m

VI Ablation Studies

VI-A Euler vs. RK4

In this section, we compare two different solvers for ordinary differential equations (ODEs): Euler and RK4. The Fig  3 and Fig 4 are the training curves of these two solvers. Other solvers, such as DOPRI, are almost the same as RK4, but RK4 has smaller computation complexity. Therefore, we only use RK4 and Euler’s method for this experiment. Contrary to what we thought, Euler’s method shows better accuracy for almost all cases, whereas RK4 should have better accuracy in terms of solving general ODE problems. For instance, the Euler method achieves a Recall/NDCG of 0.050791/0.019095 vs. 0.048531/0.018325 with RK4 on Cell Phones. However, for other datasets, we observe very similar patterns where Euler’s method consistently outperforms RK4. And Euler’s method provides a faster training speed in terms of the number of epochs than that of RK4 for all of the cases. This is because using the RK4 solver is more likely to result in overfitting the training data in this model.

Refer to caption

Figure 3: Training curve for Cell phones and Office with different solvers

Refer to caption

Figure 4: Training curve for Beauty and Health with different solvers

VI-B Impact on different number of layers

We also conducted experiments that explored the impact of varying the number of layers inside ODE on the model’s performance. The detailed results are in Table IV.

TABLE IV: Performance comparison for different numbers of layers
Dataset Beauty Health Cell Phone Office
Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20
1 layer 0.07526 0.03093 0.02901 0.01104 0.04204 0.01568 0.05647 0.02522
2 layer 0.08076 0.03407 0.03388 0.01356 0.05079 0.01909 0.05668 0.02702
3 layer 0.07919 0.03311 0.03124 0.01217 0.04760 0.01783 0.05770 0.02639
TABLE V: Performance with or without weight
Dataset Beauty Health Cell Phone Office
Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20
Without weight (W) 0.08076 0.03407 0.03403 0.01353 0.05058 0.01916 0.05586 0.02661
With weight (W) 0.08027 0.03395 0.03362 0.01356 0.05079 0.01909 0.05668 0.02702

Based on the results, we have found that a 2-layer design yields optimal results for all four cases. This is because the ODE needs to estimate the final state of the embeddings by using enough information captured by the GCN layer. In other words, the GCN layer need to provides the necessary information for the ODE to make optimal estimations.

On the other hand, a 1-layer design may not provide enough information for the ODE to make a good estimation. This is because one GCN layer may not capture all the necessary information required to estimate the final state of the embeddings.

Conversely, a 3-layer design may capture excessive information, leading to overfitting of the training data. To address this issue, we suggest incorporating the dropout technique when using a 3-layer design. Dropout can help prevent overfitting and ensure that the final state estimated by the ODE is more reliable.

VI-C Sensitivity on t

By varying tt, we also investigate how different tt affect the model accuracy. The detailed results are shown in Figure  5.

Refer to caption

Figure 5: Performance for different tt on office and Cell Phone

According to the plot, we can see that the model with only a small number of layers requires a larger value of tt to reach the optimal performance. Conversely, models with larger numbers of layers require a smaller value of tt to achieve the optimal performance. This is because models with fewer layers provide less information to the ODE, which results in the requirement of a longer time step to reach the final state. Conversely, models with a higher number of layers capture more information, which subsequently results in them being closer to the final state, thereby allowing the ODE to take smaller time steps tt. It is important to note that the number of layers plays a crucial role in the performance of the model. An interesting finding is that, for both 2-layer and 3-layer settings, there is a time point at which the model experiences a significant performance drop. However, after this time point, the performance begins to recover. The cause of this issue remains unknown, so we will leave it for future work.

VI-D Impact on weight

We are model has a learneable weight for each layer inside ODE. We are also interested in to investigate whether the weight has a significant impact on the model’s performance. The detailed results are shown in Table V. Based on the results, we found that the impact of weight is highly dependent on the dataset. For most cases, the model without weight has better performance than the model with weight on the Beauty, but the difference is not significant. However, for the Office, the model has a larger improvement when using weight. Whether or not to use weight depends highly on the specific case. It’s recommended to evaluate the impact of weight on the model for each specific dataset.

VII Conclusion

This study aims to introduce a new method for Collaborative Filtering (CF) called GODE-CF. The method is based on Graph Neural Ordinary Differential Equation (ODE) and is formulated as a graph neural ODE system suitable for collaborative filtering. We conducted an evaluation of GODE-CF on four different real-world datasets. The experimental results showed that GODE-CF outperforms various state-of-the-art baselines in terms of performance with shorter training time. Looking into the future, we plan to test more ODE solvers as we only tested two in this study, and there are a lot more ODE solvers is being developed. Additionally, we discovered that our model is prone to overfitting, which is a concern for future research. In conclusion, GODE-CF provides a new approach to CF, which has the potential to improve the performance and efficiency of collaborative filtering.

Acknowledgment

This work is supported in part by NSF under grant III-2106758.

References

  • [1] T. N. Kipf and M. Welling, “Semi-Supervised Classification with Graph Convolutional Networks,” in 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings.   OpenReview.net, 2017. [Online]. Available: https://openreview.net/forum?id=SJU4ayYgl
  • [2] J. Zhou, G. Cui, S. Hu, Z. Zhang, C. Yang, Z. Liu, L. Wang, C. Li, and M. Sun, “Graph neural networks: A review of methods and applications,” AI Open, vol. 1, pp. 57–81, 2020. [Online]. Available: https://linkinghub.elsevier.com/retrieve/pii/S2666651021000012
  • [3] S. Zhang, H. Tong, J. Xu, and R. Maciejewski, “Graph convolutional networks: a comprehensive review,” Computational Social Networks, vol. 6, no. 1, p. 11, Nov. 2019. [Online]. Available: https://doi.org/10.1186/s40649-019-0069-y
  • [4] R. v. d. Berg, T. N. Kipf, and M. Welling, “Graph Convolutional Matrix Completion,” Oct. 2017, arXiv:1706.02263 [cs, stat]. [Online]. Available: http://arxiv.org/abs/1706.02263
  • [5] M. Qu, Y. Bengio, and J. Tang, “GMNN: Graph Markov Neural Networks,” Jul. 2020, arXiv:1905.06214 [cs, stat]. [Online]. Available: http://arxiv.org/abs/1905.06214
  • [6] Q. Li, Z. Han, and X.-M. Wu, “Deeper insights into graph convolutional networks for semi-supervised learning,” in Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, ser. AAAI’18/IAAI’18/EAAI’18.   New Orleans, Louisiana, USA: AAAI Press, Feb. 2018, pp. 3538–3545.
  • [7] L. Chen, L. Wu, R. Hong, K. Zhang, and M. Wang, “Revisiting Graph Based Collaborative Filtering: A Linear Residual Graph Convolutional Network Approach,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 01, pp. 27–34, Apr. 2020. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/5330
  • [8] R. Chen, Q. Hua, Y.-S. Chang, B. Wang, L. Zhang, and X. Kong, “A Survey of Collaborative Filtering-Based Recommender Systems: From Traditional Methods to Hybrid Methods Based on Social Networks,” IEEE Access, vol. 6, pp. 64 301–64 320, 2018.
  • [9] X. Wang, X. He, M. Wang, F. Feng, and T.-S. Chua, “Neural Graph Collaborative Filtering,” in Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, ser. SIGIR’19.   New York, NY, USA: Association for Computing Machinery, Jul. 2019, pp. 165–174. [Online]. Available: https://doi.org/10.1145/3331184.3331267
  • [10] S. Rendle, C. Freudenthaler, Z. Gantner, and L. Schmidt-Thieme, “BPR: Bayesian personalized ranking from implicit feedback,” in Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, ser. UAI ’09.   Arlington, Virginia, USA: AUAI Press, Jun. 2009, pp. 452–461.
  • [11] X. He, L. Liao, H. Zhang, L. Nie, X. Hu, and T.-S. Chua, “Neural Collaborative Filtering,” in Proceedings of the 26th International Conference on World Wide Web, ser. WWW ’17.   Republic and Canton of Geneva, CHE: International World Wide Web Conferences Steering Committee, Apr. 2017, pp. 173–182. [Online]. Available: https://doi.org/10.1145/3038912.3052569
  • [12] X. Yang, Y. Guo, Y. Liu, and H. Steck, “A survey of collaborative filtering based social recommender systems,” Computer Communications, vol. 41, pp. 1–10, Mar. 2014. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0140366413001722
  • [13] T. Ebesu, B. Shen, and Y. Fang, “Collaborative Memory Network for Recommendation Systems,” in The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, ser. SIGIR ’18.   New York, NY, USA: Association for Computing Machinery, Jun. 2018, pp. 515–524. [Online]. Available: https://doi.org/10.1145/3209978.3209991
  • [14] Y. Wu, C. DuBois, A. X. Zheng, and M. Ester, “Collaborative Denoising Auto-Encoders for Top-N Recommender Systems,” in Proceedings of the Ninth ACM International Conference on Web Search and Data Mining, ser. WSDM ’16.   New York, NY, USA: Association for Computing Machinery, Feb. 2016, pp. 153–162. [Online]. Available: https://doi.org/10.1145/2835776.2835837
  • [15] X. He, K. Deng, X. Wang, Y. Li, Y. Zhang, and M. Wang, “LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation,” in Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, ser. SIGIR ’20.   New York, NY, USA: Association for Computing Machinery, Jul. 2020, pp. 639–648. [Online]. Available: https://doi.org/10.1145/3397271.3401063
  • [16] F. Wu, T. Zhang, A. H. d. Souza Jr., C. Fifty, T. Yu, and K. Q. Weinberger, “Simplifying Graph Convolutional Networks,” Jun. 2019, arXiv:1902.07153 [cs, stat]. [Online]. Available: http://arxiv.org/abs/1902.07153
  • [17] R. T. Q. Chen, Y. Rubanova, J. Bettencourt, and D. Duvenaud, “Neural ordinary differential equations,” Advances in Neural Information Processing Systems, pp. 6571–6583, 2018.
  • [18] Z. Deng, M. Nawhal, L. Meng, and G. Mori, “Continuous Graph Flow,” Sep. 2019, arXiv:1908.02436 [cs, stat]. [Online]. Available: http://arxiv.org/abs/1908.02436
  • [19] J. Choi, J. Jeon, and N. Park, “LT-OCF: Learnable-Time ODE-based Collaborative Filtering,” in Proceedings of the 30th ACM International Conference on Information & Knowledge Management, ser. CIKM ’21.   New York, NY, USA: Association for Computing Machinery, Oct. 2021, pp. 251–260. [Online]. Available: https://doi.org/10.1145/3459637.3482449
  • [20] H. Pinckaers and G. Litjens, “Neural Ordinary Differential Equations for Semantic Segmentation of Individual Colon Glands,” ArXiv, Oct. 2019. [Online]. Available: https://www.semanticscholar.org/paper/Neural-Ordinary-Differential-Equations-for-Semantic-Pinckaers-Litjens/7593ee0a8242026714b37ba2c2805d1249c82e13
  • [21] M. Zhu, B. Chang, and C. Fu, “Convolutional neural networks combined with Runge–Kutta methods,” Neural Computing and Applications, vol. 35, no. 2, pp. 1629–1643, Jan. 2023. [Online]. Available: https://doi.org/10.1007/s00521-022-07785-2
  • [22] A. Keramatfar, M. Rafiee, and H. Amirkhani, “Graph Neural Networks: A bibliometrics overview,” Machine Learning with Applications, vol. 10, p. 100401, Dec. 2022. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S2666827022000780
  • [23] L.-P. A. C. Xhonneux, M. Qu, and J. Tang, “Continuous Graph Neural Networks,” Jul. 2020, arXiv:1912.00967 [cs, stat]. [Online]. Available: http://arxiv.org/abs/1912.00967
  • [24] M. Poli, S. Massaroli, J. Park, A. Yamashita, H. Asama, and J. Park, “Graph Neural Ordinary Differential Equations,” Jun. 2021, arXiv:1911.07532 [cs, stat]. [Online]. Available: http://arxiv.org/abs/1911.07532
  • [25] K. Mao, J. Zhu, X. Xiao, B. Lu, Z. Wang, and X. He, “UltraGCN: Ultra Simplification of Graph Convolutional Networks for Recommendation,” in Proceedings of the 30th ACM International Conference on Information & Knowledge Management, ser. CIKM ’21.   New York, NY, USA: Association for Computing Machinery, Oct. 2021, pp. 1253–1262. [Online]. Available: https://doi.org/10.1145/3459637.3482291
  • [26] E. Haber and L. Ruthotto, “Stable Architectures for Deep Neural Networks,” CoRR, vol. abs/1705.03341, 2017, arXiv: 1705.03341. [Online]. Available: http://arxiv.org/abs/1705.03341
  • [27] K. Oono and T. Suzuki, “Graph neural networks exponentially lose expressive power for node classification,” ICLR, 2020.
  • [28] J. McAuley, C. Targett, Q. Shi, and A. van den Hengel, “Image-Based Recommendations on Styles and Substitutes,” in Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, ser. SIGIR ’15.   New York, NY, USA: Association for Computing Machinery, Aug. 2015, pp. 43–52. [Online]. Available: https://doi.org/10.1145/2766462.2767755
  • [29] W. Fan, X. Liu, W. Jin, X. Zhao, J. Tang, and Q. Li, “Graph Trend Filtering Networks for Recommendation,” in Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, ser. SIGIR ’22.   New York, NY, USA: Association for Computing Machinery, Jul. 2022, pp. 112–121. [Online]. Available: https://doi.org/10.1145/3477495.3531985
  • [30] Z. Fan, K. Xu, Z. Dong, H. Peng, J. Zhang, and P. S. Yu, “Graph Collaborative Signals Denoising and Augmentation for Recommendation,” Apr. 2023, arXiv:2304.03344 [cs]. [Online]. Available: http://arxiv.org/abs/2304.03344
  • [31] X. Zhou, D. Lin, Y. Liu, and C. Miao, “Layer-refined Graph Convolutional Networks for Recommendation,” Nov. 2022, arXiv:2207.11088 [cs]. [Online]. Available: http://arxiv.org/abs/2207.11088