This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

IGNN-Solver: A Graph Neural Solver for
Implicit Graph Neural Networks

Junchao Lin1, Zenan Ling1 , Zhanbo Feng2, Feng Zhou3, Jingwen Xu1 & Robert C Qiu1
1School of EIC, Huazhong University of Science and Technology
2Department of Computer Science and Engineering, Shanghai Jiao Tong University
3Center for Applied Statistics and School of Statistics, Renmin University of China
Corresponding author
Abstract

Implicit graph neural networks (IGNNs), which exhibit strong expressive power with a single layer, have recently demonstrated remarkable performance in capturing long-range dependencies (LRD) in underlying graphs while effectively mitigating the over-smoothing problem. However, IGNNs rely on computationally expensive fixed-point iterations, which lead to significant speed and scalability limitations, hindering their application to large-scale graphs. To achieve fast fixed-point solving for IGNNs, we propose a novel graph neural solver, IGNN-Solver, which leverages the generalized Anderson Acceleration method, parameterized by a small GNN, and learns iterative updates as a graph-dependent temporal process. Extensive experiments demonstrate that the IGNN-Solver significantly accelerates inference, achieving a 1.5×1.5\times to 8×8\times speedup without sacrificing accuracy. Moreover, this advantage becomes increasingly pronounced as the graph scale grows, facilitating its large-scale deployment in real-world applications.

1 Introduction

Implicit graph neural networks (IGNNs) [20; 33; 7] have emerged as a significant advancement in graph learning frameworks. Unlike traditional graph neural networks (GNNs) that stack multiple explicit layers, IGNNs utilize a single implicit layer formulated as a fixed-point equation. The solution to this fixed-point equation, known as the equilibrium, is equivalent to the output obtained by iterating an explicit layer infinitely. This allows an implicit layer to access infinite hops of neighbors, providing IGNNs with global receptive fields within just one layer [12]. As a result, IGNNs effectively address the long-standing issue of over-smoothing in conventional explicit GNNs and capture of long-range dependencies in graph-structured data [30; 1; 56].

However, existing IGNNs suffer from slow speeds and have difficulty scaling to large-scale graphs [34; 18; 52; 39]. This is primarily because IGNNs derive features by solving for fixed points, demanding substantial computational resources. For example, even on the Citeseer dataset [26] classification task with a small-scale graph, IGNNs require more than 2020 forward iterative computations to nearly converge to a fixed point  [20]. The computational overhead of solving fixed points is amplified by the task scale, resulting in notably slow inference speeds compared to explicit GNNs. This substantial drawback poses challenges for IGNNs in generalizing to large-scale graphs in practical scenarios.

In response to this challenge, we propose a novel graph neural solver for IGNNs, termed IGNN-Solver. It takes the IGNN layer as input, which includes the graph information matrix and layer parameters, and outputs the solution to the corresponding fixed-point equation. Unlike conventional solvers relying on root-finding algorithms like Broyden’s method [2], which compute output features through iterative forward passes in a potentially large GNN, IGNN-Solver offers a distinct advantage by predicting the next iteration step via a tiny graph network. The proposed IGNN-Solver remarkably accelerates the model’s inference speed without compromising accuracy and with only a slight increase in training overhead. This advantage becomes increasingly pronounced as the scale of the graph grows, making it particularly beneficial for deploying IGNNs in large-scale graph tasks.

Our IGNN-Solver comprises two components. First, we introduce a learnable initializer that estimates an optimal initial point for the optimization process. Second, we propose a generalized version of Anderson Acceleration (AA) [2], employing a tiny graph network to model the iterative updates as a sequence of graph-dependent steps. Compared to the solvers proposed for conventional Implicit Neural Networks (INNs) [4; 5; 6], we introduce novel improvements: learning solver parameters through a GNN-based method. This approach circumvents the potential loss of graph information, thereby improving the model’s performance. Moreover, IGNN-Solver has significantly fewer parameters compared to IGNN, and its training is independent of the IGNN’s inference. Consequently, the training of IGNN-Solver proceeds rapidly without sacrificing generalization.

In our experiments, we apply IGNN-Solver to 99 real-world datasets from diverse domains and scales, including 44 large-scale datasets: Amazon-all [51], Reddit [21], ogbn-arxiv [22] and ogbn-products [22]. Our results demonstrate that the IGNN-Solver achieves higher accuracy with reduced inference time, showing a 1.5×1.5\times to 8×8\times speedup, and incurs minimal additional training overhead, constituting only 1%1\% of the IGNN training time.

Our main contributions are summarized as follows:

  • We introduce IGNN-Solver, a method designed to predict the next fixed-point iteration of the IGNN layer via a tiny graph network. This innovative approach mitigates the need for extensive iterations, common in conventional IGNNs, thereby substantially accelerating inference while preserving accuracy and minimizing parameter consumption.

  • Compared to conventional solvers proposed for INNs, we have made a novel improvement in our solver by introducing a tiny GNN to learn the parameters. This not only prevents the potential loss of graph information, but also maintains the simplicity and lightweight characteristics of neural solvers.

  • We validate our approach through extensive experiments on 99 real-world datasets, including 33 large-scale. Our results demonstrate that IGNN-Solver incurs minimal computational overhead (approximately 1% of the total) and achieves up to a 88-fold increase in inference speed without compromising accuracy.

2 Related Work

The typical GNN [26] and its variants [42; 21] have been widely used for graph data modeling in various tasks. Different GNNs have been proposed to utilize attention mechanism [42], neighbors sampling [21], pseudo-coordinates [38] and graph fusion [45]. However, due to issues such as over-smoothing, depth, and bottlenecks, these models typically involve finite aggregation layers [30; 1; 56]. To address it, recent works [20; 33; 7] have developed Implicit Graph Neural Networks, encouraging these models to capture long-range dependencies on graphs. Here, we highlight the contributions of our proposed IGNN-Solver through a detailed comparison with Implicit Graph Neural Networks (IGNNs) and Deep Equilibrium Models (DEQs), both of which are closely related to our approach.

Implicit Graph Neural Networks.

Instead of stacking a series of operators hierarchically, implicit GNNs define their outputs as solutions to nonlinear dynamical systems, which is initially introduced by [20] to tackle challenges associated with learning long-range dependencies in graphs. [34] proposes a new implicit graph model enabling mini-batch training without sacrificing the ability to capture long-range dependencies. Subsequently, [43] introduces a novel approach based on implicit layer to model multi-scale structures on graphs. Additionally, [7] theoretically investigates the well-posedness of the IGNN model from a monotone operator viewpoint.

Although the aforementioned IGNN works well by alleviating the problem of over-smoothing of features by allowing meaningful fixed points to propagate implicitly, the inherent slow inference speed of implicit networks poses a major obstacle to its scalability. The main reason is that the solver of the fixed-point network is inefficient (e.g., the Picard solver used by IGNN [20] and Anderson acceleration solver used by MIGNN [7].), makes the overhead of the fixed-point solver magnified by the task scales. In comparison, our proposed IGNN-Solver accelerates the iteration process of IGNNs, which addresses the most limiting drawback compared to traditional feedforward models.

Another class of IGNNs based on Neural ODEs [13] has emerged to address issues like depth and bottlenecks. For example, [3] models continuous residual layers using GCNs. [40] proposes methods for modeling static and dynamic graphs with GCNs, along with a hybrid approach where latent states evolve continuously between RNN steps in dynamic graphs. [49] tackles the problem of continuous message passing. [10] introduces a graph neural diffusion network based on the discretization of diffusion PDEs on graphs. [41] enhances graph neural diffusion with a source term and connects the model to random walk formulation on graphs. However, they essentially view deep learning on graphs as a continuous diffusion process, which differs from IGNNs based on fixed-point networks, thus requiring manually tuning the termination time and step size of the diffusion equation. In contrast, our IGNN-Solver uses implicit formulations instead of explicit diffusion discretization, which admits an equilibrium corresponding to infinite diffusion steps and expands the receptive field.

Fixed-point Solvers for Deep Equilibrium Models.

Traditional deep learning models use multi-layered networks, while DEQs [4] find a fixed-point of a single layer, representing the network’s equilibrium state. Viewed as infinitely deep networks, DEQs use root-finding for forward propagation and implicit differentiation for backward propagation. Therefore, DEQs capture complex patterns with constant memory and computational costs [31]. Monotone operator theory has been used to guarantee the convergence of DEQs [46] and to improve the stability of implicit networks [6].

However, it is well-known that DEQs, as typical implicit models, suffer from slow training and inference speeds [32; 34], which is highly disadvantageous for large-scale scenarios like GraphGPT [58]. To alleviate this issue, recent efforts have explored certain improved solver methods for DEQs, further optimizing root-finding problems and making these models easier to solve: [23] discusses the amortization of the cost of the iterative solver that would otherwise make implicit models slow. [18] proposes a novel gradient estimate for implicit models, named phantom gradient, which significantly accelerates the backward passes in training implicit models. [6] discusses the superiority of learnable solvers over generic solvers in implicit models using a tiny neural network and significantly enhances the efficiency of such models through custom neural solvers.

In contrast, IGNN has a similar network representation to DEQ, both defining the output as the solution of the equation to obtain network outputs. But the difference lies in the fact that IGNN’s equilibrium equations encode graph structure while DEQ does not, which will undoubtedly deepen its weakness of slow inference speed and make IGNN’s solution slower. Insight on it, our proposed IGNN-Solver leverages graph information to guide solver acceleration, achieving fast and meaningful implicit graph network propagation, especially in the case of graph data being large-scale.

3 Preliminaries

Explicit GNNs.

Let 𝒢={𝑨,𝑿}\mathcal{G}=\{{\bm{A}},{\bm{X}}\} represents an undirected graph, where 𝑨n×n{\bm{A}}\in\mathbb{R}^{n\times n} is the adjacency matrix indicating the relationships between the nodes in 𝒱={𝒗1,𝒗2,,𝒗n}\mathcal{V}=\{{\bm{v}}_{1},{\bm{v}}_{2},\ldots,{\bm{v}}_{n}\}, and nn is the number of nodes. The node feature matrix 𝑿=[𝒙1,𝒙2,,𝒙n]n×d{\bm{X}}=[{\bm{x}}_{1},{\bm{x}}_{2},\ldots,{\bm{x}}_{n}]\in\mathbb{R}^{n\times d} contains the features of the nodes, with dd representing the feature dimension. Conventional (explicit) GNNs [26; 42] feature a learnable aggregation process centered on the message-passing operation within the graph [19]. This process iteratively propagates information from each node to its neighboring nodes. The formal general structure for each layer ll is defined as follows:

𝑯[l+1]=fθ(𝑨,𝑯[l]),𝑯[0]=𝑿,{\bm{H}}^{[l+1]}=f_{\theta}({{\bm{A}}},{\bm{H}}^{[l]}),\quad{{\bm{H}}^{[0]}}={\bm{X}}, (1)

where 𝑯[l]{{\bm{H}}^{[l]}} represents the hidden node representation, fθf_{\theta} denotes the parameters in ll-th layer. A commonly used GNN is Graph Convolutional Network (GCN) [26], defined as 𝑯[l+1]=σ(𝑨^𝑯[l]𝑾[l]){\bm{H}}^{[l+1]}=\sigma(\hat{{\bm{A}}}{\bm{H}}^{[l]}{\bm{W}}^{[l]}), where 𝑾[l]{{\bm{W}}^{[l]}} denotes the weight matrix of the ll-th layer, σ()\sigma(\cdot) denotes the activation function, 𝑨^=𝑫~1/2(𝑨+𝑰)𝑫~1/2{\hat{{\bm{A}}}}={\tilde{{\bm{D}}}^{-1/2}}({\bm{A}}+{\bm{I}}){\tilde{{\bm{D}}}^{-1/2}} represents the symmetric normalized graph matrix. Here, 𝑫~\tilde{{\bm{D}}} is a diagonal matrix with 𝑫~ii=1+j𝑨ij\tilde{{\bm{D}}}_{ii}=1+\sum\nolimits_{j}{\bm{A}}_{ij}. GNNs leverage the above message-passing operation in Eq. 1 to learn useful information. Still, they often involve a limited number of ll layers due to over-smoothing [1], making it challenging for GNNs to capture the long-range dependency on graphs.

Implicit GNNs.

Similar to traditional explicit GNNs, implicit GNNs [20; 33; 7; 12] also involve an aggregation process, with the distinction that the depth of layers (iteration step kk) is infinite. The aggregation process in IGNNs is typically defined as 𝒁[k+1]=σ(𝑨^𝒁[k]𝑾+bΩ(𝑿)),k=1,2,,{\bm{Z}}^{[k+1]}=\sigma(\hat{{\bm{A}}}{\bm{Z}}^{[k]}{\bm{W}}+b_{\Omega}({\bm{X}})),k=1,2,\ldots,\infty, where bΩb_{\Omega} represents affine transformation parameterized by 𝛀\bm{\Omega}, and the weight matrices 𝑾{\bm{W}} and 𝛀\bm{\Omega} are globally shared at each iteration step. The IGNN model fθf_{\theta} is formally described by

𝒁=fθ(𝒁,𝑨^,𝑿),{\bm{Z}}^{\star}=f_{{\theta}}({\bm{Z}}^{\star},\hat{{\bm{A}}},{\bm{X}}), (2)

where the representation, given as the “internal state” 𝒁{\bm{Z}}^{\star}, is obtained as the fixed-point solution of the equilibrium equation 2. Consequently, the final representation theoretically encompasses information from all neighbors in the graph. In practice, IGNNs capture long-range dependencies within the graph, offering better performance compared to GNNs with finite iterations [20; 33]. Another notable advantage of this framework is its memory efficiency, as it only needs to retain the current state 𝒁{\bm{Z}} without requiring additional intermediate representations.

For training parameters 𝜽={𝑾,𝛀}\bm{\theta}=\{{\bm{W}},\bm{\Omega}\} during the back-propagation process of neural networks, IGNNs compute the Jacobian 𝒁\nabla_{\bm{Z}}\mathcal{L} by solving the following equilibrium equation

𝒁=𝑫(𝑾𝒁𝑨^+𝑿),\nabla_{\bm{Z}}\mathcal{L}={\bm{D}}\odot({\bm{W}}^{\top}\nabla_{\bm{Z}}\mathcal{L}\hat{{\bm{A}}}^{\top}+\nabla_{\bm{X}}\mathcal{L}), (3)

where 𝑫=ϕ(𝑾𝑿𝑨^+bΩ(𝑼)){\bm{D}}=\phi^{\prime}({\bm{W}}{\bm{X}}\hat{{\bm{A}}}+b_{\Omega}({\bm{U}})) and ϕ()\phi^{\prime}(\cdot) refers to the element-wise derivative of the map ϕ\phi. Once 𝒁\nabla_{\bm{Z}}\mathcal{L} is obtained, we can use the chain rule and the implicit function theorem [28] to easily compute 𝑾\nabla_{\bm{W}}\mathcal{L} and 𝛀\nabla_{\bm{\Omega}}\mathcal{L}.

In summary, IGNNs usually require numerous iterations to obtain equilibrium, resulting in significant overhead during training and inference. Therefore, it is crucial to find a faster and more efficient method for solving the fixed-point equation of IGNNs.

Refer to caption
Figure 1: The overall architecture of the IGNN-Solver compared with the generic Anderson solver at each iteration.

4 Implicit graph neural networks Solver

Although traditional fixed-point solvers for IGNNs (as discussed in Section 1 and  2) demonstrate functionality, they are characterized by relatively slow performance, necessitate manual parameter tuning, and are not well-suited for graph-based applications. To address these limitations, we have designed a lightweight, learnable, and content-aware fixed-point solver for IGNN, which integrates a tiny graph neural network, combining the speed advantages of the solver with the information benefits of graph learning. It encodes the relationship between past residuals and the parameters α[k]\alpha^{[k]} and β\beta, allowing for more efficient adjustments of the parameters α\alpha and β\beta to improve the learning capabilities of the IGNN-Solver, as illustrated in Figure 1.

Due to the characteristics of the implicit model—where its representation capability is independent of the forward computation (i.e., the solver has no knowledge of the original task, such as predicting the node’s class, and vice versa)—, we can train this neural solver in a lightweight and unsupervised manner. After the IGNN-Solver is learned, we can utilize it to solve the fixed-point equation of the IGNN layer, thereby further updating the model. The whole training strategy with IGNN-Solver can be found in Appendix B.1 and B.2.

Algorithm 1 IGNN-Solver Iterations and Training
1:frozen IGNN model fθf_{\theta}, initializer hϕh_{\phi}, tiny-GNN predictor sξs_{\xi}, sparse subgraph 𝑨^s\hat{{\bm{A}}}_{s}, storage 𝑮(m+1)×d{\bm{G}}\in\mathbb{R}^{(m+1)\times d^{\prime}}.
2:Compute the initial value of fixed point by 𝒁[0]=hϕ(𝑿)d{\bm{Z}}^{[0]}=h_{\phi}({\bm{X}})\in\mathbb{R}^{d^{\prime}}
3:Define gθ(𝒁)=fθ(𝒁)𝒁g_{\theta}({\bm{Z}})=f_{\theta}({\bm{Z}})-{\bm{Z}} and set 𝑮[0]=gθ(𝒁[0]){\bm{G}}[0]=g_{\theta}({\bm{Z}}^{[0]})
4:for k=0,,Kk=0,\ldots,K do
5:     1) Set mk=min{m,k}m_{k}=\min\{m,k\} and 𝑮[k]=𝑮[0:(mk+1)](mk+1)×d{\bm{G}}^{[k]}={\bm{G}}\left[0:\left(m_{k}+1\right)\right]\in\mathbb{R}^{\left(m_{k}+1\right)\times d^{\prime}}
6:     2) Predict α[k],β[k]=sξ(𝑮[k],𝑨^s){\alpha}^{[k]},{\beta}^{[k]}=s_{\xi}({\bm{G}}^{[k]},\hat{{\bm{A}}}_{s}), where α[k](mk+1){\alpha}^{[k]}\in\mathbb{R}^{(m_{k}+1)}, 𝟏α[k]=1\mathbf{1}^{\top}{\alpha}^{[k]}=1 and β[k]1\beta^{[k]}\in\mathbb{R}^{1}
7:     3) Update 𝒁[k+1]=β^[k]𝟏𝑮[k]+i=0mkα^i[k]𝒁[kmk+i]{\bm{Z}}^{[k+1]}=\hat{\beta}^{[k]}\cdot\mathbf{1}^{\top}{\bm{G}}^{[k]}+\sum_{i=0}^{m_{k}}\hat{\alpha}_{i}^{[k]}{\bm{Z}}^{[k-m_{k}+i]} \triangleright (Anderson step)
8:     4) Update 𝑮=concat(𝑮[1:],[gθ(𝒁[k+1])]){\bm{G}}=\operatorname{concat}\left({\bm{G}}[1:],[g_{\theta}({\bm{Z}}^{[k+1]})]\right)
9:end for
10:if it is inference stage then
11:     return 𝒁[k+1]{\bm{Z}}^{[k+1]}
12:else
13:     return (𝒁[k+1],𝑮[k],α[k],β[k])k=0,,K({\bm{Z}}^{[k+1]},{\bm{G}}^{[k]},{\alpha}^{[k]},{\beta}^{[k]})_{k=0,\ldots,K} and 𝒁[0]{\bm{Z}}^{[0]}
14:     Compute total \mathcal{L}_{\text{total }} and back-propagate it to update IGNN-Solver {sξ,hϕ}\left\{s_{\xi},h_{\phi}\right\}
15:end if

4.1 General Formulation

For a given IGNN layer fθf_{\theta}, input features 𝑿n×d{\bm{X}}\in\mathbb{R}^{n\times d} and graph 𝑨^n×n\hat{{\bm{A}}}\in\mathbb{R}^{n\times n}, we assume that the equation 𝒁=fθ(𝒁,𝑨^,𝑿)=σ(𝑨^𝒁𝑾+bΩ(𝑿)){\bm{Z}}=f_{\theta}({\bm{Z}},\hat{{\bm{A}}},{\bm{X}})=\sigma(\hat{{\bm{A}}}{\bm{Z}}{\bm{W}}+b_{\Omega}({\bm{X}})) has an exact solution at its fixed point. This solution can be obtained using classic solvers (such as Broyden’s method) and running multiple iterations to achieve high precision.

The overall structure of the IGNN-Solver is shown in Algorithm 1. Specifically, it learns the parameters α\alpha that control the weights among past mm steps approximate solution, and β\beta that control the weights between past residuals and approximate solution, through a solver sξs_{\xi} to accelerate the approximate solution of the next update step. To train the IGNN-Solver, we minimize the joint loss function total\mathcal{L}_{\text{total}} by back-propagating through this KK-step temporal process, which is discussed in Section 4.2. It’s worth noting that the original IGNN is frozen (i.e., model parameters θ\theta are fixed), and only the IGNN-Solver parameters ξ\xi are trained here, so we do not need the ground-truth label yy that corresponds to input xx. This implies that IGNN-Solver can also be fine-tuned during inference after deployment.

4.1.1 Initializer.

To accelerate the convergence of predicted values, we propose constructing an initializer 𝒁[0]=hϕ(𝑿):dd{\bm{Z}}^{[0]}=h_{\phi}({\bm{X}}):\mathbb{R}^{d}\rightarrow\mathbb{R}^{d^{\prime}} for a rapid and reasonable input-based initial estimate value, rather than simply setting the initial values to 𝟎\mathbf{0} or random, where ϕ\phi are parameters of the initializer. We set the intermediate dimension dd^{\prime} to be significantly smaller than the feature dimension dd to reduce the training overhead of the initializer.

4.1.2 Improved Anderson Iterations with tiny-GNN.

In the original Anderson Acceleration (AA) solver, the parameter β\beta is predetermined and fixed [2], whereas the parameters α[k]\alpha^{[k]}s are determined by solving numerous linear equations using the least squares method. Although this method operates adequately, its efficiency and adaptability in optimizing IGNN models are limited. Therefore, we propose introducing a tiny and learnable graph neural network, as

α,β=sξ(𝑮,𝑨^), wheresξ:((mk+1)×d×n)((mk+1)×1),\alpha,\beta=s_{\xi}({\bm{G}},\hat{{\bm{A}}})\,\text{, where}\,s_{\xi}:(\mathbb{R}^{(m_{k}+1)\times d^{\prime}}\times\mathbb{R}^{n})\rightarrow(\mathbb{R}^{(m_{k}+1)}\times\mathbb{R}^{1}), (4)

to predict the two parameters instead of setting them as the least squares solution on the past residuals 𝑮{\bm{G}}. However, directly using the original graph for modeling would result in an oversized network [53; 59], deviating from the initial goal of using a tiny network for prediction. On the other hand, because of the curse of dimensionality, directly mapping the data from high to extremely low dimensions is not appropriate [27]. Therefore, in response to the challenges posed by large-scale graphs, we took into account both Graph Sparsification and Storage Compression in the design of the solver, as detailed below:

Graph Sparsification.

Firstly, the number of edges in graph 𝑨^\hat{{\bm{A}}} is usually large in practice, as it is affected by the scale of the input (for example, in the ogbn-products dataset, the number of edges in the graph is close to 62M). To maintain the lightweight nature of tiny-GNN and reduce the computational cost of prediction, we propose introducing graph sparsification [24; 36; 55] for a more light graph:

𝑨^s=ς(𝑨^,β),\hat{{\bm{A}}}_{s}=\varsigma(\hat{{\bm{A}}},\beta), (5)

where 𝑨^s\hat{{\bm{A}}}_{s} represents the sparse subgraph of 𝑨^\hat{{\bm{A}}}, obtained via a graph sparsification operator ς\varsigma, with the parameter β\beta controls the sparsity of 𝑨^s\hat{{\bm{A}}}_{s}. We choose the RPI-Graph approach [53] in this paper, which is a plug-and-play graph sparsification method based on the principle of relevant information. It is worth noting that this subgraph is used only in tiny-GNN predictor sξs_{\xi}, and not in IGNN model fθf_{\theta}.

Storage Compression.

Besides, considering that nn (the number of nodes) is relatively large (for example, nn is approximately 169K in the ogbn-arxiv dataset), it is not appropriate to map sξs_{\xi} from a high-dimensional space to a very low-dimensional one in equation 4 directly [45; 57; 44] owing to the curse of dimensionality [27]. Therefore, to maintain sξs_{\xi} fast and compact, we recommend compress 𝑮i[k]{\bm{G}}^{[k]}_{i} to form a smaller yet still representative version 𝑮¯i[k]\bar{{\bm{G}}}^{[k]}_{i}. Specifically, We map it from (mk+1)×d\mathbb{R}^{(m_{k}+1)\times d^{\prime}} to an appropriate space (mk+1)×p\mathbb{R}^{(m_{k}+1)\times p} by multiple layer perception (MLP). Then, the nearest mk+1m_{k}+1 sets of fixed points are merged into one feature matrix:

𝑮¯[k]=||i=kmkk𝑮¯i[k],\bar{{\bm{G}}}^{[k]}=\mathop{||}_{i=k-m_{k}}^{k}\bar{{\bm{G}}}^{[k]}_{i}, (6)

where ||i=kmkk\mathop{||}_{i=k-m_{k}}^{k} represents the concatenation operation that stack the compressed storage matrix 𝑮¯i[k](mk+1)×p\bar{{\bm{G}}}^{[k]}_{i}\in\mathbb{R}^{(m_{k}+1)\times p} in the kk-th iteration.

Once this smaller but still representative storage matrix 𝑮¯\bar{{\bm{G}}} obtained by above, we treat it as a pp-channel matrix and employ a tiny graph neural network layer

α[k],β[k]=sξ(𝑮¯[k],𝑨^s),\alpha^{[k]},\beta^{[k]}=s_{\xi}(\bar{{\bm{G}}}^{[k]},\hat{{\bm{A}}}_{s}), (7)

to predict the relative weights α[k]\alpha^{[k]} assigned to past mk+1m_{k}+1 residual, as well as the AA mixing coefficient β[k]\beta^{[k]} at the kk-th iteration.

In this way, sξs_{\xi} shall autonomously learn and adjust these parameters α[k]\alpha^{[k]} and β[k]\beta^{[k]} based on the previous solver steps and receiving gradients from subsequent iterations. We provide a comparison of different choices for sξs_{\xi} in Appendix D, demonstrating that IGNN-Solver improves the convergence path. We also introduce how to train it (see below).

4.2 Training the IGNN-Solver

Unlike the neural ODE solver, the fixed-point trajectory of the IGNN is not unique. Therefore, trajectory fitting is not applicable to the IGNN-solver training. Instead, the goal is simply to bring everything as close as possible to 𝒁{\bm{Z}}^{\star}. Formally, given an IGNN-Solver {sξ,hϕ}\left\{s_{\xi},h_{\phi}\right\} that returns (𝒁[k+1],𝑮[k],α[k],β[k])k=0,,K({\bm{Z}}^{[k+1]},{\bm{G}}^{[k]},{\alpha}^{[k]},{\beta}^{[k]})_{k=0,\ldots,K} and 𝒁[0]{\bm{Z}}^{[0]}, we introduce three objectives functions for its training. The complete training algorithm can be referred to in Algorithm 1.

Initializer Loss Functions.

To train the initializer, we minimize the distance between the initial guess and the fixed-point by

init =hϕ(𝑿)𝒁2,\mathcal{L}_{\text{init }}=\left\|h_{\phi}({\bm{X}})-{\bm{Z}}^{\star}\right\|_{2}, (8)

for facilitating the subsequent iteration process. Since the initializer directly predicts based on the input 𝑿{\bm{X}} without going through iteration, we separate this loss from other components.

Reconstruction Loss Functions.

The training of the solver does not require reference to label information or any trajectory fitting. Apart from the loss init\mathcal{L}_{\text{init}} necessary for training the initializer above, we introduce reconstruction loss

rec[k]=𝒁[k]𝒁2,\mathcal{L}_{\text{rec}}^{[k]}=\|{\bm{Z}}^{[k]}-{\bm{Z}}^{\star}\|_{2},\quad (9)

where the reconstruction loss rec [k]\mathcal{L}_{\text{rec }}^{[k]} aims to make all intermediate predictions 𝒁[k]{\bm{Z}}^{[k]} converge to the accurate fixed point 𝒁{\bm{Z}}^{\star} as closely as possible.

Auxiliary Loss Functions.

Although we utilize α[k]\alpha^{[k]} and β[k]\beta^{[k]} to improve the generic solver, we have empirically found that using an auxiliary loss to guide the solver’s prediction of α\alpha is beneficial sometimes, especially in the early stages of training. Therefore, in practice, we suggest considering this loss and gradually diminishing it as training progresses (i.e., decay the weight of this loss to 0).

α=k=0K𝑮[k]α[k]2,\mathcal{L}_{\alpha}=\sum_{k=0}^{K}\|{\bm{G}}^{[k]}\alpha^{[k]}\|_{2}, (10)

The final form of the joint loss function is as follows:

total =λ1rec [k]+λ2init +λ3α,\mathcal{L}_{\text{total }}=\lambda_{1}\mathcal{L}_{\text{rec }}^{[k]}+\lambda_{2}\mathcal{L}_{\text{init }}+\lambda_{3}\mathcal{L}_{\alpha}, (11)

where λ1\lambda_{1}, λ2\lambda_{2} and λ3\lambda_{3} control the weights of three loss functions mentioned above and elaborate more in Appendix C.3.

5 Experiments

In this section, we compare the speed/accuracy Pareto curve rather than a single point on the curve of IGNN-Solver with IGNNs and several state-of-the-art (SOTA) GNNs on various graph classification tasks at both node and graph levels. We aim to show that 1) IGNN-Solver can achieve nearly a 1.5×1.5\times to 8×8\times inference acceleration without sacrificing accuracy, and 2) IGNN-Solver are extremely compact and add little overhead to training.

Specifically, we compare our approach against 1212 representative baselines and 22 variants: IGNN w. AA (using original AA to accelerate IGNN), and IGNN w. NN (using a standard NN solver to replace our proposed graph neural solver for parameter learning) on 99 different-field and real-world node classification tasks, including 44 citation datasets: Citeseer, ACM, CoraFull, ogbn-arxiv, 33 social interaction datasets: BlogCatalog, Flickr, Reddit, and 22 product network datasets Amazon-all, ogbn-products. More details about the datasets are given in Appendix C.1. Notably, to demonstrate the scalability of the proposed model to larger datasets, we use 44 large-scale datasets: Amazon-all, Reddit, ogbn-arxiv and ogbn-products, two of which are from the Open Graph Benchmark (OGB) [22]111The code to reproduce the results in this section has been uploaded as supplementary material..

Additionally, in Section 5.2, we demonstrate that the training overhead on IGNN-Solver is minimized relative to the total time overhead on IGNN. In Section 5.3, we provide additional evidence regarding the convergence and generalizability of the IGNN-Solver. In Section 5.4, we conduct an ablation study to validate the stability of the IGNN-Solver and the effectiveness of its components. Experimental setups, descriptions and additional results are detailed in Appendices BC and D.

Refer to caption
(a) Amazon-all
Refer to caption
(b) Reddit
Refer to caption
(c) ogbn-arxiv
Refer to caption
(d) ogbn-products
Figure 2: Comparison of inference time with IGNN-Solvers on 44 large-scale datasets. More results on Citeseer, ACM, CoraFull, BlogCatalog, and Flickr can be found in Figure 6, Appendix D.1. All speed/accuracy curves within the plot are benchmarked on the same GPU with the same experimental setting, averaged over 55 independent runs.
Table 1: Node classification accuracy results on 44 large-scale real-world datasets (more results on small-scale datasets including Citeseer, ACM, CoraFull, BlogCatalog, Flickr can be found in Table 3, Appendix D.1), with experiments conducted over five trials. The mean accuracy ± standard deviation (%) is reported. The best and the runner-up results are highlighted in boldface and underline, respectively.
Type *Models Amazon-all Reddit ogbn-arxiv ogbn-products
*Nodes 334,863334,863 232,965232,965 169,343169,343 2,449,0292,449,029
*Edges 14,202,05714,202,057 11,606,91911,606,919 1,166,2431,166,243 61,859,14061,859,140
Explicit GCN [26] 79.12±1.1179.12\pm 1.11 89.65±0.1489.65\pm 0.14 71.56±0.2571.56\pm 0.25 71.91±0.2771.91\pm 0.27
GAT [42] 76.02±2.0976.02\pm 2.09 90.08±0.4490.08\pm 0.44 71.10±0.2471.10\pm 0.24 72.33±0.3272.33\pm 0.32
SGC [47] 75.56±1.7575.56\pm 1.75 91.44±0.4191.44\pm 0.41 64.66±1.0264.66\pm 1.02 70.48±0.1970.48\pm 0.19
APPNP [17] 79.80±1.2279.80\pm 1.22 91.68±0.2491.68\pm 0.24 71.28±0.2971.28\pm 0.29 74.46±0.6474.46\pm 0.64
JKNet [50] 81.19±1.0981.19\pm 1.09 91.71±0.3191.71\pm 0.31 71.08±0.3571.08\pm 0.35 73.70±0.5873.70\pm 0.58
AM-GCN [45] 81.85±1.4681.85\pm 1.46 90.20±0.3490.20\pm 0.34 68.83±0.6768.83\pm 0.67 73.19±0.4973.19\pm 0.49
DEMO-Net [48] 84.08±1.2984.08\pm 1.29 89.53±0.2989.53\pm 0.29 68.40±0.2468.40\pm 0.24 73.88±0.6273.88\pm 0.62
GCNII [11] 83.60±2.4083.60\pm 2.40 89.87±0.3889.87\pm 0.38 67.66±0.2067.66\pm 0.20 72.88±0.1872.88\pm 0.18
ACM-GCN [35] 83.04±2.6183.04\pm 2.61 90.37±0.4290.37\pm 0.42 68.32±0.1868.32\pm 0.18 71.68±0.4171.68\pm 0.41
Implicit IGNN [20] 83.90±0.5183.90\pm 0.51 92.30±1.5592.30\pm 1.55 70.49±0.7570.49\pm 0.75 74.63±0.2474.63\pm 0.24
EIGNN [33] 84.32¯±0.57\underline{84.32}\pm 0.57 92.00±0.2492.00\pm 0.24 70.59±0.3170.59\pm 0.31 74.58±0.2674.58\pm 0.26
MIGNN [7] 83.68±0.8283.68\pm 0.82 91.98±0.4291.98\pm 0.42 71.95¯±0.44\underline{71.95}\pm 0.44 74.62±0.3274.62\pm 0.32
IGNN w. AA 83.44±0.1983.44\pm 0.19 92.37±0.3592.37\pm 0.35 71.73±0.4171.73\pm 0.41 74.61±0.2874.61\pm 0.28
IGNN w. NN 84.13±1.0484.13\pm 1.04 92.42¯±0.41\underline{92.42}\pm 0.41 70.78±0.3770.78\pm 0.37 74.69¯±0.31\underline{74.69}\pm 0.31
IGNN-Solver 84.50±0.70\textbf{84.50}\pm 0.70 93.91±0.31\textbf{93.91}\pm 0.31 72.53±0.41\textbf{72.53}\pm 0.41 74.90±0.20\textbf{74.90}\pm 0.20

5.1 Performance and Efficiency Comparison

In order to demonstrate the superiority of the IGNN-Solver over IGNNs in terms of both performance and efficiency, we specifically analyze the movement of the speed/accuracy Pareto curve across various datasets rather than concentrating on a single point on the curve, which is depicted in Figure 2 and 6 (see Appendix D.1). All experiments are conducted five times, and the best results are reported. The training procedure details and hyperparameters used in each task are provided in Appendix C.

From Figure 2 (results on large-scale datasets: Amazon-all, Reddit, ogbn-arxiv and ogbn-products), we can observe that: 1) Regarding comprehensive performance and efficiency, IGNN-Solver generally outperforms all other methods, especially in large datasets with a greater graph radius, where this advantage is often more pronounced. This is attributed to the IGNN-Solver’s significant improvement in the convergence path and the enhancement of the fixed-point equation’s initial point through the initializer; 2) The rapid inference advantage of the IGNN-Solver. For instance, in the large-scale dataset Amazon-all, our IGNN-Solver achieves 8×8\times faster inference speed than IGNN while maintaining the same performance, and there is consistently at least 1.5×1.5\times acceleration, and a similar pattern has been observed as well on other datasets. Please see more results on other small-scale datasets in Figure 6 of Appendix D.1.

Furthermore, we present the node classification performance results for each method on large-scale datasets(including Amazon-all, Reddit, ogbn-arxiv and ogbn-products) in Table 1 and other small-scale datasets in Table 3 (see Appendix D.1). All experiments are conducted five times, and the average results along with the standard deviation are reported. The training procedure details and hyperparameters used in each task are provided in Appendix C. It can be observed that among all methods, implicit GNNs with infinite depth are often superior to shallow explicit GNNs in most cases. Besides, IGNN-Solver consistently shows higher accuracy percentages across most datasets compared to other SOTA explicit GNNs like DEMO-Net [48], GCNII [11], ACM-GCN [35] etc, which indicates their superior performance.

5.2 Model Size in Training

Refer to caption
(a) Overhead proportion on Citeseer.
Refer to caption
(b) Overhead proportion on Amazon-all.
Figure 3: Relative parameter size and training time of IGNN-Solver on small dataset Citeseer and large one Amazon-all, with similar patterns observed in other datasets.

To investigate the proportion of training overhead attributed to the solver throughout the entire training process, we depict in Figure 3 the percentage of training computations on IGNN-Solver relative to the total time overhead on IGNN. Our approach is not only effective but also incurs minimal training overhead for the solver: the solver module is relatively small, and requires only about 1%1\% of the original training time required by the IGNN model. Besides, this proportion will be lower for large-scale data. For instance, on the Amazon dataset, IGNN necessitates a total runtime of 33 hours, whereas the solver only requires approximately 1.61.6 minutes, indicating that our solver maintains its lightweight nature across any-scale datasets.

5.3 More Efficiency Study

Refer to caption
Figure 4: The convergence curve of IGNN-Solver and IGNN on ogbn-arxiv. The former improves the convergence rate.

We provide additional evidence on the convergence and generalizability of the neural solvers in Figure 4, where we compare the convergence of a pre-trained IGNN model under 1) canonical iteration in IGNN [20]; and 2) IGNN-Solver with K=6K=6.

From Figure 4, we first note that the IGNN-Solver trained with KK unrolling steps can generalize beyond KK, and both solvers eventually reach a stable state, demonstrating good convergence. Secondly, thanks to the fixed-point solution of the graph neural parameterization, we observe that our IGNN-Solver continuously enhances the convergence of the canonical iterators while being more computationally efficient (i.e., each step of the neural solver is cheaper than the standard iterator step). This explains the observed improvement in inference efficiency of at least 1.5×1.5\times in Section 5.1.

Refer to caption
Figure 5: Ablation studies for IGNN-Solver.

5.4 Ablation Study

Here, we analyze the effect of different loss components of the IGNN-Solver. For the main convergence loss conv\mathcal{L}_{\text{conv}} of fixed-point iterations in equation 11, we designed two contrasting schemes concerning its weight λ1[k]\lambda_{1}^{[k]}s: setting them to a constant value (i.e., λ1[k]=λ1\lambda_{1}^{[k]}=\lambda_{1} for all kk), or setting all values except the KK-th term to zero (i.e., λ1[k]=λ1\lambda_{1}^{[k]}=\lambda_{1} if k=Kk=K else 0). The results of IGNN-Solver and its variants on Citeseer are shown in Figure 5. The patterns observed in other datasets and solvers exhibit similarities.

It is observed that the two variant solvers still perform well, yet our suggested monotonically increasing scheme (i.e., emphasizing the later steps to a greater extent) exhibits the best performance. Additionally, removing the initializer or the alpha loss α\mathcal{L}_{\alpha} will correspondingly affect performance, with the former having a more detrimental impact on the solver. Nonetheless, the performance of all these ablation settings remains significantly superior to methods without a solver, indicating the advantages of employing a custom learnable solver for the IGNN model.

6 Conclusion

This paper introduces IGNN-Solver, a novel graph neural solver tailored for achieving fast fixed-point solving in IGNNs. By leveraging the generalized Anderson Acceleration method and parameterizing it with a tiny GNN, IGNN-Solver learns iterative updates as a graph-dependent temporal process. Our extensive experiments demonstrate the significant acceleration in inference achieved by IGNN-Solvers, with a speedup ranging from 1.5×1.5\times to 8×8\times, all while maintaining accuracy. Notably, this acceleration is particularly pronounced as the scale of the graph increases. These findings underscore the potential of IGNN-Solver for large-scale end-to-end deployment in real-world applications.

Limitations.

In contrast to previous works on IGNNs [20; 33; 7], the theoretical guarantee of fixed-point existence and stability in IGNN-Solver remains uncertain, notwithstanding favorable empirical findings in Figure D.2. This is because we make no specific constraint on the formulation of IGNNs, a departure from the approach adopted in previous IGNN-related studies.

Acknowledgments

R. C. Qiu would like to acknowledge the National Natural Science Foundation of China (via NSFC-12141107), the Interdisciplinary Research Program of HUST (2023JCYJ012), the Key Research and Development Program of Guangxi (GuiKe-AB21196034). F. Zhou was supported by the National Natural Science Foundation of China Project (via NSFC-62106121) and the MOE Project of Key Research Institute of Humanities and Social Sciences (22JJD110001).

Z. Ling is supported by the National Natural Science Foundation of China (via NSFC-62406119), the Natural Science Foundation of Hubei Province (2024AFB074), and the Guangdong Provincial Key Laboratory of Mathematical Foundations for Artificial Intelligence (2023B1212010001).

References

  • Alon & Yahav [2020] Uri Alon and Eran Yahav. On the bottleneck of graph neural networks and its practical implications. In International Conference on Learning Representations, 2020.
  • Anderson [1965] Donald G Anderson. Iterative procedures for nonlinear integral equations. Journal of the ACM (JACM), 12(4):547–560, 1965.
  • Avelar et al. [2019] Pedro HC Avelar, Anderson R Tavares, Marco Gori, and Luis C Lamb. Discrete and continuous deep residual learning over graphs. arXiv preprint arXiv:1911.09554, 2019.
  • Bai et al. [2019] Shaojie Bai, J Zico Kolter, and Vladlen Koltun. Deep equilibrium models. Advances in neural information processing systems, 32, 2019.
  • Bai et al. [2020] Shaojie Bai, Vladlen Koltun, and J Zico Kolter. Multiscale deep equilibrium models. In Proceedings of the 34th International Conference on Neural Information Processing Systems, pp.  5238–5250, 2020.
  • Bai et al. [2021] Shaojie Bai, Vladlen Koltun, and J Zico Kolter. Neural deep equilibrium solvers. In International Conference on Learning Representations, 2021.
  • Baker et al. [2023] Justin Baker, Qingsong Wang, Cory D Hauck, and Bao Wang. Implicit graph neural networks: a monotone operator viewpoint. In International Conference on Machine Learning, pp. 1521–1548. PMLR, 2023.
  • Bojchevski & Günnemann [2018] Aleksandar Bojchevski and Stephan Günnemann. Deep gaussian embedding of graphs: Unsupervised inductive learning via ranking. In International Conference on Learning Representations, pp. 1–13, 2018.
  • Broyden [1965] Charles G Broyden. A class of methods for solving nonlinear simultaneous equations. Mathematics of computation, 19(92):577–593, 1965.
  • Chamberlain et al. [2021] Ben Chamberlain, James Rowbottom, Maria I Gorinova, Michael Bronstein, Stefan Webb, and Emanuele Rossi. Grand: Graph neural diffusion. In International conference on machine learning, pp. 1407–1418. PMLR, 2021.
  • Chen et al. [2020] Ming Chen, Zhewei Wei, Zengfeng Huang, Bolin Ding, and Yaliang Li. Simple and deep graph convolutional networks. In International conference on machine learning, pp. 1725–1735. PMLR, 2020.
  • Chen et al. [2022] Qi Chen, Yifei Wang, Yisen Wang, Jianlong Chang, Qi Tian, Jiansheng Yang, and Zhouchen Lin. Efficient and scalable implicit graph neural networks with virtual equilibrium. In 2022 IEEE International Conference on Big Data (Big Data), pp.  864–873. IEEE, 2022.
  • Chen et al. [2018] Ricky TQ Chen, Yulia Rubanova, Jesse Bettencourt, and David K Duvenaud. Neural ordinary differential equations. Advances in neural information processing systems, 31, 2018.
  • Cho et al. [2014] Kyunghyun Cho, Bart van Merriënboer, Dzmitry Bahdanau, and Yoshua Bengio. On the properties of neural machine translation: Encoder–decoder approaches. In Dekai Wu, Marine Carpuat, Xavier Carreras, and Eva Maria Vecchi (eds.), Proceedings of SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, pp.  103–111, Doha, Qatar, October 2014. Association for Computational Linguistics. doi: 10.3115/v1/W14-4012. URL https://aclanthology.org/W14-4012.
  • De Boer et al. [2005] Pieter-Tjerk De Boer, Dirk P Kroese, Shie Mannor, and Reuven Y Rubinstein. A tutorial on the cross-entropy method. Annals of operations research, 134:19–67, 2005.
  • Feng et al. [2020] Wenzheng Feng, Jie Zhang, Yuxiao Dong, Yu Han, Huanbo Luan, Qian Xu, Qiang Yang, Evgeny Kharlamov, and Jie Tang. Graph random neural networks for semi-supervised learning on graphs. Advances in neural information processing systems, 33:22092–22103, 2020.
  • Gasteiger et al. [2018] Johannes Gasteiger, Aleksandar Bojchevski, and Stephan Günnemann. Predict then propagate: Graph neural networks meet personalized pagerank. In International Conference on Learning Representations, 2018.
  • Geng et al. [2021] Zhengyang Geng, Xin-Yu Zhang, Shaojie Bai, Yisen Wang, and Zhouchen Lin. On training implicit models. Advances in Neural Information Processing Systems, 34:24247–24260, 2021.
  • Gilmer et al. [2017] Justin Gilmer, Samuel S Schoenholz, Patrick F Riley, Oriol Vinyals, and George E Dahl. Neural message passing for quantum chemistry. In International conference on machine learning, pp. 1263–1272. PMLR, 2017.
  • Gu et al. [2020] Fangda Gu, Heng Chang, Wenwu Zhu, Somayeh Sojoudi, and Laurent El Ghaoui. Implicit graph neural networks. Advances in Neural Information Processing Systems, 33:11984–11995, 2020.
  • Hamilton et al. [2017] Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017.
  • Hu et al. [2020] Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118–22133, 2020.
  • Huang et al. [2021] Zhichun Huang, Shaojie Bai, and J Zico Kolter. (implicit)2(implicit)^{2}: Implicit layers for implicit representations. Advances in Neural Information Processing Systems, 34:9639–9650, 2021.
  • Hui et al. [2023] Bo Hui, Da Yan, Xiaolong Ma, and Wei-Shinn Ku. Rethinking graph lottery tickets: Graph sparsity matters. arXiv preprint arXiv:2305.02190, 2023.
  • Kingma [2014] Diederik P Kingma. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • Kipf & Welling [2017] Thomas Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net, 2017.
  • Köppen [2000] Mario Köppen. The curse of dimensionality. In 5th online world conference on soft computing in industrial applications (WSC5), volume 1, pp.  4–8, 2000.
  • Krantz & Parks [2002] Steven George Krantz and Harold R Parks. The implicit function theorem: history, theory, and applications. Springer Science & Business Media, 2002.
  • Lea et al. [2017] Colin Lea, Michael D Flynn, Rene Vidal, Austin Reiter, and Gregory D Hager. Temporal convolutional networks for action segmentation and detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  156–165, 2017.
  • Li et al. [2018] Qimai Li, Zhichao Han, and Xiao Ming Wu. Deeper insights into graph convolutional networks for semi-supervised learning. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, pp.  3538–3545. AAAI press, 2018.
  • Ling et al. [2023] Zenan Ling, Xingyu Xie, Qiuhao Wang, Zongpeng Zhang, and Zhouchen Lin. Global convergence of over-parameterized deep equilibrium models. In International Conference on Artificial Intelligence and Statistics, pp.  767–787. PMLR, 2023.
  • Ling et al. [2024] Zenan Ling, Longbo Li, Zhanbo Feng, Yixuan Zhang, Feng Zhou, Robert C Qiu, and Zhenyu Liao. Deep equilibrium models are almost equivalent to not-so-deep explicit models for high-dimensional Gaussian mixtures. In Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp.  30585–30609. PMLR, 21–27 Jul 2024. URL https://proceedings.mlr.press/v235/ling24a.html.
  • Liu et al. [2021] Juncheng Liu, Kenji Kawaguchi, Bryan Hooi, Yiwei Wang, and Xiaokui Xiao. Eignn: Efficient infinite-depth graph neural networks. Advances in Neural Information Processing Systems, 34:18762–18773, 2021.
  • Liu et al. [2024] Juncheng Liu, Bryan Hooi, Kenji Kawaguchi, Yiwei Wang, Chaosheng Dong, and Xiaokui Xiao. Scalable and effective implicit graph neural networks on large graphs. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=QcMdPYBwTu.
  • Luan et al. [2021] Sitao Luan, Chenqing Hua, Qincheng Lu, Jiaqi Zhu, Mingde Zhao, Shuyuan Zhang, Xiao-Wen Chang, and Doina Precup. Is heterophily a real nightmare for graph neural networks to do node classification? arXiv preprint arXiv:2109.05641, 2021.
  • Lutzeyer et al. [2022] Johannes F Lutzeyer, Changmin Wu, and Michalis Vazirgiannis. Sparsifying the update step in graph neural networks. In Topological, Algebraic and Geometric Learning Workshops 2022, pp.  258–268. PMLR, 2022.
  • Meng et al. [2019] Zaiqiao Meng, Shangsong Liang, Hongyan Bao, and Xiangliang Zhang. Co-embedding attributed networks. In Proceedings of the twelfth ACM international conference on web search and data mining, pp.  393–401, 2019.
  • Monti et al. [2017] Federico Monti, Davide Boscaini, Jonathan Masci, Emanuele Rodola, Jan Svoboda, and Michael M Bronstein. Geometric deep learning on graphs and manifolds using mixture model cnns. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.  5115–5124, July 2017.
  • Nastorg et al. [2023] Matthieu Nastorg, Michele-Alessandro Bucci, Thibault Faney, Jean-Marc Gratien, Guillaume Charpiat, and Marc Schoenauer. An implicit gnn solver for poisson-like problems. arXiv preprint arXiv:2302.10891, 2023.
  • Poli et al. [2019] Michael Poli, Stefano Massaroli, Junyoung Park, Atsushi Yamashita, Hajime Asama, and Jinkyoo Park. Graph neural ordinary differential equations. arXiv preprint arXiv:1911.07532, 2019.
  • Thorpe et al. [2022] Matthew Thorpe, Tan Nguyen, Hedi Xia, Thomas Strohmer, Andrea Bertozzi, Stanley Osher, and Bao Wang. Grand++: Graph neural diffusion with a source term. ICLR, 2022.
  • Veličković et al. [2018] Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph Attention Networks. International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rJXMpikCZ.
  • Liu et al. [2022] Juncheng Liu, Bryan Hooi, Kenji Kawaguchi, and Xiaokui Xiao. Mgnni: Multiscale graph neural networks with implicit layers. Advances in Neural Information Processing Systems, 35:21358–21370, 2022.
  • Poli et al. [2020] Michael Poli, Stefano Massaroli, Atsushi Yamashita, Hajime Asama, and Jinkyoo Park. Hypersolvers: Toward fast continuous-depth models. Advances in Neural Information Processing Systems, 33:21105–21117, 2020.
  • Wang et al. [2020] Xiao Wang, Meiqi Zhu, Deyu Bo, Peng Cui, Chuan Shi, and Jian Pei. Am-gcn: Adaptive multi-channel graph convolutional networks. In Proceedings of the 26th ACM SIGKDD International conference on knowledge discovery & data mining, pp.  1243–1253, 2020.
  • Winston & Kolter [2020] Ezra Winston and J Zico Kolter. Monotone operator equilibrium networks. Advances in neural information processing systems, 33:10718–10728, 2020.
  • Wu et al. [2019a] Felix Wu, Amauri Souza, Tianyi Zhang, Christopher Fifty, Tao Yu, and Kilian Weinberger. Simplifying graph convolutional networks. In International conference on machine learning, pp. 6861–6871. PMLR, 2019a.
  • Wu et al. [2019b] Jun Wu, Jingrui He, and Jiejun Xu. Demo-net: Degree-specific graph neural networks for node and graph classification. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp.  406–415, 2019b. doi: 10.1145/3292500.3330950. URL https://doi.org/10.1145/3292500.3330950.
  • Xhonneux et al. [2020] Louis-Pascal Xhonneux, Meng Qu, and Jian Tang. Continuous graph neural networks. In International conference on machine learning, pp. 10432–10441. PMLR, 2020.
  • Xu et al. [2018] Keyulu Xu, Chengtao Li, Yonglong Tian, Tomohiro Sonobe, Ken-ichi Kawarabayashi, and Stefanie Jegelka. Representation learning on graphs with jumping knowledge networks. In International conference on machine learning, pp. 5453–5462. PMLR, 2018.
  • Yang & Leskovec [2012] Jaewon Yang and Jure Leskovec. Defining and evaluating network communities based on ground-truth. In Proceedings of the ACM SIGKDD Workshop on Mining Data Semantics, pp.  1–8, 2012.
  • Yang et al. [2021] Yongyi Yang, Tang Liu, Yangkun Wang, Zengfeng Huang, and David Wipf. Implicit vs unfolded graph neural networks. arXiv preprint arXiv:2111.06592, 2021.
  • Yu et al. [2022] Shujian Yu, Francesco Alesiani, Wenzhe Yin, Robert Jenssen, and Jose C. Principe. Principle of relevant information for graph sparsification. In James Cussens and Kun Zhang (eds.), Proceedings of the Thirty-Eighth Conference on Uncertainty in Artificial Intelligence, volume 180 of Proceedings of Machine Learning Research, pp.  2331–2341. PMLR, 01–05 Aug 2022. URL https://proceedings.mlr.press/v180/yu22c.html.
  • Yun et al. [2019] Seongjun Yun, Minbyul Jeong, Raehyun Kim, Jaewoo Kang, and Hyunwoo J Kim. Graph transformer networks. In Advances in Neural Information Processing Systems, volume 32, pp.  11960–11970. Curran Associates, Inc., 2019. doi: 10.5555/3454287.3455360. URL https://proceedings.neurips.cc/paper/2019/file/9d63484abb477c97640154d40595a3bb-Paper.pdf.
  • Zhang et al. [2023a] Shuai Zhang, Meng Wang, Pin-Yu Chen, Sijia Liu, Songtao Lu, and Miao Liu. Joint edge-model sparse learning is provably efficient for graph neural networks. arXiv preprint arXiv:2302.02922, 2023a.
  • Zhang et al. [2023b] Xu Zhang, Yonghui Xu, Wei He, Wei Guo, and Lizhen Cui. A comprehensive review of the oversmoothing in graph neural networks. In CCF Conference on Computer Supported Cooperative Work and Social Computing, pp.  451–465. Springer, 2023b.
  • Zhao et al. [2021] Jianan Zhao, Xiao Wang, Chuan Shi, Binbin Hu, Guojie Song, and Yanfang Ye. Heterogeneous graph structure learning for graph neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pp.  4697–4705, 2021.
  • Zhao et al. [2023] Qifang Zhao, Weidong Ren, Tianyu Li, Xiaoxiao Xu, and Hong Liu. Graphgpt: Graph learning with generative pre-trained transformers. arXiv preprint arXiv:2401.00529, 2023.
  • Zheng et al. [2020] Cheng Zheng, Bo Zong, Wei Cheng, Dongjin Song, Jingchao Ni, Wenchao Yu, Haifeng Chen, and Wei Wang. Robust graph representation learning via neural sparsification. In International Conference on Machine Learning, pp. 11458–11468. PMLR, 2020.

Supplementary Material
IGNN-Solver: A Graph Neural Solver for Implicit Graph Neural Networks

Appendix A Anderson Acceleration Solver for Deep Equilibriums (DEQs)

Algorithm 2 Anderson Acceleration Solver
1:fixed-point function fθ:nnf_{\theta}:\mathbb{R}^{n}\rightarrow\mathbb{R}^{n}, max storage size mm, residual control parameter β\beta
2:Set initial point value 𝒁[0]n{\bm{Z}}^{[0]}\in\mathbb{R}^{n} \triangleright Set 𝟎\mathbf{0} or random value normally
3:for k=0k=0, \ldots, K1K-1 do
4:     1) Set mk=min{m,k}m_{k}=\min\{m,k\} \triangleright Storage of the most recent steps
5:     2) Solve α[k]=argminαmk+1𝑮[k]α2, s.t. 𝟏α[k]=1\alpha^{[k]}=\arg\min_{\alpha\in\mathbb{R}^{m_{k}+1}}\left\|{\bm{G}}^{[k]}\alpha\right\|_{2}\text{, s.t. }\mathbf{1}^{\top}\alpha^{[k]}=1 \triangleright Compute weights
6:     3) 𝒁[k+1]=βi=0mkαi[k]fθ(𝒁[kmk+i])+(1β)i=0mkαi[k]𝒁[kmk+i]{\bm{Z}}^{[k+1]}=\beta\sum_{i=0}^{m_{k}}\alpha_{i}^{[k]}f_{\theta}({\bm{Z}}^{[k-m_{k}+i]})+(1-\beta)\sum_{i=0}^{m_{k}}\alpha_{i}^{[k]}{\bm{Z}}^{[k-m_{k}+i]}\quad \triangleright AA update
7:end for

Given an implicit network layer fθf_{\theta} and an input 𝑿{\bm{X}}, DEQs aim to find the fixed point 𝒁{\bm{Z}}^{*} in the system that represents the equilibrium state of the network output by solving a root-finding problem:

gθ(𝒁,𝑿):=fθ(𝒁,𝑿)𝒁=0.g_{\theta}\left({\bm{Z}}^{\star},{\bm{X}}\right):=f_{\theta}\left({\bm{Z}}^{\star},{\bm{X}}\right)-{\bm{Z}}^{\star}=0. (12)

In practical solving of DEQs, it is often necessary to utilize fixed-point solvers during the training process, such as the classical Broyden method [9] or Anderson Acceleration (AA) [2], which can directly seek the equilibrium point 𝒁{\bm{Z}}^{\star} through quasi-Newton methods. This approach demonstrates super-linear convergence properties.

We briefly introduce AA-solver here, as our approach relies on it. Algorithm 2 is a pseudocode example illustrating it, where 𝑮[k]=[gθ(𝐳[kmk])gθ(𝐳[k])]{\bm{G}}^{[k]}=[g_{\theta}(\mathbf{z}^{[k-m_{k}]})\ldots g_{\theta}(\mathbf{z}^{[k]})] is the past residuals. It can be observed that the basic idea of Anderson Acceleration (AA) is to accelerate the generation of the next approximate solution by utilizing the information stored from the history mm iteration steps, constructing a normalized linear combination with weights α[k]\alpha^{[k]}. It computes the weights greedily to minimize the linear combination of past AA update steps. This approach is typically particularly effective for iteratively slow-converging processes, such as solving implicit functions.

Subsequently, during the backward process, the implicit function theorem [28] can be employed to implicitly differentiate through the equilibrium point and generate gradients with respect to the model parameters θ\theta by solving linear equations based on the Jacobian matrix. Since the computation process relies only on the final output and does not require storing any intermediate activation values [4], the memory consumption during the training process remains constant, which is also a key reason why DEQs are attractive.

Appendix B Training Strategies for IGNN using IGNN-Solver.

B.1 IGNN Training with frozen IGNN-Solver

During the training processing of the IGNN model, we assume that both the initializer hϕh_{\phi} and the neural solver sξs_{\xi} of IGNN-Solver have approached the desired state via the training method outlined in Section 4.1, and they will no longer be updated in the subsequent model training. For a given IGNN layer fθf_{\theta}, as well as the input graph 𝑨{\bm{A}} and feature matrix 𝑿{\bm{X}}, we define:

gθ(𝒁,𝑨^,𝑿):=fθ(𝒁,𝑨^,𝑿)𝒁=0,g_{\theta}({\bm{Z}}^{\star},\hat{{\bm{A}}},{\bm{X}}):=f_{\theta}({\bm{Z}}^{\star},\hat{{\bm{A}}},{\bm{X}})-{\bm{Z}}^{\star}=0, (13)

and utilize the IGNN-Solver proposed in Section 4.1 to solve equation 13, obtaining the fixed point 𝒁{\bm{Z}}^{\star}. Furthermore, the fixed point 𝒁{\bm{Z}}^{\star} is passed through a linear layer to obtain an embedding 𝑯=Linear(𝒁){\bm{H}}=\text{Linear}({\bm{Z}}^{\star}). Then we jointly with downstream tasks, compute the Binary Cross Entropy (BCE) loss [15] to back-propagate and update the parameters of fθf_{\theta} in a supervised way. The complete process is illustrated in Algorithm 3.

Algorithm 3 IGNN model training
1:graph matrix 𝑨^\hat{{\bm{A}}}, feature matrix 𝑿{\bm{X}}, fixed IGNN-Solver sξs_{\xi} and initializer hϕh_{\phi}
2:Initialize 𝒁[0]=hϕ(𝑿){\bm{Z}}^{[0]}=h_{\phi}({\bm{X}}) and randomly initialize fθf_{\theta}
3:while Stopping condition is not met do
4:     𝒁Solvefθ(𝒁,𝑨^,𝑿)𝒁=0{\bm{Z}}^{\star}\leftarrow\text{Solve}\ f_{\theta}({\bm{Z}}^{\star},\hat{{\bm{A}}},{\bm{X}})-{\bm{Z}}^{\star}=0 \triangleright via frozen sξs_{\xi}, 𝑨^\hat{{\bm{A}}} and initial value 𝒁[0]{\bm{Z}}^{[0]}
5:     𝑯Linear(𝒁){\bm{H}}\leftarrow\text{Linear}({\bm{Z}}^{\star})
6:     taskBCE(𝑯,𝒀)\mathcal{L}_{\text{task}}\leftarrow\text{BCE}({\bm{H}},{\bm{Y}}) \triangleright computing the label loss
7:     Back-propagate task\mathcal{L}_{\text{task}} to update fθf_{\theta}
8:end while
9:return fθf_{\theta}

B.2 Advanced IGNN Training Strategy with IGNN-Solver

Given the fact that model parameter θ\theta gradually updates only with the model training iterations (i.e., we assume θt+1θt\left\|\theta_{t+1}-\theta_{t}\right\| is only related to θt\theta_{t} at training step tt), we propose training the lightweight IGNN-Solver sξ,hϕs_{\xi},h_{\phi} and the large model fθf_{\theta} in an alternating way. Specifically, we adopt the following procedure:

  1. (i)

    Warm up and train the IGNN model and its solver (IGNN-Solver) for some steps.

  2. (ii)

    Fix the IGNN-Solver sξ,hϕs_{\xi},h_{\phi} and solving the fixed points of fθf_{\theta} in IGNN via Algorithm 3 for T1T_{1} steps.

  3. (iii)

    Fix the current model parameters θ\theta and start fine-tuning the IGNN-Solver sξ,hϕs_{\xi},h_{\phi} via Algorithm 1 over some T2T_{2} steps.

  4. (iv)

    Repeat steps (ii) and (iii) until reaching the maximum training steps for the IGNN model.

Here T1T_{1} is approximately 4% to 10% of T2T_{2}, adjusted as needed. The sum of all T2T_{2} values is referred to as epochmax\text{epoch}_{\text{max}}. It is reassuring that the additional cost of fine-tuning the IGNN-Solver (step (iii)) is sufficiently offset by the substantial benefits of accelerated solving (step (ii)). Moreover, we are surprised to find that the proposed IGNN-Solver does not exhibit a decline in expressive capability for excessively small or large T1T_{1} values. This indicates that, thanks to its robust stability, accidentally setting T1T_{1} high does not significantly affect the normal training of the IGNN. Conversely, if T1T_{1} is set slightly lower, the model still demonstrates good generalization capabilities. Thus, the IGNN-Solver shows low sensitivity to T1T_{1}.

Appendix C Experimental Details

The experiments are conducted on 99 public datasets of different scales, including 55 widely adopted benchmark datasets Citeseer, ACM, CoraFull, BlogCatalog, Flickr, and 44 large-scale benchmark datasets Amazon-all, Reddit, ogbn-arxiv and ogbn-products. We endeavor to conduct inference testing under almost identical hyperparameter conditions as previous work [7; 33; 20; 45; 26; 42; 47; 17; 50], including performance and efficiency comparisons. All the Experiments are conducted independently five times (i.e., using 55 random seeds) on a machine with Intel(R) Xeon(R) Gold 6138 CPU @ 2.00GHz with a single 3090 GPU. A detailed description of all tasks and additional details are provided below.

C.1 Datasets

  • Citeseer [26]: This dataset is a citation network of research papers, which are divided into six categories. The citation network takes 3,327 scientific papers as nodes and 4,732 citation links as edges. The feature of each node in the dataset is a word vector to describe whether the paper has the corresponding words or not.

  • ACM [54]: It is a citation network dataset, where nodes represent papers and node features are constructed by the keywords. Papers are divided into 3 categories according to their types of conferences.

  • CoraFull [8]: Similar to the Citeseer dataset, CoraFull is a well-known citation network labeled based on the paper topic, which contains 19,793 scientific publications. CoraFull is classified into one of 70 categories, where nodes represent papers and the edges represent citations.

  • BlogCatalog [37]: This dataset is a social relationship network. The graph is composed of bloggers and their social relationships (such as friends). Node attributes are constructed by keywords in the user profile. The labels represent bloggers’ interests. All nodes are divided into six categories.

  • Flickr [48]: It is a graphic social network where nodes represent users and edges correspond to the friendships among users. All the nodes are divided into 9 classes according to the interest groups of users.

  • Amazon-all [51]: The widely-used benchmark dataset Amazon-all encompasses the Amazon product co-purchasing network dataset. This dataset represents products as nodes and co-purchases as edges. It includes 58 product types, each with over 5,000 products, selected from a total pool of 75,149 product types.

  • Reddit [16]: The Reddit dataset is a social network where nodes represent posts, and edges indicate that the same user commented on two connected posts. Each node contains 602 dimensional features.

  • ogbn-arxiv [22]: The ogbn-arxiv dataset is a citation network between all Computer Science (CS) arXiv papers. Each node is an arXiv paper, and each directed edge indicates that one paper cites another. Each paper comes with a 128-dimensional feature vector obtained by averaging the embeddings of words in its title and abstract. The task is to predict the 40 subject areas of arXiv CS papers.

  • ogbn-products [22]: The ogbn-products dataset contains an undirected and unweighted graph, representing an Amazon product co-purchasing network. Nodes represent products sold on Amazon, and edges between two products indicate that the products are purchased together. The task is to predict the category of a product in a multi-class classification setup, where the 47 top-level categories are used for target labels.

Table 2: Hyper-paramaters setting on training. *In addition, we decay the loss weight λ1\lambda_{1} from 0 to the setting value on a linear schedule over all IGNN-Solver training steps.
Scale Datasets nhid dropout lr KK epochmax\text{epoch}_{\text{max}} training test λ1\lambda_{1}* λ2\lambda_{2} λ3\lambda_{3}
Small Citeseer 128 0.5 0.002 10 100 360 1000 0.1 5 1e-4
ACM 128 0.5 0.001 10 100 180 1000 0.3 5 1e-4
CoraFull 512 0.5 0.002 15 300 4200 1000 0.3 5 1e-4
BlogCatalog 512 0.5 0.001 10 100 360 1000 0.5 5 1e-4
Flickr 512 0.5 0.002 15 200 540 1000 0.1 5 1e-5
Large Amazon-all 128 0.5 0.005 20 1000 16970 28285 0.3 5 1e-5
Reddit 512 0.5 0.005 15 2000 139,779 46,593 0.5 5 1e-5
ogbn-arxiv 128 0.5 0.001 20 2000 90,941 48,603 0.5 5 1e-5
ogbn-pruducts 128 0.5 0.001 20 2000 196,615 2,213,091 0.5 5 1e-5

C.2 Baselines

We compare IGNN-Solver with 1212 state-of-the-art methods, including 33 implicit IGNNs, i.e., MIGNN [7], EIGNN [33], IGNN [20], and 99 explicit/traditional GNNs, i.e., AM-GCN [45], GCN [26], GAT [42], SGC [47], APPNP [17], JKNet [50], DEMO-Net [48], GCNII [11], ACM-GCN [35]. The specific parameter settings are as follows:

For the configuration of hidden layers, for the sake of fairness, we set the same hidden layers for all baselines. For example, in the Flickr dataset, the dimension of the hidden layers is uniformly set to 512512 for all methods (including ours, see in Table 2). In addition, for EIGNN, the arbitrary gamma is set to 0.80.8, which is consistent with the original paper. For AM-GCN, the number of nearest neighbors in KNN graph is set from 5,6,75,6,7. For GCN, the hidden layer setup is same as others. For GAT, three attention headers are used for each layer. For SGC, the power of self-loops in the graph adjacency matrix is set to 33. For APPNP, we set the number of iterations to 33 and teleport probability α\alpha to 0.50.5. For JKNet, we set layers to 11 in small datasets and set to 88 in large ones. For DEMO-Net, the regularization parameter is set as 5e45e-4, the learning rate is set as 5e35e-3, and the hash dimension is set as 256. For GCNII, we set α=0.1\alpha_{\ell}=0.1 and L2L_{2} regularization to 5e45e-4, consistent with the original paper. For ACM-GCN, we set layers to 11 in small datasets and set to 44 in large ones. For our IGNN-Solver and its variants, we employ the same settings as the basic IGNN model. Furthermore, we adjust these parameters affecting the convergence and performance of IGNN through a combined approach of hierarchical grid search and manual tuning. The Adam optimizer [25] is used for optimization.

C.3 Hyper-parameters Setting in IGNN-Solver

We present in Table 2 all the hyper-parameters preset by IGNN-Solver across all datasets, where KK represents the threshold of maximum iteration in IGNN-Solver, epochmax\text{epoch}_{\text{max}} means the maximum training steps in IGNN. It is worth noting that we set the value of λ1\lambda_{1} in a linearly increasing manner to penalize the intermediate estimation error when the training step KK is large. Similarly, we set the value of λ3\lambda_{3} in a linearly decreasing manner to prevent overfitting α\alpha during later training epochs.

Refer to caption
(a) Citeseer
Refer to caption
(b) ACM
Refer to caption
(c) CoraFull
Refer to caption
(d) BlogCatalog
Refer to caption
(e) Flickr
Figure 6: Comparison of inference time with IGNN-Solvers on 55 small-scale datasets. All speed/accuracy curves within the same plot are benchmarked on the same GPU with the same experimental setting, averaged over 55 independent runs.
Table 3: Node Classification accuracy results on 55 small-scale real-world datasets, with experiments conducted over five trials. The mean accuracy ± standard deviation (%) is reported. The best and the runner-up results are highlighted in boldface and underline, respectively.
Type *Models Citeseer ACM CoraFull BlogCatalog Flickr
*Nodes 3,3273,327 3,0253,025 19,79319,793 5,1965,196 7,5757,575
*Edges 4,7324,732 13,12813,128 65,31165,311 171,743171,743 239,738239,738
Explicit GCN [26] 64.80±4.1564.80\pm 4.15 83.78±3.9583.78\pm 3.95 35.98±9.4435.98\pm 9.44 71.30±1.1271.30\pm 1.12 76.98±1.8376.98\pm 1.83
GAT [42] 71.24±2.8871.24\pm 2.88 80.60±5.1280.60\pm 5.12 50.24±1.8750.24\pm 1.87 76.24±2.7676.24\pm 2.76 72.64±3.2372.64\pm 3.23
SGC [47] 70.32±2.7570.32\pm 2.75 85.40±1.2385.40\pm 1.23 46.56±4.4346.56\pm 4.43 69.76±1.1169.76\pm 1.11 71.88±3.4771.88\pm 3.47
APPNP [17] 61.56±8.9261.56\pm 8.92 84.16±4.2584.16\pm 4.25 21.24±4.7621.24\pm 4.76 61.34±9.3961.34\pm 9.39 73.42±3.8073.42\pm 3.80
JKNet [50] 63.78±8.7663.78\pm 8.76 64.96±5.4764.96\pm 5.47 23.04±6.0123.04\pm 6.01 72.62±6.8372.62\pm 6.83 75.68±1.1575.68\pm 1.15
AM-GCN [45] 73.10±1.6273.10\pm 1.62 89.56±0.3089.56\pm 0.30 53.40±1.5953.40\pm 1.59 73.86±1.1073.86\pm 1.10 76.86±2.0276.86\pm 2.02
DEMO-Net [48] 68.34±2.9468.34\pm 2.94 84.38±2.1984.38\pm 2.19 61.74±3.6561.74\pm 3.65 74.26±2.7074.26\pm 2.70 75.60±3.9575.60\pm 3.95
GCNII [11] 71.98±0.8071.98\pm 0.80 85.36±1.0585.36\pm 1.05 57.64±3.3457.64\pm 3.34 74.94±3.8174.94\pm 3.81 79.92¯±2.14\underline{79.92}\pm 2.14
ACM-GCN [35] 72.38±1.4672.38\pm 1.46 88.98±0.4188.98\pm 0.41 59.88±1.5959.88\pm 1.59 78.18±1.75\textbf{78.18}\pm 1.75 74.82±3.7874.82\pm 3.78
Implicit IGNN [20] 72.96±1.8372.96\pm 1.83 90.88±0.9590.88\pm 0.95 65.52±0.5165.52\pm 0.51 75.68±0.5575.68\pm 0.55 75.80±0.2975.80\pm 0.29
EIGNN [33] 72.38±1.3672.38\pm 1.36 88.36±1.0388.36\pm 1.03 61.80±0.6061.80\pm 0.60 75.34±0.3875.34\pm 0.38 75.66±0.9475.66\pm 0.94
MIGNN [7] 73.79±0.9473.79\pm 0.94 89.59±1.6189.59\pm 1.61 62.94±0.4662.94\pm 0.46 76.68±1.4976.68\pm 1.49 74.96±0.4974.96\pm 0.49
IGNN w. AA 75.28¯±0.38\underline{75.28}\pm 0.38 91.34±0.46\textbf{91.34}\pm 0.46 65.88¯±0.34\underline{65.88}\pm 0.34 76.82±0.3476.82\pm 0.34 75.84±0.2775.84\pm 0.27
IGNN w. NN 74.78±0.4274.78\pm 0.42 91.08±0.5691.08\pm 0.56 65.62±0.1665.62\pm 0.16 76.88±0.7476.88\pm 0.74 78.55±0.4978.55\pm 0.49
IGNN-Solver 75.60±0.22\textbf{75.60}\pm 0.22 91.20¯±1.15\underline{91.20}\pm 1.15 66.08±1.23\textbf{66.08}\pm 1.23 77.64¯±0.54\underline{77.64}\pm 0.54 80.14±0.23\textbf{80.14}\pm 0.23

Appendix D More Results and Additional Experiments

D.1 More Performance and Efficiency Comparison on other small-scale tasks

We further test IGNN-Solver for a few small-scale graph node classification tasks, including Citeseer, ACM, CoraFull, BlogCatalog and Flickr. We employ the training procedure details in Appendix C and hyperparameters used outlined in Appendix B, and report the mean accuracy ± standard deviation in Table 3 and Figure 6. Generally, learning LRD is not crucial for these tasks, as the diameter of the graphs is quite small [7]. However, as seen in Table 3, even for these small-scale node classification tasks, the IGNN-Solver can still surpass IGNN and even outperforms many explicit GNNs and other enhanced implicit ones. As shown in Figure 6, even in small-scale tasks, the solver still improves upon the convergence path of IGNN. These results and which in Section 5.1 confirm the expressive power of the IGNN-Solver using learnable graph neural, even exceeding that of many explicit and improved implicit GNNs.

D.2 Other options for graph neural layers in IGNN-Solver.

Refer to caption
(a) The relative error curve during warm-up.
Refer to caption
(b) The relative error curve during inference.
Figure 7: The proposed IGNN-Solver (GN) improves the convergence path, which is attributed to its lightweight nature and capability to leverage rich graph information.

We provide additional evidence on the convergence and generalization of IGNN-Solver, where we tried other options besides our proposed graph neural (GN) layers, including neural network (NN), temporal convolutional network (TCN) [29] and gated recurrent unit (GRU) [14]. We present the relative residual curves on the Citeseer dataset during the warm-up stage and the final inference stage. As illustrated in Figure 7, TCN and GRU, despite having more parameters and more complex structures, actually perform worse than NN. Note that IGNN-Solver (GN) improves the convergence path of all other solvers and exhibits the fastest rate of decrease, proving to be the most suitable solver for IGNN, which is attributed to its capability to leverage rich graph information effectively.

D.3 Training Dynamics of IGNNs with IGNN-Solver

Refer to caption
Figure 8: Training time and inference time(s) per epoch of IGNN and IGNN-Solver on Amazon-all datasets.

In this section, we present the time cost trends of IGNN and IGNN-Solver during training and inference on the Amazon dataset. From Figure 8, it is evident that IGNN starts relatively fast overall at the beginning of training. However, as the training of IGNN progresses, the model becomes increasingly complex, leading to a sharp rise in the time cost for fixed-point computation. After 100 epochs, this cost remains persistently high and fluctuates continuously. This issue arises because the IGNN reaches the maximum number of iterations without achieving the preset minimum error.

On the contrary, our IGNN-Solver demonstrates significantly lower training and inference time consumption compared to the former and maintains a stable state throughout the process. This is attributed to the solver’s consistent memory consumption and its rapid fixed-point computation capability, which drives the unique advantage of IGNN-Solver.