This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\label

key

\catchline

SStaGCN: Simplified stacking based graph convolutional networks

Jia Cai School of Digital Economy, Guangdong University of Finance &\& Economics
Guangzhou, 510320, P. R. China.
[email protected]
   Zhilong Xiong School of Statistics and Mathematics, Guangdong University of Finance &\& Economics
Guangzhou, 510320, P. R. China.
[email protected]
   Shaogao Lv111Corresponding author. Department of Statistics and Data Science, Nanjing Audit University
Nanjing, 211815, P. R. China.
[email protected]
((Day Month Year); (Day Month Year))
Abstract

Graph Convolutional Networks (GCNs) are powerful models extensively studied for various graph-structured data learning tasks. However, designing GCNs that effectively mitigate the over-smoothing phenomenon remains a crucial challenge. In this paper, we propose a novel Simplified Stacking-based GCN (SStaGCN) framework, leveraging stacking and aggregation techniques to address different types of graph-structured data. Specifically, we first employ stacking base models to extract node features from the graph. Next, we apply aggregation methods—such as mean, attention, and voting techniques—to further enhance feature extraction capabilities. These refined node features are then fed into a vanilla GCN model. Additionally, we provide a theoretical analysis of the generalization bound for the proposed model. Extensive experiments on three public citation networks and three heterogeneous tabular datasets demonstrate the effectiveness and efficiency of the SStaGCN approach over several state-of-the-art GCNs. Notably, SStaGCN effectively mitigates the over-smoothing issue common in GCNs.

keywords:
Graph convolutional network; stacking; aggregation.
{history}
\ccode

Mathematics Subject Classification 2010: 68T05, 68W40

1 Introduction

Recent research on learning from graph-structured data has gained considerable attention across diverse fields within artificial intelligence. Graph Neural Networks (GNNs), particularly Graph Convolutional Networks (GCNs) [4, 14, 15], have achieved remarkable success in modeling graph-structured data and have been widely applied to recommendation systems [38], computer vision [5], molecular design [37], natural language processing [46], node classification [19], and clustering tasks [49].

Despite these successes, real-world applications present diverse types of graph-structured data, raising the question: can we design a general framework to handle distinct types of graph-structured data in a simpler, more effective manner? Additionally, as discussed in [24], the graph convolution operation in GCNs is a specific form of Laplacian smoothing, which mixes node features across different clusters and can lead to over-smoothing [24].

Sun et al. [39] addressed this issue by developing an RNN-like GCN with AdaBoost, enabling the extraction of knowledge from high-order neighbors of current nodes. However, the relation A=BA^{\ell}=B^{\ell} does not necessarily imply A=BA=B. A simple example illustrating this is the case where =2\ell=2 (two layers), with

A=(0110),B=(1001).A=\left(\begin{matrix}0&1\\ 1&0\end{matrix}\right),\qquad B=\left(\begin{matrix}1&0\\ 0&1\end{matrix}\right).

As a result, this RNN-like GCN inevitably loses crucial information about the original graph structure, leaving the challenge of designing a simple GCN model to tackle the over-smoothing issue unresolved.

The stacking method [9], also known as stacked generalization, is an ensemble approach that combines multiple classifiers using base models and a meta-model. Stacking has shown success in feature extraction tasks across various data types, including disordered or irregular data, due to its advantageous properties:

By combining distinct learners in the base models, stacking effectively captures discernible features. Base models are trained on the entire training dataset, allowing for robust performance evaluation on test data. A meta-model then aggregates these outputs to make final predictions on the test data. Combining stacking with GCN models can yield significant benefits. In this paper, we propose a novel simplified stacking-based architecture, Simplified Stacking-based GCN (SStaGCN), to address the over-smoothing issue in GCNs. SStaGCN integrates the feature extraction strengths of stacking with GCNs’ graph learning capabilities. This synergy enables SStaGCN to harness the strengths of both stacking and GCNs.

The main contributions of this paper are as follows:

{itemlist}

We propose a novel, versatile architecture that combines simplified stacking and GCNs, designed to adapt flexibly to diverse types of graph-structured data, including heterogeneous tabular data222Heterogeneous tabular data includes a mix of discrete and continuous variables..

We present a generalization bound analysis to elucidate the contributions of stacking and aggregation from a learning theory perspective.

We conduct extensive evaluations of our approach against strong baselines on node prediction tasks. The experimental results demonstrate significant performance improvements in both homogeneous and heterogeneous node classification tasks across various real-world graph-structured datasets, effectively alleviating the over-smoothing phenomenon.

The remainder of the paper is organized as follows. In Section 2, we provide a brief review of related work. Section 3 presents the theoretical analysis and proposed algorithm for GCNs. In Section 4, we detail experiments conducted on three public citation networks and three heterogeneous tabular datasets. Section 5 offers discussions and concluding remarks. The proof of the main result is included in the Appendix.

2 Related Work

Graph Convolutional Networks

GCNs are commonly understood as extensions of traditional convolutional neural networks applied to graph domains. Generally, there are two types of GCNs [3]: spatial GCNs and spectral GCNs. Spatial GCNs construct new feature vectors for each node by leveraging neighborhood information, where convolution acts as a ”patch operator.” Spectral GCNs, on the other hand, define convolution by decomposing a graph signal in the spectral domain and applying a spectral filter (such as Fourier or wavelet filters, [4, 45, 23]) to the spectral components [32, 36]. However, spectral GCNs require computation of the Laplacian eigenvectors, which becomes impractical for large-scale graphs due to its high computational cost.

To address this, Hammond et al. [16] used Chebyshev polynomials up to the KK-th order to approximate the spectral filter. Defferrard and Vandergheynst [8] proposed the KK-localized ChebyNet, and Kipf and Welling [19] introduced a simpler model by setting K=1K=1, which proved effective for semi-supervised classification tasks. Wu et al. [44] further simplified GCNs by removing nonlinearities and collapsing the weight matrices between layers. In contrast, other works like [22, 24] explored the development of deeper GCNs, while multi-scale deep GCNs were examined in [25]. Despite these advancements, over-smoothing remains a significant challenge for GCNs.

Various strategies have been proposed to address over-smoothing. Klicpera et al. [20] and Chien et al. [7] used PageRank and generalized PageRank, respectively, to update graph information. DropEdge [31] randomly removes edges in the graph to mitigate over-smoothing. Similarly, GRAND [10] introduced a random propagation strategy to augment graph data and applied consistency regularization. Chen et al. [6] developed GCNII, a GCN variant that uses initial residuals and identity mapping to address over-smoothing, while DGMLP [48] incorporated adaptive modes and residual connections.

This paper also adopts a decoupled approach, performing feature extraction followed by message propagation to classify nodes. However, our approach achieves superior performance and accuracy while providing a more flexible and general framework.

Ensemble learning based graph neural networks

Sun et al. [39] designed an RNN-like graph structure to extract information from high-order neighbors of each node, while Ivanov and Prokhorenkova [17] integrated gradient-boosted decision trees (GBDT) into GNNs, developing BGNN to handle heterogeneous tabular data. This raises a natural question: can we design a general GCN architecture that not only accommodates diverse graph data but also addresses the over-smoothing issue? This paper seeks to investigate this challenge and provides a solution to the above question.

3 The proposed approach: SStaGCN

3.1 Graph convolutional networks

Given an undirected graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) with nodes vi𝒱v_{i}\in\mathcal{V} and edges (vi,vj)(v_{i},v_{j})\in\mathcal{E}, let AN×NA\in\mathbb{R}^{N\times N} denote the adjacency matrix, where NN is the total number of nodes. The corresponding degree matrix is denoted as 𝐃\mathbf{D}, with 𝐃ii=jAij\mathbf{D}_{ii}=\sum_{j}A_{ij}. For an undirected graph, it is evident that Aij=AjiA_{ij}=A_{ji}.

In conventional GCN models for semi-supervised node classification, the node embeddings with two convolutional layers are computed as follows:

Z=A^ReLU(A^XW(0))W(1).Z=\hat{A}{\rm ReLU}(\hat{A}XW^{(0)})W^{(1)}. (1)

Here, ZN×KZ\in\mathbb{R}^{N\times K} represents the final embedding matrix (output logits) of the nodes prior to the softmax operation, where KK is the number of classes. The feature matrix XN×dX\in\mathbb{R}^{N\times d} contains node features, with dd as the input dimension. We define A^=D~12A~D~12\hat{A}=\tilde{D}^{-\frac{1}{2}}\tilde{A}\tilde{D}^{-\frac{1}{2}}, where A~=A+I\tilde{A}=A+I (with II as the identity matrix) and D~\tilde{D} is the degree matrix of A~\tilde{A}.

Furthermore, W(0)d×HW^{(0)}\in\mathbb{R}^{d\times H} is the weight matrix for the input-to-hidden layer with HH hidden features, and W(1)H×KW^{(1)}\in\mathbb{R}^{H\times K} is the weight matrix for the hidden-to-output layer. Notably, the standard GCN model in Eq. (1) applies A^\hat{A} to XX repeatedly, which leads to all node features becoming indistinguishable due to excessive smoothing.

3.2 Stacking

Stacking is a well-known and widely used ensemble machine learning algorithm that integrates models in a hierarchical framework [43]. It employs a meta-learning algorithm to optimally combine predictions from two or more base machine learning models. A traditional stacking model consists of multiple base models and a meta-model that integrates the base models’ predictions. Each base model is trained on the dataset and produces individual predictions. The meta-model then learns to best combine these predictions, typically using a straightforward approach to provide a cohesive interpretation of the base models’ outputs. As a result, linear models, such as linear regression for regression tasks and logistic regression for classification tasks, are often chosen as meta-models.

3.3 The proposed approach

As noted above, GCNs may blend node features from different clusters, which can lead to suboptimal predictions. Therefore, it is essential to aggregate node information effectively to improve predictive accuracy. Inspired by the traditional stacking approach and the work in [17, 39], we aim to reduce computational cost by using only the base models from the stacking method and then aggregating their outputs to derive the nodes’ attributes. Specifically, our proposed method operates as follows: in the first layer, we obtain XX^{\prime} by passing XN×dX\in\mathbb{R}^{N\times d} (the feature matrix) through kk base classifiers.

Xi=hi(X),i=1,,k,X^{\prime}_{i}=h_{i}(X),\qquad i=1,\cdots,k, (2)

Next, we obtain the preliminary classification results XiX^{\prime}_{i} by passing XX through each base classifier hi(X)h_{i}(X) for i=1,,ki=1,\cdots,k. Then, we apply an aggregation method to produce the final output results, i.e.,

X~=g(X1,,Xk),{\tilde{X}}=g(X^{\prime}_{1},\cdots,X^{\prime}_{k}), (3)

where g()g(\cdot) denotes the aggregation method, which is designed to combine attribute values into a single representative value. Common aggregation methods include mean, attention, or voting approaches.

Refer to caption
Figure 1: Workflow of the SStaGCN model.
Algorithm 1 SStaGCN.
0:    Feature matrix XX, normalized adjacency matrix A^\hat{A}, graph 𝒢\mathcal{G}, base classifier hi(X)h_{i}(X), i=1,,ki=1,\cdots,k, aggregation method g()g(\cdot);
0:    Final predictor f(X)f(X);
1:  Attain XiX^{\prime}_{i} (i=1,,ki=1,\cdots,k) via kk base classifiers;Xihi(X)X^{\prime}_{i}\leftarrow h_{i}(X), i=1,,ki=1,\cdots,k;
2:  Aggregate X1,,XkX^{\prime}_{1},\cdots,X^{\prime}_{k};X~g(X1,,Xk){\tilde{X}}\leftarrow g(X^{\prime}_{1},\cdots,X^{\prime}_{k});
3:  Feed X~{\tilde{X}} into vanilla GCN [19];f(X)GCN(X~)f(X)\leftarrow GCN({\tilde{X}});
4:  return  f(X)f(X);

Mean: mean operator takes the element-wise mean of the components X1,,XkX^{\prime}_{1},\cdots,X^{\prime}_{k}.

Attention [40]: The attention mechanism has been widely applied in various fields of deep learning, including image processing [29], speech recognition [1], and natural language processing [47]. This concept is inspired by the way humans focus attention selectively. Let the query vector VV represent the output of the base classifiers, and let the query vector YY denote the data label. We then compute the attention coefficients between YY and VV as follows:

ai=softmax(cos(Y,Vi)),i=1,,k,a_{i}={{\rm softmax}}({\rm cos}(Y,V_{i})),\qquad i=1,\cdots,k, (4)

where cos\cos denotes cosine similarity. Finally, the input X~{\tilde{X}} of the graph convolutional layer is aggregated by considering the following sum with attention scores aia_{i}, i=1,,ki=1,\cdots,k.

X~=i=1kaiVi.{\tilde{X}}=\sum^{k}_{i=1}a_{i}V_{i}. (5)

Voting [33]: The voting method is the most straightforward approach in ensemble learning, aiming to select one or more “winning” predictions. In this work, our goal is to select the most common prediction among the outputs of the base classifiers, using a majority voting approach. In cases where categories appear with equal frequency, a category will be chosen at random.

This brings us to a novel GCN model, SStaGCN, specifically designed to handle diverse types of graph-structured data by effectively integrating stacking, aggregation, and the vanilla GCN model [19]. In SStaGCN, the first layer leverages base models from the stacking approach, while the second layer applies an aggregation method—such as mean, attention, or voting—to enhance the feature extraction capabilities of standard GCN models. The aggregated data is then used as input to the convolutional layer, which produces the final predictions. The workflow of the proposed model is presented in Algorithm 1 and Fig. 1. This prompts the question: is there a theoretical guarantee for the proposed approach?

3.4 Generalization bound

In this section, we provide a theoretical generalization analysis of the proposed approach. In the following analysis, we assume that both the adjacency matrix AA and the feature matrix XX are fixed.

In learning theory, the risk of a function ff over the unknown population distribution \mathbb{P} is measured by

(f):=𝔼𝒳×𝒴[L(y,f(𝐱~))],\mathcal{E}(f):=\mathbb{E}_{\mathcal{X}\times\mathcal{Y}}[L(y,f(\tilde{\bf x}))],

where LL is the loss function defined as a map: L:𝒳×𝒴+L:\mathcal{X}\times\mathcal{Y}\rightarrow\mathbb{R}^{+}, Given a training data and adjacency matrix AA, the objective is to estimate parameters (W(0),W(1))(W^{(0)},W^{(1)}) from model (1) based on empirical data. Concretely, we attempt to minimize the empirical risk functional over some function class \mathcal{F},

m(f):=1mj=1mL(yj,f(𝐱~j)),\mathcal{E}_{m}(f):=\frac{1}{m}\sum_{j=1}^{m}L(y_{j},f(\tilde{\bf x}_{j})),

where {(𝐱~j,yj)j=1m}\{(\tilde{{\bf x}}_{j},y_{j})^{m}_{j=1}\} is the labeled sample achieved from the original training data {(𝐱j,yj)j=1m}\{({\bf x}_{j},y_{j})^{m}_{j=1}\} via stacking and aggregation. Typically, the clustering algorithms will only produce discernible node features. Hence, if 𝐱iR,i=1,,N\|{\bf x}_{i}\|\leq R,i=1,\cdots,N, the stacking and aggregation mechanisms will not make 𝐱~i\tilde{\bf x}_{i}, i=1,,Ni=1,\cdots,N violate the constraints, i.e., 𝐱~iR,i=1,,N\|\tilde{\bf x}_{i}\|\leq R,i=1,\cdots,N. Now we are in position to present the theoretical generalization bound.

Theorem 3.1.

Suppose 𝐱i2R,i=1,,N\|{\bf x}_{i}\|_{2}\leq R,i=1,\cdots,N, W(0)FB1\|W^{(0)}\|_{F}\leq B_{1}, W(1)FB2\|W^{(1)}\|_{F}\leq B_{2}. Denote N(v)N(v) the number of neighbors of node vΩv\in\Omega (the set of node indices with observed labels), let q=max{N(v)}q=\max\{N(v)\}, f:𝒳f:\mathcal{X}\rightarrow\mathbb{R} be any given predictor of a class for GCNs with one-hidden layer. Assume that the loss function L(y,)L(y,\cdot) is Lipschitz continuous with Lipschitz constant αL\alpha_{L}. Then, for any δ>0\delta>0, with probability at least 1δ1-\delta, we have

(f)N(f)+2αL2qKB1B2Rs=1qMsm+2log(2/δ)N,\mathcal{E}(f)\leq\mathcal{E}_{N}(f)+\frac{2\alpha_{L}\sqrt{2q}KB_{1}B_{2}R\sum^{q}_{s=1}M_{s}}{\sqrt{m}}+\sqrt{\frac{2\log(2/\delta)}{N}},

where F\|\cdot\|_{F} is the Frobenius norm, Ms=maxi[N]|A^iv|M_{s}=\max_{i\in[N]}|\hat{A}_{iv}| with vN(i)v\in N(i).

Remark 3.2.

Theorem 3.1 indicates that the dominant upper bound depends linearly on the maximum number of neighbors qq, as well as on the weight bounds B1B_{1} and B2B_{2}, which are strongly influenced by the dimension d1d_{1}. Clearly, setting d1dd_{1}\ll d results in a tighter generalization bound. In the binary classification case (K=1K=1), the result presented here is similar to that in [26]. Overall, the conditions stated in Theorem 3.1 are mild and reasonable.

4 Experiments

4.1 Datasets

To evaluate the performance of the proposed SStaGCN model across various types of graph-structured data, we utilize six real-world datasets for a semi-supervised node classification task. This includes three commonly used citation networks—Cora, CiteSeer, and Pubmed [35]—and three heterogeneous tabular datasets: House_\_class, VK_\_class, and DBLP [17]. These heterogeneous tabular datasets differ from those in [42], containing graph data with diverse edge and node types.

In the citation networks, nodes represent documents, and undirected edges denote citation relationships between documents. Node features correspond to representative words in the documents, while the label rate indicates the percentage of node labels used for training. The Cora dataset comprises 27082708 nodes, 54295429 edges, 77 classes, and 14331433 node features; CiteSeer has 33273327 nodes, 47324732 edges, 66 classes, and 37033703 node features; and Pubmed includes 1971719717 nodes, 44384438 edges, and 33 classes. We use 140140, 120120, and 6060 nodes for training in Cora, CiteSeer, and Pubmed, respectively, and allocate 10001000 nodes for testing and 500500 nodes for cross-validation. This data split matches that used in GCN, Graph Attention Network (GAT, [41]), and GWNN [45].

According to [17], House_\_class and VK_\_class datasets are derived from House and VK datasets, where target labels are converted into discrete classes due to limited availability of publicly accessible heterogeneous graph-structured data. In these heterogeneous tabular datasets, features are independently defined and vary in type, scale, and meaning. The VK dataset represents a popular social network, with node features that are both numerical (e.g., last active time) and categorical (e.g., country and university affiliation). Similar to [17], we categorize age into bins: <20<20, 202520–25, 253025–30, \cdots, 455045–50, >50>50. For the House dataset, we categorize target values within the range [1.0,1.5,2.0,2.5][1.0,1.5,2.0,2.5], resulting in five classes for House_\_class and seven for VK_\_class. In the DBLP dataset, we construct a single graph by focusing on the APA (author-paper-author) meta-path. Each dataset is divided into training, validation, and testing splits in a 0.6/0.2/0.20.6/0.2/0.2 ratio across five random seeds.

Details about the citation networks and heterogeneous tabular datasets are presented in Table 1 and Table 2, respectively.

\tbl

Summary of the citation networks. \topruleDataset Cora CiteSeer Pubmed \colruleNodes 2708 3327 19717 Edges 5429 4732 44338 Features 1433 3703 500 Classes 7 6 3 Label Rate 5.2% 3.6% 0.3% \botrule

\tbl

Summary of the heterogeneous tabular datasets. \topruleDataset House_\_class VK_\_class DBLP \colruleNodes 20640 54028 14475 Edges 182146 213644 40269 Features 6 14 5002 Classes 5 7 4 Min Target Nodes 0.14 13.48 745 Max Target Nodes 5.00 118.39 1197 \botrule

4.2 Baselines

We compare SStaGCN with four classical graph convolutional networks—ChebyNet, GCN [19], GAT, and APPNP [21]—as well as two ensemble-based GCN models: AdaGCN and BGNN.

4.3 Setting

As shown in Fig. 1, the proposed SStaGCN model consists of four layers, with the first and second layers referred to as the stacking and aggregation layers, respectively. The stacking layer utilizes base models from the stacking method, incorporating a combination of seven classical classifiers: KNN, Random Forest, Naive Bayes, Decision Tree, SVC [30], GBDT [12], and Adaboost [11]. These classifiers are widely used in classical machine learning for their strengths in handling different task types.

In the aggregation layer, we employ three aggregation methods: mean, attention, and voting. For the mean approach, we calculate the average output from the stacking layer, rounding it as needed. In the attention mechanism, we treat the label data as vector YY, use the stacking layer’s predicted output as the query vector VV, and compute attention coefficients accordingly. For the voting method, we apply hard voting from ensemble learning [34].

The output of the aggregation layer serves as the input to the first layer of a two-layer GCN. In our configuration, the GCN has 16 hidden units and uses the Adam optimizer [18] by default, with cross-entropy as the loss function. The learning rate is set to γ=0.01\gamma=0.01, the number of iterations itr=500{\rm itr}=500, weight decay at 5e45e-4, and a dropout rate of 0.

4.4 Results

\tbl

Average accuracy on 33 citation networks under 3030 runs by computing the 95%95\% confidence interval via bootstrap. \topruleMethod Cora CiteSeer Pubmed \colruleChebyNet 81.20±0.00\pm 0.00 69.80±\pm0.00 74.40±\pm0.00 GCN 81.50±\pm 0.00 70.30±\pm0.00 79.00±\pm0.00 GAT 83.00±\pm0.70 72.50±\pm0.70 79.00±\pm0.30 APPNP 85.09±\pm0.25 75.73±\pm0.30 79.73±\pm0.31 AdaGCN 85.97±\pm0.20 76.68±\pm0.20 79.95±\pm0.21 BGNN 41.97±\pm0.19 30.74±\pm0.10 10.32±\pm0.10 \colruleSim_\_Stacking 43.19±\pm2.05 62.70±\pm0.53 87.70±\pm0.23 SStaGCN (Mean) 90.35±\pm0.20 86.40±\pm0.12 82.30±\pm0.19 SStaGCN (Attention) 91.60±\pm0.18 87.20±\pm0.12 82.40±\pm0.23 SStaGCN (Voting) 93.10±\pm0.16 88.70±\pm0.14 92.07±\pm0.20 \botrule

\tbl

Average accuracy on heterogeneous tabular datasets under 3030 runs by calculating the 95%95\% confidence interval via bootstrap. \topruleMethod House_\_class VK_\_class DBLP \colruleChebyNet 54.74±\pm0.10 57.19±\pm0.36 32.14±\pm0.00 GCN 55.07±\pm0.13 56.40±\pm0.09 39.49±\pm1.37 GAT 56.50±\pm0.22 56.42±\pm0.19 76.83±\pm0.78 APPNP 57.03±\pm0.27 56.72±\pm0.11 79.47±\pm1.46 \colruleAdaGCN 26.20±\pm0.00 46.00±\pm0.00 10.06±\pm0.00 BGNN 66.70±\pm0.27 66.32±\pm0.20 86.94±\pm0.74 \colruleSim_\_Stacking 53.89±\pm0.29 56.64±\pm0.10 71.58±\pm0.64 SStaGCN (Mean) 72.35±\pm0.05 66.62±\pm0.17 82.31±\pm0.20 SStaGCN (Attention) 72.40±\pm0.12 77.64±\pm0.08 82.51±\pm0.22 SStaGCN (Voting) 76.13±\pm0.12 87.92±\pm0.07 92.60±\pm0.10 \botrule

\tbl

Average F1-score (macro) on 33 citation networks under 3030 runs by computing the 95%95\% confidence interval via bootstrap. \topruleMethod Cora CiteSeer Pubmed \colruleChebyNet 77.99±\pm0.54 63.76±\pm0.34 77.74±\pm0.42 GCN 82.89±\pm0.30 70.65±\pm0.37 78.83±\pm0.32 GAT 83.59±\pm0.25 70.62±\pm0.29 77.77±\pm0.40 APPNP 84.29±\pm0.22 71.05±\pm0.38 79.66±\pm0.31 \colruleAdaGCN 79.55±\pm0.19 63.62±\pm0.19 78.55±\pm0.21 BGNN 40.81±\pm0.25 32.73±\pm0.13 8.46±\pm0.08 \colruleSim_\_Stacking 44.02±\pm1.61 60.86±\pm0.56 87.31±\pm0.13 SStaGCN (Mean) 90.66±\pm0.18 86.42±\pm0.12 82.30±\pm0.19 SStaGCN (Attention) 91.69±\pm0.14 87.24±\pm0.14 82.45±\pm0.23 SStaGCN (Voting) 92.76±\pm0.16 88.73±\pm0.14 92.07±\pm0.20 \botrule

\tbl

Average F1-score(macro) on heterogeneous tabular datasets under 3030 runs by calculating the 95%95\% confidence interval via bootstrap. \topruleMethod House_\_class VK_\_class DBLP \colruleChebyNet 31.34±\pm0.12 57.44±\pm0.27 26.84±\pm0.62 GCN 54.95±\pm0.14 56.52±\pm0.09 38.5±\pm0.97 GAT 56.54±\pm0.68 56.41±\pm0.07 77.1±\pm1.86 APPNP 57.88±\pm0.32 56.61±\pm0.07 79.34±\pm0.23 \colruleAdaGCN 25.01±\pm0.00 37.03±\pm0.00 9.60±\pm0.00 BGNN 66.48±\pm0.22 66.18±\pm0.11 87.2±\pm0.60 \colruleSim_\_Stacking 53.32±\pm0.15 56.11±\pm0.08 71.49±\pm0.31 SStaGCN (Mean) 72.23±\pm0.04 66.74±\pm0.21 82.13±\pm0.38 SStaGCN (Attention) 72.36±\pm0.09 77.62±\pm0.10 82.68±\pm0.12 SStaGCN (Voting) 75.45±\pm0.82 87.84±\pm0.04 92.64±\pm0.06 \botrule

\tbl

p-values of the paired t-test of SStaGCN (Voting) with competitors on 66 different datasets (Cora, Citeseer, Pubmed, House_\_class, VK_\_class, and DBLP). \topruleModels Cora CiteSeer Pubmed House_\_class VK_\_class DBLP \colruleChebyNet 2.59e-06 2.77e-08 1.17e-06 3.51e-10 1.11e-11 6.97e-09 GCN 4.19e-16 5.15e-17 1.84e-15 2.36e-08 2.25e-11 2.48e-07 GAT 2.10e-19 1.54e-20 8.84e-19 3.35e-08 1.38e-09 4.95e-06 APPNP 6.86e-42 7.12e-42 6.25e-41 7.05e-15 1.39e-21 2.26e-09 AdaGCN 1.82e-19 1.23e-20 8.42e-19 3.05e-38 7.94e-43 1.69e-35 BGNN 1.02e-24 2.33e-21 4.74e-22 6.54e-07 1.07e-08 0.11e-03 Sim_\_Stacking 5.61e-12 3.79e-15 2.52e-09 4.56e-08 6.07e-11 1.07e-06 \botrule

The results of the comparative evaluation for node classification are summarized in Tables 3-11. In these tables, SStaGCN (Mean) refers to the use of the mean aggregation mechanism in the second layer of SStaGCN, while SStaGCN (Attention) and SStaGCN (Voting) correspond to the attention and voting mechanisms, respectively. We report the accuracy, macro F1-score, and training time on the test set for the proposed SStaGCN model and other methods. The experimental results demonstrate a significant improvement of the SStaGCN model over the baselines. Specifically, for the three public citation networks, SStaGCN (Voting) achieves an improvement in accuracy (and F1-score) of approximately 8%8\% (8%8\%), 12%12\% (17%17\%), and 13%13\% (12%12\%) for the Cora, CiteSeer, and Pubmed datasets, respectively. For the heterogeneous tabular datasets, SStaGCN (Voting) shows an improvement in accuracy (and F1-score) of about 9%9\% (9%9\%), 21%21\% (21%21\%), and 6%6\% (5%5\%) for the House_\_class, VK_\_class, and DBLP datasets, respectively.

AdaGCN performs poorly on the heterogeneous tabular datasets, possibly because AdaGCN is designed for deeper GCN architectures, which may mix node features from different clusters as the GCN layers deepen. SStaGCN, by contrast, enhances the performance of GCN, providing better results across distinct types of graph-structured data. Additionally, the paired t-test results in Table 7 indicate that the proposed SStaGCN model significantly outperforms the simplified stacking and other GCN models.

This impressive improvement can be explained as follows:

{itemlist}

The stacking and aggregation steps in SStaGCN provide a dimensionality reduction effect, making the graph data more discernible. For example, in the Cora dataset, the feature size reduces from 2708×14332708\times 1433 to 2708×72708\times 7 after stacking and aggregation. This results in a relatively smaller d1d_{1}, as discussed in Remark 1, significantly enhancing both the predictive power and computational efficiency of the subsequent graph convolution model.

In the aggregation step of the SStaGCN model, the mean and attention mechanisms somewhat disrupt the pre-classification results, making them less suitable for feature extraction, whereas the voting mechanism preserves these results. Consequently, experimental results demonstrate that SStaGCN (Voting) is more effective across the six datasets.

Simplified stacking can effectively extract useful attributes but overlooks graph structure information, while GCN models are limited in feature extraction. Therefore, the SStaGCN model combines the strengths of simplified stacking and GCN, achieving both higher classification accuracy and reduced computational cost.

Tables 8 and 9 present a comparison of training times between SStaGCN and other methods. BGNN achieves the fastest runtime on the three citation networks, followed closely by our SStaGCN method. However, SStaGCN runs faster than other methods on the heterogeneous tabular datasets, with the exception of the DBLP dataset. We attribute this to the extra computation time required for the stacking and aggregation steps, which ultimately enhance efficiency when feeding the feature outputs into the GCN model.

\tbl

Average training time(s) on 33 citation networks by computing the 95%95\% confidence interval via bootstrap. \topruleMethod Cora CiteSeer Pubmed \colruleChebyNet 22.74±\pm1.24 30.87±\pm1.21 124.99±\pm1.77 GCN 13.41±\pm0.16 99.21±\pm0.98 55.61±\pm0.73 GAT 20.98±\pm0.46 30.74±\pm1.40 126.33±\pm1.80 APPNP 203.75±\pm0.15 55.40±\pm0.40 457.62±\pm12.77 \colruleAdaGCN 772.26±\pm83.56 2129.02±\pm148.97 2098.10±\pm275.88 BGNN 1.33±\pm0.00 2.40±\pm0.00 2.54±\pm0.00 \colruleSim_\_Stacking 11.9±\pm0.08 27.9±\pm0.24 79.1±\pm1.40 SStaGCN (Mean) 10.9±\pm0.13 17.2±\pm0.13 89.6±\pm2.20 SStaGCN (Attention) 11.2±\pm0.24 17.6±\pm0.41 87.6±\pm0.96 SStaGCN (Voting) 16.2±\pm0.19 29.6±\pm1.40 13.1±\pm2.61 \botrule

\tbl

Average training time(s) on heterogeneous tabular datasets by calculating the 95%95\% confidence interval via bootstrap. \topruleMethod House_\_class VK_\_class DBLP \colruleChebyNet 833.82±\pm0.00 1394.68±\pm0.00 8890.74±\pm0.00 GCN 46.06±\pm0.80 120.1±\pm3.35 268.5±\pm5.35 GAT 197.6±\pm3.08 410.9±\pm9.95 205.3±\pm2.00 APPNP 129.7±\pm3.26 383.8±\pm12.02 176.8±\pm5.58 \colruleAdaGCN 607.31±\pm0.00 511.41±\pm0.00 590.90±\pm0.00 BGNN 26.37±\pm1.65 93.47±\pm6.23 50.25±\pm3.04 \colruleSim_\_Stacking 16.41±\pm0.36 43.58±\pm2.78 380.8±\pm12.82 SStaGCN (Mean) 52.94±\pm0.51 132.9±\pm1.61 188.7±\pm0.54 SStaGCN (Attention) 57.38±\pm1.75 133.2±\pm1.11 246.2±\pm0.37 SStaGCN (Voting) 59.69±\pm0.82 154.1±\pm1.02 310.7±\pm0.49 \botrule

To demonstrate the effect of stacking and aggregation steps in the proposed model, we provide a t-SNE visualization [27] in Figs. 2(a) and 3(a). These figures show that the combination of stacking and aggregation effectively extracts features, enhancing the discriminative power of the graph data.

Refer to caption
(a) GCN
Refer to caption
(b) SStaGCN
Figure 2: Visualization of classification features by the GCN (left) and the features after conducting stacking and aggregation steps in the SStaGCN model (right) on CiteSeer dataset, node colors denote classes.

Refer to caption
(a) GCN
Refer to caption
(b) SStaGCN
Figure 3: Visualization of classification features by the GCN (left) and the features after conducting stacking and aggregation steps in the SStaGCN model (right) on DBLP dataset, node colors denote classes.

Table 10 suggests that using all seven classifiers is unnecessary. For example, on the Cora dataset, the combination of KNN, Random Forest, and Naive Bayes achieves the highest accuracy with minimal computational cost (only 3 seconds). This indicates that each of the seven classifiers has unique strengths suited to different tasks. However, selecting the optimal combination of base models remains an open area for theoretical analysis.

\tbl

Accuracy and training time(s) on Cora dataset by combinations of different classifiers based on SStaGCN model. \topruleKNN Random Forest Naive Bayes Decision Tree GBDT Adaboost SVC Accuracy Training Time \colrule \surd \surd 91.2 13.90 \surd \surd \surd 93.6 16.60 \surd \surd 84.2 567.9 \surd \surd \surd 92.9 144.7 \surd \surd \surd 93.1 149.5 \surd \surd \surd 92.8 15.90 \surd \surd \surd \surd 93.4 18.80 \surd \surd \surd \surd \surd 92.9 568.3 \surd \surd \surd \surd \surd \surd 92.9 570.7 \surd \surd \surd \surd \surd \surd \surd 92.8 708.5 \botrule

To further illustrate the impact of simplified stacking on the over-smoothing problem, we conduct an additional experiment on this topic. In Fig. 4(a), we observe that a conventional GCN tends to blend node features from different clusters as the number of convolutional layers increases. However, as shown in Fig. 5(a) and Table 11333The number of layers excludes those included in the stacking and aggregation parts of SStaGCN, which consist of only two layers., the proposed SStaGCN effectively mitigates the over-smoothing phenomenon and enhances accuracy.

Refer to caption
(a) 2-layer
Refer to caption
(b) 3-layer
Refer to caption
(c) 4-layer
Refer to caption
(d) 5-layer
Refer to caption
(e) 6-layer
Refer to caption
(f) 7-layer
Figure 4: Visualization of final classification features via GCN on Cora dataset with 22, 33, 44, 55, 66, 77 layers, node colors denote classes.

Refer to caption
(a) 2-layer
Refer to caption
(b) 3-layer
Refer to caption
(c) 4-layer
Refer to caption
(d) 5-layer
Refer to caption
(e) 6-layer
Refer to caption
(f) 7-layer
Figure 5: Visualization of final classification features via SStaGCN on Cora dataset with 22, 33, 44, 55, 66, 77 layers, node colors denote classes.

\tbl

Accuracy comparison between GCN and SStaGCN models on Cora dataset using distinct number of layers. \topruleMethod 2-layer 3-layer 4-layer 5-layer 6-layer 7-layer \colruleGCN 80.5 80.4 75.8 71.9 72.6 60.8 SStaGCN 93.3 88.8 87.5 86.4 84.8 84.3 \botrule

Overall, these experiments demonstrate the superiority of SStaGCN model over competitors.

4.5 Visualization

Refer to caption
(a) GCN
Refer to caption
(b) AdaGCN
Refer to caption
(c) BGNN
Refer to caption
(d) SStaGCN
Figure 6: Visualization of final classification features via (a). GCN , (b). AdaGCN , (c). BGNN , and (d). SStaGCN model on Citeseer dataset, node colors denote classes.

Refer to caption
(a) GCN
Refer to caption
(b) AdaGCN
Refer to caption
(c) BGNN
Refer to caption
(d) SStaGCN
Figure 7: Visualization of final classification features via (a). GCN , (b). AdaGCN , (c). BGNN, and (d). SStaGCN model on DBLP dataset, node colors denote classes.

To further illustrate the performance of SStaGCN, we plot the final classification features of GCN, AdaGCN, BGNN, and our SStaGCN. Fig. 6(a) (for CiteSeer) and Fig. 7(a) (for DBLP) display the final classification features of each method on these datasets. As shown in these figures, the proposed SStaGCN misclassifies relatively fewer points, while GCN, AdaGCN, and BGNN result in more misclassified classes.

5 Conclusion

Traditional GCNs often face the over-smoothing problem. In this work, we propose a novel GCN architecture, SStaGCN, which leverages stacking and aggregation techniques to capture pre-classified data features, followed by GCN for predictions on distinct graph-structured data. SStaGCN effectively explores and utilizes features for heterogeneous graph data in a stacked manner. By incorporating classical machine learning methods, we design a GCN model that provides a versatile framework for handling diverse types of graph-structured data, offering new insights into understanding GCNs. Extensive experiments demonstrate that the proposed model outperforms several state-of-the-art competitors in terms of accuracy, F1-score, and training time. The framework presented here could also be extended to regression tasks. Additionally, we believe this method can address various types of heterogeneous graph data beyond tabular data. A promising future research direction is to investigate deeper GCNs within our framework, as suggested by [39]. The source code of SStaGCN will be issued soon 444https://github.com/dragon0916/SStaGCN..

Appendix

In this part, we provide a detailed proof of Theorem 3.1. Before proceeding with the proof, we introduce several supporting lemmas. First, we present the contraction inequality for Rademacher complexity in vector form.

Lemma 5.1.

[28] Let 𝒳{\cal X} be any set, (𝐱1,,𝐱N)𝒳({\bf x}_{1},\cdots,{\bf x}_{N})\in{\cal X} and {\cal F} be a class of functions f:𝒳Kf:{\cal X}\to{\mathbb{R}}^{K} ad let τi:K\tau_{i}:{\mathbb{R}}^{K}\to{\mathbb{R}} have Lipschitz constant MM. Then

𝔼supfi=1Nσiτif(𝐱i)2M𝔼supfi=1Nk=1Kσikfk(𝐱i),{\mathbb{E}}\sup_{f\in{\cal F}}\sum^{N}_{i=1}\sigma_{i}\tau_{i}f({\bf x}_{i})\leq\sqrt{2}M{\mathbb{E}}\sup_{f\in{\cal F}}\sum^{N}_{i=1}\sum^{K}_{k=1}\sigma_{ik}f_{k}({\bf x}_{i}),

where σik\sigma_{ik} is an independent doubly indexed Rademacher sequence and fk(𝐱i)f_{k}({\bf x}_{i}) is the kk-th component of f(𝐱i)f({\bf x}_{i}).

Lemma 5.2.

[2] Consider a loss function L:𝒳×𝒴+L:{\cal X}\times{\cal Y}\to{\mathbb{R}}^{+}. Denote ={L(y,f()),f}{\cal H}=\{L(y,f(\cdot)),f\in{\cal F}\}, and let (𝐱i,yi)i=1N({\bf x}_{i},y_{i})^{N}_{i=1} be independently selected according to the probability measure \mathbb{P}. Then for any 0<δ<10<\delta<1, with probability at least 1δ1-\delta,

(f)N(f)+2^()+2log(2/δ)N,f.{\cal E}(f)\leq{\cal E}_{N}(f)+2{\cal{\widehat{R}}}({\cal H})+\sqrt{\frac{2\log(2/\delta)}{N}},~{}~{}\forall f\in{\cal H}.

We first give a lemma which plays essential role in the proof of Theorem 3.1.

Lemma 5.3.

Let q=max{N(v)}q=\max\{N(v)\} for each node vΩv\in\Omega, then

maxvjN(v)A^vj𝐱j22R2q.\max_{v}\|\sum_{j\in N(v)}{\hat{A}}_{vj}{\bf x}_{j}\|^{2}_{2}\leq R^{2}q.
Proof 5.4.

Denote A^vqv×qv{\hat{A}}_{v}\in{\mathbb{R}}^{q_{v}\times q_{v}} as the sub-matrix of A^\hat{A} whose row and column indices belong to the set jN(v)j\in N(v). Hence the size of A^v{\hat{A}}_{v} depends on the node vv. Obviously, q=maxqvq=\max{q_{v}} for vΩv\in\Omega. Let X~v=(𝐱1T,,𝐱qT)Tq×d{\tilde{X}}_{v}=({\bf x}^{T}_{1},\cdots,{\bf x}^{T}_{q})^{T}\in{\mathbb{R}}^{q\times d} be the feature matrix of the nodes in 𝒢v{\cal G}_{v} (subgraph of 𝒢\cal G). Hence,

maxvjN(v)A^vj𝐱j22=maxvA^vX~v22maxvA^v22X~v22,\max_{v}\Big{\|}\sum_{j\in N(v)}{\hat{A}}_{vj}{\bf x}_{j}\Big{\|}^{2}_{2}=\max_{v}\|{\hat{A}}_{v\cdot}{\tilde{X}}_{v}\|^{2}_{2}\leq\max_{v}\|{\hat{A}}_{v\cdot}\|^{2}_{2}\|{\tilde{X}}_{v}\|^{2}_{2},

where A^v\hat{A}_{v\cdot} is the vv-\!th row of the matrix A^\hat{A} with column index jj belong to the set N(v)N(v). Notice that

X~v2=supt2=1X~vt2s=1qv𝐱s22RqvRq,\|{\tilde{X}}_{v}\|_{2}=\sup_{\|t\|_{2}=1}\|{\tilde{X}}_{v}t\|_{2}\leq\sqrt{\sum^{q_{v}}_{s=1}\|{\bf x}_{s}\|^{2}_{2}}\leq R\sqrt{q_{v}}\leq R\sqrt{q},

and A^21\|\hat{A}\|_{2}\leq 1. Therefore,

maxvjN(v)A^vjxj22A^22X~v22R2q.\max_{v}\Big{\|}\sum_{j\in N(v)}{\hat{A}}_{vj}x_{j}\Big{\|}^{2}_{2}\leq\|\hat{A}\|^{2}_{2}\|{\tilde{X}}_{v}\|^{2}_{2}\leq R^{2}q.

Now we are in position to give the proof of Theorem 3.1.

Proof 5.5 (Proof of Theorem 3.1).

To allow a slight abuse of notations, we will use XjX_{j} to denote X~j\tilde{X}_{j} due to the explanation on page 3. Denote h(W(0))=ReLUh(W^{(0)})={\rm ReLU}(A^XW(0))({\hat{A}}XW^{(0)}), f(W(1))f(W^{(1)})== softmax{\rm softmax} (h^(W(0))W(1))({\hat{h}}(W^{(0)})W^{(1)})== (f1(W(1)),,fm(W(1)))T(f_{1}(W^{(1)}),\cdots,f_{m}(W^{(1)}))^{T} with h^(W(0))=A^h(W(0)){\hat{h}}(W^{(0)})={\hat{A}}h(W^{(0)}). Applying Proposition 4 in [13] to the case λ=1\lambda=1, we can attain the Lipschitz constant for standard softmax function is M=1M=1. Let A^i\hat{A}_{i\cdot} stands for the ii-th row of the matrix A^\hat{A}, for function set

B1,B2={fi=softmax(A^iReLU(A^XW(0))W(1)),i=1,,m,\displaystyle{\cal F}_{B_{1},B_{2}}=\{f_{i}={\rm softmax}({\hat{A}}_{i\cdot}{\rm ReLU}({\hat{A}}XW^{(0)})W^{(1)}),i=1,\cdots,m,
W(0)FB1,W(1)FB2},\displaystyle\|W^{(0)}\|_{F}\leq B_{1},\|W^{(1)}\|_{F}\leq B_{2}\},

the empirical Rademacher complexity is defined as

^(B1,B2)=𝔼σ[1msupW(1)FB2W(0)FB1i=1mσifi(W(1))],{\cal{\hat{R}}}({{\cal F}}_{B_{1},B_{2}})={\mathbb{E}}_{\sigma}\Big{[}\frac{1}{m}\sup_{\stackrel{{\scriptstyle\|W^{(0)}\|_{F}\leq B_{1}}}{{\|W^{(1)}\|_{F}\leq B_{2}}}}\sum^{m}_{i=1}\sigma_{i}f_{i}(W^{(1)})\Big{]},

where {σi}i=1m\{\sigma_{i}\}^{m}_{i=1} is an i.i.d. family of Rademacher variables independent of 𝐱i{\bf x}_{i}. By the contraction property of Rademacher complexity,

^()αL^(B1,B2),{\cal{\hat{R}}}({\cal H})\leq\alpha_{L}{\cal{\hat{R}}}({{\cal F}}_{B_{1},B_{2}}),

and notice Lemma 5.2, we only need to bound ^(B1,B2){\cal\hat{R}}({\cal F}_{B_{1},B_{2}}). Therefore, we have the following estimate by utilizing Lemma 5.1.

^(B1,B2)\displaystyle{\cal\hat{R}}({\cal F}_{B_{1},B_{2}}) =𝔼σ[1msupW(1)FB2W(0)FB1i=1mσifi(W(1))]\displaystyle={\mathbb{E}}_{\sigma}\Big{[}\frac{1}{m}\sup_{\stackrel{{\scriptstyle\|W^{(0)}\|_{F}\leq B_{1}}}{{\|W^{(1)}\|_{F}\leq B_{2}}}}\sum^{m}_{i=1}\sigma_{i}f_{i}(W^{(1)})\Big{]}
2m𝔼σ[supW(1)FB2W(0)FB1i=1mk=1Kσikh^i(W(0))𝐰k(1)],\displaystyle\leq\frac{\sqrt{2}}{m}{\mathbb{E}}_{\sigma}\Big{[}\sup_{\stackrel{{\scriptstyle\|W^{(0)}\|_{F}\leq B_{1}}}{{\|W^{(1)}\|_{F}\leq B_{2}}}}\sum^{m}_{i=1}\sum^{K}_{k=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)}){\bf w}^{(1)}_{k}\Big{]},

where W(1)=(𝐰1(1),,𝐰K(1))W^{(1)}=({\bf w}^{(1)}_{1},\cdots,{\bf w}^{(1)}_{K}), notice the property of inner product, the above estimate can be further bounded as

^(B1,B2)\displaystyle{\cal\hat{R}}({\cal F}_{B_{1},B_{2}}) 2m𝔼σ[supW(1)FB2W(0)FB1k=1Kmaxk[K]𝐰k(1)2i=1mσikh^i(W(0))2],\displaystyle\leq\frac{\sqrt{2}}{m}{\mathbb{E}}_{\sigma}\Big{[}\sup_{\stackrel{{\scriptstyle\|W^{(0)}\|_{F}\leq B_{1}}}{{\|W^{(1)}\|_{F}\leq B_{2}}}}\sum^{K}_{k=1}\max_{k\in[K]}\|{\bf w}^{(1)}_{k}\|_{2}\big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\big{\|}_{2}\Big{]},
2m𝔼σ[supW(1)FB2W(0)FB1W(1)Fk=1Ki=1mσikh^i(W(0))2],\displaystyle\leq\frac{\sqrt{2}}{m}{\mathbb{E}}_{\sigma}\Big{[}\sup_{\stackrel{{\scriptstyle\|W^{(0)}\|_{F}\leq B_{1}}}{{\|W^{(1)}\|_{F}\leq B_{2}}}}\|W^{(1)}\|_{F}\sum^{K}_{k=1}\big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\big{\|}_{2}\Big{]},
2B2m𝔼σ[supW(0)FB1k=1Ki=1mσikh^i(W(0))2],\displaystyle\leq\frac{\sqrt{2}B_{2}}{m}{\mathbb{E}}_{\sigma}\Big{[}\sup_{\|W^{(0)}\|_{F}\leq B_{1}}\sum^{K}_{k=1}\big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\big{\|}_{2}\Big{]},
2B2m𝔼σ[k=1KsupW(0)FB1i=1mσikh^i(W(0))2],\displaystyle\leq\frac{\sqrt{2}B_{2}}{m}{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|W^{(0)}\|_{F}\leq B_{1}}\big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\big{\|}_{2}\Big{]},

the last inequality follows by the property that sup(sas)ssup(as)\sup(\sum_{s}a_{s})\leq\sum_{s}\sup(a_{s}). Now the key point is how to estimate the term supW(0)FB1i=1mσikh^i(W(0))2\sup_{\|W^{(0)}\|_{F}\leq B_{1}}\big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\big{\|}_{2}. We will employ the idea introduced in [26] (in the proof of Theorem 1) to remove the “sup” term. Let hv(W(0))h_{v}(W^{(0)}) == (hv1(𝐰1(0))\Big{(}h^{1}_{v}({\bf w}^{(0)}_{1}), hv2(𝐰2(0))h^{2}_{v}({\bf w}^{(0)}_{2}), \cdots, hvH(𝐰H(0)))h^{H}_{v}({\bf w}^{(0)}_{H})\Big{)} with W(0)W^{(0)} == (𝐰1(0)({\bf w}^{(0)}_{1},\cdots, wH(0))w^{(0)}_{H}) and notice that h^(W(0))=A^h(W(0)){\hat{h}}(W^{(0)})={\hat{A}}h(W^{(0)}), then we have

i=1mσikh^i(W(0))22\displaystyle\Big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\Big{\|}^{2}_{2} =t=1H(i=1mσikvN(i)A^ivhvt(𝐰t(0)))22\displaystyle=\sum^{H}_{t=1}\Big{(}\sum^{m}_{i=1}\sigma_{ik}\sum_{v\in N(i)}{\hat{A}}_{iv}h^{t}_{v}({\bf w}^{(0)}_{t})\Big{)}^{2}_{2}
=t=1H𝐰t(0)22(i=1mσikvN(i)A^ivhvt(𝐰t0/𝐰t(0)2))2.\displaystyle=\sum^{H}_{t=1}\|{\bf w}^{(0)}_{t}\|^{2}_{2}\Big{(}\sum^{m}_{i=1}\sigma_{ik}\sum_{v\in N(i)}{\hat{A}}_{iv}h^{t}_{v}({\bf w}^{0}_{t}/\|{\bf w}^{(0)}_{t}\|_{2})\Big{)}^{2}.

By the definition of Frobenius norm WF2=t=1H𝐰t22\|W\|^{2}_{F}=\sum^{H}_{t=1}\|{\bf w}_{t}\|^{2}_{2}, the supremum of the above quantity under the constraint WFR\|W\|_{F}\leq R must be obtained when 𝐰t02=B1\|{\bf w}_{t_{0}}\|_{2}=B_{1} for some t0[H]t_{0}\in[H], and 𝐰t2=0\|{\bf w}_{t}\|_{2}=0 for all tt0t\neq t_{0}. Hence

supW(0)FB1i=1mσikh^i(W(0))2\displaystyle\sup_{\|W^{(0)}\|_{F}\leq B_{1}}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\Big{\|}_{2} =sup𝐰2=B1(i=1mσikvN(i)A^ivhv(𝐰)).\displaystyle=\sup_{\|{\bf w}\|_{2}=B_{1}}\Big{(}\sum^{m}_{i=1}\sigma_{ik}\sum_{v\in N(i)}{\hat{A}}_{iv}h_{v}({\bf w})\Big{)}.

Let ns(j)n_{s}(j) be the ss-th neighbor number of node jj (s[q]s\in[q], j[m]j\in[m]). Recall q:=max|N(j)|,j[m]q:=\max|N(j)|,j\in[m], Ms=maxi[m]|A^iv|M_{s}=\max_{i\in[m]}|\hat{A}_{iv}| with vN(i)v\in N(i), therefore

𝔼σ[k=1KsupW(0)FB1i=1mσikh^i(W(0))2]\displaystyle{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|W^{(0)}\|_{F}\leq B_{1}}\big{\|}\sum^{m}_{i=1}\sigma_{ik}{\hat{h}}_{i}(W^{(0)})\big{\|}_{2}\Big{]}
=𝔼σ[k=1Ksup𝐰2=B1(i=1mσikvN(i)A^ivhv(𝐰))]\displaystyle={\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|{\bf w}\|_{2}=B_{1}}\Big{(}\sum^{m}_{i=1}\sigma_{ik}\sum_{v\in N(i)}{\hat{A}}_{iv}h_{v}({\bf w})\Big{)}\Big{]}
𝔼σ[k=1Ksup𝐰2=B1(s=1qMsi=1mσikhns(i)(𝐰))].\displaystyle\leq{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|{\bf w}\|_{2}=B_{1}}\Big{(}\sum^{q}_{s=1}M_{s}\sum^{m}_{i=1}\sigma_{ik}h_{n_{s}(i)}({\bf w})\Big{)}\Big{]}.

Applying the conclusion sup(sas)ssup(as)\sup(\sum_{s}a_{s})\leq\sum_{s}\sup(a_{s}) and contraction property of Rademacher complexity again, we have

𝔼σ[k=1Ksup𝐰2=B1(s=1qMsi=1mσikhns(i)(𝐰))]\displaystyle{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|{\bf w}\|_{2}=B_{1}}\Big{(}\sum^{q}_{s=1}M_{s}\sum^{m}_{i=1}\sigma_{ik}h_{n_{s}(i)}({\bf w})\Big{)}\Big{]}
𝔼σ[k=1Ks=1qMssup𝐰2=B1i=1mσikhns(i)(𝐰)]\displaystyle\leq{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sum^{q}_{s=1}M_{s}\sup_{\|{\bf w}\|_{2}=B_{1}}\sum^{m}_{i=1}\sigma_{ik}h_{n_{s}(i)}({\bf w})\Big{]}
s=1qMs𝔼σ[k=1Ksup𝐰2=B1i=1mσik(jN(ns(i))A^ns(i)j𝐱j,𝐰)]\displaystyle\leq\sum^{q}_{s=1}M_{s}{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|{\bf w}\|_{2}=B_{1}}\sum^{m}_{i=1}\sigma_{ik}\Big{(}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}\langle{\bf x}_{j},{\bf w}\rangle\Big{)}\Big{]}
=s=1qMs𝔼σ[k=1Ksup𝐰2=B1i=1mσikjN(ns(i))A^ns(i)j𝐱j,𝐰]\displaystyle=\sum^{q}_{s=1}M_{s}{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\sup_{\|{\bf w}\|_{2}=B_{1}}\Big{\langle}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j},{\bf w}\Big{\rangle}\Big{]}
B1s=1qMs𝔼σ[k=1Ki=1mσikjN(ns(i))A^ns(i)j𝐱j2].\displaystyle\leq B_{1}\sum^{q}_{s=1}M_{s}{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}_{2}\Big{]}.

Therefore, we only need to estimate the term

𝔼σ[k=1Ki=1mσikjN(ns(i))A^ns(i)j𝐱j2].{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}_{2}\Big{]}.

Applying Cauchy-Schwartz inequality yields that

𝔼σ[k=1Ki=1mσikjN(ns(i))A^ns(i)j𝐱j2]\displaystyle{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}_{2}\Big{]} 𝔼σ(k=1Ki=1mσikjN(ns(i))A^ns(i)j𝐱j2)2\displaystyle\leq\sqrt{{\mathbb{E}}_{\sigma}\Big{(}\sum^{K}_{k=1}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}_{2}\Big{)}^{2}}
𝔼σKk=1Ki=1mσikjN(ns(i))A^ns(i)j𝐱j22\displaystyle\leq\sqrt{{\mathbb{E}}_{\sigma}K\sum^{K}_{k=1}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}^{2}_{2}}
Ki=1mjN(ns(i))A^ns(i)j𝐱j22,\displaystyle\leq K\sqrt{\sum^{m}_{i=1}\Big{\|}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}_{2}^{2}},

where the last inequality is due to the i.i.d condition of Rademacher sequences. Plugging the conclusion of Lemma 5.3 into the above term leads to

𝔼σ[k=1Ki=1mσikjN(ns(i))A^ns(i)j𝐱j2]KRqm,{\mathbb{E}}_{\sigma}\Big{[}\sum^{K}_{k=1}\Big{\|}\sum^{m}_{i=1}\sigma_{ik}\sum_{j\in N(n_{s}(i))}{\hat{A}}_{n_{s}(i)j}{\bf x}_{j}\Big{\|}_{2}\Big{]}\leq KR\sqrt{qm},

and

^(B1,B2)2qKB1B2Rs=1qMsm.{\cal{\hat{R}}}({{\cal F}}_{B_{1},B_{2}})\leq\frac{\sqrt{2q}KB_{1}B_{2}R\sum^{q}_{s=1}M_{s}}{\sqrt{m}}.

This completes the proof by combining with Lemma 5.2.

Acknowledgement

The work described in this paper was supported partially by the National Natural Science Foundation of China (12271111, 11871277), Special Support Plan for High-Level Talents of Guangdong Province (2019TQ05X571), Guangdong Basic and Applied Basic Research Foundation (2022A1515011726). The authors would like to thank Prof. Hong Chen from Huazhong Agricultural University for useful discussions about the theoretical analysis, which have helped to improve the presentation of the paper.

References

  • [1] D. Bahdanau, K. Cho, and Y. Bengio, Neural machine translation by jointly learning to align and translate, in Int. Conf. on Learning Representations (ICLR), San Diego, United States (May 2015).
  • [2] P. L. Bartlett and S. Mendelson, Rademacher and Gaussian complexities: risk bounds and structural results, in Int. Conf. on Computational Learning Theory & and European Conference on Computational Learning Theory, Amsterdam, Netherlands, (July 2001), pp. 224-240.
  • [3] M. M. Bronstein, J. Bruna, Y. LeCun, A. Szlam, and P. Vandergheynst, Geometric deep learning: going beyond euclidean data, IEEE Sigal Proc. Mag. 34(4) (2017) 18-42.
  • [4] J. Bruna, W. Zaremba, A. Szlam, and Y. Lecun, Spectral networks and locally connected networks on graphs, in Int. Conf. on Learning Representations (ICLR), Banff, Canada (April 2014).
  • [5] S. Casas, C. Gulino, R. Liao, and R. Urtasun, Spagnn: spatially-aware graph neural networks for relational behavior forecasting from sensor data, in IEEE Int. Conf. on Robotics and Automation (ICRA), Paris, France (May 2020), pp. 9491-9497.
  • [6] M. Chen, Z. Wei, Z. Huang, B. Ding, and Y. Li, Simple and deep graph convolutional networks, in Int. Conf. on Machine Learning (ICML), Virtual, (July 2020), pp. 1725-1735.
  • [7] E. Chien, J. Peng, P. Li, and O. Milenkovic, Adaptive universal generalized pagerank graph neural network, in Int. Conf. on Learning Representations (ICLR), Virtual (May 2021).
  • [8] M. Defferrard, X. Bresson, and P. Vandergheynst, Convolutional neural networks on graphs with fast localized spectral filtering, in Advances in Neural Information Processing Systems(NeurIPS), Barcelona, Spain (December 2016).
  • [9] S. Dz̆eroski and B. Z̆enko, Is combining classifiers with stacking better than selecting the best one? Mach. Learn. 54 (2004) 255-273.
  • [10] W. Feng, J. Zhang, Y. Dong, Y. Han, H. Luan, Q. Xu, Q. Yang, E. Kharlamov, and J. Tang, Graph random neural networks for semi-supervised learning on graphs, in Advances in Neural Information Processing Systems(NeurIPS), Virtual (December 2020), pp. 22092-22103.
  • [11] Y. Freund and R. E. Schapire, A short introduction to boosting, Journal of Japanese Society for Artificial Intelligence 14 (1999) 771-780.
  • [12] J. Friedman, Greedy function approximation: a gradient boosting machine, Ann. Stat. 29 (2001) 1189-1232.
  • [13] B. Gao and L. Pavel, On the properties of the softmax function with application in game theory and reinforcement learning, Technical Report, arXiv:1704.00805 (2017).
  • [14] M. Gori, G. Monfardini, and F. Scarselli, A new model for learning in graph domains, in Int. Joint Conf. on Neural Networks (IJCNN), Montreal, Canada (July 2005).
  • [15] W. L. Hamilton, Z. Ying, and J. Leskovec, Inductive representation learning on large graphs, in Advances in Neural Information Processing Systems(NeurIPS), Long Beach, United States (December 2017).
  • [16] D. K. Hammond, P. Vandergheynst, and R. Gribonval, Wavelets on graphs via spectral graph theory, Appl. Comput. Harmon. A., 30 (2011) 129-150.
  • [17] S. Ivanov and L. Prokhorenkova, Boost then convolve: gradient boosting meets graph neural networks, in Int. Conf. on Learning Representations (ICLR), Virtual (May 2021).
  • [18] D. Kingma and J. Ba, Adam: a method for stochastic optimization, in Int. Conf. on Learning Representations (ICLR), San Diego, United States (May 2015).
  • [19] T. Kipf and M. Welling, Semi-supervised classification with graph convolutional networks, in Int. Conf. on Learning Representations (ICLR), Toulon, France (April 2017).
  • [20] J. Klicpera, A. Bojchevski, and S. Günnemann, Predict then propagate: graph neural networks meet personalized pagerank, in Int. Conf. on Learning Representations (ICLR), New Orleans, United States (April 2019).
  • [21] J. Klicpera, A. Bojchevski, and S. Günnemann, Predict then propagate: graph neural networks meet personalized pagerank, in Int. Conf. on Learning Representations (ICLR), New Orleans, United States (April 2019).
  • [22] G. Li, M. Mueller, A. Thabet, and B. Ghanem, Deepgcns: can gcns go as deep as cnns? in Int. Conf. on Computer Vision (ICCV), Seoul, Korea (October 2019).
  • [23] M. Li, Z. Ma, Y. G. Wang, and X. Zhuang, Fast haar transforms for graph neural networks, Neural Networks, 128(2020) 188-198.
  • [24] Q. Li, Z. Han, and X. M. Wu, Deeper insights into graph convolutional networks for semi-supervised learning, in AAAI Conf. on Artificial Intelligence, New Orleans, United States (February 2018).
  • [25] S. Luan, M. Zhao, X. W. Chang, and D. Precup, Break the ceiling: stronger multi-scale deep graph convolutional networks, in Advances in Neural Information Processing Systems(NeurIPS), Vancouver, Canada (December 2019).
  • [26] S. Lv, Generalization bounds for graph convolutional neural networks via Rademacher complexity, Technical Report, arXiv:2102.10234 (2021).
  • [27] L. V. D. Maaten and G. E. Hinton, Visualizing data using t-sne, J. Mach. Learn. Res. 9(2008) 2579-2605.
  • [28] A. Maurer, A vector-contraction inequality for Rademacher complexities, in Int. Conf. on Algorithmic Learning Theory, Bari, Italy (October 2016).
  • [29] V. Mnih, N. Heess, A. Graves, and K. Kavukcuoglu, Recurrent models of visual attention, in Advances in Neural Information Processing Systems(NeurIPS), Montreal, Canada (December 2014).
  • [30] K. Pal and B. Patel, Data classification with k-fold cross validation and holdout accuracy estimation methods with 5 different machine learning techniques, in 2020 fourth Int. Conf. on Computing Methodologies and Communication (ICCMC), Erode, India (March 2020), pp. 83-87.
  • [31] Y. Rong, W. Huang, T. Xu, and J. Huang, Dropedge: towards deep graph convolutional networks on node classification, in Int. Conf. on Learning Representations(ICLR), Virtual, (April 2020).
  • [32] A. Sandryhaila and J. Moura, Discrete signal processing on graphs, IEEE Trans. Signal Processing 61(7) (2013) 1644-1656.
  • [33] R. Schapire, Y. Freund, P. Barlett, and W. S. Lee, Boosting the margin: a new explanation for the effectiveness of voting methods, in Int. Conf. on Machine Learning (ICML), San Francisco, United States (July 1997).
  • [34] F. Schwenker, Ensemble methods: foundations and algorithms, IEEE Comput. Intell. M. 8(2013) 77-79.
  • [35] P. Sen, G. Namata, M. Bilgic, L. Getoor, B. Gallagher, and T. Eliassi-Rad, Collective classification in network data, AI Mag., 29(2008) 93-106.
  • [36] D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega, and P. Vandergheynst, The emerging field of signal processing on graphs: extending high-dimensional data analysis to networks and other irregular domains, IEEE Signal Proc. Mag. 30(3) (2013) 83-98.
  • [37] J. M. Stokes, K. Yang, K. Swanson, W. Jin, and J. J. Collins, A deep learning approach to antibiotic discovery, Cell 180(4) (2020) 688-702.e13.
  • [38] J. Sun, W. Guo, D. Zhang, Y. Zhang, F. Regol, Y. Hu, H. Guo, R. Tang, H. Yuan, X. He, and M. Coates, A framework for recommending accurate and diverse items using bayesian graph convolutional neural networks, in Proc. of the 26th ACM SIGKDD Int. Conf. on Knowledge Discovery & Data Mining, New York, United States (July 2020).
  • [39] K. Sun, Z. Lin, and Z. Zhu, Adagcn: adaboosting graph convolutional networks into deep models, in Int. Conf. on Learning Representations (ICLR), Virtual (May 2021).
  • [40] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, u. Kaiser, and I. Polosukhin, Attention is all you need, in Advances in Neural Information Processing Systems(NeurIPS), Long Beach, United States (December 2017), pp. 6000-6010.
  • [41] P. Velickovic, G. Cucurull, A. Casanova, A. Romero, P. Lio’, and Y. Bengio, Graph attention networks, in Int. Conf. on Learning Representations (ICLR), Vancouver, Canada (April 2018).
  • [42] X. Wang, H. Ji, C. Shi, B. Wang, Y. Ye, P. Cui, and P. S. Yu, Heterogeneous graph attention network, in The World Wide Web Conference, San Francisco, United States (May 2019 ), pp. 2022–2032.
  • [43] D. Wolpert, Stacked generalization, Neural Netw. 5(1992) 241-259.
  • [44] F. Wu, T. Zhang, A. Souza, C. Fifty, T. Yu, and K. Q. Weinberger, Simplifying graph convolutional networks, in Int. Conf. on Machine Learning(ICML), Long Beach, United States (June 2019).
  • [45] B. Xu, H. Shen, Q. Cao, Y. Qiu, and X. Cheng, Graph wavelet neural network, in Int. Conf. on Learning Representations (ICLR), New Orleans, United States (May 2019).
  • [46] L. Yao, C. Mao, and Y. Luo, Graph convolutional networks for text classification, in AAAI Conf. on Artificial Intelligence, Honolulu, USA (January 2019).
  • [47] W. Yin, H. Schütze, B. Xiang, and B. Zhou, Abcnn: attention-based convolutional neural network for modeling sentence pairs, Transactions of the Association for Computational Linguistics, 4(2016) 259-272.
  • [48] W. Zhang, Z. Sheng, Y. Jiang, Y. Xia, J. Gao, Z. Yang, and B. Cui, Evaluating deep graph neural networks, Technical Report, arXiv:2108.00955 (2021).
  • [49] J. Zhu, Max-margin nonparametric latent feature models for link prediction, in Int. Conf. on Machine Learning (ICML), Edinburgh, UK (June 2012).