Node Duplication Improves Cold-start Link Prediction
Abstract
Graph Neural Networks (GNNs) are prominent in graph machine learning and have shown state-of-the-art performance in Link Prediction (LP) tasks. Nonetheless, recent studies show that GNNs struggle to produce good results on low-degree nodes despite their overall strong performance. In practical applications of LP, like recommendation systems, improving performance on low-degree nodes is critical, as it amounts to tackling the cold-start problem of improving the experiences of users with few observed interactions. In this paper, we investigate improving GNNs’ LP performance on low-degree nodes while preserving their performance on high-degree nodes and propose a simple yet surprisingly effective augmentation technique called NodeDup. Specifically, NodeDup duplicates low-degree nodes and creates links between nodes and their own duplicates before following the standard supervised LP training scheme. By leveraging a “multi-view” perspective for low-degree nodes, NodeDup shows significant LP performance improvements on low-degree nodes without compromising any performance on high-degree nodes. Additionally, as a plug-and-play augmentation module, NodeDup can be easily applied on existing GNNs with very light computational cost. Extensive experiments show that NodeDup achieves 38.49%, 13.34%, and 6.76% improvements on isolated, low-degree, and warm nodes, respectively, on average across all datasets compared to GNNs and state-of-the-art cold-start methods.
1 Introduction
Link prediction (LP) is a fundamental task of graph-structured data (Liben-Nowell & Kleinberg, 2007; Trouillon et al., 2016), which aims to predict the likelihood of the links existing between two nodes in the network. It has wide-ranging real-world applications across different domains, such as friend recommendations in social media (Sankar et al., 2021; Tang et al., 2022; Fan et al., 2022), product recommendations in e-commerce platforms (Ying et al., 2018; He et al., 2020), knowledge graph completion (Li et al., 2023; Vashishth et al., 2020; Zhang et al., 2020), and chemical interaction prediction (Stanfield et al., 2017; Kovács et al., 2019; Yang et al., 2021; Guo et al., 2022a).

In recent years, graph neural networks (GNNs) (Kipf & Welling, 2016a; Veličković et al., 2017; Hamilton et al., 2017) have been widely applied to LP, and a series of cutting-edge models have been proposed (Zhang & Chen, 2018; Zhang et al., 2021; Zhu et al., 2021; Zhao et al., 2022b). Most GNNs follow a message-passing scheme (Gilmer et al., 2017) in which information is iteratively aggregated from neighbors and used to update node representations accordingly. Consequently, the success of GNNs usually heavily relies on having sufficient high-quality neighbors for each node (Zheng et al., 2021; Liu et al., 2021). However, real-world graphs often exhibit long-tailed distribution in terms of node degrees, where a significant fraction of nodes have very few neighbors (Tang et al., 2020b; Ding et al., 2021; Hao et al., 2021). For example, Figure 1 shows the long-tailed degree distribution of the Citeseer dataset. Moreover, LP performances w.r.t. node degrees on this dataset also clearly indicate that GNNs struggle to generate satisfactory results for nodes with low or zero degrees. For simplicity, in this paper, we refer to the nodes with low or zero degrees as cold nodes and the nodes with higher degrees as warm nodes.
To boost GNNs’ performance on cold nodes, recent studies have proposed various training strategies (Liu et al., 2020, 2021; Zheng et al., 2021; Hu et al., 2022) and augmentation strategies (Hu et al., 2022; Rong et al., 2019; Zhao et al., 2022b) to improve representation learning quality. For instance, ColdBrew (Zheng et al., 2021) posits that training a powerful MLP can rediscover missing neighbor information for cold nodes; TailGNN (Liu et al., 2021) utilizes a cold-node-specific module to accomplish the same objective. However, such advanced training strategies (e.g., ColdBrew and TailGNN) share a notable drawback: they are trained with a bias towards cold nodes, which then sacrifices performance on warm nodes (empirically validated in Table 1). However, in real-world applications, both cold nodes and warm nodes are critical (Clauset et al., 2009). On the other hand, while augmentation methods such as LAGNN (Liu et al., 2022b) do not have such bias, they primarily focus on improving the overall performance of GNNs in LP tasks, which may be dominated by warm nodes due to their higher connectivity. Additionally, the augmentation methods usually introduce a significant amount of extra computational costs (empirically validated in Figure 5). In light of the existing work discussed above on improving LP performance for cold nodes, we are naturally motivated to explore the following crucial but rather unexplored research question:
Can we improve LP performance on cold nodes without compromising warm node performance?
We observe that cold node LP performance usually suffers because they are under-represented in standard supervised LP training due to their few (if any) connections. Given this observation, in this work, we introduce a simple yet effective augmentation method, NodeDup, for improving LP performance on cold nodes. Specifically, NodeDup duplicates cold nodes and establishes edges between each original cold node and its corresponding duplicate. Subsequently, we conduct standard supervised end-to-end training of GNNs on the augmented graph. To better understand why NodeDup is able to improve LP performance for cold nodes, we thoroughly analyze it from multiple perspectives, during which we discover that this simple technique effectively offers a “multi-view” perspective of cold nodes during training. This “multi-view” perspective of the cold nodes acts similarly to an ensemble and drives performance improvements for these nodes. Additionally, our straightforward augmentation method provides valuable supervised training signals for cold nodes and especially isolated nodes. Furthermore, we also introduce NodeDup(L), a lightweight variation of NodeDup that adds only self-loop edges into training edges for cold nodes. NodeDup(L) empirically offers up to a 1.3 speedup over NodeDup for the training process and achieves significant speedup over existing augmentation baselines. In our experiments, we comprehensively evaluate our method on seven benchmark datasets. Compared to GNNs and state-of-the-art cold-start methods, NodeDup achieves 38.49%, 13.34%, and 6.76% improvements on isolated, low-degree, and warm nodes, respectively, on average across all datasets. NodeDup also greatly outperforms augmentation baselines on cold nodes, with comparable warm node performance. Finally, as plug-and-play augmentation methods, our methods are versatile and effective with different LP encoders/decoders. They also achieve significant performance in a more realistic inductive setting.
2 Preliminaries
2.1 Notation
Let an attributed graph be , where is the set of nodes and is the edges where each indicates nodes and are linked. Let be the node attribute matrix, where is the attribute dimension. Let be the set of neighbors of node , i.e., , and the degree of node is . We separate the set of nodes into three disjoint sets , , and by their degrees based on the threshold hyperparameter 111This threshold is set as 2 in our experiments, based on observed performance gaps in LP on various datasets, as shown in Figure 1 and Figure 6. Further reasons for this threshold are detailed in Section D.1.. For each node , if ; if ; if . For ease of notation, we also use to denote the cold nodes, which is the union of Isolated and Low-degree nodes.
2.2 LP with GNNs
In this work, we follow the commonly-used encoder-decoder framework for GNN-based LP (Kipf & Welling, 2016b; Berg et al., 2017; Schlichtkrull et al., 2018; Ying et al., 2018; Davidson et al., 2018; Zhu et al., 2021; Yun et al., 2021; Zhao et al., 2022b), where a GNN encoder learns the node representations and the decoder predicts the link existence probabilities given each pair of node representations. Most GNNs follow the message passing design (Gilmer et al., 2017) that iteratively aggregate each node’s neighbors’ information to update its embeddings. Without the loss of generality, for each node , the -th layer of a GNN can be defined as
(1) |
where is the -th layer’s output representation of node , , is the (typically permutation-invariant) aggregation function, and is the update function that combines node ’s neighbor embedding and its own embedding from the previous layer. For any node pair and , the decoding process can be defined as where is the GNN’s output representation for node and is the Sigmoid function. Following existing literature, we use inner product (Wang et al., 2021; Zheng et al., 2021) as the default DECODER.
The standard supervised LP training optimizes model parameters w.r.t. a training set, which is usually the union of all observed edges and no-edge node pairs (as training with all no-edges is infeasible in practice), where is the negative sampling rate ( usually). We use to denote the training set labels, where if and 0 otherwise.
2.3 The Cold-Start Problem
The cold-start problem is prevalent in various domains and scenarios. In recommendation systems (Chen et al., 2020; Lu et al., 2020; Hao et al., 2021; Zhu et al., 2019; Volkovs et al., 2017; Liu & Zheng, 2020), cold-start refers to the lack of sufficient interaction history for new users or items, which makes it challenging to provide accurate recommendations. Similarly, in the context of GNNs, the cold-start problem refers to performance in tasks involving cold nodes, which have few or no neighbors in the graph. As illustrated in Figure 1, GNNs usually struggle with cold nodes in LP tasks due to unreliable or missing neighbors’ information. In this work, we focus on enhancing LP performance for cold nodes, specifically predicting the presence of links between a cold node and target node (w.l.o.g.). Additionally, we aim to maintain satisfactory LP performance for warm nodes. Prior studies on cold-start problems (Tang et al., 2020b; Liu et al., 2021; Zheng et al., 2021) inspired this research direction.
3 Node Duplication to Improve Cold-start Performance
We make a simple but powerful observation: cold nodes are strongly under-represented in the LP training. Given that they have few or even no directly connected neighbors, they hardly participate in the standard supervised LP training as described in Section 2. For example, a model will not see an isolated node unless it is randomly sampled as a negative training edge for another node. In light of such observations, our proposed augmentation technique is simple: we duplicate under-represented cold nodes. By both training and aggregating with the edges connecting the cold nodes with their duplications, cold nodes are able to gain better visibility in the training process, which allows the GNN-based LP models to learn better representations. In this section, we introduce NodeDup in detail, followed by comprehensive analyses of why it works from different perspectives.
3.1 Proposed Method
The implementation of NodeDup can be summarized into four simple steps: (1): duplicate all cold nodes to generate the augmented node set , whose node feature matrix is then . (2): for each cold node and its duplication , add an edge between them and get the augmented edge set . (3): include the augmented edges into the training set and get . (4): proceed with the standard supervised LP training on the augmented graph with augmented training set . We also summarize this whole process of NodeDup in Algorithm 1 in Appendix C.
Time Complexity. We discuss complexity of our method in terms of the training process on the augmented graph. We use GSage (Hamilton et al., 2017) and inner product decoder as the default architecture when demonstrating the following complexity (w.l.o.g). With the augmented graph, GSage has a complexity of , where represents the number of sampled neighbors for each node, is the number of GSage layers (Wu et al., 2020), and denotes the size of node representations. In comparison to the non-augmented graph, NodeDup introduces an extra time complexity of . For the inner product decoder, we incorporate additionally positive edges and also sample negative edges into the training process, resulting in the extra time complexity of the decoder as . Given that all cold nodes have few ( in our experiments) neighbors, and GSage is also always shallow (so is small) (Zhao & Akoglu, 2019), the overall extra complexity introduced by NodeDup is .


3.2 How does node duplication help cold-start LP?
In this subsection, we analyze how such a simple method can improve cold-start LP from two perspectives: the neighborhood aggregation in GNNs and the supervision signal during training. In short, NodeDup leverages the extra information from an additional “view”. The existing view is when a node is regarded as the anchor node during message passing, whereas the additional view is when that node is regarded as one of its neighbors thanks to the duplicated node from NodeDup.
Aggregation. As described in Section 2.2, when UPDATE and AGG do not share the transformation for node features, GNNs would have separate weights for self-representation and neighbor representations. The separate weights enable the neighbors and the node itself to play distinct roles in the UPDATE step. By leveraging this property, with NodeDup, the model can leverage the two “views” for each node: first, the existing view is when a node is regarded as the anchor node during message passing, and the additional view is when that node is regarded as one of its neighbors thanks to the duplicated node from NodeDup. Taking the official PyG (Fey & Lenssen, 2019) implementation of GSage (Hamilton et al., 2017) as an example, it updates node representations using . Here, and correspond to the self-representation and neighbors’ representations, respectively. Without NodeDup, isolated nodes have no neighbors, which results with . Thus, the representations of all are only updated by . With NodeDup, the updating process for isolated node becomes . It indicates that is also incorporated into the node updating process for isolated nodes, which offers an additional perspective for isolated nodes’ representation learning. Similarly, GAT (Veličković et al., 2017) updates node representations with , where . Attention scores in partially correspond to the self-representation and partially to neighbors’ representation . In this case, neighbor information also offers a different perspective compared to self-representation. Such “multi-view” enriches the representations learned for the isolated nodes in a similar way to how ensemble methods work (Allen-Zhu & Li, 2020). Moreover, apart from addressing isolated nodes, the same mechanism and multi-view perspective also apply to Low-degree nodes.
Supervision. For LP tasks, besides the aggregation, edges also serve as supervised training signals. Cold nodes have few or no positive training edges connecting to them, potentially leading to out-of-distribution (OOD) issues (Wu et al., 2022), especially for isolated nodes. The additional edges, added by NodeDup to connect cold nodes with their duplicates, serve as additional positive supervision signals for LP. More supervision signals for cold nodes usually lead to better-quality embeddings.
Ablation Study. Figure 2 shows an ablation study on these two designs where NodeDup w/o Step (3) indicates only using the augmented nodes and edges in aggregation but not supervision; NodeDup w/o Step (2) indicates only using the augmented edges in supervision but not aggregation. We can observe that using augmented nodes/edges either in supervision or aggregation can significantly improve the LP performance on Isolated nodes, and NodeDup, by combining them, results in larger improvements. Besides, NodeDup also achieves improvements on Low-degree nodes while not sacrificing Warm nodes’ performance.
3.3 Relation between NodeDup and Self-distillation
Recently, Allen-Zhu & Li (2020) showed that the success of self-distillation, similar to our method, contributes to ensemble learning by providing models with different perspectives on the knowledge. Building on this insight, we show an interesting interpretation of NodeDup, as a simplified and enhanced version of self-distillation for LP tasks for cold nodes, illustrated in Figure 3, in which we draw a connection between self-distillation and NodeDup.

In self-distillation, a teacher GNN is first trained to learn the node representations from original features . We denote the original graph as , and we denote the graph, where we replace the node features in with , as in Figure 3. The student GNN is then initialized with random parameters and trained with the sum of two loss functions: , where denotes the supervised training loss with and denotes the knowledge distillation loss with . Figure 4 shows that self-distillation outperforms the teacher GNN across all settings.
The effect of is similar to that of creating an additional link connecting nodes in to their corresponding nodes in when optimizing with . This is illustrated by the red dashed line in Figure 3. For better clarity, we show the similarities between these two when we use the inner product as the decoder for LP with the following example. Given a node with normalized teacher embedding and normalized student embedding , the additional loss term that would be added for distillation with cosine similarity is . On the other hand, for the dashed line edges in Figure 3, we add an edge between the node and its corresponding node in with embedding . When trained with an inner product decoder and binary cross-entropy loss, it results in the following: . Since we always add the edge , we know , and can simplify the loss as follows: . Here, we can observe that and are positively correlated as is a monotonically increasing function.
To further improve this step and mitigate potential noise in , we explore a whole graph duplication technique, where is replaced with an exact duplicate of to train the student GNN. The results in Figure 4 demonstrate significant performance enhancement achieved by whole graph duplication compared to self-distillation. NodeDup is a lightweight variation of the whole graph duplication technique, which focuses on duplicating only the cold nodes and adding edges connecting them to their duplicates. From the results, it is evident that NodeDup consistently outperforms the teacher GNN and self-distillation in all scenarios. Additionally, NodeDup exhibits superior performance on isolated nodes and is much more efficient compared to the whole graph duplication approach.
3.4 NodeDup(L): An Efficient Variant of NodeDup
Inspired by the above analysis, we further introduce a lightweight variant of NodeDup for better efficiency, NodeDup(L). To provide above-described “multi-view” information as well as the supervision signals for cold nodes, NodeDup(L) simply add additional self-loop edges for the cold nodes into the edge set , that is, . NodeDup(L) preserves the two essential designs of NodeDup while avoiding the addition of extra nodes, which further saves time and space complexity. Moreover, NodeDup differs from NodeDup(L) since each duplicated node in NodeDup will provide another view for itself because of dropout layers, which leads to different performance as shown in Section 4.2.
NodeDup(L) vs. Self-loop. We remark upon a resemblance between NodeDup(L) and self-loops in GNNs (e.g., the additional self-connection in the normalized adjacency matrix by GCN) as they both add self-loop edges. However, they differ in two aspects. During aggregation: NodeDup(L) intentionally incorporates the self-representation into the aggregated neighbors’ representation by adding additional edges. Taking GSage as an example, the weight matrix would serve an extra “view” of when updating , whereas the default self-loops only use information from . Additionally, in the supervision signal: different from the normal self-loops and the self-loops introduced in the previous works (Cai et al., 2019; Wang et al., 2020), where self-loops are solely for aggregation, the edges added by NodeDup(L) also serve as additional positive training samples for the cold nodes.
4 Experiments
GSage | Imbalance | TailGNN | Cold-brew | NodeDup(L) | NodeDup | ||
---|---|---|---|---|---|---|---|
Cora | Isolated | 32.203.58 | 34.511.11 | 36.951.34 | 28.170.67 | 39.761.32 | 44.273.82 |
Low-degree | 59.451.09 | 59.421.21 | 61.350.79 | 57.270.63 | 62.531.03 | 61.981.14 | |
Warm | 61.140.78 | 59.540.46 | 60.610.90 | 56.280.81 | 62.070.37 | 59.070.68 | |
Overall | 58.310.68 | 57.550.67 | 59.020.71 | 54.440.53 | 60.490.49 | 58.920.82 | |
Citeseer | Isolated | 47.132.43 | 46.260.86 | 37.843.36 | 37.784.23 | 52.461.16 | 57.541.04 |
Low-degree | 61.880.79 | 61.900.60 | 62.061.73 | 59.129.97 | 73.711.22 | 75.500.39 | |
Warm | 71.450.52 | 71.540.86 | 71.321.83 | 65.127.82 | 74.990.37 | 74.680.67 | |
Overall | 63.770.83 | 63.660.43 | 62.021.89 | 58.037.72 | 70.340.35 | 71.730.47 | |
CS | Isolated | 56.411.61 | 46.601.66 | 55.701.38 | 57.700.81 | 65.181.25 | 65.871.70 |
Low-degree | 75.950.25 | 75.530.21 | 73.600.70 | 73.990.34 | 81.460.57 | 81.120.36 | |
Warm | 84.370.46 | 83.700.46 | 79.860.35 | 78.230.28 | 85.480.26 | 84.760.41 | |
Overall | 83.330.42 | 82.560.40 | 79.050.36 | 77.630.23 | 84.900.29 | 84.230.39 | |
Physics | Isolated | 47.411.38 | 55.010.58 | 52.541.34 | 64.380.85 | 65.040.63 | 66.650.95 |
Low-degree | 79.310.28 | 79.500.27 | 75.950.27 | 75.860.10 | 82.700.22 | 84.040.22 | |
Warm | 90.280.23 | 89.850.09 | 85.930.40 | 78.480.14 | 90.440.23 | 90.330.05 | |
Overall | 89.760.22 | 89.380.09 | 85.480.38 | 78.340.13 | 90.090.22 | 90.030.05 | |
Computers | Isolated | 9.321.44 | 10.140.59 | 10.631.59 | 9.751.24 | 17.111.62 | 19.622.63 |
Low-degree | 57.910.97 | 56.190.82 | 51.211.58 | 49.030.94 | 62.141.06 | 61.160.92 | |
Warm | 66.870.47 | 65.620.21 | 62.770.44 | 57.520.28 | 68.020.41 | 68.100.25 | |
Overall | 66.670.47 | 65.420.20 | 62.550.45 | 57.350.28 | 67.860.41 | 67.940.25 | |
Photos | Isolated | 9.252.31 | 10.801.72 | 13.621.00 | 12.862.58 | 21.502.14 | 17.843.53 |
Low-degree | 52.610.88 | 50.680.57 | 42.752.50 | 43.140.64 | 55.701.38 | 54.131.58 | |
Warm | 67.640.55 | 64.540.50 | 61.630.73 | 58.060.56 | 69.680.87 | 68.680.49 | |
Overall | 67.320.54 | 64.240.49 | 61.290.75 | 57.770.56 | 69.400.86 | 68.390.48 | |
IGB-100K | Isolated | 75.920.52 | 77.320.79 | 77.290.34 | 82.310.30 | 87.430.44 | 88.040.20 |
Low-degree | 79.380.23 | 79.190.09 | 80.570.14 | 83.840.16 | 88.370.24 | 88.980.17 | |
Warm | 86.420.24 | 86.010.19 | 85.350.19 | 82.440.21 | 88.540.31 | 88.280.20 | |
Overall | 84.770.21 | 84.470.14 | 84.190.18 | 82.680.17 | 88.470.28 | 88.390.18 |
4.1 Experimental Settings
Datasets and Evaluation Settings. We conduct experiments on 7 benchmark datasets: Cora, Citeseer, CS, Physics, Computers, Photos and IGB-100K, with their details specified in Appendix B. We randomly split edges into training, validation, and testing sets. We allocated 10% for validation and 40% for testing in Computers and Photos, 5%/10% for testing in IGB-100K, and 10%/20% in other datasets. We follow the standard evaluation metrics used in the Open Graph Benchmark (Hu et al., 2020) for LP, in which we rank missing references higher than 500 negative reference candidates for each node. The negative references are randomly sampled from nodes not connected to the source node. We use Hits@10 as the main evaluation metric (Han et al., 2022) and also report MRR performance in Appendix D. We follow Guo et al. (2022b) and Shiao et al. (2022) for the inductive settings, where new nodes appear after the training process.
Baselines. Both NodeDup and NodeDup(L) are flexible to integrate with different GNN encoder architectures and LP decoders. For our experiments, we use GSage (Hamilton et al., 2017) encoder and the inner product decoder as the default base LP model. To comprehensively evaluate our work, we compare NodeDup against three categories of baselines. (1) Base LP models. (2) Cold-start methods: TailGNN (Liu et al., 2021) and Cold-brew (Zheng et al., 2021) primarily aim to enhance the performance on cold nodes. We also compared with Imbalance (Lin et al., 2017), viewing cold nodes as an issue of the imbalance concerning node degrees. (3) Graph data augmentation methods: Augmentation frameworks including DropEdge (Rong et al., 2019), TuneUP (Hu et al., 2022), and LAGNN (Liu et al., 2022b) typically improve the performance while introducing additional preprocessing or training time.

4.2 Performance Compared to Base GNN LP Models
Isolated and Low-degree Nodes. We compare our methods with base GNN LP models that consist of a GNN encoder in conjunction with an inner product decoder and are trained with a supervised loss. From Table 1, we observe consistent improvements for both NodeDup(L) and NodeDup over the base GSage model across all datasets, particularly in the Isolated and Low-degree node settings. Notably, in the Isolated setting, NodeDup achieves an impressive 29.6% improvement, on average, across all datasets. These findings provide clear evidence that our methods effectively address the issue of sub-optimal LP performance on cold nodes.
Warm Nodes and Overall. It is encouraging to see that NodeDup(L) consistently outperforms GSage across all the datasets in the Warm nodes and Overall settings. NodeDup also outperforms GSage in 13 out of 14 cases under both settings. These findings support the notion that our methods can effectively maintain and enhance the performance of Warm nodes.
NodeDup vs. NodeDup(L). Furthermore, we observe that NodeDup achieves greater improvements over NodeDup(L) for Isolated nodes. However, NodeDup(L) outperforms NodeDup on 6 out of 7 datasets for Warm nodes. The additional improvements achieved by NodeDup for Isolated nodes can be attributed to the extra view provided to cold nodes through node duplication during aggregation. On the other hand, the impact of node duplication on the original graph structure likely affects the performance of Warm nodes, which explains the superior performance of NodeDup(L) in this setting compared to NodeDup.
4.3 Performance Compare to Cold-start Methods
Table 1 presents the LP performance of various cold-start baselines. For both Isolated and Low-degree nodes, we consistently observe substantial improvements of our NodeDup and NodeDup(L) methods compared to other cold-start baselines. Specifically, NodeDup and NodeDup(L) achieve 38.49% and 34.74% improvement for Isolated nodes on average across all datasets, respectively.
In addition, our methods consistently outperform cold-start baselines for Warm nodes across all the datasets, where NodeDup(L) and NodeDup achieve 6.76% and 7.95% improvements on average, respectively. This shows that our methods can successfully overcome issues with degrading performance on Warm nodes in cold-start baselines.
GSage | NodeDup(L) | NodeDup | ||
---|---|---|---|---|
Citeseer | Isolated | 58.420.49 | 62.421.88 | 62.941.91 |
Low-degree | 67.751.06 | 69.931.18 | 72.051.23 | |
Warm | 72.981.15 | 75.041.03 | 74.402.43 | |
Overall | 66.980.61 | 69.650.83 | 70.261.16 | |
Physics | Isolated | 85.620.23 | 85.940.15 | 86.900.35 |
Low-degree | 80.870.43 | 81.230.56 | 85.560.25 | |
Warm | 90.220.36 | 90.370.25 | 90.540.14 | |
Overall | 89.400.33 | 89.570.23 | 89.980.13 | |
IGB-100K | Isolated | 84.330.87 | 92.940.11 | 93.950.06 |
Low-degree | 93.190.06 | 93.330.11 | 94.000.09 | |
Warm | 90.760.13 | 91.210.07 | 91.200.08 | |
Overall | 90.310.18 | 91.920.05 | 92.210.04 |
GAT | NodeDup(L) | NodeDup | JKNet | NodeDup(L) | NodeDup | ||
---|---|---|---|---|---|---|---|
Citeseer | Isolated | 37.782.36 | 38.952.75 | 44.041.03 | 37.780.63 | 49.060.60 | 55.150.87 |
Low-degree | 58.042.40 | 61.931.66 | 66.730.96 | 60.741.18 | 71.780.64 | 75.261.16 | |
Warm | 56.372.15 | 64.551.74 | 66.611.67 | 71.610.76 | 74.660.47 | 75.810.89 | |
Overall | 53.421.59 | 58.890.89 | 62.410.78 | 61.730.57 | 68.910.38 | 71.750.82 | |
Physics | Isolated | 38.191.23 | 39.951.48 | 45.892.82 | 42.571.93 | 55.472.25 | 61.112.27 |
Low-degree | 74.190.31 | 74.770.46 | 76.360.25 | 75.360.23 | 79.550.21 | 81.140.28 | |
Warm | 85.840.32 | 86.020.45 | 85.840.15 | 88.240.32 | 89.420.16 | 89.240.16 | |
Overall | 85.270.30 | 85.470.45 | 85.370.14 | 87.640.31 | 88.960.15 | 88.870.15 | |
IGB-100K | Isolated | 75.870.48 | 78.170.58 | 80.180.31 | 69.290.73 | 86.600.46 | 86.850.41 |
Low-degree | 77.050.15 | 78.500.31 | 81.000.12 | 76.900.27 | 86.940.15 | 87.650.20 | |
Warm | 81.400.07 | 81.950.25 | 81.190.20 | 84.930.30 | 87.410.13 | 86.190.12 | |
Overall | 80.420.07 | 81.190.25 | 81.110.19 | 82.910.28 | 87.290.13 | 86.470.13 |
4.4 Performance Compared to Augmentation Methods
Effectiveness Comparison. Since NodeDup and NodeDup(L) use graph data augmentation techniques, we compare them to other data augmentation baselines. The performance and time consumption results are presented in Figure 5 for three datasets (Citeseer, Physics, and IGB-100K), while the results for the remaining datasets are provided in Appendix D due to the page limit. From Figure 5, we consistently observe that NodeDup outperforms all the graph augmentation baselines for Isolated and Low-degree nodes across all three datasets. Similarly, NodeDup(L) outperforms graph data augmentation baselines on 17/18 cases for Isolated and Low-degree nodes. Not only did our methods perform better for Isolated and Low-degree nodes, NodeDup and NodeDup(L) also perform on par or above baselines for Warm nodes.
Efficiency Comparison. Augmentation methods often come with the trade-off of adding additional run time before or during model training. For example, LAGNN (Liu et al., 2022b) requires extra preprocessing time to train the generative model prior to GNN training. It also takes additional time to generate extra features for each node during training. Although Dropedge (Rong et al., 2019) and TuneUP (Hu et al., 2022) are free of preprocessing, they require additional time to drop edges in each training epoch compared to base GNN training. Furthermore, the two-stage training employed by TuneUP doubles the training time compared to one-stage training methods. For NodeDup methods, duplicating nodes and adding edges is remarkably swift and consumes significantly less preprocessing time than other augmentation methods. As an example, NodeDup(L) and NodeDup are 977.0 and 488.5 faster than LAGNN in preprocessing Citeseer, respectively. We also observe that NodeDup(L) has the least training time among all augmentation methods and datasets, while NodeDup also requires less training time in 8/9 cases. Additionally, NodeDup(L) achieves significant efficiency benefits compared to NodeDup in Figure 5, especially when the number of nodes in the graph increases substantially. Taking the IGB-100K dataset as an example, NodeDup(L) is 1.3 faster than NodeDup for the entire training process.
4.5 Performance under the Inductive Setting
Under the inductive setting (Guo et al., 2022b; Shiao et al., 2022), which closely resembles real-world LP scenarios, the presence of new nodes after the training stage adds an additional challenge compared to the transductive setting. We evaluate and present the effectiveness of our methods under this setting in Table 2 for Citeseer, Physics, and IGB-100K datasets. Additional results for other datasets can be found in Appendix D. In Table 2, we observe that our methods consistently outperform base GSage across all of the datasets. We also observe significant improvements in the performance of our methods on Isolated nodes, where NodeDup and NodeDup(L) achieve 5.50% and 3.57% improvements averaged across the three datasets, respectively. Additionally, NodeDup achieves 5.09% improvements on Low-degree nodes. NodeDup leads to more pronounced improvements on Low-degree/Isolated nodes, making it particularly beneficial for the inductive setting.
4.6 Performance with Different Encoders/Decoders
As a simple plug-and-play augmentation method, NodeDup can work with different GNN encoders and LP decoders. In Tables 3 and 4, we present results with GAT (Veličković et al., 2017) and JKNet (Xu et al., 2018) as encoders, along with a MLP decoder. Due to the space limit, we only report the results of three datasets here and leave the remaining in Appendix D. When applying NodeDup to base LP training, with GAT or JKNet as the encoder and inner product as the decoder, we observe significant performance improvements across the board. Regardless of the encoder choice, NodeDup consistently outperforms the base models, particularly for Isolated and Low-degree nodes. Meanwhile, we note that the improvement achieved with our methods is more pronounced when using JKNet as the encoder compared to GAT. This can be attributed to the more informative multi-view provided by JKNet during the aggregation step, as discussed in Section 3.2.
MLP-Dec. | NodeDup(L) | NodeDup | ||
---|---|---|---|---|
Citeseer | Isolated | 17.161.14 | 37.843.06 | 51.172.19 |
Low-degree | 63.821.58 | 68.491.19 | 71.981.29 | |
Warm | 72.931.25 | 75.330.54 | 75.720.55 | |
Overall | 59.491.21 | 66.070.74 | 69.890.65 | |
Physics | Isolated | 11.591.88 | 60.252.54 | 59.501.87 |
Low-degree | 76.370.64 | 81.740.77 | 82.580.79 | |
Warm | 91.540.33 | 91.960.36 | 91.590.22 | |
Overall | 90.780.33 | 91.510.38 | 91.130.23 | |
IGB-100K | Isolated | 3.510.32 | 82.711.05 | 82.020.73 |
Low-degree | 75.250.49 | 85.960.42 | 86.040.26 | |
Warm | 85.060.08 | 87.890.13 | 86.870.48 | |
Overall | 80.160.16 | 87.350.21 | 86.540.40 |
In Table 4, we present the results of our methods applied to the base LP training, where GSage serves as the encoder and MLP as the decoder. Regardless of the decoder, we observe better performance with our methods. These improvements are significantly higher compared to the improvements observed with the inner product decoder. The primary reason for this discrepancy is the inclusion of additional supervised training signals for isolated nodes in our methods, as discussed in Section 3.2. These signals play a crucial role in training the MLP decoder, making it more responsive to the specific challenges presented by isolated nodes.
5 Conclusion
GNNs in LP encounter difficulties when dealing with cold nodes that lack sufficient or absent neighbors. To address this challenge, we presented a simple yet effective augmentation method (NodeDup) specifically tailored for the cold-start LP problem, which can effectively enhance the prediction capabilities of GNNs for cold nodes while maintaining overall performance. Extensive evaluations demonstrated that both NodeDup and its lightweight variant, NodeDup(L), consistently outperformed baselines on both cold node and warm node settings across 7 benchmark datasets. NodeDup also achieved better runtime efficiency compared to the augmentation baselines.
Impact Statement. In this work, our simple but effective method enhances the link prediction performance on cold-start nodes, which mitigates the degree bias and advances the fairness of graph machine learning. It can be widely used and beneficial for various real-world applications, such as recommendation systems, social network analysis, and bioinformatics. We do not foresee any negative societal impact or ethical concerns posed by our method. Nonetheless, we note that both positive and negative societal impacts can be made by applications of graph machine learning techniques, which may benefit from the improvements induced by our work. Care must be taken, in general, to ensure positive societal and ethical consequences of machine learning.
References
- Allen-Zhu & Li (2020) Allen-Zhu, Z. and Li, Y. Towards understanding ensemble, knowledge distillation and self-distillation in deep learning. arXiv preprint arXiv:2012.09816, 2020.
- Berg et al. (2017) Berg, R. v. d., Kipf, T. N., and Welling, M. Graph convolutional matrix completion. arXiv preprint arXiv:1706.02263, 2017.
- Cai & Ji (2020) Cai, L. and Ji, S. A multi-scale approach for graph link prediction. In Proceedings of the AAAI conference on artificial intelligence, 2020.
- Cai et al. (2019) Cai, L., Yan, B., Mai, G., Janowicz, K., and Zhu, R. Transgcn: Coupling transformation assumptions with graph convolutional networks for link prediction. In Proceedings of the 10th international conference on knowledge capture, pp. 131–138, 2019.
- Cai et al. (2021) Cai, L., Li, J., Wang, J., and Ji, S. Line graph neural networks for link prediction. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.
- Chamberlain et al. (2022) Chamberlain, B. P., Shirobokov, S., Rossi, E., Frasca, F., Markovich, T., Hammerla, N., Bronstein, M. M., and Hansmire, M. Graph neural networks for link prediction with subgraph sketching. arXiv preprint arXiv:2209.15486, 2022.
- Chen et al. (2020) Chen, Z., Xiao, R., Li, C., Ye, G., Sun, H., and Deng, H. Esam: Discriminative domain adaptation with non-displayed items to improve long-tail performance. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 579–588, 2020.
- Chien et al. (2021) Chien, E., Chang, W.-C., Hsieh, C.-J., Yu, H.-F., Zhang, J., Milenkovic, O., and Dhillon, I. S. Node feature extraction by self-supervised multi-scale neighborhood prediction. arXiv preprint arXiv:2111.00064, 2021.
- Clauset et al. (2009) Clauset, A., Shalizi, C. R., and Newman, M. E. Power-law distributions in empirical data. SIAM review, 2009.
- Davidson et al. (2018) Davidson, T. R., Falorsi, L., De Cao, N., Kipf, T., and Tomczak, J. M. Hyperspherical variational auto-encoders. arXiv preprint arXiv:1804.00891, 2018.
- Ding et al. (2021) Ding, H., Ma, Y., Deoras, A., Wang, Y., and Wang, H. Zero-shot recommender systems. arXiv preprint arXiv:2105.08318, 2021.
- Ding et al. (2022) Ding, K., Xu, Z., Tong, H., and Liu, H. Data augmentation for deep graph learning: A survey. ACM SIGKDD Explorations Newsletter, 2022.
- Dong et al. (2022) Dong, K., Tian, Y., Guo, Z., Yang, Y., and Chawla, N. Fakeedge: Alleviate dataset shift in link prediction. In Learning on Graphs Conference, pp. 56–1. PMLR, 2022.
- Fan et al. (2022) Fan, W., Liu, X., Jin, W., Zhao, X., Tang, J., and Li, Q. Graph trend filtering networks for recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, 2022.
- Feng et al. (2020) Feng, W., Zhang, J., Dong, Y., Han, Y., Luan, H., Xu, Q., Yang, Q., Kharlamov, E., and Tang, J. Graph random neural networks for semi-supervised learning on graphs. Advances in neural information processing systems, 33:22092–22103, 2020.
- Fey & Lenssen (2019) Fey, M. and Lenssen, J. E. Fast graph representation learning with pytorch geometric. arXiv preprint arXiv:1903.02428, 2019.
- Gilmer et al. (2017) Gilmer, J., Schoenholz, S. S., Riley, P. F., Vinyals, O., and Dahl, G. E. Neural message passing for quantum chemistry. In International conference on machine learning. PMLR, 2017.
- Guo et al. (2022a) Guo, Z., Guo, K., Nan, B., Tian, Y., Iyer, R. G., Ma, Y., Wiest, O., Zhang, X., Wang, W., Zhang, C., et al. Graph-based molecular representation learning. arXiv preprint arXiv:2207.04869, 2022a.
- Guo et al. (2022b) Guo, Z., Shiao, W., Zhang, S., Liu, Y., Chawla, N., Shah, N., and Zhao, T. Linkless link prediction via relational distillation. arXiv preprint arXiv:2210.05801, 2022b.
- Hamilton et al. (2017) Hamilton, W., Ying, Z., and Leskovec, J. Inductive representation learning on large graphs. Advances in neural information processing systems, 2017.
- Han et al. (2022) Han, X., Zhao, T., Liu, Y., Hu, X., and Shah, N. Mlpinit: Embarrassingly simple gnn training acceleration with mlp initialization. arXiv preprint arXiv:2210.00102, 2022.
- Hao et al. (2021) Hao, B., Zhang, J., Yin, H., Li, C., and Chen, H. Pre-training graph neural networks for cold-start users and items representation. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, 2021.
- He et al. (2020) He, X., Deng, K., Wang, X., Li, Y., Zhang, Y., and Wang, M. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, 2020.
- Hu et al. (2020) Hu, W., Fey, M., Zitnik, M., Dong, Y., Ren, H., Liu, B., Catasta, M., and Leskovec, J. Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 2020.
- Hu et al. (2022) Hu, W., Cao, K., Huang, K., Huang, E. W., Subbian, K., and Leskovec, J. Tuneup: A training strategy for improving generalization of graph neural networks. arXiv preprint arXiv:2210.14843, 2022.
- Kang et al. (2019) Kang, B., Xie, S., Rohrbach, M., Yan, Z., Gordo, A., Feng, J., and Kalantidis, Y. Decoupling representation and classifier for long-tailed recognition. arXiv preprint arXiv:1910.09217, 2019.
- Khatua et al. (2023) Khatua, A., Mailthody, V. S., Taleka, B., Ma, T., Song, X., and Hwu, W.-m. Igb: Addressing the gaps in labeling, features, heterogeneity, and size of public graph datasets for deep learning research, 2023. URL https://arxiv.org/abs/2302.13522.
- Kipf & Welling (2016a) Kipf, T. N. and Welling, M. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016a.
- Kipf & Welling (2016b) Kipf, T. N. and Welling, M. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308, 2016b.
- Kovács et al. (2019) Kovács, I. A., Luck, K., Spirohn, K., Wang, Y., Pollis, C., Schlabach, S., Bian, W., Kim, D.-K., Kishore, N., Hao, T., et al. Network-based prediction of protein interactions. Nature communications, 2019.
- Li et al. (2023) Li, J., Shomer, H., Ding, J., Wang, Y., Ma, Y., Shah, N., Tang, J., and Yin, D. Are message passing neural networks really helpful for knowledge graph completion? ACL, 2023.
- Liben-Nowell & Kleinberg (2007) Liben-Nowell, D. and Kleinberg, J. The link-prediction problem for social networks. Journal of the American Society for Information Science and Technology, 2007.
- Lin et al. (2017) Lin, T.-Y., Goyal, P., Girshick, R., He, K., and Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, pp. 2980–2988, 2017.
- Liu et al. (2022a) Liu, G., Zhao, T., Xu, J., Luo, T., and Jiang, M. Graph rationalization with environment-based augmentations. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 1069–1078, 2022a.
- Liu & Zheng (2020) Liu, S. and Zheng, Y. Long-tail session-based recommendation. In Proceedings of the 14th ACM Conference on Recommender Systems, pp. 509–514, 2020.
- Liu et al. (2022b) Liu, S., Ying, R., Dong, H., Li, L., Xu, T., Rong, Y., Zhao, P., Huang, J., and Wu, D. Local augmentation for graph neural networks. In International Conference on Machine Learning, pp. 14054–14072. PMLR, 2022b.
- Liu et al. (2020) Liu, Z., Zhang, W., Fang, Y., Zhang, X., and Hoi, S. C. Towards locality-aware meta-learning of tail node embeddings on networks. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pp. 975–984, 2020.
- Liu et al. (2021) Liu, Z., Nguyen, T.-K., and Fang, Y. Tail-gnn: Tail-node graph neural networks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, 2021.
- Liu et al. (2023) Liu, Z., Nguyen, T.-K., and Fang, Y. On generalized degree fairness in graph neural networks. arXiv preprint arXiv:2302.03881, 2023.
- Lu et al. (2020) Lu, Y., Fang, Y., and Shi, C. Meta-learning on heterogeneous information networks for cold-start recommendation. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2020.
- Luo et al. (2022) Luo, Y., McThrow, M., Au, W. Y., Komikado, T., Uchino, K., Maruhash, K., and Ji, S. Automated data augmentations for graph classification. arXiv preprint arXiv:2202.13248, 2022.
- Park et al. (2021) Park, H., Lee, S., Kim, S., Park, J., Jeong, J., Kim, K.-M., Ha, J.-W., and Kim, H. J. Metropolis-hastings data augmentation for graph neural networks. Advances in Neural Information Processing Systems, 34:19010–19020, 2021.
- Provost (2000) Provost, F. Machine learning from imbalanced data sets 101. 2000.
- Ren et al. (2020) Ren, J., Yu, C., Ma, X., Zhao, H., Yi, S., et al. Balanced meta-softmax for long-tailed visual recognition. Advances in neural information processing systems, 33:4175–4186, 2020.
- Rong et al. (2019) Rong, Y., Huang, W., Xu, T., and Huang, J. Dropedge: Towards deep graph convolutional networks on node classification. arXiv preprint arXiv:1907.10903, 2019.
- Sankar et al. (2021) Sankar, A., Liu, Y., Yu, J., and Shah, N. Graph neural networks for friend ranking in large-scale social platforms. In Proceedings of the Web Conference 2021, 2021.
- Schlichtkrull et al. (2018) Schlichtkrull, M., Kipf, T. N., Bloem, P., Berg, R. v. d., Titov, I., and Welling, M. Modeling relational data with graph convolutional networks. In European semantic web conference, pp. 593–607. Springer, 2018.
- Shiao et al. (2022) Shiao, W., Guo, Z., Zhao, T., Papalexakis, E. E., Liu, Y., and Shah, N. Link prediction with non-contrastive learning. arXiv preprint arXiv:2211.14394, 2022.
- Stanfield et al. (2017) Stanfield, Z., Coşkun, M., and Koyutürk, M. Drug response prediction as a link prediction problem. Scientific reports, 2017.
- Tan et al. (2020) Tan, J., Wang, C., Li, B., Li, Q., Ouyang, W., Yin, C., and Yan, J. Equalization loss for long-tailed object recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11662–11671, 2020.
- Tang et al. (2020a) Tang, K., Huang, J., and Zhang, H. Long-tailed classification by keeping the good and removing the bad momentum causal effect. Advances in Neural Information Processing Systems, 33:1513–1524, 2020a.
- Tang et al. (2020b) Tang, X., Yao, H., Sun, Y., Wang, Y., Tang, J., Aggarwal, C., Mitra, P., and Wang, S. Investigating and mitigating degree-related biases in graph convoltuional networks. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, 2020b.
- Tang et al. (2022) Tang, X., Liu, Y., He, X., Wang, S., and Shah, N. Friend story ranking with edge-contextual local graph convolutions. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, 2022.
- Trouillon et al. (2016) Trouillon, T., Welbl, J., Riedel, S., Gaussier, É., and Bouchard, G. Complex embeddings for simple link prediction. In International conference on machine learning. PMLR, 2016.
- Vashishth et al. (2020) Vashishth, S., Sanyal, S., Nitin, V., and Talukdar, P. Composition-based multi-relational graph convolutional networks. In International Conference on Learning Representations, 2020.
- Veličković et al. (2017) Veličković, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., and Bengio, Y. Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.
- Volkovs et al. (2017) Volkovs, M., Yu, G., and Poutanen, T. Dropoutnet: Addressing cold start in recommender systems. Advances in neural information processing systems, 30, 2017.
- Wang et al. (2020) Wang, Z., Lei, Y., and Li, W. Neighborhood attention networks with adversarial learning for link prediction. IEEE Transactions on Neural Networks and Learning Systems, 32(8):3653–3663, 2020.
- Wang et al. (2021) Wang, Z., Zhou, Y., Hong, L., Zou, Y., Su, H., and Chen, S. Pairwise learning for neural link prediction. arXiv preprint arXiv:2112.02936, 2021.
- Wu et al. (2019) Wu, J., He, J., and Xu, J. Net: Degree-specific graph neural networks for node and graph classification. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 406–415, 2019.
- Wu et al. (2022) Wu, Q., Zhang, H., Yan, J., and Wipf, D. Handling distribution shifts on graphs: An invariance perspective. arXiv preprint arXiv:2202.02466, 2022.
- Wu et al. (2020) Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., and Philip, S. Y. A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems, 2020.
- Xu et al. (2018) Xu, K., Li, C., Tian, Y., Sonobe, T., Kawarabayashi, K.-i., and Jegelka, S. Representation learning on graphs with jumping knowledge networks. In International conference on machine learning, pp. 5453–5462. PMLR, 2018.
- Yan et al. (2021) Yan, Z., Ma, T., Gao, L., Tang, Z., and Chen, C. Link prediction with persistent homology: An interactive view. In International conference on machine learning, pp. 11659–11669. PMLR, 2021.
- Yang et al. (2021) Yang, C., Xiao, C., Ma, F., Glass, L., and Sun, J. Safedrug: Dual molecular graph encoders for recommending effective and safe drug combinations. In IJCAI, pp. 3735–3741, 2021.
- Yang et al. (2016) Yang, Z., Cohen, W., and Salakhudinov, R. Revisiting semi-supervised learning with graph embeddings. In International conference on machine learning, pp. 40–48. PMLR, 2016.
- Ying et al. (2018) Ying, R., He, R., Chen, K., Eksombatchai, P., Hamilton, W. L., and Leskovec, J. Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, 2018.
- Yun et al. (2021) Yun, S., Kim, S., Lee, J., Kang, J., and Kim, H. J. Neo-gnns: Neighborhood overlap-aware graph neural networks for link prediction. Advances in Neural Information Processing Systems, 2021.
- Zhang et al. (2020) Zhang, C., Yao, H., Huang, C., Jiang, M., Li, Z., and Chawla, N. V. Few-shot knowledge graph completion. In Proceedings of the AAAI Conference on Artificial Intelligence, 2020.
- Zhang & Chen (2018) Zhang, M. and Chen, Y. Link prediction based on graph neural networks. Advances in neural information processing systems, 2018.
- Zhang et al. (2021) Zhang, M., Li, P., Xia, Y., Wang, K., and Jin, L. Labeling trick: A theory of using graph neural networks for multi-node representation learning. Advances in Neural Information Processing Systems, 2021.
- Zhao & Akoglu (2019) Zhao, L. and Akoglu, L. Pairnorm: Tackling oversmoothing in gnns. arXiv preprint arXiv:1909.12223, 2019.
- Zhao et al. (2021) Zhao, T., Liu, Y., Neves, L., Woodford, O., Jiang, M., and Shah, N. Data augmentation for graph neural networks. In Proceedings of the aaai conference on artificial intelligence, 2021.
- Zhao et al. (2022a) Zhao, T., Jin, W., Liu, Y., Wang, Y., Liu, G., Günneman, S., Shah, N., and Jiang, M. Graph data augmentation for graph machine learning: A survey. arXiv preprint arXiv:2202.08871, 2022a.
- Zhao et al. (2022b) Zhao, T., Liu, G., Wang, D., Yu, W., and Jiang, M. Learning from counterfactual links for link prediction. In International Conference on Machine Learning. PMLR, 2022b.
- Zheng et al. (2021) Zheng, W., Huang, E. W., Rao, N., Katariya, S., Wang, Z., and Subbian, K. Cold brew: Distilling graph node representations with incomplete or missing neighborhoods. arXiv preprint arXiv:2111.04840, 2021.
- Zhu et al. (2019) Zhu, Y., Lin, J., He, S., Wang, B., Guan, Z., Liu, H., and Cai, D. Addressing the item cold-start problem by attribute-driven active learning. IEEE Transactions on Knowledge and Data Engineering, 2019.
- Zhu et al. (2021) Zhu, Z., Zhang, Z., Xhonneux, L.-P., and Tang, J. Neural bellman-ford networks: A general graph neural network framework for link prediction. Advances in Neural Information Processing Systems, 2021.
Appendix A Related Work
LP with GNNs. Over the past few years, GNN architectures (Kipf & Welling, 2016a; Gilmer et al., 2017; Hamilton et al., 2017; Veličković et al., 2017; Xu et al., 2018) have gained significant attention and demonstrated promising outcomes in LP tasks. There are two primary approaches to applying GNNs in LP. The first approach involves a node-wise encoder-decoder framework, which we discussed in Section 2. The second approach reformulates LP tasks as enclosing subgraph classification tasks (Zhang & Chen, 2018; Cai & Ji, 2020; Cai et al., 2021; Dong et al., 2022). Instead of directly predicting links, these methods perform graph classification tasks on the enclosing subgraphs sampled around the target link. These methods can achieve even better results compared to node-wise encoder-decoder frameworks by assigning node labels to indicate different roles within the subgraphs. However, constructing subgraphs poses challenges in terms of efficiency and scalability, requiring substantial computational resources. Our work focuses on the encoder-decoder framework for LP, circumventing the issues associated with subgraph construction.
Methods for Cold-start Nodes. Recently, several GNN-based methods (Wu et al., 2019; Liu et al., 2020; Tang et al., 2020b; Liu et al., 2021; Zheng et al., 2021) have explored degree-specific transformations to address robustness and cold-start node issues. Tang et al.(Tang et al., 2020b) introduced a degree-related graph convolutional network to mitigate degree-related bias in node classification tasks. Liu et al.(Liu et al., 2021) proposed a transferable neighborhood translation model to address missing neighbors for cold-start nodes. Zheng et al.(Zheng et al., 2021) tackled the cold-start nodes problem by recovering missing latent neighbor information. These methods require cold-start-node-specific architectural components, unlike our approach, which does not necessitate any architectural modifications. Additionally, other studies have focused on long-tail scenarios in various domains, such as cold-start recommendation(Chen et al., 2020; Lu et al., 2020; Hao et al., 2021). Imbalance tasks present another common long-tail problem, where there are long-tail instances within small classes (Lin et al., 2017; Ren et al., 2020; Tan et al., 2020; Kang et al., 2019; Tang et al., 2020a). Approaches like (Lin et al., 2017; Ren et al., 2020; Tan et al., 2020) address this issue by adapting the loss for different samples. However, due to the different problem settings, it is challenging to directly apply these methods to our tasks. We only incorporate the balanced cross entropy introduced by Lin et al. (Lin et al., 2017) as one of our baselines.
Graph Data Augmentation. Graph data augmentation expands the original data by perturbing or modifying the graphs to enhance the generalizability of GNNs (Zhao et al., 2022a; Ding et al., 2022). Existing methods primarily focus on semi-supervised node-level tasks(Rong et al., 2019; Feng et al., 2020; Zhao et al., 2021; Park et al., 2021) and graph-level tasks (Liu et al., 2022a; Luo et al., 2022). However, the exploration of graph data augmentation for LP remains limited (Zhao et al., 2022b). CFLP (Zhao et al., 2022b) proposes the creation of counterfactual links to learn representations from both observed and counterfactual links. Nevertheless, this method encounters scalability issues due to the high computational complexity associated with finding counterfactual links. Moreover, there exist general graph data augmentation methods (Liu et al., 2022b; Hu et al., 2022) that can be applied to various tasks. LAGNN (Liu et al., 2022b) proposed to use a generative model to provide additional neighbor features for each node. TuneUP (Hu et al., 2022) designs a two-stage training strategy, which trains GNNs twice to make them perform well on both warm nodes and cold-start nodes. These augmentation methods come with the trade-off of introducing extra runtime either before or during the model training. Unlike TLC-GNN (Yan et al., 2021), which necessitates extracting topological features for each node pair, and GIANT (Chien et al., 2021), which requires pre-training of the text encoder to improve node features, our methods are more streamlined and less complex.
Appendix B Additional Datasets Details
This section provides detailed information about the datasets used in our experiments. We consider various types of networks, including citation networks, collaboration networks, and co-purchase networks. The datasets we utilize are as follows:
-
•
Citation Networks: Cora and Citeseer originally introduced by Yang et al. (2016), consist of citation networks where the nodes represent papers and the edges represent citations between papers. IGB-100K (Khatua et al., 2023) is a recently-released benchmark citation network with high-quality node features and a large dataset size.
-
•
Collaboration Networks: CS and Physics are representative collaboration networks. In these networks, the nodes correspond to authors and the edges represent collaborations between authors.
-
•
Co-purchase Networks: Computers and Photos are co-purchase networks, where the nodes represent products and the edges indicate the co-purchase relationship between two products.
Why there are no OGB (Hu et al., 2020) datasets applied? OGB benchmarks that come with node features, such as OGB-collab and OGB-citation2, lack a substantial number of isolated or low-degree nodes, which makes it challenging to yield convincing results for experiments focusing on the cold-start problem. This is primarily due to the split setting adopted by OGB, where the evaluation is centered around a set of the most recent papers with high degrees. Besides, considering these datasets have their fixed splitting settings based on time, it will lead to inconsistent problems to compared with the leaderboard results if we use our own splitting method to ensure we have a reasonable number of isolated/low-degree nodes. Given these constraints, we opted for another extensive benchmark dataset, IGB-100K (Khatua et al., 2023), to test and showcase the effectiveness of our methods on large-scale graphs. We further conducted the experiments on IGB-1M, which are shown in Section D.8.
Transductive Setting | ||||||||
---|---|---|---|---|---|---|---|---|
Datasets | Original Graph | Testing Isolated | Testing Low-degree | Testing Warm | ||||
#Nodes | #Edges | #Nodes | #Edges | #Nodes | #Edges | #Nodes | #Edges | |
Cora | 2,708 | 5,278 | 135 | 164 | 541 | 726 | 662 | 1,220 |
Citeseer | 3,327 | 4,552 | 291 | 342 | 492 | 591 | 469 | 887 |
CS | 18,333 | 163,788 | 309 | 409 | 1,855 | 2,687 | 10,785 | 29,660 |
Physics | 34,493 | 495,924 | 275 | 397 | 2,062 | 3,188 | 25,730 | 95,599 |
Computers | 13,752 | 491,722 | 218 | 367 | 830 | 1,996 | 11,887 | 194,325 |
Photos | 7,650 | 238,162 | 127 | 213 | 516 | 1,178 | 6,595 | 93,873 |
IGB-100K | 100,000 | 547,416 | 1,556 | 1,737 | 6,750 | 7,894 | 23,949 | 35,109 |
Inductive Setting | ||||||||
Datasets | Original Graph | Testing Isolated | Testing Low-degree | Testing Warm | ||||
#Nodes | #Edges | #Nodes | #Edges | #Nodes | #Edges | #Nodes | #Edges | |
Cora | 2,708 | 5,278 | 149 | 198 | 305 | 351 | 333 | 505 |
Citeseer | 3,327 | 4,552 | 239 | 265 | 272 | 302 | 239 | 339 |
CS | 18,333 | 163,788 | 1,145 | 1,867 | 1,202 | 1,476 | 6,933 | 13,033 |
Physics | 34,493 | 495,924 | 2,363 | 5,263 | 1,403 | 1,779 | 17,881 | 42,548 |
Computers | 13,752 | 491,722 | 1,126 | 4,938 | 239 | 302 | 9,235 | 43,928 |
Photos | 7,650 | 238,162 | 610 | 2,375 | 169 | 212 | 5,118 | 21,225 |
IGB-100K | 100,000 | 547,416 | 5,507 | 9,708 | 8,706 | 13,815 | 24,903 | 41,217 |
B.1 Transductive Setting
For the transductive setting, we randomly split the edges into training, validation, and testing sets based on the splitting ratio specified in Section 4.1. The nodes in training/validation/testing are all visible during the training process. However, the positive edges in validation/testing sets are masked out for training. After the split, we calculate the degrees of each node using the validation graph. The dataset statistics are shown in Table 5.
B.2 Inductive Setting
The inductive setting is considered a more realistic setting compared to the transductive setting, where new nodes appear after the training process. Following the inductive setting introduced in Guo et al. (2022b) and Shiao et al. (2022), we perform node splitting to randomly sample 10% nodes from the original graph as the new nodes appear after the training process. The remaining nodes are considered observed nodes during the training. Next, we group the edges into three sets: observed-observed, observed-new, and new-new node pairs. We select 10% of observed-observed, 10% of observed-new, and 10% of new-new node pairs as the testing edges. We consider the remaining observed-new and new-new node pairs, along with an additional 10% of observed-observed node pairs, as the newly visible edges for the testing inference. The datasets statistics are shown in Table 5.
Appendix C NodeDup Algorithm
In this section, we provide a detailed description of our algorithm, which is outlined in Algorithm 1. Compared to the default training of GNNs for LP tasks, NodeDup incorporates additional augmentation steps, denoted as L1-L5 in Algorithm 1.
Appendix D Further Experimental Results
D.1 Selection of the threshold .
Our decision to set the threshold at 2 is grounded in data-driven analysis, as illustrated in Figure 1 and Figure 6. These figures reveal that nodes with degrees not exceeding 2 consistently perform below the average Hits@10 across all datasets, and higher than 2 will outperform the average results. Besides, our choice aligns with methodologies in previous studies (Liu et al., 2020, 2021), where cold nodes are identified using a fixed threshold across all the datasets. In addition, we conduct experiments with different thresholds on Cora and Citeseer datasets. The results are shown in Table 6. Our findings were consistent across different thresholds, with similar observations at = 1, = 2, and = 3. This indicates that our method’s effectiveness is not significantly impacted by changes in this threshold.




= 1 | = 2 | = 3 | |||||
Gsage | NodeDup | Gsage | NodeDup | Gsage | NodeDup | ||
Cora | Isolated | 31.345.60 | 42.202.30 | 32.203.58 | 44.273.82 | 31.951.26 | 43.172.94 |
Low-degree | 53.981.20 | 57.991.34 | 59.451.09 | 61.981.14 | 59.641.01 | 62.680.63 | |
Warm | 61.680.29 | 61.170.43 | 61.140.78 | 59.070.68 | 61.030.79 | 59.910.44 | |
Overall | 58.010.57 | 59.160.44 | 58.310.68 | 58.920.82 | 58.080.74 | 59.990.53 | |
Citeseer | Isolated | 47.251.82 | 56.491.72 | 47.132.43 | 57.541.04 | 47.312.17 | 56.901.12 |
Low-degree | 54.100.85 | 71.090.47 | 61.880.79 | 75.500.39 | 62.970.83 | 75.450.40 | |
Warm | 72.410.35 | 74.571.04 | 71.450.52 | 74.680.67 | 73.570.46 | 75.020.84 | |
Overall | 64.270.45 | 70.530.91 | 63.770.83 | 71.730.47 | 64.050.42 | 71.800.40 |
D.2 MRR results compared with the base GNN model and cold-start baselines
Table 7 presents the performance of our methods, evaluated using MRR, compared against the base GNN model and cold-start baselines. We can observe that NodeDup(L) consistently achieves significant improvements over the baseline methods for Isolated and Low-degree nodes across all datasets. NodeDup also outperforms baselines in 13 out of 14 cases on the cold nodes. This further demonstrates the superior effectiveness of our methods in addressing cold node scenarios. Furthermore, we can also observe that our methods consistently perform on par or above baseline methods in Warm nodes and the overall setting. This observation further supports the effectiveness of our methods in maintaining and even improving the performance of Warm nodes.
GSage | Imbalance | TailGNN | Cold-brew | NodeDup(L) | NodeDup | ||
---|---|---|---|---|---|---|---|
Cora | Isolated | 16.731.50 | 17.120.77 | 20.880.97 | 15.961.60 | 22.830.48 | 25.611.41 |
Low-degree | 38.460.62 | 37.931.17 | 40.190.96 | 35.200.55 | 40.201.02 | 39.780.97 | |
Warm | 36.970.60 | 34.940.87 | 36.390.51 | 31.970.31 | 36.990.41 | 35.340.32 | |
Overall | 35.910.51 | 34.590.81 | 36.490.59 | 31.840.17 | 36.890.47 | 35.820.34 | |
Citeseer | Isolated | 29.362.30 | 28.351.02 | 22.491.67 | 21.915.24 | 34.190.77 | 38.261.26 |
Low-degree | 44.130.38 | 44.670.44 | 43.921.55 | 34.6510.10 | 51.580.56 | 53.710.64 | |
Warm | 46.680.48 | 46.951.01 | 45.931.17 | 36.457.50 | 48.890.53 | 48.050.42 | |
Overall | 42.600.59 | 42.720.52 | 40.871.34 | 33.137.90 | 47.000.44 | 48.050.54 | |
CS | Isolated | 35.540.74 | 29.611.62 | 30.320.92 | 32.350.77 | 42.221.41 | 44.940.60 |
Low-degree | 56.180.81 | 57.440.68 | 46.660.61 | 42.670.26 | 61.200.64 | 61.650.84 | |
Warm | 58.180.84 | 57.030.77 | 48.320.44 | 43.710.41 | 59.940.54 | 58.670.72 | |
Overall | 57.730.83 | 56.720.73 | 47.960.45 | 43.480.38 | 59.830.52 | 58.740.70 | |
Physics | Isolated | 27.731.10 | 33.610.34 | 23.170.74 | 30.620.30 | 41.121.10 | 45.622.45 |
Low-degree | 61.400.52 | 62.740.27 | 47.050.39 | 41.950.15 | 64.040.43 | 65.940.21 | |
Warm | 66.720.47 | 66.030.09 | 55.770.49 | 46.060.12 | 66.940.49 | 66.830.04 | |
Overall | 66.390.47 | 65.800.09 | 55.360.49 | 45.860.12 | 66.740.49 | 66.720.04 | |
Computers | Isolated | 4.500.75 | 5.010.71 | 4.880.54 | 4.070.46 | 8.591.45 | 9.651.10 |
Low-degree | 26.650.62 | 26.850.31 | 21.220.56 | 23.400.59 | 28.851.13 | 29.780.32 | |
Warm | 34.110.40 | 33.770.17 | 31.020.34 | 28.750.23 | 35.110.31 | 35.630.14 | |
Overall | 33.980.40 | 33.650.16 | 30.880.34 | 28.640.23 | 35.000.31 | 35.520.13 | |
Photos | Isolated | 3.990.52 | 4.791.38 | 5.780.94 | 6.490.98 | 8.231.10 | 7.901.55 |
Low-degree | 25.101.35 | 24.601.20 | 20.411.29 | 21.540.35 | 27.900.90 | 26.901.29 | |
Warm | 34.900.57 | 33.030.47 | 30.790.63 | 29.400.23 | 36.840.55 | 35.690.43 | |
Overall | 34.710.57 | 32.870.47 | 30.600.63 | 29.260.22 | 36.660.54 | 35.520.43 | |
IGB-100K | Isolated | 53.200.24 | 50.810.41 | 45.250.26 | 48.420.25 | 59.340.51 | 61.750.47 |
Low-degree | 55.930.28 | 55.790.30 | 51.110.29 | 51.920.15 | 62.350.49 | 63.910.26 | |
Warm | 61.310.49 | 60.630.40 | 55.910.18 | 50.880.20 | 61.560.48 | 61.240.19 | |
Overall | 60.050.43 | 59.400.36 | 54.650.20 | 50.970.17 | 61.610.48 | 61.730.21 |
D.3 Additional results compared with augmentation baselines
Figure 7 presents the performance compared with augmentation methods on the remaining datasets. On Cora and CS datasets, we can consistently observe that NodeDup and NodeDup(L) outperform all the graph augmentation baselines for Isolated and Low-degree nodes. Moreover, for Warm nodes, NodeDup can also perform on par or above baselines. On the Computers and Photos datasets, our methods generally achieve comparable or superior performance compared to the baselines, except in comparison to TuneUP. However, it is worth noting that both NodeDup and NodeDup(L) exhibit more than 2 faster execution speed than TuneUP on these two datasets.

D.4 Additional results under the inductive setting
We further evaluate and present the effectiveness of our methods under the inductive setting on the remaining datasets in Table 8. We can observe that both NodeDup and NodeDup(L) consistently outperform GSage for Isolated, Low-degree, and Warm nodes. Compared to NodeDup(L), NodeDup is particularly beneficial for this inductive setting.
GSage | NodeDup(L) | NodeDup | ||
---|---|---|---|---|
Cora | Isolated | 43.641.84 | 45.310.83 | 46.060.66 |
Low-degree | 60.060.62 | 60.460.91 | 61.942.22 | |
Warm | 60.591.13 | 60.951.40 | 62.531.23 | |
Overall | 57.230.33 | 57.650.82 | 59.241.02 | |
CS | Isolated | 74.340.56 | 75.420.36 | 77.800.68 |
Low-degree | 75.750.48 | 77.020.65 | 81.330.60 | |
Warm | 82.550.27 | 83.520.67 | 83.550.50 | |
Overall | 81.000.28 | 82.010.59 | 82.700.52 | |
Computers | Isolated | 66.810.72 | 67.030.51 | 69.820.63 |
Low-degree | 64.172.01 | 65.101.76 | 66.360.69 | |
Warm | 68.760.40 | 68.780.39 | 70.490.41 | |
Overall | 68.540.42 | 68.590.39 | 70.400.42 | |
Photos | Isolated | 68.290.67 | 69.600.75 | 70.460.53 |
Low-degree | 63.021.51 | 64.251.31 | 68.492.39 | |
Warm | 70.170.57 | 71.050.70 | 71.610.81 | |
Overall | 69.920.57 | 70.840.63 | 71.470.77 |
D.5 Performance compared with Heuristic Methods
We compare our method with traditional link prediction baselines, such as common neighbors (CN), Adamic-Adar(AA), Resource allocation (RA). The results are shown in Table 9. We observe that NodeDup can consistently outperform these heuristic methods across all the datasets, with particularly significant improvements observed on Isolated nodes.
CN | AA | RA | DegFairGNN | GSage | NodeDup | ||
---|---|---|---|---|---|---|---|
Cora | Isolated | 0.00 | 0.00 | 0.00 | 18.701.53 | 32.203.58 | 44.273.82 |
Low-degree | 20.30 | 20.14 | 20.14 | 38.430.14 | 59.451.09 | 61.981.14 | |
Warm | 38.33 | 38.90 | 38.90 | 42.491.82 | 61.140.78 | 59.070.68 | |
Overall | 25.27 | 25.49 | 25.49 | 39.241.10 | 58.310.68 | 58.920.82 | |
Citeseer | Isolated | 0.00 | 0.00 | 0.00 | 15.501.27 | 47.132.43 | 57.541.04 |
Low-degree | 26.86 | 27.00 | 27.00 | 45.060.96 | 61.880.79 | 75.500.39 | |
Warm | 37.30 | 39.02 | 39.02 | 55.471.08 | 71.450.52 | 74.680.67 | |
Overall | 30.81 | 31.85 | 31.85 | 44.581.03 | 63.770.83 | 71.730.47 | |
CS | Isolated | 0.00 | 0.00 | 0.00 | 17.931.35 | 56.411.61 | 65.871.70 |
Low-degree | 39.60 | 39.60 | 39.60 | 49.830.68 | 75.950.25 | 81.120.36 | |
Warm | 72.73 | 72.74 | 72.72 | 61.720.37 | 84.370.46 | 84.760.41 | |
Overall | 69.10 | 69.11 | 69.10 | 60.200.37 | 83.330.42 | 84.230.39 | |
Physics | Isolated | 0.00 | 0.00 | 0.00 | 19.482.94 | 47.411.38 | 66.650.95 |
Low-degree | 46.08 | 46.08 | 46.08 | 47.630.52 | 79.310.28 | 84.040.22 | |
Warm | 85.48 | 85.74 | 85.70 | 62.790.82 | 90.280.23 | 90.330.05 | |
Overall | 83.87 | 84.12 | 84.09 | 62.130.76 | 89.760.22 | 90.030.05 | |
Computers | Isolated | 0.00 | 0.00 | 0.00 | 9.361.81 | 9.321.44 | 19.622.63 |
Low-degree | 28.31 | 28.31 | 28.31 | 18.900.81 | 57.910.97 | 61.160.92 | |
Warm | 59.67 | 63.50 | 62.84 | 31.442.25 | 66.870.47 | 68.100.25 | |
Overall | 59.24 | 63.03 | 62.37 | 31.272.22 | 66.670.47 | 67.940.25 | |
Photos | Isolated | 0.00 | 0.00 | 0.00 | 12.991.51 | 9.252.31 | 17.843.53 |
Low-degree | 28.44 | 28.78 | 28.78 | 20.180.21 | 52.610.88 | 54.131.58 | |
Warm | 64.53 | 67.26 | 66.88 | 42.720.89 | 67.640.55 | 68.680.49 | |
Overall | 63.94 | 66.64 | 66.26 | 42.370.87 | 67.320.54 | 68.390.48 | |
IGB-100K | Isolated | 0.00 | 0.00 | 0.00 | 57.0921.08 | 75.920.52 | 88.040.20 |
Low-degree | 12.26 | 12.26 | 12.26 | 59.4521.84 | 79.380.23 | 88.980.17 | |
Warm | 30.65 | 30.65 | 30.65 | 65.5720.43 | 86.420.24 | 88.280.20 | |
Overall | 26.22 | 26.22 | 26.22 | 64.1620.70 | 84.770.21 | 88.390.18 |
D.6 Performance compared with the methods tackling degree bias in GNNs
We have conducted the experiments to compare with the existing methods, TailGNN (Liu et al., 2021) and Cold-brew (Zheng et al., 2021), which also tackle degree bias in GNNs. The results, presented in Table 1, consistently show our methods outperforming these approaches. DegFairGNN (Liu et al., 2023) introduces a learnable debiasing function in the GNN architecture to produce fair representations for nodes, aiming for similar predictions for nodes within the same class, regardless of their degrees. Unfortunately, we’ve found that this approach is not well-suited for link prediction tasks for several reasons: (1) This method is designed specifically for node classification tasks. For example, the fairness loss, which ensures prediction distribution uniformity among low and high-degree node groups, is not suitable for link prediction because there is no defined node class in link prediction tasks. (2) This approach achieves significant performance in node classification tasks by effectively mitigating degree bias. However, in the context of link prediction, the degree trait is crucial. Applying DegFairGNN (Liu et al., 2023) would compromise the model’s ability to learn from structural information, such as isomorphism and common neighbors. This, in turn, would negatively impact link prediction performance, as evidenced by references (Zhang & Chen, 2018; Chamberlain et al., 2022).
D.7 Performance compared with upsampling
In Section 3, we discussed the issue of under-representation of cold nodes during the training of LP, which is the main cause of their unsatisfactory performance. To tackle this problem, one straightforward and naive approach is upsampling (Provost, 2000), which involves increasing the number of samples in the minority class. In order to further demonstrate the effectiveness of our methods, we conducted experiments where we doubled the edge sampling probability of code nodes, aiming to enhance their visibility. The results are presented in Table 10. We can observe that NodeDup(L) consistently outperforms upsampling, and NodeDup outperforms upsampling in almost all the cases, except for Warm nodes on Cora.
Dataset | Method | Isolated | Low-degree | Warm | Overall |
---|---|---|---|---|---|
Cora | Upsampling | 32.812.75 | 59.570.60 | 60.490.81 | 57.900.65 |
NodeDup(L) | 39.761.32 | 62.531.03 | 62.070.37 | 60.490.49 | |
NodeDup | 44.273.82 | 61.981.14 | 59.070.68 | 58.920.82 | |
Citeseer | Upsampling | 46.880.45 | 62.321.57 | 71.331.35 | 63.810.81 |
NodeDup(L) | 52.461.16 | 73.711.22 | 74.990.37 | 70.340.35 | |
NodeDup | 57.541.04 | 75.500.39 | 74.680.67 | 71.730.47 | |
CS | Upsampling | 49.632.24 | 75.620.13 | 83.400.73 | 82.340.64 |
NodeDup(L) | 65.181.25 | 81.460.57 | 85.480.26 | 84.900.29 | |
NodeDup | 65.871.70 | 81.120.36 | 84.760.41 | 84.230.39 | |
Physics | Upsampling | 52.010.97 | 79.630.13 | 89.410.32 | 89.330.46 |
NodeDup(L) | 65.040.63 | 82.700.22 | 90.440.23 | 90.090.22 | |
NodeDup | 66.650.95 | 84.040.22 | 90.330.05 | 90.030.05 | |
Computers | Upsampling | 11.360.72 | 58.230.88 | 67.070.49 | 66.870.48 |
NodeDup(L) | 17.111.62 | 62.141.06 | 68.020.41 | 67.860.41 | |
NodeDup | 19.622.63 | 61.160.92 | 68.100.25 | 67.940.25 | |
Photos | Upsampling | 10.922.15 | 51.670.98 | 65.750.73 | 65.450.71 |
NodeDup(L) | 21.502.14 | 55.701.38 | 69.680.87 | 69.400.86 | |
NodeDup | 17.843.53 | 54.131.58 | 68.680.49 | 68.390.48 | |
IGB-100K | Upsampling | 75.490.90 | 79.470.11 | 86.540.19 | 84.870.14 |
NodeDup(L) | 87.430.44 | 88.370.24 | 88.540.31 | 88.470.28 | |
NodeDup | 88.040.20 | 88.980.17 | 88.280.20 | 88.390.18 |
GSage | NodeDup | ||
---|---|---|---|
IGB-1M | Isolated | 82.100.06 | 87.810.40 |
Low-degree | 84.730.06 | 90.840.03 | |
Warm | 89.980.02 | 91.310.02 | |
Overall | 89.800.02 | 91.290.02 |
D.8 Performance on Large-scale datasets
As outlined in Section 3.1, our methods incur a minimal increase in time complexity compared to base GNNs, with the increase being linearly proportional to the number of cold nodes. This ensures scalability. Besides, the effectiveness of our method is also insensitive to dataset size. We extend our experiments to the IGB-1M dataset, featuring 1 million nodes and 12 million edges. The findings, which we detail in Table 11, affirm the effectiveness of our methods in handling large-scale datasets, consistent with observations from smaller datasets.
GAT | NodeDup(L) | NodeDup | JKNet | NodeDup(L) | NodeDup | ||
---|---|---|---|---|---|---|---|
Cora | Isolated | 25.611.78 | 30.732.54 | 36.831.76 | 30.121.02 | 37.442.27 | 43.903.66 |
Low-degree | 54.880.84 | 55.760.50 | 56.720.81 | 59.560.66 | 61.931.64 | 62.891.43 | |
Warm | 55.311.14 | 55.361.28 | 53.701.26 | 58.640.12 | 59.361.00 | 57.671.60 | |
Overall | 52.850.91 | 53.580.80 | 53.430.49 | 56.740.27 | 58.540.83 | 58.401.33 | |
CS | Isolated | 33.741.98 | 34.770.90 | 41.762.99 | 54.431.77 | 56.382.14 | 64.791.68 |
Low-degree | 70.200.47 | 70.900.32 | 71.920.36 | 73.970.72 | 76.640.38 | 77.770.43 | |
Warm | 78.390.28 | 78.670.33 | 77.690.89 | 82.380.67 | 83.290.37 | 79.200.13 | |
Overall | 77.160.24 | 77.490.30 | 77.200.80 | 81.350.62 | 82.410.32 | 78.910.13 | |
Computers | Isolated | 12.042.08 | 16.842.34 | 17.172.22 | 9.923.07 | 23.812.02 | 25.501.32 |
Low-degree | 53.601.51 | 53.621.00 | 53.652.35 | 62.291.08 | 67.210.99 | 68.490.70 | |
Warm | 60.191.19 | 58.640.81 | 58.551.01 | 69.960.33 | 70.900.40 | 70.660.25 | |
Overall | 60.031.19 | 58.500.80 | 58.771.93 | 69.770.32 | 70.780.40 | 70.550.25 | |
Photos | Isolated | 15.313.46 | 18.032.50 | 18.773.33 | 12.772.40 | 19.441.31 | 20.561.61 |
Low-degree | 43.119.93 | 43.409.61 | 44.219.25 | 57.272.06 | 59.861.09 | 60.930.74 | |
Warm | 56.178.28 | 56.758.33 | 56.108.35 | 68.350.81 | 69.560.69 | 69.600.50 | |
Overall | 55.919.22 | 56.488.26 | 55.938.28 | 68.090.82 | 69.330.68 | 69.380.49 |
MLP-Dec. | NodeDup(L) | NodeDup | ||
---|---|---|---|---|
Cora | Isolated | 16.832.61 | 37.323.87 | 38.411.22 |
Low-degree | 58.831.77 | 64.462.13 | 64.021.02 | |
Warm | 58.840.86 | 61.570.98 | 58.660.61 | |
Overall | 55.571.10 | 60.680.66 | 58.930.25 | |
CS | Isolated | 5.601.14 | 58.680.95 | 60.200.68 |
Low-degree | 71.461.08 | 78.820.68 | 79.580.31 | |
Warm | 84.540.32 | 85.880.22 | 85.200.24 | |
Overall | 82.480.32 | 84.960.25 | 84.420.22 | |
Computers | Isolated | 6.133.63 | 27.743.38 | 26.703.98 |
Low-degree | 62.561.34 | 62.603.38 | 63.353.64 | |
Warm | 69.721.31 | 70.012.41 | 68.432.50 | |
Overall | 69.531.30 | 69.913.11 | 68.302.51 | |
Photos | Isolated | 6.342.67 | 18.152.02 | 18.971.71 |
Low-degree | 55.636.21 | 56.136.36 | 55.937.27 | |
Warm | 70.406.84 | 70.676.30 | 69.975.07 | |
Overall | 69.896.81 | 69.936.24 | 69.695.07 |
D.9 Ablation study
Performance with various encoders and decoders. For the ablation study, we further explored various encoders and decoders on the remaining datasets. The results are shown in Table 12 and Table 13. From these two tables, we can observe that regardless of the encoders or decoders, both NodeDup and NodeDup(L) consistently outperform the base model for Isolated and Low-degree nodes, which further demonstrates the effectiveness of our methods on cold nodes. Furthermore, NodeDup(L) consistently achieves better performance compared to the base model for Warm nodes.
SEAL | SEAL + NodeDup | ||
---|---|---|---|
Cora | Isolated | 62.201.06 | 70.730.61 |
Low-degree | 66.802.83 | 67.704.11 | |
Warm | 56.692.36 | 54.871.61 | |
Overall | 60.602.38 | 60.892.36 | |
Citeseer | Isolated | 56.925.53 | 66.371.01 |
Low-degree | 64.132.56 | 65.541.69 | |
Warm | 58.813.22 | 60.732.75 | |
Overall | 60.182.98 | 63.351.43 |
Performance with SEAL (Zhang & Chen, 2018). Considering our methods are flexible to integrate with GNN-based link prediction structures, we conduct the experiments on top of SEAL (Zhang & Chen, 2018) on the Cora and Citeseer datasets. The results are shown in Table 14. We can observe that adding NodeDup on top of SEAL can consistently improve link prediction performance in the Isolated and Low-degree node settings on these two datasets.
The influence of duplication times and nodes. In our experiments, we duplicate cold nodes once and add one edge for each cold node in NodeDup. In Figure 8, we present the results of our ablation study, focusing on the effects of duplication times and duplicated nodes on the performance of NodeDup in terms of Isolated, Low-degree, and Overall settings. The numbers displayed in each block represent the differences compared to duplicating cold nodes once. We observe that increasing the duplication times does not necessarily lead to improvements across all settings, except when duplicating all nodes for Isolated nodes performance. We also notice that duplicating all nodes multiple times can significantly enhance the performance on Isolated nodes. However, this strategy negatively impacts the overall performance due to the increased number of isolated nodes in the graph. Furthermore, we observe that duplicating cold nodes once remains the optimal strategy since it consistently delivers strong performance across all settings.
To make our analysis more comprehensive, we further conducted the experiments to show the results with duplicating warm nodes, mid-warm nodes, and randomly sampled nodes for one time, respectively, on Citeseer. The results are shown in Table 15. From the table, we can observe that duplicating mid-warm and warm nodes are not useful for the LP performance for all the settings. It’s probably because for the mid-warm and warm nodes, the neighbors’ information and supervised training signals are informative enough, therefore NodeDup cannot contribute more. We can also observe that duplicating random nodes is more effective than duplicating warm nodes but less effective than duplicating cold nodes and duplicating all nodes.



Isolated | Low-degree | Warm | Overall | |
---|---|---|---|---|
Supervised | 47.13 | 61.88 | 71.45 | 63.77 |
D_Isolated | 54.04 | 72.28 | 74.53 | 69.95 |
D_Cold | 57.54 | 75.50 | 74.68 | 71.73 |
D_Mid-warm | 46.93 | 61.34 | 71.84 | 63.75 |
D_Warm | 47.49 | 62.20 | 71.54 | 63.99 |
D_Random | 54.10 | 72.39 | 75.05 | 70.06 |
D_All | 58.87 | 76.09 | 76.01 | 72.44 |
Appendix E Implementation Details
In this section, we introduce the implementation details of our experiments. Our implementation can be found at https://github.com/snap-research/NodeDup/tree/main.
Parameter Settings. We use 2-layer GNN architectures with 256 hidden dimensions for all GNNs and datasets. The dropout rate is set as 0.5. We report the results over 10 random seeds. Hyperparameters were tuned using an early stopping strategy based on performance on the validation set. We manually tune the learning rate for the final results. For the results with the inner product as the decoder, we tune the learning rate over range: . For the results with MLP as the decoder, we tune the learning rate over range: .
Hardware and Software Configuration All methods were implemented in Python 3.10.9 with Pytorch 1.13.1 and PyTorch Geometric (Fey & Lenssen, 2019). The experiments were all conducted on an NVIDIA P100 GPU with 16GB memory.
Appendix F Limitations
In our work, NodeDup and NodeDup(L) are specifically proposed for LP tasks. Although cold-start is a widespread issue in all graph learning tasks, our proposed methods might not be able to generalize to other tasks, such as node classification, due to their unique design.