Pre-Trained Models for Heterogeneous Information Networks
Abstract
In network representation learning we learn how to represent heterogeneous information networks in a low-dimensional space so as to facilitate effective search, classification, and prediction solutions. Previous network representation learning methods typically require sufficient task-specific labeled data to address domain-specific problems. The trained model usually cannot be transferred to out-of-domain datasets. We propose a self-supervised pre-training and fine-tuning framework, PF-HIN, to capture the features of a heterogeneous information network. Unlike traditional network representation learning models that have to train the entire model all over again for every downstream task and dataset, PF-HIN only needs to fine-tune the model and a small number of extra task-specific parameters, thus improving model efficiency and effectiveness. During pre-training, we first transform the neighborhood of a given node into a sequence. PF-HIN is pre-trained based on two self-supervised tasks, masked node modeling and adjacent node prediction. We adopt deep bi-directional transformer encoders to train the model, and leverage factorized embedding parameterization and cross-layer parameter sharing to reduce the parameters. In the fine-tuning stage, we choose four benchmark downstream tasks, i.e., link prediction, similarity search, node classification, and node clustering. PF-HIN consistently and significantly outperforms state-of-the-art alternatives on each of these tasks, on four datasets.
Index Terms:
Heterogeneous information network; Pre-training; Transformer1 Introduction
Complex information often involves multiple types of objects and relations. Such information can be represented via heterogeneous information networks (HINs) [1]. In a heterogeneous information network (HIN) different types of nodes (objects) are connected by edges (relations) [2]. Compared to homogeneous networks that only feature a single type of node, HINs provide a richer modeling tool, leading to more effective solutions for search, classification, and prediction tasks [3].
In order to mine the rich information captured by a HIN, network representation learning (NRL) embeds a network into a low-dimensional space. network representation learning (NRL) has drawn a significant amount of interest from the research community. Classical network embedding models like DeepWalk [4], LINE [5] and node2vec [6] have been devised for homogeneous networks, using random walks to capture the structure of networks. However, these methods lack the ability to capture a heterogeneous information network with multiple types of objects and relations. Hence, models designed specifically for HINs have been proposed [7, 8, 9]. A central concept here is that of a metapath, which is a sequence of node types with edge types in between. To leverage the relationship between nodes and metapaths, different mechanisms have been proposed, such as the heterogeneous SkipGram [7], proximity distance [8], and the Hardmard function [9]. Because of the limited ability of metapaths to capture the neighborhood structure of a node, the performance of these NRL methods is limited.
Recently, graph neural networks have shown promising results on modeling the structure of a network [10, 11, 12]. GNNs usually involve encoders that are able to explore and capture the neighborhood structure around a node, thus improving the performance on representing an HIN. However, GNNs need to be trained in an end-to-end manner with supervised information for a task, and the model learned on one dataset cannot easily be transferred to other, out-of-domain datasets. For different datasets and tasks, the methods listed above need to be re-trained all over. Additionally, especially on large-scale datasets, the amount of available labeled data is rarely sufficient for effective training.
Inspired by advances in pre-training frameworks in language technology [13, 14, 15], there is a trend to investigate pre-trained models for NRL. In particular, graph contrastive coding (GCC) [16] and GPT-GNN [17] are the most advanced solutions in this stream.111The experiments on downstream tasks using GPT-GNN [17] were conducted on the same dataset that had been employed for pre-training. Nevertheless, they are mainly proposed for generic NRL, meaning that they overlook the heterogeneous features of HINs; while they are generally applicable to HINs, they tend to fall short when handling HINs (as demonstrated empirically in Section 5 below).
We aim to overcome the shortcomings listed above, and propose 1. to pre-train a model on large datasets using self-supervision tasks, and 2. for a specific downstream task on a specific dataset, to use fine-tuning techniques with few task-specific parameters, so as to improve the model efficiency and effectiveness. We refer to this two-stage (Pre-training and Fine-tuning) framework for exploring the features of a HIN as PF-HIN.
Given a node in a HIN, we first explore the node’s neighborhood by transforming it into a sequence to better capture the features of neighboring structure. Then, a ranking of all the nodes is established based on their betweenness centrality, eigencentrality and closeness centrality. We use rank-guided heterogeneous walks to generate the sequence and group different types of nodes into so-called mini-sequences, that is, sequences of nodes of the same type [12].
We design two tasks for pre-training PF-HIN. One is the masked node modeling (MNM) task, in which a certain percentage of nodes in the mini-sequences are masked and we need to predict those masked nodes. This operation is meant to help PF-HIN learn type-specific node features. The other task is the adjacent node prediction (ANP) task, which is meant to capture the relationship between nodes. Given a node having sequence , our aim is to predict whether the node with sequence is an adjacent node. We adopt two strategies to reduce the parameters to further improve the efficiency of PF-HIN, i.e., factorized embedding parameterization and cross-layer parameter sharing. The large-scale dataset we use for pre-training is the open academic graph (OAG), containing 179 million nodes and 2 billion edges.
During fine-tuning, we choose four benchmark downstream tasks: 1. link prediction, 2. similarity search, 3. node classification, and 4. node clustering. In link prediction and similarity search, we use node sequence pairs as input, and identify whether there is a link between two nodes and measure the similarity between two nodes, respectively. In the node classification and node clustering tasks, we use a single node sequence as input, employing a softmax layer for classification and a k-means algorithm for clustering, respectively.
In our experiments, which are meant to demonstrate that PF-HIN is transferable across datasets, besides a subset of OAG denoted as OAG-mini, we include three other datasets for downstream tasks: DBLP, YELP and YAGO. PF-HIN consistently and significantly outperforms the state-of-the-art on these downstream tasks.
Our main contributions can be summarized as follows:
-
•
We propose a pre-training and fine-tuning framework PF-HIN to mine information contained in a HIN; PF-HIN is transferable to different downstream tasks and to datasets of different domains.
-
•
We adopt deep bi-directional transformer encoders to capture the structural features of a HIN; the architecture of PF-HIN is a variant of a GNN.
-
•
We use type-based masked node modeling and adjacent node prediction tasks to pre-train PF-HIN; both help PF-HIN to capture heterogeneous node features and relationships between nodes.
-
•
We show that PF-HIN significantly outperforms the state of the art on four benchmark downstream tasks across datasets.
2 Related work
2.1 Network representation learning
Research on NRL traces back to dimensionality reduction techniques [18, 19, 20, 21], which utilize feature vectors of nodes to construct an affinity graph and then calculate eigenvectors. Graph factorization models [22] represent a graph as an adjacency matrix, and generate a low-dimensional representation via matrix factorization. Such models suffer from high computational costs and data sparsity, and cannot capture the global network structure [5].
Random walks or paths in a network are being used to help preserve the local and global structure of a network. DeepWalk [4] leverages random walks and applies the SkipGram word2vec model to learn network embeddings. node2vec [6] extends DeepWalk; it adopts a biased random walk strategy to explore the network structure. LINE [5] harnesses first- and second-order proximities to encode local and neighborhood structure information.
The aforementioned approaches are designed for homogeneous networks; other methods have been introduced for heterogeneous networks. PTE [23] defines the conditional probability of nodes of one type generated by nodes of another, and forces the conditional distribution to be close to its empirical distribution. metapath2vec [7] has a heterogeneous SkipGram with its context window restricted to one specific type. HINE [8] uses a metapath based proximity and minimizes the distance between nodes’ joint probability and empirical probabilities. HIN2Vec [9] uses Hadamard multiplication of nodes and metapaths to capture features.
What we contribute to NRL on top of the work listed above is an efficient and effective method for representation learning for HINs based on graph neural networks (GNNs).
2.2 Graph neural networks
GNN models have shown promising results for representing networks. Efforts have been devoted to generalizing convolutional operations from visual data to graph data. Bruna et al. [24] propose a spectral graph theory based graph convolution operation. \AcpGCN [10] adopt localized first-order approximations of spectral graph convolutions to improve scalability. There is a line of research to improve spectral GNN models [25, 26, 27, 28], but it processes the whole graph simultaneously, leading to efficiency bottlenecks. To address the problem, spatial GNN models have been proposed [29, 30, 31, 32]. GraphSAGE [32] leverages a sampling strategy to iteratively sample neighboring nodes instead of the whole graph. Gao et al. [31] utilize a sub-graph training method to reduce memory and computational cost.
GNN fuse neighboring nodes or walks in graphs so as to learn a new node representation [11, 33, 34]. The main difference with convolution based models is that graph attention networks introduce attention mechanisms to assign higher weights to more important nodes or walks. GAT [11] harnesses masked self-attention layers to apply different weights to different nodes in a neighborhood, to improve efficiency on graph-structured data. GIN [35] models injective multiset functions for neighbor aggregation by parameterizing universal multiset functions with neural networks.
The above GNN models have been devised for homogeneous networks as they aggregate neighboring nodes or walks regardless of their types. Targeting HINs, HetGNN [12] first samples a fixed number of neighboring nodes of a given node and then groups these based on their types. Then, it uses a neural network architecture with two modules to aggregate the feature information of the neighboring nodes. One module is used to encode features of each type of node, the other to aggregate features of different types. HGT [36] uses node- and edge-type dependent parameters to describe heterogeneous attention over each edge; it also uses a heterogeneous mini-batch graph sampling algorithm for training.
What we contribute to GNNs is that the traditional NRL and GNN models listed above need to be re-trained all over again for different datasets and tasks, while our proposal PF-HIN only needs fine-tuning using a small number of task-specific parameters for a specific task and dataset, after pre-training via self-supervision tasks.
2.3 Graph pre-training
There exist relatively few approaches for pre-training a GNN model for downstream tasks. InfoGraph [37] maximizes the mutual information between graph-level embeddings and sub-structure embeddings. Hu et al. [38] pre-train a GNN at the level of nodes and graphs to learn local and global features, showing performance improvements on various graph classification tasks. Our proposed model, PF-HIN, differs as we focus on node-level transfer learning and pre-train our model on a single (large-scale) graph.
Hu et al. [39] design three pre-training tasks: denoising link reconstruction, centrality score ranking, and cluster preserving. GPT-GNN [17] adopts HGT [36] as its base GNN and uses attribute generation and edge generation as pre-training tasks. Hu et al. [17] only conduct their downstream tasks on the same dataset that was used for pre-training. GCC [16] designs subgraph instance discrimination as a pre-training task and uses contrastive learning to train GNNs, with its base GNNs as GIN; then it transfers its pre-trained model to different datasets. However, it is designed for homogeneous networks, not apt to exploit heterogeneous networks. Hwang et al. [40] proposes to adopt meta-path prediction as a pre-training task, which is to predict if two nodes are connected via a meta-path. The authors have not assigned a name for this method, so use MPP to refer to it in this paper.
What we contribute to graph pre-training on top of the work listed above is that we are able to not only fine-tune the proposed model across different tasks and different datasets, but can also deal with heterogeneous networks.

3 The Proposed Model PF-HIN
3.1 Preliminaries
A heterogeneous information network (HIN) is a graph , where denotes the set of nodes and denotes the set of edges between nodes. Each node and edge is associated with a type mapping function, and , respectively. and denote the sets of node and edge types. A HIN is a network where and/or .
A visual presentation of the proposed model, pre-training and fin-tuning heterogeneous information network (PF-HIN), is given in Figure 1. Below, we describe the node sequence generation procedure, the input representation, followed by the pre-training and fine-tuning stages of PF-HIN.
3.2 Heterogeneous node sequence generation
We first transform the structure of a node’s neighborhood to a sequence of length . To measure the importance of nodes based on the structural roles in the graph, node centrality is proposed in [41]. This framework makes use of three centrality metrics222This framework is extensible in the sense that additional metric that is of particular interest to the user can be explicitly supplemented after the three metrics, as long as the combination exhibits better performance in downstream tasks. In our implementation, we stick to the basic version consisted of three metrics to demonstrate the effectiveness of the framework., i.e., 1. betweenness centrality, 2. eigencentrality, and 3. closeness centrality. Betweenness centrality is calculated as the fraction of shortest paths that pass through a given node. Eigencentrality measures the influence of a node on its neighbors. Closeness centrality computes the total length of the shortest paths between the given node and others. We assign learnable weights to these metrics.
To capture heterogeneous features of a node’s neighbor, we adopt a so-called rank-guided heterogeneous walk to form a sequence. The walk is with restart; it will iteratively travel from a node to its neighbors. It starts from node , and it first reaches out to a node with a higher rank, which is what makes the walk rank-guided. This walk will not stop until it collects a pre-determined number of nodes. In order to assign the model with a sense of heterogeneity, we constrain the number of different types to be collected in the sequence so that every type of node can be included. We group nodes into mini-sequences, where a mini-sequence is a sequence of nodes having the same type [12]. In each mini-sequence, the nodes are sorted based on each node’s rank, which serve as a kind of position information.
Importantly, unlike traditional sampling strategies like random walks, breadth first search or depth first search, our sampling strategy is able to extract important and influential neighboring nodes for each node by selecting nodes with a higher rank; this allows us to capture more representative structural information of a neighborhood. Our sampling strategy collects all types of node for each node while traditional strategies ignore the nodes’ types. Nodes of the same type are grouped in mini-sequences so that further type-based analysis can be conducted to capture the heterogeneous features of a HIN. More empirical results with analysis are provided in Section 5.3.
3.3 Input embeddings learned via Bi-LSTMs
After generating the sequences, we learn the input embeddings of each node in the sequence using a Bi-LSTM layer. Given the input sequence , in which , a Bi-LSTM is used to capture the interaction relationships between nodes. The Bi-LSTM is composed of a forward and a backward LSTM layer. The LSTM layer is defined as follows:
(1) | ||||
(2) | ||||
(3) | ||||
(4) | ||||
(5) | ||||
(6) |
where is the output hidden state of node , represents the element-wise product, and are learnable parameters, which denote weight and bias, respectively; , , are the input gate vector, forget gate vector and output vector, respectively. We concatenate the hidden state of the forward and backward LSTM layers to form the final hidden state of the Bi-LSTM layer. For each type of node, we adopt different Bi-LSTMs so as to extract type-specific features.
3.4 Masked node modeling
After generating the input embeddings via a Bi-LSTM, we adopt masked node modeling (MNM) as our pre-training task. We randomly mask a percentage of the input nodes and then predict those masked nodes. We conduct this task on the type-based mini-sequences generated by the aforementioned rank guided heterogeneous walk. For each group of nodes with the same type, we randomly mask nodes in the mini-sequence. Given the mini-sequence of type , denoted as , we randomly choose 15% of the nodes to be replaced. And for a chosen node , we replace its token with the actual [MASK] token with 80% probability, another random node token with 10% probability and the unchanged with 10% probability. The masked sequence is fed into the bi-directional transformer encoders. The embeddings generated via the Bi-LSTM are used as token embeddings, while the rank information is transferred as position embeddings. After the transformer module, the final hidden state corresponding to the [MASK] token is fed to a feedforward layer. The output is used to predict the target node via a softmax classification layer:
(7) | ||||
(8) |
where is the output of the feedforward layer, is the classification weight shared with the input node embedding matrix, is the number of nodes in the -type mini-sequence, is the dimension of the hidden state size, is the predicted distribution of over all nodes.
For training, we use the cross-entropy between the one-hot label and the prediction :
(9) |
where and are the -th components of and , respectively. We adopt a smoothing strategy by setting for the target node and for each of the other nodes. By doing so, we loosen the restriction that a one-hot label corresponds to only one answer.
3.5 Adjacent node prediction
After the masked node modeling module, we design another pre-training task, i.e., adjacent node prediction (ANP), to further capture the relationship between nodes. Unlike the MNM task, which operates on type-based mini-sequences, we perform the ANP task on full sequences, and we compare two full sequences to see whether their starting nodes are adjacent or not. The reason that we do not perform the ANP task on type-based mini-sequences is that given types of nodes, there will be mini-sequence pairs to be analyzed, which is very time-consuming.
In our setting, for node with sequence and node with sequence , 50% of the time we choose to be the actual adjacent node of (labeled as IsAdjacent), and 50% of the time we randomly choose from the corpus (labeled as NotAdjacent). Given the classification layer weights , the scoring function of whether the node pair is adjacent is shown as follows:
(10) |
where is a binary vector with and , denotes the hidden vector of classification label used in a transformer architecture [42]. Considering the positive adjacent node pair and a negative adjacent node pair , we calculate a cross-entropy loss as follows:
(11) |
where is the label (positive or negative) of that node pair.
3.6 Transformer architecture
Our two pre-training tasks share the same transformer architecture. To increase the training speed of our model, we adopt two parameter reduction techniques to lower memory requirements, inspired by the ALBERT architecture [43]. Instead of setting the node embedding size to be equal to the hidden layer size like BERT does, we make more efficient use of the total number of model parameters, dictating that . We adopt factorized embedding parameterization, which decomposes the parameters into two smaller matrices. Rather than mapping the one-hot vectors directly to a hidden space with size , we first map them to a low-dimensional embedding space with size , and then map it to the hidden space. Additionally, we adopt cross-layer parameter sharing to further boost efficiency. Traditional sharing mechanisms either only share the feed forward network parameters across layers or only the attention parameters. We share all parameters across layers.
We denote the number of transformer layers as , and the number of self-attention heads as . For our parameter settings we follow the configuration of ALBERT [43], where is set to 12, to 768, to 12, to 128, and the total number of parameters is equal to 12M. For the procedure of our pre-training task, see Algorithm 1.
3.7 Fine-tuning PF-HIN
The self-attention mechanism in the transformer allows PF-HIN to model many downstream tasks. Fine-tuning can be realized by simply swapping out the proper inputs and outputs, regardless of the single node sequence or sequence pairs used. For each downstream task, the task-specific inputs and outputs are simply plugged into PF-HIN and all parameters are fine-tuned end-to-end. Here, we introduce four tasks: 1. link prediction, 2. similarity search, 3. node classification, and 4. node clustering as downstream tasks.
Specifically, in link prediction, we predict whether there is a link between two nodes, and the inputs are the node sequence pairs. To generate output, we feed the classification label into the sigmoid layer, so as to predict the existence of a link between two nodes. The only new parameters are the classification layer weights , where is the size of hidden state.
In similarity search, in order to measure the similarity between two nodes, we use the node sequence pairs as input. We leverage the token-level output representations to compute the similarity score of two nodes.
In node classification, we only use a single node sequence as input and generate the classification label via a softmax layer. To calculate the classification loss, we only need to add classification layer weights as new parameters, where is the number of classification labels and is the size of hidden state.
In node clustering, we also use a one node sequence as input and then put the token-level output embeddings to a clustering model, so as to cluster the data.
Experimental details for these downstream tasks are introduced in Section 4 below.
4 Experimental Setup
We detail our datasets, baseline models, and parameter settings.
4.1 Datasets
We adopt the open academic graph OAG333https://www.openacademic.ai/oag/ as our pre-training dataset, which is a heterogeneous academic dataset. It contains over 178 million nodes and 2.223 billion edges with five types of node: 1. author, 2. paper, 3. venue, 4. field, and 5. institute.
For downstream tasks, we transfer our pre-trained model to four datasets: 1. OAG-mini, 2. DBLP, 3. YELP, and 4. YAGO. OAG-mini is a subset extracted from OAG; the authors are split into four areas: machine learning, data mining, database, and information retrieval. DBLP444http://dblp.uni-trier.de is also an academic dataset with four types of node: 1. author, 2. paper, 3. venue, and 4. topic; the authors are split into the same areas as those in OAG-mini. YELP555https://www.yelp.com/dataset\_challenge is a social media dataset, with restaurants reviews and four types of node: 1. review, 2. customer, 3. restaurant, and 4. food-related keywords. The restaurants are separated into 1. Chinese food, 2. fast food, and 3. sushi bar. YAGO666https://old.datahub.io/dataset/yago is a knowledge base and we extracted a subset of it containing movie information, having five types of node: 1. movie, 2. actor, 3. director, 4. composer, and 5. producer. The movies are split into five types: 1. action, 2. adventure, 3. sci-fi, 4. crime, and 5. horror. The dataset statistics are shown in Table I.
Dataset | #nodes | #edges | #node types |
---|---|---|---|
OAG | 178,663,927 | 2,236,196,802 | 5 |
OAG-mini | 473,324 | 2,343,578 | 5 |
DBLP | 301,273 | 1,382,587 | 4 |
YELP | 201,374 | 872,432 | 4 |
YAGO | 52,384 | 143,173 | 5 |
4.2 Algorithms used for comparison
We first choose network embedding methods to directly train the downstream datasets for specific tasks as baselines: DeepWalk [4], LINE [5] and node2vec [6]; they were originally applied to homogeneous information networks. DeepWalk and node2vec leverage random walks, while node2vec uses a biased walk strategy to capture the network structure. LINE uses the local and neighborhood structural information via first-order and second-order proximities.
We include three state-of-the-art algorithms devised for HINs: metapath2vec [7], HINE [8], HIN2Vec [9]. They are all based on metapaths, but differ in the way they use metapath features: metapath2vec adopts heterogeneous SkipGrams, HINE proposes a metapath-based notion of proximity, and HIN2Vec utilizes the Hadamard multiplication of nodes and metapaths.
We also include other GNN models, i.e., GCN [10], GAT [11], GraphSAGE [32] and GIN [35], which were originally devised for homogeneous information networks. GCN and GraphSAGE are based on convolutional operations, while GCN requires the Laplacian of the full graph, and GraphSAGE only needs a node’s local neighborhood. GAT employs an attention mechanism to capture the correlation between central node and neighboring nodes. GIN uses parameterizing universal multiset functions with neural networks to model injective multiset functions for neighbor aggregation.
We also select HetGNN [12], HGT [36] as models for comparison; both have been devised for HIN embeddings. HetGNN samples heterogeneous neighbors, grouping them based on their node types, and then aggregates feature information of the sampled neighboring nodes. HGT has node- and edge-type dependent parameters to characterize heterogeneous attention over each edge.
Aside from those network embedding methods, for a fair comparison, we also choose GPT-GNN [17], GCC [16] and MPP [40] to run the entire pre-training and fine-tuning pipeline. GPT-GNN utilizes attribute generation and edge generation tasks to pre-train GNN, with HGT as its base GNN. GCC adopts subgraph instance discrimination as a pre-training task, taking GIN as its base GNN. MPP proposes to adopt meta-path prediction as the pre-training task.
4.3 Parameters
For pre-training, we set the generated sequence length to 20. The dimension of the node embedding is set to 128 and the size of hidden state is set to 768. On transformer layers, we use 0.1 as the dropout probability. The Adam learning rate is initiated as 0.001 with a linear decay. We use 256 sequences to form a batch and the training epoch is set to 20. The training loss is the sum of the mean masked node modeling likelihood and the mean adjacent node prediction likelihood.
In fine-tuning, most parameters remain the same as in pre-training, except the learning rate, batch size and number of epochs. We use grid search to determine the best configuration. The learning rate is chosen from . The training epoch is chosen from . The batch size is chosen from . The optimal parameters are task-specific. For the other models, we adopt the best configurations reported in the source publications.
We report on statistical significance with a paired two-tailed t-test and we mark a significant improvement of PF-HIN over GPT-GNN for with ▲.
5 Results and Analysis
We present the results of fine-tuning PF-HIN on four downstream tasks: 1. link prediction, 2. similarity search, 3. node classification, and 4. node clustering. We analyze the computational costs, conduct an ablation analysis, and study the parameter sensitivity.
5.1 Downstream tasks
We evaluate PF-HIN over four downstream tasks, i.e., link prediction, similarity search, node classification and node clustering.
5.1.1 Link prediction
This task is to predict which links would occur in the future. Unlike previous work [6] that randomly samples a certain percentage of links as the training dataset and uses the remaining links as the evaluation dataset, we adopt a sequential split of training and test data. We first train a binary logistic classifier on the graph of training data, and then use the test dataset with the same number of random negative (non-existent) links to evaluate the trained classifier. We only consider the new links in the training dataset and remove duplicate links from the evaluation. We adopt AUC and F1 score as evaluation metrics.
OAG-mini | DBLP | YELP | YAGO | |||||
---|---|---|---|---|---|---|---|---|
Model | AUC | F1 | AUC | F1 | AUC | F1 | AUC | F1 |
DeepWalk | 0.378 | 0.266 | 0.583 | 0.351 | 0.602 | 0.467 | 0.735 | 0.525 |
LINE | 0.382 | 0.257 | 0.274 | 0.357 | 0.605 | 0.463 | 0.739 | 0.531 |
node2vec | 0.392 | 0.273 | 0.584 | 0.355 | 0.609 | 0.471 | 0.742 | 0.534 |
metapath2vec | 0.412 | 0.282 | 0.604 | 0.367 | 0.618 | 0.473 | 0.744 | 0.541 |
HINE | 0.423 | 0.288 | 0.607 | 0.369 | 0.621 | 0.482 | 0.763 | 0.548 |
HIN2Vec | 0.426 | 0.291 | 0.611 | 0.376 | 0.625 | 0.493 | 0.768 | 0.578 |
GCN | 0.437 | 0.294 | 0.623 | 0.392 | 0.638 | 0.516 | 0.779 | 0.583 |
GraphSage | 0.445 | 0.293 | 0.627 | 0.395 | 0.641 | 0.525 | 0.783 | 0.592 |
GAT | 0.451 | 0.294 | 0.631 | 0.392 | 0.644 | 0.537 | 0.781 | 0.596 |
GIN | 0.456 | 0.299 | 0.636 | 0.394 | 0.647 | 0.539 | 0.785 | 0.598 |
HetGNN | 0.467 | 0.317 | 0.642 | 0.402 | 0.663 | 0.544 | 0.793 | 0.602 |
HGT | 0.473 | 0.321 | 0.648 | 0.407 | 0.672 | 0.549 | 0.799 | 0.605 |
GPT-GNN | 0.513 | 0.371 | 0.678 | 0.423 | 0.679 | 0.558 | 0.811 | 0.617 |
GCC | 0.507 | 0.352 | 0.669 | 0.417 | 0.668 | 0.552 | 0.805 | 0.609 |
MPP | 0.514 | 0.375 | 0.671 | 0.425 | 0.682 | 0.559 | 0.812 | 0.619 |
PF-HIN | 0.519▲ | 0.383▲ | 0.692▲ | 0.442▲ | 0.691▲ | 0.565▲ | 0.822▲ | 0.624▲ |
We present the link prediction results in Table II, with the highest results set in bold. Scores increase as the dataset size decreases. Traditional homogeneous models (DeepWalk, LINE, node2vec) perform worse than traditional heterogeneous metapath based models (metapath2vec, HINE, HIN2Vec); metapaths capture the network structure better than random walks. However, homogeneous GNN models (GCN, GraphSAGE, GAT, GIN) achieve even better results than traditional heterogeneous methods. Deep neural networks explore the entire network more effectively, generating better representations for link prediction. HetGNN and HGT outperform the homogeneous GNN models, since they take the node types into consideration. GPT-GNN, GCC and MPP outperform all of the above methods including their base GNN (HGT and GIN). Adopting pre-training tasks can boost the downstream task performance.
PF-HIN outperforms GCC. This is because our pre-training on (type-based) mini-sequences helps to explore the HIN, while GIN is designed for homogeneous information. PF-HIN outperforms GPT-GNN due to our choice of pre-training task, as the ANP task is more effective on predicting links between nodes than the edge generation task used in GPT-GNN. By deciding whether two nodes are adjacent, ANP can tell if a link connects them. PF-HIN also outperforms MPP. This is because our heterogeneous node sequence contains more information than the meta-paths used by MPP.
5.1.2 Similarity search
In this task, we aim to find nodes that are similar to a given node. To evaluate the similarity between two nodes, we calculate the cosine similarity based on the node representations. It is hard to rank all pairs of nodes explicitly, so we provide an estimation based on the grouping label , in which similar nodes are gathered in one group. Given a node , if we rank other nodes based on the similarity score, intuitively, nodes from the same group (similar ones) should be at the top of the ranked list while dissimilar ones should be ranked at the bottom. We define the AUC value as:
(12) |
OAG-mini | DBLP | YELP | YAGO | |
---|---|---|---|---|
Model | AUC | AUC | AUC | AUC |
DeepWalk | 0.478 | 0.511 | 0.553 | 0.656 |
LINE | 0.482 | 0.506 | 0.558 | 0.661 |
node2vec | 0.483 | 0.513 | 0.559 | 0.653 |
metapath2vec | 0.494 | 0.545 | 0.578 | 0.673 |
HINE | 0.506 | 0.554 | 0.583 | 0.683 |
HIN2Vec | 0.512 | 0.556 | 0.587 | 0.687 |
GCN | 0.509 | 0.553 | 0.581 | 0.682 |
GraphSage | 0.513 | 0.557 | 0.586 | 0.689 |
GAT | 0.510 | 0.555 | 0.584 | 0.691 |
GIN | 0.514 | 0.559 | 0.587 | 0.690 |
HetGNN | 0.527 | 0.563 | 0.592 | 0.694 |
HGT | 0.532 | 0.568 | 0.591 | 0.697 |
GPT-GNN | 0.563 | 0.596 | 0.603 | 0.708 |
GCC | 0.554 | 0.587 | 0.597 | 0.702 |
MPP | 0.559 | 0.591 | 0.593 | 0.705 |
PF-HIN | 0.574▲ | 0.612▲ | 0.613▲ | 0.719▲ |
Table III displays the results on the similarity search task. The highest scores are set in bold. The traditional heterogeneous models and homogeneous GNN models achieve comparable results; metapath based mechanisms and deep neural networks can generate expressive node embeddings for similarity search. HetGNN and HGT outperform the above methods, which shows the power of combining GNN and type features. GPT-GNN is the best performing baseline as it leverages both pre-training power and type features. PF-HIN performs the best in all cases, illustrating the effectiveness of our pre-training and fine-tuning framework on learning heterogeneous node representations for similarity search.
5.1.3 Node classification
Next we report on the results for the multi-label node classification task. The size (ratio) of the training data is set to 30% and the remaining nodes are used for testing. We adopt micro-F1 (MIC) and macro-F1 (MAC) as our evaluation metrics.
OAG-mini | DBLP | YELP | YAGO | |||||
---|---|---|---|---|---|---|---|---|
Model | MIC | MAC | MIC | MAC | MIC | MAC | MIC | MAC |
DeepWalk | 0.175 | 0.173 | 0.193 | 0.191 | 0.163 | 0.145 | 0.328 | 0.265 |
LINE | 0.177 | 0.172 | 0.184 | 0.179 | 0.274 | 0.276 | 0.366 | 0.320 |
node2vec | 0.180 | 0.178 | 0.201 | 0.198 | 0.194 | 0.151 | 0.332 | 0.280 |
metapath2vec | 0.195 | 0.194 | 0.209 | 0.207 | 0.264 | 0.269 | 0.370 | 0.332 |
HINE | 0.198 | 0.192 | 0.234 | 0.230 | 0.276 | 0.284 | 0.401 | 0.363 |
HIN2Vec | 0.204 | 0.201 | 0.246 | 0.241 | 0.291 | 0.306 | 0.428 | 0.394 |
GCN | 0.209 | 0.205 | 0.257 | 0.256 | 0.302 | 0.311 | 0.459 | 0.447 |
GraphSage | 0.211 | 0.207 | 0.267 | 0.269 | 0.305 | 0.318 | 0.464 | 0.456 |
GAT | 0.213 | 0.211 | 0.271 | 0.273 | 0.303 | 0.315 | 0.469 | 0.462 |
GIN | 0.218 | 0.223 | 0.274 | 0.277 | 0.308 | 0.321 | 0.474 | 0.466 |
HetGNN | 0.234 | 0.237 | 0.285 | 0.282 | 0.309 | 0.324 | 0.478 | 0.471 |
HGT | 0.239 | 0.241 | 0.283 | 0.278 | 0.313 | 0.327 | 0.484 | 0.483 |
GPT-GNN | 0.276 | 0.274 | 0.304 | 0.299 | 0.322 | 0.339 | 0.496 | 0.493 |
GCC | 0.266 | 0.264 | 0.295 | 0.291 | 0.315 | 0.332 | 0.489 | 0.486 |
MPP | 0.269 | 0.268 | 0.298 | 0.294 | 0.318 | 0.336 | 0.493 | 0.488 |
PF-HIN | 0.294▲ | 0.292▲ | 0.318▲ | 0.311▲ | 0.330▲ | 0.354▲ | 0.516▲ | 0.509▲ |
Table IV provides the results on the node classification task; the highest scores are set in bold. GNN based models perform relatively well, showing the benefits of using deep neural networks for exploring features of the network data for classification. PF-HIN achieves the highest scores thanks to our fine-tuning framework, which aggregates the full sequence information for node classification.
5.1.4 Node clustering
Finally, we report on the outcomes on the node clustering task. We feed the generated node embeddings of each model into a clustering model. Here, we choose a k-means algorithm to cluster the data. The size (ratio) of the training data is set to 30% and the remaining nodes are used for testing. We use normalized mutual information (NMI) and adjusted rand index (ARI) as evaluation metrics.
OAG-mini | DBLP | YELP | YAGO | |||||
---|---|---|---|---|---|---|---|---|
Model | NMI | ARI | NMI | ARI | NMI | ARI | NMI | ARI |
DeepWalk | 0.601 | 0.613 | 0.672 | 0.686 | 0.713 | 0.744 | 0.856 | 0.886 |
LINE | 0.598 | 0.609 | 0.678 | 0.693 | 0.705 | 0.739 | 0.861 | 0.894 |
node2vec | 0.604 | 0.616 | 0.673 | 0.689 | 0.719 | 0.748 | 0.867 | 0.899 |
metapath2vec | 0.628 | 0.632 | 0.711 | 0.738 | 0.748 | 0.785 | 0.896 | 0.917 |
HINE | 0.634 | 0.638 | 0.718 | 0.741 | 0.753 | 0.786 | 0.899 | 0.921 |
HIN2Vec | 0.637 | 0.640 | 0.721 | 0.744 | 0.749 | 0.789 | 0.902 | 0.923 |
GCN | 0.616 | 0.641 | 0.701 | 0.719 | 0.744 | 0.775 | 0.881 | 0.907 |
GraphSage | 0.619 | 0.643 | 0.705 | 0.722 | 0.746 | 0.778 | 0.885 | 0.911 |
GAT | 0.622 | 0.645 | 0.709 | 0.728 | 0.748 | 0.782 | 0.893 | 0.915 |
GIN | 0.625 | 0.650 | 0.711 | 0.735 | 0.752 | 0.785 | 0.898 | 0.919 |
HetGNN | 0.644 | 0.678 | 0.729 | 0.748 | 0.759 | 0.792 | 0.904 | 0.926 |
HGT | 0.648 | 0.682 | 0.734 | 0.753 | 0.763 | 0.794 | 0.909 | 0.931 |
GPT-GNN | 0.673 | 0.712 | 0.756 | 0.761 | 0.769 | 0.808 | 0.919 | 0.937 |
GCC | 0.652 | 0.707 | 0.741 | 0.756 | 0.765 | 0.799 | 0.913 | 0.933 |
MPP | 0.662 | 0.709 | 0.747 | 0.758 | 0.766 | 0.802 | 0.915 | 0.935 |
PF-HIN | 0.691▲ | 0.728▲ | 0.770 | 0.773▲ | 0.781▲ | 0.816▲ | 0.927▲ | 0.946▲ |
Table V shows the performance on the node clustering task, with the highest scores set in bold. Despite the strong ability of homogeneous GNN models to capture structural information of a network, they perform slightly worse than traditional heterogeneous models. Taking node type information into account makes a real difference when clustering nodes. GPT-GNN is the best performing baseline, combining type information and pre-training power. PF-HIN outperforms all the baselines, proving that PF-HIN is able to generate effective node embeddings for node clustering.
5.2 Computational costs
To evaluate the efficiency of our fine-tuning framework compared to other models, we conduct an analysis of the computational costs. Specifically, we analyze the running time of each model on each task, using the early stopping mechanism. Due to space limitations we only report the results on the DBLP dataset; the results for the remaining datasets are qualitatively similar. See Table VI. We use standard hardware (Intel (R) Core (TM) i7-10700K CPU + GTX-2080 GPU); the time reported is wall-clock time, averaged over 10 runs.
LP | SS | MC | NC | |
---|---|---|---|---|
Model | Time (h) | Time (h) | Time (h) | Time (h) |
DeepWalk | 1.09 | 1.15 | 0.74 | 0.53 |
LINE | 1.19 | 1.31 | 0.76 | 0.62 |
node2vec | 1.46 | 1.54 | 0.97 | 0.74 |
metapath2vec | 1.68 | 1.90 | 1.35 | 1.05 |
HINE | 1.84 | 2.11 | 1.38 | 1.08 |
HIN2Vec | 2.02 | 2.53 | 1.56 | 1.26 |
GCN | 3.45 | 4.14 | 2.66 | 2.08 |
GraphSage | 2.75 | 3.25 | 2.17 | 1.46 |
GAT | 2.55 | 2.94 | 1.73 | 1.55 |
GIN | 3.59 | 4.54 | 3.21 | 2.33 |
HGT | 3.11 | 3.80 | 2.53 | 2.19 |
HetGNN | 3.15 | 3.68 | 2.51 | 1.80 |
GPT-GNN | 1.64 | 1.92 | 1.35 | 1.03 |
GCC | 1.55 | 2.28 | 1.56 | 1.13 |
PF-HIN | 1.52 | 1.71 | 1.18 | 0.82 |
PF-HIN’s running time is longer than of the three traditional homogeneous models DeepWalk, LINE and node2vec, which are based on random walks; it is not as high as any of the other models. GNN based models like GCN, GraphSAGE, GAT, GIN and HetGNN have a higher running time than all other models, since the complexity of traditional deep neural networks is much higher than other algorithms. GPT-GNN, GCC and PF-HIN have relatively short running times; this is because pre-trained parameters help the loss function converge much faster. Among the pre-train and fine-tune models, PF-HIN is the most efficient one since the complexity of the transformer encoder we use is lower than that of HGT in GPT-GNN and that of GIN in GCC.
5.3 Ablation analysis
We analyze the effect of the pre-training tasks, the bi-directional transformer encoders, the components of the input representation, and the rank-guided heterogeneous walk sampling strategy.
5.3.1 Effect of pre-training task
To evaluate the effect of the pre-training tasks, we introduce two variants of PF-HIN, i.e., PF-HINMNM and PF-HINANP. PF-HINMNM is like PF-HIN but excludes pre-training on the masked node modeling task, PF-HINANP is like PF-HIN but excludes pre-training on the adjacent node prediction task. Due to space limitations, we only report the experimental outcomes on the DBLP dataset; the findings on the other datasets are qualitatively similar. Table VII shows the experimental results of the ablation analysis over pre-training tasks. For the link prediction task, ANP is more important than MNM since predicting if two nodes are adjacent could also tell if they are connected by a link. In similarity search, those two tasks have a comparable effect. For the node classification and node clustering tasks, MNM plays a more important role, and this is because MNM directly models node features and hence PF-HIN is better able to explore them.
LP | SS | MC | NC | ||||
---|---|---|---|---|---|---|---|
Model | AUC | F1 | AUC | MIC | MAC | NMI | ARI |
PF-HIN | 0.692 | 0.442 | 0.612 | 0.318 | 0.311 | 0.770 | 0.773 |
PF-HINANP | 0.421 | 0.221 | 0.356 | 0.217 | 0.257 | 0.548 | 0.629 |
PF-HINMNM | 0.498 | 0.283 | 0.332 | 0.172 | 0.174 | 0.422 | 0.438 |
5.3.2 Effect of the bi-directional transformer encoder
Our bi-directional transformer encoder is a variant of GNN applied to HINs, aggregating neighborhood information. For our analysis, we replace the transformer encoders with a CNN, a bi-directional LSTM, and an attention mechanism. Specifically, the model using the CNN encoder is denoted as PF-HIN(CNN), the model using bi-directional LSTM as PF-HIN(LSTM), and the model using an attention mechanism as PF-HIN(attention). Again, we report on experiments on the DBLP dataset only. Table VIII presents the experimental results of different encoders. The CNN, LSTM and attention mechanism based models achieve comparable results on the four tasks. PF-HIN consistently outperforms all three models, which shows the importance of our bi-directional transformer encoder for mining the information contained in a HIN.
LP | SS | MC | NC | ||||
---|---|---|---|---|---|---|---|
Model | AUC | F1 | AUC | MIC | MAC | NMI | ARI |
PF-HIN | 0.692 | 0.442 | 0.612 | 0.318 | 0.311 | 0.770 | 0.773 |
PF-HIN(CNN) | 0.643 | 0.418 | 0.578 | 0.296 | 0.287 | 0.743 | 0.747 |
PF-HIN(LSTM) | 0.657 | 0.421 | 0.567 | 0.293 | 0.279 | 0.754 | 0.753 |
PF-HIN(attention) | 0.629 | 0.411 | 0.572 | 0.299 | 0.291 | 0.748 | 0.757 |
5.3.3 Effect of pre-training strategy
In PF-HIN, the MNM task is pre-trained on (type-based) mini-sequences, while the ANP task is pre-trained on two full sequences generated via two starting nodes. Here we analyze the effect of this strategy. We consider three variants of PF-HIN. The first is to pre-train the two tasks on two full sequences without considering the heterogeneous features of the network, denoted as PF-HIN(full-full). In the second, we try to assign the ANP task with a sense of heterogeneous features. As explained in Section 3.5, it is too time-consuming to pre-train all mini-sequence pairs for the ANP task, so in the second model we choose the two longest mini-sequences to train the ANP task, while the MNM task is trained on full sequences, denoted as PF-HIN(full-mini). The third is that the ANP and MNM tasks are both trained on (type-based) mini-sequences, denoted as PF-HIN(mini-mini). Table IX shows the results of the ablation analysis. PF-HIN(full-full) outperforms PF-HIN(full-mini), which illustrates that despite taking heterogeneous features into consideration, only choosing two mini-sequences for the ANP task may harm the performance as information is missed. However, PF-HIN(mini-mini) outperforms PF-HIN(full-full), showing that considering heterogeneous features on the MNM task may help boost the model performance. The strategy selected for PF-HIN achieves the best results.
LP | SS | MC | NC | ||||
---|---|---|---|---|---|---|---|
Model | AUC | F1 | AUC | MIC | MAC | NMI | ARI |
PF-HIN | 0.692 | 0.442 | 0.612 | 0.318 | 0.311 | 0.770 | 0.773 |
PF-HIN(full-full) | 0.671 | 0.419 | 0.589 | 0.295 | 0.294 | 0.747 | 0.757 |
PF-HIN(full-mini) | 0.659 | 0.407 | 0.577 | 0.287 | 0.283 | 0.733 | 0.751 |
PF-HIN(mini-mini) | 0.679 | 0.429 | 0.596 | 0.307 | 0.301 | 0.759 | 0.766 |
LP | SS | MC | NC | ||||
---|---|---|---|---|---|---|---|
Model | AUC | F1 | AUC | MIC | MAC | NMI | ARI |
PF-HIN | 0.692 | 0.442 | 0.612 | 0.318 | 0.311 | 0.770 | 0.773 |
PF-HIN(BFS) | 0.659 | 0.417 | 0.577 | 0.279 | 0.292 | 0.743 | 0.758 |
PF-HIN(DFS) | 0.643 | 0.406 | 0.565 | 0.269 | 0.283 | 0.741 | 0.752 |
PF-HIN(random) | 0.668 | 0.423 | 0.589 | 0.292 | 0.297 | 0.657 | 0.764 |
5.3.4 Effect of rank-guided heterogeneous walk sampling
In this paper, we have adopted a rank-guided heterogeneous walk sampling strategy to sample nodes to form input sequences. Here we introduce three variants. The first is to only use a breadth first search (BFS) sampling strategy, denoted as PF-HIN(BFS); the second is to only use a depth first search (DFS) sampling strategy, denoted as PF-HIN(DFS); and the last is to randomly choose neighboring nodes to form the node sequence, denoted as PF-HIN(random).
Table X shows the experimental results. PF-HIN(BFS) outperforms PF-HIN(DFS) and PF-HIN(random); aggregating a node’s closest neighborhood’s information is more informative than choosing far-away nodes or randomly chosen neighboring nodes. PF-HIN outperforms PF-HIN(BFS); choosing nodes with a higher importance leads to better performing feature representations of a HIN.
5.3.5 Effect of centrality metric
In this paper, we have three centrality metrics to measure the importance of a node. Here we introduce three variants. The first is to remove the betweenness centrality denoted as PF-HINbetweenness; the second is to remove the eigencentrality denoted as PF-HINeigen; the third is to remove the closeness centrality denoted as PF-HINcloseness. Each variant assigns equal weights for the left two metrics.
LP | SS | MC | NC | ||||
---|---|---|---|---|---|---|---|
Model | AUC | F1 | AUC | MIC | MAC | NMI | ARI |
PF-HIN | 0.692 | 0.442 | 0.612 | 0.318 | 0.311 | 0.770 | 0.773 |
PF-HINbetweenness | 0.674 | 0.423 | 0.582 | 0.304 | 0.298 | 0.757 | 0.759 |
PF-HINeigen | 0.668 | 0.416 | 0.586 | 0.297 | 0.292 | 0.753 | 0.751 |
PF-HINcloseness | 0.671 | 0.420 | 0.577 | 0.301 | 0.302 | 0.761 | 0.754 |
Table XI presents the experimental results. Removing a metric will influence the experimental results which illustrates that each metric is necessary for ranking the nodes. In addition, each metric has different influence on different tasks, so it is reasonable for us to adopt the learnable weights on them.
5.4 Parameter sensitivity
We conduct a sensitivity analysis of the hyper-parameters of PF-HIN. We choose two parameters for analysis: the maximum length of the input sequence and the dimension of the node embedding. For each downstream task, we only choose one metric for evaluation: AUC for link prediction, AUC for similarity search, micro-F1 value for node classification, and NMI value for node clustering. Figures 2 and 3 show the results of our parameter analysis.








Figure 2 shows the results for the maximum length of the input sequence. The performance improves rapidly when the length increases from 0 to 20. A short node sequence is not able to fully express neighborhood information. When the length reaches 20 or longer, the performance stabilizes, and longer sequence lengths may even hurt the performance. Given a node, its neighboring information can be well represented by its direct neighborhood, however, including far-away nodes may introduce noise. According to this analysis, we choose the length of the input sequence to be 20 so as to balance effectiveness and efficiency.
As to the dimension of node embeddings, Figure 3 shows that the performance improves as the dimension increases, for all tasks and datasets. The higher dimensions are able to capture more features. PF-HIN is not very sensitive to the dimension we choose, especially once it is at least 128. The performance gap is not very large between dimension 128 and 256. Thus, we choose 128 as our setting for the dimension of node embeddings for efficiency considerations.
6 Conclusions
We have considered the problem of network representation learning for heterogeneous information networks (HINs). We propose a novel model, PF-HIN, to mine the information captured by a HIN. PF-HIN is a self-supervised pre-training and fine-tuning framework. In the pre-training stage, we first use rank-guided heterogeneous walks to generate input sequences and group them into (type-based) mini-sequences. The pre-training tasks we utilize are masked node modeling (MNM) and adjacent node prediction (ANP). Then we leverage bi-directional transformer layers to pre-train the model. We adopt factorized embedding parameterization and cross-layer parameter sharing strategies to reduce the number of parameters. We fine-tune PF-HIN on four tasks: link prediction, similarity search, node classification, and node clustering. PF-HIN significantly and consistently outperforms state-of-the-art models on the above tasks on four real-life datasets.
In future work, we plan to conduct further graph learning tasks in the context of a diverse range of information retrieval tasks, including but not limited to academic search, financial search, product search, and social media search. It is also of interest to see how to model a dynamic HIN that is constantly evolving, using a pre-training and fine-tuning framework.
Acknowledgment
This work was partially supported by NSFC under grants No. 61872446, NSF of Hunan Province under grant No. 2019JJ20024, and The Science and Technology Innovation Program of Hunan Province under grant No. 2020RC4046, and by the Hybrid Intelligence Center, a 10-year program funded by the Dutch Ministry of Education, Culture and Science through the Netherlands Organisation for Scientific Research, https://hybrid-intelligence-centre.nl. All content represents the opinion of the authors, which is not necessarily shared or endorsed by their respective employers and/or sponsors.
References
- Reinanda et al. [2020] R. Reinanda, E. Meij, and M. de Rijke, “Knowledge graphs: An information retrieval perspective,” Foundations and Trends in Information Retrieval, vol. 14, no. 4, pp. 289–444, October 2020.
- Sun and Han [2012] Y. Sun and J. Han, Mining Heterogeneous Information Networks: Principles and Methodologies, ser. Synthesis Lectures on Data Mining and Knowledge Discovery. Morgan & Claypool Publishers, 2012.
- Cai et al. [2018] H. Cai, V. W. Zheng, and K. C. Chang, “A comprehensive survey of graph embedding: Problems, techniques, and applications,” IEEE Trans. Knowl. Data Eng., vol. 30, no. 9, pp. 1616–1637, 2018.
- Perozzi et al. [2014] B. Perozzi, R. Al-Rfou, and S. Skiena, “Deepwalk: online learning of social representations,” in KDD, 2014, pp. 701–710.
- Tang et al. [2015a] J. Tang, M. Qu, M. Wang, M. Zhang, J. Yan, and Q. Mei, “LINE: large-scale information network embedding,” in WWW, 2015, pp. 1067–1077.
- Grover and Leskovec [2016] A. Grover and J. Leskovec, “node2vec: Scalable feature learning for networks,” in KDD, 2016, pp. 855–864.
- Dong et al. [2017] Y. Dong, N. V. Chawla, and A. Swami, “metapath2vec: Scalable representation learning for heterogeneous networks,” in KDD, 2017, pp. 135–144.
- Huang and Mamoulis [2017] Z. Huang and N. Mamoulis, “Heterogeneous information network embedding for meta path based proximity,” CoRR, vol. abs/1701.05291, 2017.
- Fu et al. [2017] T. Fu, W. Lee, and Z. Lei, “Hin2vec: Explore meta-paths in heterogeneous information networks for representation learning,” in CIKM, 2017, pp. 1797–1806.
- Kipf and Welling [2017] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” in ICLR, 2017.
- Velickovic et al. [2018] P. Velickovic, G. Cucurull, A. Casanova, A. Romero, P. Liò, and Y. Bengio, “Graph attention networks,” in ICLR, 2018.
- Zhang et al. [2019] C. Zhang, D. Song, C. Huang, A. Swami, and N. V. Chawla, “Heterogeneous graph neural network,” in KDD, 2019, pp. 793–803.
- Dai and Le [2015] A. M. Dai and Q. V. Le, “Semi-supervised sequence learning,” in NeurIPS, 2015, pp. 3079–3087.
- Peters et al. [2018] M. E. Peters, M. Neumann, M. Iyyer, M. Gardner, C. Clark, K. Lee, and L. Zettlemoyer, “Deep contextualized word representations,” in NAACL-HLT, 2018, pp. 2227–2237.
- Howard and Ruder [2018] J. Howard and S. Ruder, “Universal language model fine-tuning for text classification,” in ACL, 2018, pp. 328–339.
- Qiu et al. [2020] J. Qiu, Q. Chen, Y. Dong, J. Zhang, H. Yang, M. Ding, K. Wang, and J. Tang, “GCC: graph contrastive coding for graph neural network pre-training,” in KDD, 2020, pp. 1150–1160.
- Hu et al. [2020a] Z. Hu, Y. Dong, K. Wang, K. Chang, and Y. Sun, “GPT-GNN: generative pre-training of graph neural networks,” in KDD, 2020, pp. 1857–1867.
- Belkin and Niyogi [2001] M. Belkin and P. Niyogi, “Laplacian eigenmaps and spectral techniques for embedding and clustering,” in NeurIPS, 2001, pp. 585–591.
- Cox and Cox [2008] M. A. A. Cox and T. F. Cox, Multidimensional Scaling. Springer Berlin Heidelberg, 2008.
- Roweis and Saul [2000] S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by locally linear embedding,” Science, vol. 290, no. 5500, pp. 2323–2326, 2000.
- Wang et al. [2016] D. Wang, P. Cui, and W. Zhu, “Structural deep network embedding,” in KDD, 2016, pp. 1225–1234.
- Ahmed et al. [2013] A. Ahmed, N. Shervashidze, S. M. Narayanamurthy, V. Josifovski, and A. J. Smola, “Distributed large-scale natural graph factorization,” in WWW, 2013, pp. 37–48.
- Tang et al. [2015b] J. Tang, M. Qu, and Q. Mei, “PTE: predictive text embedding through large-scale heterogeneous text networks,” in KDD, 2015, pp. 1165–1174.
- Bruna et al. [2014] J. Bruna, W. Zaremba, A. Szlam, and Y. LeCun, “Spectral networks and locally connected networks on graphs,” in ICLR, 2014.
- Defferrard et al. [2016] M. Defferrard, X. Bresson, and P. Vandergheynst, “Convolutional neural networks on graphs with fast localized spectral filtering,” in NeurIPS, 2016, pp. 3837–3845.
- Henaff et al. [2015] M. Henaff, J. Bruna, and Y. LeCun, “Deep convolutional networks on graph-structured data,” CoRR, vol. abs/1506.05163, 2015.
- Li et al. [2018] R. Li, S. Wang, F. Zhu, and J. Huang, “Adaptive graph convolutional neural networks,” in AAAI, 2018, pp. 3546–3553.
- Levie et al. [2019] R. Levie, F. Monti, X. Bresson, and M. M. Bronstein, “Cayleynets: Graph convolutional neural networks with complex rational spectral filters,” IEEE Trans. Signal Process., vol. 67, no. 1, pp. 97–109, 2019.
- Monti et al. [2017] F. Monti, D. Boscaini, J. Masci, E. Rodolà, J. Svoboda, and M. M. Bronstein, “Geometric deep learning on graphs and manifolds using mixture model cnns,” in CVPR, 2017, pp. 5425–5434.
- Niepert et al. [2016] M. Niepert, M. Ahmed, and K. Kutzkov, “Learning convolutional neural networks for graphs,” in ICML, 2016, pp. 2014–2023.
- Gao et al. [2018] H. Gao, Z. Wang, and S. Ji, “Large-scale learnable graph convolutional networks,” in KDD, 2018, pp. 1416–1424.
- Hamilton et al. [2017] W. L. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” in NeurIPS, 2017, pp. 1024–1034.
- Zhang et al. [2018] J. Zhang, X. Shi, J. Xie, H. Ma, I. King, and D. Yeung, “Gaan: Gated attention networks for learning on large and spatiotemporal graphs,” in UAI, 2018, pp. 339–349.
- Lee et al. [2018] J. B. Lee, R. A. Rossi, and X. Kong, “Graph classification using structural attention,” in KDD, 2018, pp. 1666–1674.
- Xu et al. [2019] K. Xu, W. Hu, J. Leskovec, and S. Jegelka, “How powerful are graph neural networks?” in ICLR, 2019.
- Hu et al. [2020b] Z. Hu, Y. Dong, K. Wang, and Y. Sun, “Heterogeneous graph transformer,” in WWW, 2020, pp. 2704–2710.
- Sun et al. [2020] F. Sun, J. Hoffmann, V. Verma, and J. Tang, “Infograph: Unsupervised and semi-supervised graph-level representation learning via mutual information maximization,” in ICLR, 2020.
- Hu et al. [2020c] W. Hu, B. Liu, J. Gomes, M. Zitnik, P. Liang, V. S. Pande, and J. Leskovec, “Strategies for pre-training graph neural networks,” in ICLR, 2020.
- Hu et al. [2019] Z. Hu, C. Fan, T. Chen, K. Chang, and Y. Sun, “Pre-training graph neural networks for generic structural feature extraction,” CoRR, vol. abs/1905.13728, 2019.
- Hwang et al. [2020] D. Hwang, J. Park, S. Kwon, K. Kim, J. Ha, and H. J. Kim, “Self-supervised auxiliary learning with meta-paths for heterogeneous graphs,” in NeurIPS, 2020.
- Bonacich and Phillip [1987] Bonacich and Phillip, “Power and centrality: A family of measures,” American Journal of Sociology, vol. 92, no. 5, pp. 1170–1182, 1987.
- Devlin et al. [2019] J. Devlin, M. Chang, K. Lee, and K. Toutanova, “BERT: pre-training of deep bidirectional transformers for language understanding,” in NAACL-HLT, 2019, pp. 4171–4186.
- Lan et al. [2020] Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut, “ALBERT: A lite BERT for self-supervised learning of language representations,” in ICLR, 2020.
![]() |
Yang Fang received the Master degree from National University of Defense Technology (NUDT), China, in 2018. He is currently working toward the Ph.D. degree at NUDT, and his research mainly focuses on graph/network representation learning. |
![]() |
Xiang Zhao received the PhD degree from The University of New South Wales, Australia, in 2014. He is currently an associate professor at National University of Defense Technology, China, and he is the head of the knowledge systems engineering group. His research interests include graph data management and mining, with a special focus on knowledge graphs. |
![]() |
Yifan Chen received the PhD degree from the University of Amsterdam (UvA), the Netherlands, in 2019, and from the National University of Defense Technology (NUDT), China, in 2020. His research interests include graph data mining and recommender systems. |
![]() |
WEIDONG XIAO was born in Harbin, China,in 1968. He received the Ph.D. degree from the National University of Defense Technology(NUDT), China. He is currently a Full Professor in Science and Technology on Information Systems Engineering Laboratory, National University of Defense, China. His research interests include big data analytics and social computing. |
![]() |
Maarten de Rijke is a Distinguished University Professor in Artificial Intelligence and Information Retrieval at the University of Amsterdam. He is the scientific director of the national Innovation Center for Artificial Intelligence and an elected member of the Royal Netherlands Academy of Arts and Sciences. Together with a team of PhD students and postdocs he carries out research on intelligent technology to connect people to information. |