Local-Global History-aware Contrastive Learning for Temporal Knowledge Graph Reasoning
Abstract
Temporal knowledge graphs (TKGs) have been identified as a promising approach to represent the dynamics of facts along the timeline. The extrapolation of TKG is to predict unknowable facts happening in the future, holding significant practical value across diverse fields. Most extrapolation studies in TKGs focus on modeling global historical fact repeating and cyclic patterns, as well as local historical adjacent fact evolution patterns, showing promising performance in predicting future unknown facts. Yet, existing methods still face two major challenges: (1) They usually neglect the importance of historical information in KG snapshots related to the queries when encoding the local and global historical information; (2) They exhibit weak anti-noise capabilities, which hinders their performance when the inputs are contaminated with noise. To this end, we propose a novel Local-global history-aware Contrastive Learning model (LogCL) for TKG reasoning, which adopts contrastive learning to better guide the fusion of local and global historical information and enhance the ability to resist interference. Specifically, for the first challenge, LogCL proposes an entity-aware attention mechanism applied to the local and global historical facts encoder, which captures the key historical information related to queries. For the latter issue, LogCL designs a local-global query contrast module, effectively improving the robustness of the model. The experimental results on four benchmark datasets demonstrate that LogCL delivers better and more robust performance than the state-of-the-art baselines.
Index Terms:
Temporal knowledge graph, Graph convolutional network, Contrastive learning
I Introduction
Temporal knowledge graphs (TKGs), which represent dynamic facts as quadruples in the form of (subject, relation, object, time), are actually sequences of KG snapshots with respective timestamps. Reasoning on TKGs aims to predict the unknown facts by modeling the historical KGs snapshots, which involves two reasoning settings: interpolation and extrapolation. The interpolation setting focuses on completing the missing facts in history while the extrapolation setting aims to predict the facts happening in the future. For TKGs, the extrapolation task is much more challenging and has great practical significance for various downstream applications, such as medical aided diagnosis system [1, 2], traffic flow prediction [3] and stock prediction [4]. Therefore, this paper focuses on extrapolation tasks that explore future unknown facts forecasting.
Accurate predictions of future facts require a comprehensive understanding of the patterns of development of historical facts. According to human cognition of the development of historical facts[5, 6, 7], predicting future facts involves the exploration of two historical patterns: the repetition or cycling of historical facts and the evolution of recent adjacent facts. Many efforts for TKG extrapolation have been made toward learning the facts repeating and cyclic patterns by global historical information and modeling the adjacent fact evolution patterns by local historical information.
The studies of the historical fact repeating or cyclic pattern such as CyGNet [8], aims to extract global repetition historical information for different queries in a heuristic way. This method leads to the narrow results that the predictions often lean towards the most frequently occurring facts. The research on historical adjacent fact evolution pattern focuses on local historical facts temporal dependency modeling such as RE-GCN[9] and TANGO-Tucker[10], but lacks the capture of global historical information. Recently, some methods have tried to consider capturing both global and local historical patterns, such as TiRGN[11] and HIPNet[12], which integrate global and local final prediction results to restrict the prediction range and achieve promising performance. Yet, the performance of these methods is limited due to the following two challenges:


The importance of historical information related to the query is neglected during the process of encoding historical facts. Most extrapolation methods on TKGs model the evolution of entities and relations by the order of time and argue that the KG snapshot facts closer to the query time are more important to predict the query. However, the entities in the query may not appear in all historical KG snapshots, which results in each snapshot taking on a different role in predicting the query. As shown in Fig. 1, for the query , the entity does not appear and is not directly related to other facts in KG snapshot, which provides little help for the prediction of the query. The facts containing the entity in KG snapshot appear earlier but are more helpful for predicting the query than KG snapshot. With the example in Fig. 1, we find that the historical facts in each KG snapshot are not always crucial for prediction. Existing approaches lack essential patterns for capturing the important historical facts related to queries in the KG snapshot, impairing the ability to accurately predict future facts. So finding a way to filter the irrelevant KG snapshots based on the queries is much critical to improve the performance of TKG reasoning.
The weak anti-noise ability exists in existing methods that cannot effectively guarantee accurate predictions when the inputs adjoint noise. Mainstream TKG reasoning methods focus on how to model historical information and improve the accuracy of predictions. Nevertheless, the robustness of the model as a key important property to ensure the correct prediction results is rarely considered for TKG reasoning. In practice, during the process of training, the input data may encounter different noise interference, which leads to a significant degradation in prediction performance and even generates wrong prediction results. To further elaborate on the challenge mentioned above, we conduct experiments to simulate scenarios where the inputs are perturbed by adding Gaussian noise on both the classical RE-GCN model and TiRGN model. The results shown in Fig. 2, it is observed that both REGCN and TiRGN models suffer varying degrees of performance degradation when Gaussian noise is added, especially the MRR of REGCN model on ICEWS14 and ICEWS18 datasets are reduced by 63.8% and 66.4%, respectively. This indicates that the existing TKG method is weak in resisting the interference of noise. Therefore, how to effectively improve the robustness of the model to ensure correct prediction results is of great significance.
To address the above challenges, we propose a novel Local-global history-aware Contrastive Learning (LogCL) based on encoder-decoder structure for TKG reasoning, which leverages contrastive learning to guide the fusion of local and global historical information and effectively enhance the robustness of the model. Specifically, during the model encoding phase, we propose an entity-aware attention to flexibly learn query-related local and global historical information, thus forming a local entity-aware attention recurrent encoder and a global entity-aware attention encoder (First challenge). Inspired by the unsupervised contrastive learning [13], we design a local-global query contrast module to alleviate the interference brought by external noise to the model, which greatly improves the anti-noise ability of the model (Second challenge). As shown in Fig. 2, LogCL performs better than REGCN and TiRGN in the face of noise interference. In the final decoding part, we achieve entity prediction by fusing local and global historical information.
In general, this work presents the following contributions:
-
•
We propose a novel TKG extrapolation model that uses contrastive learning to better guide the fusion of global and local historical information, and enhance the robustness of LogCL. To the best of our knowledge, LogCL is the first model to leverage local-global history-aware contrastive learning to improve model robustness in TKG reasoning.
-
•
We propose entity-aware attention applied to encode local and global historical information in an elegant way, which captures the importance of historical facts related to queries in KG snapshots.
-
•
Extensive experiments on four public datasets demonstrate that our proposed method shows better and more robust performance than the state-of-the-art baselines.
The rest of this paper is organized as follows. We first review the related work in Section II. Then we describe the notation and problem statement, and introduce our proposed LogCL model in Section III. Next, we report and analyze the experimental results in Section IV. Finally, Section V concludes our paper.
II Related Work
In this section, a review of the related work is presented, which includes KG reasoning methods and contrastive learning.
II-A Reasoning on KGs
Reasoning on KGs has been a hot research topic that infers unknown facts based on known facts. Existing reasoning methods over KGs can be divided into two categories: static KG (SKG) reasoning and temporal KG reasoning.
II-A1 Static KG reasoning
Reasoning on SKG focus on static knowledge modeling, which can be roughly divided into four categories [14, 15, 16]: translation-based methods, matrix factorization methods, convolutional neural network (CNN)-based methods and graph neural network (GNN)-based methods. Translation-based methods such as TransE [17] and TransH [18], which learn embedded representations of entities and relations in a low-dimensional vector space. Matrix factorization methods including DistMult [19] and CompIEx [20], which adopt tensor/matrix factorization to obtain the latent representation of entities and relations. CNN-based methods such as ConvE [21] and Conv-TransE [22], which utilize CNNs to capture the interactions between entities and relations. GNN-based methods, such as R-GCN [23], CompGCN [24] and KBGAT [25], which utilize GNNs to capture the structured semantic features of graphs. SKG reasoning methods have high efficiency in processing static knowledge. However, SKG reasoning methods do not take temporal information into consideration, which makes it difficult to model dynamic facts.
II-A2 Temporal KG reasoning
Reasoning over TKGs involves two settings: interpolation and extrapolation. The interpolation of TKGs infers the missing facts at historical timestamps. Most research works are based on the extension of SKG reasoning methods. For example, TTransE [26] extends the scoring function based on the idea of TransE[17] and adds temporal constraints between facts that share entities. HyTE[27] considers timestamps as time-related hyperplanes and generates time-aware representations by projecting the entities and relations into time-related hyperplanes. TNTComplEx[28] extends the idea of tensor factorization and proposes a 4th-order tensor factorization to obtain the time-aware representations of entities. Since these methods do not involve the evolution of facts over time, they are not suitable for modeling the entity representations in future unseen timestamps.
TKG extrapolation aims to predict facts at future timestamps. According to the law of human development of history cognition mainly involves two historical patterns: the repetition or cycling of historical facts and the evolution of recent adjacent facts. The former [8, 29] explores the repetition of facts related to queries from a global perspective. For example, CyGNet [8] and CENET [29] adopt a copy-generation mechanism to obtain the global repetition of one-hop facts, facilitating future entity reasoning. The latter [30, 10, 31, 32, 9, 33] aims to study the local historical facts temporal dependency. For example, RE-GCN [9] models the evolution of entities and relations at each timestamp by using the local historical dependency. CEN [33] utilizes online learning strategies to deal with time-variability issues. GHT [34] designs two Transformer modules [35] to capture instantaneous structural information and temporal evolution information, respectively, and presents a novel relational continuous-time encoding function to facilitate feature evolution with Hawkes processes. TECHS [36] encodes the topology and temporal dynamics by using heterogeneous attention and leverages both propositional and first-order reasoning to make predictions. Recent methods such as TiGRN [11] and HIPNet [12], attempt to integrate local and global historical information to obtain more accurate results. Nevertheless, these approaches fail to model the importance of historical facts related to queries in KG snapshots. This makes it difficult to distinguish the significance of each KG snapshot for predicting the query. To address this dilemma, we propose an entity-aware attention mechanism to flexibly model the KG snapshots that are relevant and irrelevant to the queries.
II-B Contrastive Learning
Contrastive Learning as a self-supervised learning paradigm, due to its impressive performance in distinguishing instances of different categories, has gained significant attention in recent years. The basic idea of contrastive learning [37, 13] is to learn the similarity between different views of the same instances through training, while maximizing the dissimilarity between different views of instances. In self-supervised contrastive learning, the augmented instances are derived by randomly sampling a mini-batch of instances. The origin instances and the augmented instances are used to optimize the following loss function given a pair of positive instances . The contrastive loss is expressed as:
(1) |
where is the sum of the number of the original and augmented instances, is the projection embedding of the instance , is a temperature parameter helping the model learn from hard negatives, and is dot product that is used to compute the similarity of instances between different views. Recently, contrastive learning has been applied to static KG reasoning to alleviate the long-tail data problem but is rarely explored in TKG reasoning. The recent CENET [29] is a single-view historical contrastive learning method designed to enhance the representation of entities with less historical interactions. Different from CENET, we propose a local-global query contrast module to mitigate noise interference and enhance model robustness.

III Our Approach
In this section, we will first introduce basic notations and definitions used in this paper and present the problem statement for the extrapolation of TKG, and then describe the framework and training optimizations of LogCL in detail.
Notations | Descriptions |
A TKG | |
A KG snapshot composed of facts at | |
Global historical query subgraph | |
Set of valid facts at | |
Set of entities, relations and times in | |
The number of entities, relations and times | |
Embedding dimensionality | |
Length of the local KG snapshot sequence | |
Randomly initialized embedding matrices of entities | |
Randomly initialized embedding matrices of realtions | |
Embedding matrices of entities at | |
Embedding matrices of relations at | |
Global embedding matrices of entities |
III-A Notations and Definitions
A TKG is formalized as a sequence of KG snapshots, i.e., . Each KG snapshot actually is a directed multi-relational graph that is a set of valid facts at . A fact (or an event) is denoted as a quadruple () where the subject entity and the object entity is connected by a relation at time . The primary notations and their meanings used in this paper are described in Table I.
TKG extrapolation task involves forecasting a missing object entity (or subject entity ) given a query (or ) according to previous historical KG snapshots . Without loss of generality, the inverse relation quadruples are added to the TKG dataset. So the TKG extrapolation task can be reduced to object entities predictions.
III-B Architecture Overview
As shown in Fig. 3, the overall framework of LogCL is composed of three components: global entity-aware attention encoder, local entity-aware recurrent encoder and local-global query contrast module. Local entity-aware attention recurrent encoder adopts an entity-aware attention mechanism to capture the necessary information related to queries for prediction. If facts in KG snapshots at the latest timestamps contain semantic information relevant to the queries, these facts will help to improve the probability of entity prediction. Global entity-aware attention encoder takes into account the global historical facts related to queries and avoids overlooking important historical facts that do not appear in the recent local KG snapshots. Local-global query contrast module is introduced to guide a better integration of global and local historical information and to enhance the robustness of LogCL. Finally, LogCL combines local and global historical embedding representations for entity prediction.

III-C Local Entity-Aware Attention Recurrent Encoder
As discussed in the introduction, each KG snapshot at the latest timestamp (i.e. ) doesn’t provide equally contributions for predicting the query . The importance of KG snapshot historical information relevant to the query needs to be modeled to further optimize the evolution of entity and relation in adjacent time. To this end, we propose a local entity-aware attention recurrent encoder that efficiently captures historical information in KG snapshots relevant to the query. The local entity-aware attention recurrent encoder involves a KG snapshot aggregation pipeline and a KG snapshot sequences evolution pipeline. Specifically, the aggregation of the KG snapshot is to learn the spatial semantic structure of concurrent facts by the Graph Convolution Network (GCN) from the spatial view, while the evolution of KG snapshot sequences is to capture sequential dependencies of entities and relations through the recurrent mechanism from the temporal view. Note that, the inputs to entities and relations on the first timestamp are randomly initialized.
III-C1 KG Snapshot Aggregation
For KG snapshots on each timestamp, we update the representation of entities by capturing the spatial structural semantics information among concurrent facts. Considering that some facts in KG snapshots occur cyclically, such as periodic meetings, we first follow [38] to encode the time numerical information to obtain the dynamic entities embedding. Formally, the dynamic entities embedding as follows:
|
(2) |
|
(3) |
where is the time interval that is modeled by rescaling a learnable time unit with a time bias . is a periodic activation function, is the vector concatenation operation, is a linear transformation matrix, and is the dynamic embedding of entity at timestamp . Then, we utilize an R-GCN [23] to capture the structure dependencies among concurrent facts. Note that, the R-GCN can be replaced by other relation-aware GCNs, such as CompGCN[24] and KBGAT[25]. The entity-aggregating R-GCN aggregator is defined as:
|
(4) |
where denote the embedding of entities and relations in the -th layers of the R-GCN in each KG snapshot at -th timestamp, separately. For simplicity, the final layer of the output of R-GCN is denoted as . is a normalization constant that equals the in-degree of entity, and are the parameters for aggregating features and self-loop in the -th layer, and is the RReLU activation function.
III-C2 KG Snapshot Sequence Evolution
After obtaining the structural semantic embedding of entities in each KG snapshot, it needs to further model the sequential dependencies of entities and relations in KG snapshot sequences at the latest timestamps. We progressively update the representations of entities by a gated recurrent unit (GRU) [39] which is a flexible and efficient recurrent mechanism modeling short sequences. The entity-oriented GRU can be represented as:
|
(5) |
where is the entity embedding matrix after KG snapshot aggeration at . As for relation, we follow [9] that considers -related entities and the corresponding relation embedding to obtain at KG snapshot. Thus, a time gate unit is adopted to update the relation embedding at . Formally,
|
(6) |
|
(7) |
|
(8) |
where is the mean pooling operation, is the embedding of entities connected to at , consists of at , is the finally updated through a time gate , is the learnable weight matrice of the time gate, and is the sigmoid activate function.
After the above operations, we obtain the entity evolution representations and the relation evolution representations at the latest timestamps. However, these evolutionary representations do not take into account the contribution of historical facts related to queries in the KG snapshots for predicting the query. Thus, we propose an entity-aware attention mechanism to distinguish the importance of different KG snapshots for queries. We first exploit the mean pooling applied on all relation embedding associated with the entity in at timestamps , and then compute the final local entity representation via the entity-aware attention mechanism. The operation is as follows,
(9) |
|
(10) |
|
(11) |
where is the attention score of KG snapshots, and are the learnable weight matrices, is the softmax acitvate function, and is the final local presentation at timestamp after local entity-aware attention recurrent encoder. In this way, LogCL can effectively model KG snapshot information related to queries at the latest timestamps, providing more accurate prediction results.
III-D Global Entity-Aware Attention Encoder
The global entity-aware attention encoder aims to capture the longer historical facts that are not considered in the local KG snapshot sequence. For each query , unlike CyGNet [8] and TiRGN[11] that only use one-hop target object entities associated with queries to learn the repetition of facts in further history, we consider candidate multi-hop historical facts that provide more semantic information of historical facts to help entities prediction. For example, meetings that are held periodically are preceded by different hosting processes, and these different hosting processes can be of great help to future meetings. Therefore, we construct the historical query subgraph by sampling the historical facts relevant to queries to obtain rich semantic information.
For a TKG , we first sample the one-hop historical facts containing the query subject entity in the given query. Next, we extract the one-hop target object entity associated with the query entity-relation pair, and then proceed to sample the one-hop facts containing the one-hop target object entity. Finally, we integrate the two collected historical fact sets to obtain the most relevant historical subgraph to the query, . It is worth noting that, the historical query subgraph is a static KG that does not involve temporal information and can change along the query time.
After obtaining the historical query subgraph, we update the global entity representation by modeling the structural semantic representation of the historical query subgraph. To encode the historical query subgraph that is formed by sampling two-hop neighbors, we adopt another R-GCN to perform message aggregation. Since the historical query subgraph does not consider temporal information, we directly use the randomly initialized embedding of entities and relations as input to R-GCN. The specific formula is expressed as follows:
|
(12) |
where denotes the output embedding of entities in the -th layers of the R-GCN in the historical query subgraph. For simplicity, the final layer of the output of R-GCN is denoted as . For global historical information, we also use the entity-aware attention mechanism to learn the historical facts related to queries. The specific formula is expressed as follows:
|
(13) |
|
(14) |
where the is the attention score, is the learnable weight matrice, and denotes the final global representation of entities through global entity-aware attention encoder.
III-E Local and Global Historical Contrastive Learning
Clearly, the local and global encoders defined above well capture the local and global dependency for queries. However, during the actual training process, the input data is disturbed by extraneous noise, leading to a sharp decline in reasoning performance. Existing methods cannot address this challenge well. Inspired by the unsupervised contrastive learning [13, 29], we propose a local-global query contrast module that identifies highly correlated local and global features of entities and relations at the query level, enabling noise filtering. Specifically, for each query at timestamp , The purpose of the local-global query contrast module is to learn the local and global contrastive representations of queries by minimizing the supervised contrastive loss, which makes the local and global representations of the same query to be as close as possible in the semantic space, while different queries are separated. Formally, Formally, the query embedding representations of local and global are obtained by using the concatenation operation:
|
(15) |
|
(16) |
where is the representation of the local query at timestamp , is the representation of the global query, and is used to normalize and project the embedding of queries onto the unit sphere for further contrastive training. It is worth noting that since the historical query subgraph does not consider temporal information, the origin relation embeddings are used to encode the global embedding representation of queries.
In our work, the global and local query representations can be considered as the two-view representations derived from encoding historical facts in TKGs. Thus, if taking the query representation generated by the local encoder as the anchor, the global encoder can be viewed as an augmented encoding process. The local and global representations of the same query are used as positive pairs at timestamp , i.e.,(, ), the local and global of representation of different queries are used as negative pairs i.e.,(, ). Formally, the computation of the supervised contrastive loss at timestamp is as follows:
(17) |
where and denote the minibatch that is the set of queries and the number of queries at timestamp , separately; is the temperature parameter. The objective of is to make the representations of the same category closer, enhancing the common essential features of the local and global encoders required for predicting the query, thus alleviating the influence of noise and improving the robustness of the model. Similarly, the supervised loss can be obtained by using the representation generated by the global encoder as an anchor. To make the LogCL further distinguish different semantic representations of queries in the semantic space, we adopt Eq. 17 to constrain the local and global query representations and obtain two supervised losses and . In this way, we can obtain the four supervised contrastive losses: , , and . Thus, the final supervised contrastive loss is computed as .
III-F Prediction and Optimization
ConvTransE[22] is a strong score function that is widely adopted for the recent TKG reasoning task. In this work, we also adopt ConvTransE to perform the entity prediction task at timestamp . Therefore, the entity prediction score of ConvTransE is as follows:
(18) |
(19) |
where is a variable factor that is set at [0,1], which is used to trade off global and local representations of entities. In this paper, the entity prediction task can be considered as a multi-label learning problem. Thus, the loss of entity prediction is formalized as:
(20) |
where is the entity prediction probabilistic scores. is the label of which the element is 1 if the fact occurs, otherwise 0. The final loss function is computed as:
|
(21) |
In the work, the contrastive supervised loss and the entity prediction of loss are trained simultaneously.
Note that, since reverse quadruples are added in the training or testing process, the inverse query sets are generated by the original query sets at timestamp . During each epoch, the KG snapshots in history are built using the combination of the original quadruples and the inverse quadruples. If the combination sets are directly used for training, the entity-aware attention will perceive the target subject entity and target object entity, resulting in data leakage. To avoid such potential data leakage, we present a two-phase forward propagation strategy. In detail, the original query set is first considered in the forward propagation, and then the second forward propagation is performed on the query inverse sets during each epoch. In this way, the prediction scores and loss can be obtained for training or testing from the two-phase forward propagation. The detailed training procedure of LogCL is summarized in Algorithm 1.
III-G Complexity Analysis
We analyze the computational complexity of our proposed LogCL model. For the local entity-aware attention recurrent encoder, the time complexities of entity and relation evolution are and , respectively, where is the complexity of entity attention mechanism at timestamp. For the global entity-aware encoder, the time complexity of generating the historical query subgraph is , the time complexity of historical query subgraph aggregation is . The time complexity of the local-global query contrast module is . Thus, the time complexity of LogCL is .
Dataset | ICEWS14 | ICEWS18 | ICEWS05-15 | GDELT |
Entities | 6,869 | 10,094 | 23,033 | 7,691 |
Relations | 230 | 256 | 251 | 240 |
Training | 74,845 | 373,018 | 368,868 | 1,734,399 |
Validation | 8,514 | 45,995 | 46,302 | 238,765 |
Test | 7,371 | 49,545 | 46,159 | 305,241 |
Time granularity | 24 hours | 24 hours | 24 hours | 15 mins |
Snapshot numbers | 365 | 365 | 4017 | 2975 |
Model | ICEWS14 | ICEWS18 | ICEWS05-15 | GDELT | |||||||||||||
MRR | Hits@1 | Hits@3 | Hits@10 | MRR | Hits@1 | Hits@3 | Hits@10 | MRR | Hits@1 | Hits@3 | Hits@10 | MRR | Hits@1 | Hits@3 | Hits@10 | ||
Static | DisMult (2014) | 15.44 | 10.91 | 17.24 | 23.92 | 11.51 | 7.03 | 12.87 | 20.86 | 17.95 | 13.12 | 20.71 | 29.32 | 8.68 | 5.58 | 9.96 | 17.13 |
ComplEx (2016) | 32.54 | 23.43 | 36.13 | 50.73 | 22.94 | 15.19 | 27.05 | 42.11 | 32.63 | 24.01 | 37.50 | 52.81 | 16.96 | 11.25 | 19.52 | 32.35 | |
ConvE (2018) | 35.09 | 25.23 | 39.38 | 54.68 | 24.51 | 16.23 | 29.25 | 44.51 | 33.81 | 24.78 | 39.00 | 54.95 | 16.55 | 11.02 | 18.88 | 31.60 | |
Conv-TransE (2019) | 33.80 | 25.40 | 38.54 | 53.99 | 22.11 | 13.94 | 26.44 | 42.28 | 33.03 | 24.15 | 38.07 | 54.32 | 16.20 | 10.85 | 18.38 | 30.86 | |
RotatE (2019) | 21.31 | 10.26 | 24.35 | 44.75 | 12.78 | 4.01 | 14.89 | 31.91 | 24.71 | 13.22 | 29.04 | 48.16 | 13.45 | 6.95 | 14.09 | 25.99 | |
Interpolation | TTransE (2016) | 13.72 | 2.98 | 17.70 | 35.74 | 8.31 | 1.92 | 8.56 | 21.89 | 15.57 | 4.80 | 19.24 | 38.29 | 5.50 | 0.47 | 4.94 | 15.25 |
TA-DisMult (2018) | 25.80 | 16.94 | 29.74 | 42.99 | 16.75 | 8.61 | 18.41 | 33.59 | 24.31 | 14.58 | 27.92 | 44.21 | 12.00 | 5.76 | 12.94 | 23.54 | |
DE-SimIE (2020) | 33.36 | 24.85 | 37.15 | 49.82 | 19.30 | 11.53 | 21.86 | 34.80 | 35.02 | 25.91 | 38.99 | 52.75 | 19.70 | 12.22 | 21.39 | 33.70 | |
TNTCompIEx (2020) | 34.05 | 25.08 | 38.50 | 50.92 | 21.23 | 13.28 | 24.02 | 36.91 | 27.54 | 9.52 | 30.80 | 42.86 | 19.53 | 12.41 | 20.75 | 33.42 | |
Extrapolation | RE-NET (2020) | 36.93 | 26.83 | 39.51 | 54.78 | 28.81 | 19.05 | 32.44 | 47.51 | 43.32 | 33.43 | 47.77 | 63.06 | 19.62 | 12.42 | 21.00 | 34.01 |
CyGNet (2020) | 35.05 | 25.73 | 39.01 | 53.55 | 24.93 | 15.90 | 28.28 | 42.61 | 36.81 | 26.61 | 41.63 | 56.22 | 18.48 | 11.52 | 19.57 | 31.98 | |
TANGO-Tucker (2021) | 36.80 | 27.43 | 40.89 | 54.93 | 28.68 | 19.35 | 32.17 | 47.04 | 42.86 | 32.72 | 48.14 | 62.34 | 19.53 | 12.43 | 20.79 | 33.19 | |
xERTE (2021) | 40.02 | 32.06 | 44.63 | 56.17 | 29.98 | 22.05 | 33.46 | 44.83 | 46.62 | 37.84 | 52.31 | 63.92 | 18.09 | 12.30 | 20.06 | 30.34 | |
TITer (2021) | 40.87 | 32.28 | 45.45 | 57.10 | 29.98 | 22.05 | 33.46 | 44.83 | 47.69 | 37.95 | 52.92 | 65.81 | 15.46 | 10.98 | 15.61 | 24.31 | |
RE-GCN (2021) | 40.39 | 30.66 | 44.96 | 59.21 | 30.58 | 21.01 | 34.34 | 48.75 | 48.03 | 37.33 | 53.85 | 68.27 | 19.64 | 12.42 | 20.90 | 33.69 | |
CEN (2022) | 42.20 | 32.08 | 47.46 | 61.31 | 31.50 | 21.70 | 35.44 | 50.59 | 46.84 | 36.38 | 52.45 | 67.01 | 20.39 | 12.96 | 21.77 | 34.97 | |
TiRGN (2022) | 44.04 | 33.83 | 48.95 | 63.84 | 33.66 | 23.19 | 37.99 | 54.22 | 50.04 | 39.25 | 56.13 | 70.71 | 21.67 | 13.63 | 23.27 | 37.60 | |
HisMatch (2022) | 46.42 | 35.91 | 51.63 | 66.84 | 33.99 | 23.91 | 37.90 | 53.94 | 52.85 | 42.01 | 59.05 | 73.28 | 22.01 | 14.45 | 23.80 | 36.61 | |
RETIA (2023) | 42.76 | 32.28 | 47.77 | 62.75 | 32.43 | 22.23 | 36.48 | 52.94 | 47.26 | 36.64 | 52.90 | 67.76 | 20.12 | 12.76 | 21.45 | 34.49 | |
CENET (2023) | 39.02 | 29.62 | 43.23 | 57.49 | 27.85 | 18.15 | 31.63 | 46.98 | 41.95 | 32.17 | 46.93 | 60.43 | 20.23 | 12.69 | 21.70 | 34.92 | |
LogCL | 48.87 | 37.76 | 54.71 | 70.26 | 35.67 | 24.53 | 40.32 | 57.74 | 57.04 | 46.07 | 63.72 | 77.87 | 23.75 | 14.64 | 25.60 | 42.33 |
IV Experiments
In this section, we conduct many experiments to evaluate LogCL on four public datasets, and analyze the experimental results in detail.
IV-A Datasets and Baselines
IV-A1 Datasets
To evaluate LogCL on entity prediction task, we adopt four benchmark datasets that are widely used for TKG extrapolation, including ICEWS14, ICEWS18, ICEWS05-15 [40] and GDELT [41]. ICEWS14, ICEWS18 and ICEWS05-15 are subsets generated from Integrated Crisis Early Warning System (ICEWS) 111http://www.icews.com/, which contains a large number of political events with specific timestamps. GDELT 222https://www.gdeltproject.org/ is a subset generated from the Global Database of Events, Language datasets, which contains 20 of types events (e.g. Make a public statement, Appeal and consult). We follow the preprocessing strategy [30, 9] to split all datasets into training, validation, and test sets with the proportions of 80%/10%/10%. The statistics of all datasets are provided in Table II.
IV-A2 Baselines
To demonstrate the effectiveness of LogCL for TKG reasoning, we compare LogCL with 20 up-to-date KG reasoning methods that cover SKG and TKG reasoning methods (including interpolation and extrapolation). The SKG methods include DisMult [19], ConvE [21], ComplEx [20], Conv-TransE [22], RotatE [42]. The interpolation methods include TTransE[26], TA-DisMult [43], De-SimIE[44], TNTComIEx[28], TANGO-Tucker [10]. The extrapolation methods include xERTE[32], TITer[31], CyGNet[8], RE-NET[30], RE-GCN[9], CEN[33], TiRGN[11], HisMatch [38], RETIA [45], and CENET[29].
IV-B Experimental Settings
IV-B1 Evaluation Metrics
We adopt two evaluation metrics: mean reciprocal rank (MRR) and Hits@k (k=1,3,10), which are widely used to evaluate the effectiveness of TKG reasoning methods. MRR is the average reciprocal value that is used to compute the ranks of the ground truth for all queries, and Hits@k represents the proportion of times that the true entity candidates appear in the top-k of the ranked candidates. Some recent works[10, 38] point out that the traditional static filtered setting is not suitable for extrapolation on TKG reasoning due to ignoring the time dimension of facts. Actually, only the facts occurring at the same time should be filtered. Therefore, we report the experimental results with the time-aware filtered setting that is widely used in recent works, which only filters out the quadruples occurring at the query time. In addition, all experimental results on MRR and Hits@1/3/10 in this paper are reported as percentages.
IV-B2 Implementation Details
For all datasets, the embedding size is set to 200, the learning rate is set to 0.001 and the batch size is set as the number of quadruples in each timestamp. The parameters of LogCL are optimized by using adam [46] during training. The layer number of R-GCN on both local entity-aware attention recurrent encoder and global entity-aware attention encoder is set to 2 and the dropout rate for each layer is set to 0.2. The optimal local historical KG snapshots sequences lengths of ICEWS14, ICEWS18, ICEWS05-15, GDELT are set to 7, 7, 9 and 7, respectively. We follow works[9, 11, 45] that add static KG information on ICEWS14, ICEWS18 and ICEWS05-15 datasets. The prediction weight of all datasets is set to 0.9. The optimal temperature coefficient of ICEWS14, ICEWS18, ICEWS05-15, GDELT are set to 0.03, 0.03, 0.07 and 0.07. For the decoder on all datasets, the number of kernels is set to 50, the kernel size is set to 23, and the dropout rate is set to 0.2.
For SKG reasoning methods, the time dimension is removed on all TKG datasets. Some of the extrapolation baselines, including xERTE[32], TANGO-Trucker[10], TITer[31], REGCN[9], CEN[33], TiRGN [11] and RETIA[45], publish their open source code, and their results with the time-aware filter setting are reported under the default parameters. The others are taken from the Hismathch [38]. For fairness of comparison, results of CEN and RETIA are reported under the offline setting that is adopted to other baselines. We also compare our LogCL with CEN and RETIA models under the online setting and results are reported in the following section H.
Model | ICEWS14 | ICEWS18 | ICEWS05-15 | |||||||||
MRR | Hits@1 | Hits@3 | Hits@10 | MRR | Hits@1 | Hits@3 | Hits@10 | MRR | Hits@1 | Hits@3 | Hits@10 | |
LogCL | 48.87 | 37.76 | 54.71 | 70.26 | 35.67 | 24.53 | 40.32 | 57.74 | 57.04 | 46.07 | 63.72 | 77.87 |
LogCL- | 44.74 | 33.53 | 50.29 | 66.72 | 30.21 | 18.96 | 34.45 | 52.69 | 51.92 | 40.31 | 58.37 | 74.66 |
LogCL- | 46.81 | 35.52 | 52.54 | 68.91 | 35.31 | 24.14 | 39.92 | 57.40 | 56.78 | 45.73 | 63.50 | 77.59 |
LogCL-w/o-eatt | 40.34 | 30.29 | 44.82 | 60.05 | 31.01 | 20.99 | 34.88 | 50.71 | 46.25 | 35.76 | 51.71 | 66.50 |
LogCL--w/o-eatt | 38.61 | 29.04 | 42.79 | 57.42 | 27.83 | 18.56 | 31.11 | 46.30 | 41.40 | 31.20 | 46.44 | 60.93 |
LogCL--w/o-eatt | 39.86 | 29.90 | 44.40 | 59.48 | 30.95 | 21.01 | 34.85 | 50.55 | 46.16 | 35.56 | 51.58 | 66.70 |
LogCL-w/o-cl | 46.84 | 35.62 | 52.44 | 69.05 | 35.32 | 24.04 | 40.00 | 57.62 | 56.85 | 45.83 | 63.61 | 77.81 |





IV-C Results of TKG Extrapolation
The overall experimental results of LogCL and baselines on four benchmark datasets are reported in Table III. The best results are marked in bold, and the second-best results are reported using underlining. From the experimental results, we have the following observation:
-
•
The performance of LogCL consistently is better than the state-of-the-art methods on four benchmark datasets. Specifically, when comparing the second-best results, LogCL achieves the most significant improvements of 5.2%, 4.9%, 7.9%, and 7.9% in MRR on ICEWS14, ICEWS18, ICEWS05-15 and GDELT datasets.
-
•
For the state-of-the-art models Hismatch and TIRGN, although HisMatch also considers the local historical information related to the query, it fails to effectively consider the importance of the local historical fact information related to the query and does not consider the global historical information, resulting in weaker performance than LogCL. TiRGN takes into account both global and local historical information, but ignoring query-related historical information leads to lower performance than Hismath. Although CENET considers historical contrastive learning with our model LogCL, its performance is lower than LogCL due to the lack of evolutionary modeling of facts. In addition, CENET uses historical contrastive learning to enhance entities with less historical interactions by contrast, while we use contrastive learning from a local and global view to improve the robustness of the model.
-
•
An interesting phenomenon on all models is that the results on ICEWS14 and ICEWS18 datasets perform better than GDELT and ICEWS05-15 datasets. The reason for such phenomenon may be that more complex dynamic interactions among entities and relations exist in ICEWS18 and GDELT datasets. Such complex dynamic facts association are difficult to be modeled and lead to the worse results on ICEWS18 and GDELT datasets. The excellent performance exhibited by LogCL on ICEWS18 and GDELT datasets demonstrates the ability of LogCL to model complex temporal facts interactions.
-
•
Extrapolation models on all datasets perform better than interpolation and static models, mainly because static models do not consider temporal information and are difficult to capture the dynamic changes of entities and relations. For interpolation models, they only focus on the completion of historical missing facts and lack the ability to model the evolution of entities and relationships over time and predict future unseen facts.


IV-D Ablation Study
To further better analyze each part of LogCL that contributes to the prediction results, we conduct ablation studies on all datasets in terms of MRR and Hits@1/5/10. The results of the LogCL variants are presented in Table IV.
IV-D1 Impact of Global Entity-Aware Attention Encoder
The results denoted as LogCL- in Table IV demonstrate the performance of LogCL without considering the global entity-aware attention encoder. It can be seen that the performances of LogCL- consistently performs poorly compared to the LogCL on ICEWS14, ICEWS18 and GDELT datasets, indicating that modeling the longer facts related to the queries is beneficial for TKG reasoning. We also conduct experiments to investigate the impact of varying the number of layers in the R-GCN in the global entity-aware attention encoder, and the results are shown in Fig. 6. It can be observed that demonstrate that the two-hop results are indeed slightly better than the one-hop results on both datasets. However, increasing the number of hops beyond two does not significantly improve performance on the ICEWS14 dataset and leads to decreased performance on the ICEWS18 dataset. A similar phenomenon also can be found in the local entity-aware attention recurrent encoder.
IV-D2 Impact of Local Entity-Aware Attention Recurrent Encoder
LogCL- in Table IV is a variant of LogCL without modeling the recent local history at the adjacent timestamps. More specifically, the local entity-aware attention recurrent encoder is removed in LogCL-. The scores of all entities are obtained by using the representation of entities from the global entity-aware attention encoder and the static relation embedding as the representation of the query relation. The results of LogCL- in Table IV can be observed that ignoring the local entity-aware attention recurrent encoder can generate a great impact on the performances. Mainly because the recent local history contains rich information describing the behavior trends related to the query, which helps in the selection of the correct answer. Another observation is that the variant LogCL- is superior to the variant LogCL-, indicating that modeling the recent local entities and relations evolution is more effective than long historical information.




IV-D3 Impact of Entity-Aware Attention
To demonstrate how the entity-aware attention contributes to the prediction results, the entity-aware attention is removed in the local entity-aware attention recurrent encoder and global entity-aware attention encoder, which is denoted as ”-w/o-eatt” in Table IV. It can be intuitively observed that the performance degradation is great after removing the entity-aware attention on ICEWS14, ICEWS18 and ICEWS05-15 datasets. The main reason is that the entity attention mechanism can effectively capture the facts that the local KG snapshot and global historical subgraph are related to the query.
IV-D4 Impact of Contrastive Learning
LogCL-w/o-cl denotes a variant of LogCL that removes the local-global contrastive learning. It can be observed that the performance of LogCL-w/o-cl is worse than that of LogCL, especially on the ICEWS14 dataset. It is because contrastive learning plays a role in guiding LogCL to fuse different historical characteristics. We also conduct experiments to investigate the impact of four query contrast strategies. The results are presented in Fig. 7. LogCL-gl, LogCL-lg, LogCL-gg, and LogCL-ll denote variants using four different query losses, respectively. It can be observed that LogCL-gl and Log-lg demonstrate slightly superior performance compared to LogCL-gg and LogCL-ll. This suggests that contrastive training by emphasizing the distinction between local and global query representations yields greater benefits.
To further analyze the ability of the local-global query contrast module of LogCL to resist noise interference, we conduct experiments by adding Gaussian noise with different intensities to LogCL and LogCL-w/o-cl. Here, Gaussian noise generated by the Gaussian distribution is added to the entity representation as the initial input of the model. As our primary objective is to predict future Gaussian entities, we solely introduce noise to the entities and not the relations. We control the intensity of Gaussian noises by varying the variance in the Gaussian formula. The experimental results are shown in Fig. 5. When faced with the same intensity of noise, a larger span of Y-axis values on the same color indicates a larger performance degradation. We can intuitively observe that the LogCL model always performs better than the LogCL-w/o-cl model under the same intensity of noise on ICEWS14, ICEWS18 and ICEWS05-15 datasets. This shows that local-global contrastive learning can effectively improve the anti-noise ability of the model. In addition, we also observe another interesting phenomenon, although the performance of LogCL model and LogCL-w/o-cl model gradually decreases with the increase of noise intensity, it is worth noting that the performance of LogCL model decreases less than that of LogCL-w/o-cl model when facing stronger noise, especially on ICEWS05-15 dataset. LogCL shows relatively strong performance even when subjected to strong noise interference.
Model | ICEWS14 | ICEWS18 | ICEWS05-15 | |||
MRR | Hits@1 | MRR | Hits@1 | MRR | Hits@1 | |
LogCL (RGCN) | 48.87 | 37.76 | 35.67 | 24.53 | 57.04 | 46.07 |
LogCL (CompGCN-sub) | 49.25 | 36.84 | 35.33 | 24.26 | 56.93 | 45.92 |
LogCL (CompGCN-mult) | 47.92 | 36.85 | 35.32 | 24.05 | 56.40 | 45.46 |
LogCL (KBAT) | 48.46 | 37.17 | 35.70 | 24.41 | 56.01 | 45.14 |


Query and Answer | LogCL | LogCL-w/o-eatt | LogCL-w/o-cl |
Query: ( China, Sign_ formal_agreement, ?, 2014-12-4 ) Answer: South_Africa | South_Africa 0.6136 | Kazakhstan 0.109 | South_Africa 0.501 |
Ashraf_Ghani_Ahmadzai 0.088 | France 0.083 | Ashraf Ghani Ahmadzai 0.244 | |
Malaysia 0.084 | South_Korea 0.076 | South Korea 0.058 | |
South_Korea 0.045 | Vietnama 0.069 | Malaysia 0.050 | |
European_Parliament 0.043 | Iraq 0.049 | European_Parliament 0.029 | |
Query: ( Iran, Engage_in_ diplomatic_cooperation, ?, 2014-11-30 ) Answer: Oman | Oman 0.687 | Oman 0.1098 | Oman 0.668 |
Portugal 0.122 | Iraq 0.1095 | Portugal 0.129 | |
Guinea 0.077 | Food_and_Agriculture_Organization 0.0813 | Iraq 0.045 | |
Qatqr 0.034 | China 0.0562 | Qatar 0.041 | |
China 0.021 | Tajikistan 0.0434 | Guinea 0.012 |
Model | ICEWS14 | ICEWS18 | ICEWS05-15 | |||
MRR | Hits@1 | MRR | Hits@1 | MRR | Hits@1 | |
LogCL | 48.87 | 37.76 | 35.67 | 24.53 | 57.04 | 46.07 |
LogCL- | 50.69 | 39.66 | 37.38 | 25.96 | 58.69 | 47.79 |
LogCL- | 47.04 | 35.87 | 33.89 | 22.97 | 55.38 | 44.34 |
IV-E Sensitivity Analysis
In this section, we conduct experiments on ICEWS14 and 18 datasets to further analyze the impact of parameters in LogCL, including the parameter in the prediction part, the temperature coefficient in the local-global query contrast module, and GNN aggregation encoder.
IV-E1 Analysis of the Parameter
To explore the parameter that is used to trade off global and local representations of entities, we conduct experiments with different values of from 0 to 1 with other optimal hyperparameters fixed. The results of MRR and Hist@3 are shown in Fig.8. A larger value of indicates a higher proportion of the local entity-aware attention recurrent encoder. It can be seen that the performance of LogCL shows a trend of first upward and then falling as increases on ICEWS14 and ICEWS18 datasets, which indicates that only considering local historical information or global historical information cannot effectively predict the results. This trend further illustrates that an appropriate parameter is important for combining local and global historical patterns.
IV-E2 Analysis of Temperature Coefficient
We perform a batch of experiments with different temperature coefficients with other optimal hyperparameters fixed. The results of MRR and Hits@3 are shown in Fig.9. Different datasets are affected by the temperature coefficient differently, and choosing the appropriate temperature coefficient is helpful for TKG reasoning.
IV-F Study On Different GNN Aggregation
To further study the impact of different kinds of GNNs in the local entity-aware attention recurrent encoder and the global entity-aware attention encoder, R-GCN is replaced in these two encoders with CompGCN[24] and KBGAT[25]. The MRR and Hits@1 results of ICEWS14, ICEWS18, and ICEWS05-15 are reported in Table V. It can be seen that LogCL (RGCN) gets the best results on ICEWS05-15 dataset and shows strong performance in ICEWS14 and ICEWS18 datasets.
IV-G Study On the Two-Phase Propagation
To investigate the role of the two-phase propagation in our LogCL, we manage experiments to evaluate the results of two forward propagations separately. Log- denotes a variant that predicts the object entities on the origin query set and only performs the first phase propagation. Log- that predicts the subjected entities on the reverse query set and only considers the second phase propagation. The results are shown in Table VII. It can be seen that the performance of Log- is superior to the Log-. The reason behind this phenomenon is that the dataset composed of inverse relations may introduce a certain bias, which subsequently impacts the performance of the model. In contrast, the original dataset provides a more comprehensive reflection of the real relations between entities, resulting in better performance in the experimental results.
IV-H Study On the Online Training
Since the evolutionary pattern changes with emerging facts, exploring how to adjust the model to emerging facts is crucial. Following the CEN [33] and RETIA [45], we perform a batch of experiments to learn emerging facts on ICEWS14, ICEWS18 and ICEWS05-15 datasets under the online setting. The results are shown in the Fig.10. The results of CEN, RETIA and LogCL under the online setting outperform the results in Table III under the offline setting. It is because that historical facts in the test set are updated under the online setting. In addition, LogCL achieves greater improvements than the CEN and RETIA under the online setting. This proves that LogCL can effectively solve domain obstacles such as time-varying problems.
IV-I Case Study
We provide the case study for LogCL, LogCL-w/o-eatt, and LogCL-w/o-cl on ICEWS14 dataset. Two queries with the top-5 prediction entities and scores are reported in Table VI.
For the query (China, Sign_formal_agreement, ?, 2014-12-4), we can observe that LogCL and logCL-w/o-cl can predict the correct answer, while the top-5 results given by LogCL-w/o-eatt do not contain the correct answer. For the query (Iran, Engage_in_diplomatic_cooperation, ?, 2014-11-30), although LogCL, LogCL-w/o-eatt, and LogCL-w/o-eatt can give answers prepared and the answers are top-1, the prediction scores of LogCl and LogCL-w/o-eatt for correct answers are much higher than that of LogCL-w/o-eatt. The reason for the above phenomenon is that the reasoning of LogCL-w/o-eatt primarily relies on historical facts that are in close proximity in time and weakly associated with the query. As a result, LogCL-w/o-eatt fails to capture the historical facts that have a strong relevance to the query, leading to lower accurate prediction results. These two cases can further prove the effectiveness of the entity-aware attention mechanism and the strong reasoning ability of LogCL.
V Conclusion
In this paper, we propose a novel TKG reasoning model, namely LogCL, which uses contrastive learning to better guide the model to integrate local and global historical information, thereby improving the robustness of the model. We propose a local entity-aware attention recurrent encoder, which effectively captures the importance of query-related historical information in each KG snapshot at the most recent time. We propose a global entity-aware attention encoder, which learns the importance of the global historical information in the history subgraph that is built by sampling the global historical facts based on queries. We design a local-global query contrast module to focus on the important common features of the local encoder and global encoder, so as to enhance the anti-noise ability of LogCL. Extensive experiments on four public datasets demonstrate that our proposed LogCL performs significant improvements and has a more robust performance than the state-of-the-art baselines.
References
- [1] P. Ernst, C. Meng, A. Siu, and G. Weikum, “Knowlife: A knowledge graph for health and life sciences,” in IEEE 30th International Conference on Data Engineering, Chicago, ICDE 2014, IL, USA, March 31 - April 4, 2014, I. F. Cruz, E. Ferrari, Y. Tao, E. Bertino, and G. Trajcevski, Eds. IEEE Computer Society, 2014, pp. 1254–1257.
- [2] L. Cui, H. Seo, M. Tabar, F. Ma, S. Wang, and D. Lee, “DETERRENT: knowledge guided graph attention network for detecting healthcare misinformation,” in KDD ’20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, August 23-27, 2020. ACM, 2020, pp. 492–502.
- [3] C. Yang and G. Qi, “An urban traffic knowledge graph-driven spatial-temporal graph convolutional network for traffic flow prediction,” in Proceedings of the 11th International Joint Conference on Knowledge Graphs, IJCKG 2022, Hangzhou, China, October 27-28, 2022. ACM, 2022, pp. 110–114.
- [4] F. Feng, X. He, X. Wang, C. Luo, Y. Liu, and T. Chua, “Temporal relational ranking for stock prediction,” ACM Trans. Inf. Syst., vol. 37, no. 2, pp. 27:1–27:30, 2019.
- [5] G. W. Trompf, The idea of historical recurrence in Western thought: from antiquity to the Reformation. Univ of California Press, 1979, vol. 1.
- [6] J. S. B. Evans, “Heuristic and analytic processes in reasoning,” British Journal of Psychology, vol. 75, no. 4, pp. 451–468, 1984.
- [7] S. A. Sloman, “The empirical case for two systems of reasoning.” Psychological bulletin, vol. 119, no. 1, p. 3, 1996.
- [8] C. Zhu, M. Chen, C. Fan, G. Cheng, and Y. Zhang, “Learning from history: Modeling temporal knowledge graphs with sequential copy-generation networks,” vol. 35, no. 5, pp. 4732–4740, 2021, number: 5.
- [9] Z. Li, X. Jin, W. Li, S. Guan, J. Guo, H. Shen, Y. Wang, and X. Cheng, “Temporal knowledge graph reasoning based on evolutional representation learning,” in Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval. Association for Computing Machinery, 2021, pp. 408–417.
- [10] Z. Han, Z. Ding, Y. Ma, Y. Gu, and V. Tresp, “Learning neural ordinary equations for forecasting future links on temporal knowledge graphs,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021. Association for Computational Linguistics, 2021, pp. 8352–8364.
- [11] Y. Li, S. Sun, and J. Zhao, “Tirgn: time-guided recurrent graph network with local-global historical patterns for temporal knowledge graph reasoning,” in Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022. ijcai. org, 2022, pp. 2152–2158.
- [12] Y. He, P. Zhang, L. Liu, Q. Liang, W. Zhang, and C. Zhang, “HIP network: Historical information passing network for extrapolation reasoning on temporal knowledge graph,” in Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event / Montreal, Canada, 19-27 August 2021. ijcai.org, 2021, pp. 1915–1921.
- [13] P. Khosla, P. Teterwak, C. Wang, A. Sarna, Y. Tian, P. Isola, A. Maschinot, C. Liu, and D. Krishnan, “Supervised contrastive learning,” in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020.
- [14] Q. Wang, Z. Mao, B. Wang, and L. Guo, “Knowledge graph embedding: A survey of approaches and applications,” IEEE Transactions on Knowledge and Data Engineering, vol. 29, no. 12, pp. 2724–2743, 2017.
- [15] S. Liang, J. Shao, D. Zhang, J. Zhang, and B. Cui, “DRGI: deep relational graph infomax for knowledge graph completion: (extended abstract),” in 38th IEEE International Conference on Data Engineering, ICDE 2022, Kuala Lumpur, Malaysia, May 9-12, 2022. IEEE, 2022, pp. 1499–1500.
- [16] B. Liu, X. Wang, P. Liu, S. Li, Q. Fu, and Y. Chai, “Unikg: A unified interoperable knowledge graph database system,” in 37th IEEE International Conference on Data Engineering, ICDE 2021, Chania, Greece, April 19-22, 2021. IEEE, 2021, pp. 2681–2684.
- [17] A. Bordes, N. Usunier, A. García-Durán, J. Weston, and O. Yakhnenko, “Translating embeddings for modeling multi-relational data,” in The 27th Annual Conference on Neural Information Processing Systems 2013, December 5-8, 2013, Lake Tahoe, Nevada, United States, 2013, pp. 2787–2795.
- [18] Z. Wang, J. Zhang, J. Feng, and Z. Chen, “Knowledge graph embedding by translating on hyperplanes,” in Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, July 27 -31, 2014, Québec City, Québec, Canada. AAAI Press, 2014, pp. 1112–1119.
- [19] B. Yang, W. Yih, X. He, J. Gao, and L. Deng, “Embedding entities and relations for learning and inference in knowledge bases,” in 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.
- [20] T. Trouillon, J. Welbl, S. Riedel, É. Gaussier, and G. Bouchard, “Complex embeddings for simple link prediction,” in Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016, ser. JMLR Workshop and Conference Proceedings, vol. 48. JMLR.org, 2016, pp. 2071–2080.
- [21] T. Dettmers, P. Minervini, P. Stenetorp, and S. Riedel, “Convolutional 2d knowledge graph embeddings,” in Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018. AAAI Press, 2018, pp. 1811–1818.
- [22] C. Shang, Y. Tang, J. Huang, J. Bi, X. He, and B. Zhou, “End-to-end structure-aware convolutional networks for knowledge base completion,” in The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019. AAAI Press, 2019, pp. 3060–3067.
- [23] M. S. Schlichtkrull, T. N. Kipf, P. Bloem, R. van den Berg, I. Titov, and M. Welling, “Modeling relational data with graph convolutional networks,” in The Semantic Web - 15th International Conference, ESWC 2018, Heraklion, Crete, Greece, June 3-7, 2018, Proceedings, ser. Lecture Notes in Computer Science, vol. 10843. Springer, 2018, pp. 593–607.
- [24] S. Vashishth, S. Sanyal, V. Nitin, and P. P. Talukdar, “Composition-based multi-relational graph convolutional networks,” in 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020.
- [25] D. Nathani, J. Chauhan, C. Sharma, and M. Kaul, in Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28- August 2, 2019, Volume 1: Long Papers. Association for Computational Linguistics, 2019, pp. 4710–4723.
- [26] J. Leblay and M. W. Chekol, “Deriving validity time in knowledge graph,” in Companion of the The Web Conference 2018 on The Web Conference 2018, WWW 2018, Lyon , France, April 23-27, 2018, P. Champin, F. Gandon, M. Lalmas, and P. G. Ipeirotis, Eds. ACM, 2018, pp. 1771–1776.
- [27] S. S. Dasgupta, S. N. Ray, and P. P. Talukdar, “Hyte: Hyperplane-based temporally aware knowledge graph embedding,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, 2018, pp. 2001–2011.
- [28] T. Lacroix, G. Obozinski, and N. Usunier, “Tensor decompositions for temporal knowledge base completion,” in 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020.
- [29] Y. Xu, J. Ou, H. Xu, and L. Fu, “Temporal knowledge graph reasoning with historical contrastive learning,” in Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence, IAAI 2023, Thirteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2023, Washington, DC, USA, February 7-14, 2023. AAAI Press, 2023, pp. 4765–4773.
- [30] W. Jin, M. Qu, X. Jin, and X. Ren, “Recurrent event network: Autoregressive structure inferenceover temporal knowledge graphs,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, 2022-11-24, pp. 6669–6683.
- [31] H. Sun, J. Zhong, Y. Ma, Z. Han, and K. He, “TimeTraveler: Reinforcement learning for temporal knowledge graph forecasting,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2021, pp. 8306–8319.
- [32] Z. Han, P. Chen, Y. Ma, and V. Tresp, “Explainable subgraph reasoning for forecasting on temporal knowledge graphs,” in 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021.
- [33] Z. Li, S. Guan, X. Jin, W. Peng, Y. Lyu, Y. Zhu, L. Bai, W. Li, J. Guo, and X. Cheng, “Complex evolutional pattern learning for temporal knowledge graph reasoning,” in Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Association for Computational Linguistics, 2022, pp. 290–296.
- [34] H. Sun, S. Geng, J. Zhong, H. Hu, and K. He, “Graph hawkes transformer for extrapolated reasoning on temporal knowledge graphs,” in Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022. Association for Computational Linguistics, 2022, pp. 7481–7493.
- [35] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, 2017, pp. 5998–6008.
- [36] Q. Lin, J. Liu, R. Mao, F. Xu, and E. Cambria, “TECHS: temporal logical graph networks for explainable extrapolation reasoning,” in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023. Association for Computational Linguistics, 2023, pp. 1281–1293.
- [37] T. Chen, S. Kornblith, M. Norouzi, and G. E. Hinton, “A simple framework for contrastive learning of visual representations,” in Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, ser. Proceedings of Machine Learning Research, vol. 119. PMLR, 2020, pp. 1597–1607.
- [38] Z. Li, Z. Hou, S. Guan, X. Jin, W. Peng, L. Bai, Y. Lyu, W. Li, J. Guo, and X. Cheng, “Hismatch: Historical structure matching based temporal knowledge graph reasoning,” in Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022. Association for Computational Linguistics, 2022, pp. 7328–7338.
- [39] K. Cho, B. van Merrienboer, Ç. Gülçehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio, “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, October 25-29, 2014, Doha, Qatar,. ACL, 2014, pp. 1724–1734.
- [40] E. Boschee, J. Lautenschlager, S. O’Brien, S. Shellman, J. Starz, and M. Ward, “Icews coded event data. harvard dataverse,” V4 (http://dx. doi. org/10.7910/DVN/28075), 2015.
- [41] K. Leetaru and P. A. Schrodt., “Gdelt: Global data on events, location, and tone, 1979-2012,” ISA Annual Convention, vol. 4, pp. 1–49, 2013.
- [42] Z. Sun, Z. Deng, J. Nie, and J. Tang, “Rotate: Knowledge graph embedding by relational rotation in complex space,” in 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
- [43] A. García-Durán, S. Dumancic, and M. Niepert, “Learning sequence encoders for temporal knowledge graph completion,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018. Association for Computational Linguistics, 2018, pp. 4816–4821.
- [44] R. Goel, S. M. Kazemi, M. A. Brubaker, and P. Poupart, “Diachronic embedding for temporal knowledge graph completion,” in The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020. AAAI Press, 2020, pp. 3988–3995.
- [45] K. Liu, F. Zhao, G. Xu, X. Wang, and H. Jin, “Retia: relation-entity twin-interact aggregation for temporal knowledge graph extrapolation,” in IEEE International Conference on Data Engineering. IEEE, 2023.
- [46] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015.