SEG:Seeds-Enhanced Iterative Refinement Graph Neural Network for Entity Alignment
Abstract
Entity alignment is crucial for merging knowledge across knowledge graphs, as it matches entities with identical semantics. The standard method matches these entities based on their embedding similarities using semi-supervised learning. However, diverse data sources lead to non-isomorphic neighborhood structures for aligned entities, complicating alignment, especially for less common and sparsely connected entities. This paper presents a soft label propagation framework that integrates multi-source data and iterative seed enhancement, addressing scalability challenges in handling extensive datasets where scale computing excels. The framework uses seeds for anchoring and selects optimal relationship pairs to create soft labels rich in neighborhood features and semantic relationship data. A bidirectional weighted joint loss function is implemented, which reduces the distance between positive samples and differentially processes negative samples, taking into account the non-isomorphic neighborhood structures. Our method outperforms existing semi-supervised approaches, as evidenced by superior results on multiple datasets, significantly improving the quality of entity alignment.
Index Terms:
Entity Alignment; GCN; Entity embedding;I INTRODUCTION

Knowledge graph is a large-scale structured semantic knowledge base used to describe concepts in the physical world and their interrelationships, aggregate information and data, and link relationships on the Web into knowledge [1, 2, 3, 4, 5, 6, 7, 8]. Knowledge graphs contain rich information such as entity attributes, semantic relationships, and entity types. There are many large knowledge graphs, such as DBpedia [9]and YAGO[10], which have been widely used in natural language processing, information retrieval, and intelligent question answering. Due to the frequent inclusion of many entities and relationships in knowledge graphs, there is also a close connection with scale computing. However, knowledge graphs are inherently incomplete. If downstream applications need support, integrating various KGs for knowledge fusion effectively solves this problem. Researchers have proposed many entity alignment methods. People have gradually shifted from traditional TransE [11, 12, 13, 14, 15, 16, 17, 18] methods to deep learning-based modeling of the graph structure of knowledge graphs. The methods based on deep learning [19, 20, 21, 22, 23, 24, 25] mainly follow the learning embedding-similarity calculation-greedy matching mode, which is to use seed entities to automatically extract implicit features of entities from different KGs using neural networks and then embed them into the same implicit space and calculate the similarity between different entity embeddings. However, in actual knowledge graphs, due to the heterogeneity of data sources, entities usually exhibit non-isomorphic neighborhood structures, and people have begun to realize that more is needed to consider entity embeddings for alignment. [26, 27, 28, 29, 30] considered the positive interaction between entity alignment and relationships in a semi-supervised context and mined relational information to assist in alignment. [31, 32] considers the simultaneous roles of entities and relationships.
Although these scholars have attached varying degrees of importance to the importance of relationships in alignment, most methods still ignore two issues:
First, the difference between entities and relations is ignored. In the mainstream methods, embedding-similarity-matching, similarity methods can also be divided into two types: based on equivalent entities having similar attributes[33] and based on equivalent entities having similar adjacent entities. However, in these methods, the propagation of alignment information focuses more on the propagation from the similarity of entities to relations. Not only does it largely ignore the adequate auxiliary alignment information of the relational semantics itself, but there is also a problem of information loss. More importantly, it ignores the differences between entities and relations, which leads to inconsistent information. As shown in Fig 1, if only considering the propagation of similarity in entities, then Alan Turing and Alan Turing, Stephen Hawking, and Stephen Hawking can be fully aligned, but in relations, alma-mater, and graduate, student and worker cannot be aligned. GCN-based frameworks also use relations as weights[34] or information. However, noise is inevitably mixed in when the same model is used to train the representation of entities and relations jointly. The integration of relations into entity representation will cause the entity representation to be blurred and smooth.
Second, there is an obvious problem in current research: over-emphasis on reducing the distance between positive samples while ignoring the distance between negative samples, especially when facing pseudo-aligned entities with high similarity; there needs to be more differentiated treatment of these pseudo-entities. In cross-knowledge graph alignment, effectively dealing with the non-isomorphic neighborhood differences between positive and negative samples is a significant challenge.
In order to overcome the limitations of existing methods, this paper proposes a soft label propagation framework based on multi-source fusion combined with an iterative process of seed enhancement. First, we conduct dual-angle modeling based on entities and relations and fuse the best-matching relationship pairs selected in both entity and relationship modes to generate soft labels. These soft labels contain rich entity neighbor features and integrate auxiliary information of relationship semantics, avoiding the problem of auxiliary information loss. Secondly, the soft labels fused by multi-source information propagate alignment information, and the topological structure and neighborhood differences are effectively captured. On this basis, a bidirectional adaptive weighted joint loss function is introduced, which not only shortens the distance between positive samples through commonalities but also widens the distance between high-similarity negative samples to varying degrees and more comprehensively considers the commonalities and differences in the alignment process. Some contributions are as follows:
-
•
Obtaining soft labels from the perspectives of entities and relations can reveal shared entity features and mine the rich information of the relationship semantics itself, thereby improving alignment accuracy.
-
•
The difference problem of non-isomorphic neighborhoods is solved by effectively shortening the distance between positive samples and using a weighted joint loss function to differentiate and distance negative samples.
-
•
Our method, tested on real datasets, significantly boosts model performance, proving highly feasible and valuable for entity alignment across knowledge graphs.
The remainder of this paper is organized as follows: Section II reviews related work, Section III provides a detailed description of the proposed method, Section IV presents the experimental setup and results, and finally, Section V offers a brief conclusion.

II RELATED WORK
In this part, we examine existing scholarly contributions about entity alignment, encompassing conventional translation-based tactics as well as contemporary graph representational techniques that leverage Graph Convolutional Networks (GCNs)[35][36].
II-A Conventional Approaches Utilizing Translation Techniques
Translation-based entity alignment. By continuously adjusting the vector representation, the entity vector + the relation vector = the entity vector. In order to solve the shortcomings of TransE. MTransE[37] embeds entities and relations in different knowledge graphs into different vector spaces and maps them to a unified space through distance calibration and linear transformation strategies to achieve cross-knowledge graph alignment. BootEA[33] is a Bootstrap-based entity alignment method that converts entity alignment into a classification problem. It completes the alignment task by iteratively marking possible entity alignments as training data and finding the maximum possibility of alignment.
II-B GCN-Driven Graph Representation Techniques
GNN-based entity alignment. GCN-align can embed entities from different KGs into a unified vector space. Embedding can be learned from the structure and attribute information of the entity. HMAN[38] embeds learning entity representation from three perspectives: structural information, relationship features, and attribute characteristics.RD-GCN[39] uses the graph attention mechanism to encourage dual relations between the interaction graph and the original graph and then feeds the vertex representation in the original graph into the GCN layer with highway neural network gating to capture the structural information of the neighbors.
II-C Semi-supervised entity alignment
III PROPOSED METHOD
We introduce the SEG methodology, whose structural outline is depicted in Fig 2. Entity encoding and alignment candidate generation.Soft label screening: The soft labels of the two modalities are fused, and the screening strategy is used to obtain the final accurate soft labels.Pruning when performing neighborhood aggregation based on soft labels: Propagate the alignment relationship in the soft labels to the neighborhood structure of the candidate alignment.Bidirectional weighted joint mechanism: Use a weighted method to differentiate negative samples to guide positive and negative sample learning between two knowledge graphs.Bidirectional iterative collaborative training.
In this study, we consider the scenario where two distinct knowledge graphs are provided as inputs. A knowledge graph is structured as , where denotes the set of entities, represents the set of relations, and is a collection of triples defined as . The source knowledge graph is denoted by ; similarly, the target knowledge graph is denoted by . The set of seed alignment entities is given by , indicating that entity from and entity from are identical, with signifying equivalence. To initiate the process, we employ a pre-trained model to assign initial embeddings to the entities.
III-A Graph representation learning
The Graph Attention Network serves as a feature extraction mechanism, introducing an innovative neural network framework for processing data structured as a graph. This architecture comprises an input layer, an intermediary hidden layer, and a final output layer. Within each hidden layer, the representation of a node is determined by its interaction with the surrounding nodes’ information, necessitating ongoing updates to the hidden layer’s state. The computation is based on the following formula:
(1) |
where and represent the embedding of the entity at the layer, represents the number of layers, and represents the weight parameter of the layer, represents the ReLU activation function.
(2) |
(3) |
where represents the set of relation triplets with as the head entity and as the tail entity, is the attention parameter, are vector concatenation, element multiplication and LeakyReLU respectively function.
(4) |
Where and represent the size of the head entity set and the tail entity set connected by , respectively, is the attention parameter.
(5) |
and are trainable weight matrices, and their output matrices are represented as .
Alignment candidates. After obtaining the embedding representation of the entity through , the embedding representation of the two heterogeneous knowledge graphs is obtained respectively. , and then the candidate entity pairs need to be matched, that is, , cosine similarity is used to evaluate the similarity between entities. The similarity calculation results of all embeddings in and can be calculated according to the cosine similarity, and the calculation results are saved in the similarity matrix cosine_distances. The calculation formula is as follows:
(6) |
, Only the similarity in the matrix 0.95 is selected as the Candidate pre-selected set. However, these candidate sets only represent possible matching entity pairs, and there may even be one-to-many situations that cannot guarantee the accuracy of the final alignment.
III-B Soft Labels
Based on Entity propagate similarity. taking the aligned seeds of and as anchors, for any seed pair , we construct the neighbourhood node sets of and , and , or . represents the -th neighbor node of , and represents the connection between the two. represents the -th neighbor node of , , , is the same. Here, the above cosine similarity is also used to calculate the similarity of the nodes in the neighborhood of the seed pair. In order to ensure that the relationship pairs screened by the neighborhood candidate alignment set are highly accurate and convenient for subsequent training, two standards are set here: First, the neighborhood alignment nodes need to meet the set similarity threshold and the matching number threshold of the derived relationship. Second, we must ensure that the neighborhood nodes are aligned one-to-one. Only the neighborhood pairs with the highest similarity are selected when encountering a one-to-many or many-to-one situation.
(7) |
represents the soft label set screened based on the entity embedding model, represents the cosine similarity of the neighborhood node pair, and match is the number of repetitions of the relationship between the entities. For example: and ,that is, keep this relationship pair in the set .
Based on relation propagate similarity. In this section, the Bert model is introduced to encode the relations of the KG graph, and the semantic features of the relations are fully captured using the name and description of the relations.
(8) |
RD: Relation Description. Similarly, setting the cosine similarity and matching threshold also requires one-to-one matching.
(9) |
represents the number of soft labels selected based on the relation embedding model, represents the cosine similarity of the relation pair, and is the number of times this relation pair appears repeatedly. , has a higher priority than .
III-C Pruning Based on Soft Labels
Soft labels enhance alignment precision by guiding neighborhood matching and are selectively retained based on match quality. GCN integrates this neighborhood data, while Highway networks combine entity names with neighborhood information to refine the final embeddings.
III-D Bidirectional Weighting Mechanism
Inspired by contrastive learning, we have engineered a bidirectional weighting mechanism based on similarity aimed at optimizing the weight distribution of samples during the entity alignment process. This mechanism adjusts the weights of negative samples differentially to achieve a precise approximation of positive samples and effectively optimize negative samples. Positive samples are actual alignments , while negative samples are selected from the top 50 nodes with the highest similarity to the positive samples in either or . Notably, due to the most significant risk of misalignment, the node with the highest similarity is assigned the highest weight penalty.
(10) |
Method | DBPJA-EN | DBPFR-EN | DBPZH-EN | |||||||||
Hit@1 | Hit@5 | MRR | Hit@1 | Hit@5 | MRR | Hit@1 | Hit@5 | MRR | ||||
SEA | 29.12 | 55.64 | 0.416 | 32.64 | 60.28 | 0.455 | 34.56 | 59.54 | 0.464 | |||
BootEA | 52.71 | 71.89 | 0.616 | 57.61 | 77.27 | 0.666 | 55.45 | 73.72 | 0.639 | |||
RAGA | 79.29 | 89.12 | 0.838 | 85.27 | 93.17 | 0.889 | 68.72 | 82.55 | 0.750 | |||
RANM | 90.56 | 94.30 | 0.923 | 90.96 | 94.74 | 0.927 | 77.64 | 85.66 | 0.813 | |||
SEG | 91.56 | 95.2 | 0.933 | 92.26 | 95.44 | 0.937 | 78.55 | 86.53 | 0.822 | |||
SEG( w / 0 BWM) | 91.0 | 94.07 | 0.924 | 92.18 | 95.47 | 0.937 | 78.2 | 86.16 | 0.819 | |||
SEG( w / 0 SL) | 91.53 | 95.10 | 0.932 | 91.77 | 95.8 | 0.936 | 78.04 | 85.89 | 0.817 |
P represents the number of positive samples, while N denotes the number of negative samples corresponding to each positive sample (n_negatives_per_positive). represents the distance of features. The feature vector of the j-th negative sample corresponds to the i-th positive sample, where , and are hyperparameters.
Alignment learning. In this paper, a margin-based loss function is used. Moreover, The entity alignment training optimization in this paper adopts a semi-supervised learning method and expands the training data through an iterative strategy to reduce the dependence on manually labelled seeds and the consumption of workforce and material resources.
(11) |
(12) |
represents the alignment seed positive sample set, and is the same, represents the alignment seed negative sample set,represents the function between two vectors, is a marginal hyperparameter.
IV EXPERIMENTS
This section presents the experimental settings, including experimental datasets, evaluation metrics, model variants, and experimental implementation. The model in this paper is tested on the DBP15K datasets to study further the essential components of this model and their contributions.
IV-A Datasets
DBP-15K is built on DBpedia and includes three cross-language DBpedia datasets: ZH-EN (Chinese-English), JA-EN (Japanese-English), and FR-EN (French-English).
Datasets | Graph basic information statistics | ||
Entity | Relations | Rel.triple | |
JA-EN | 19,814 | 1,299 | 77,214 |
19,780 | 1,153 | 93,484 | |
FR-EN | 19,661 | 903 | 105,998 |
19,663 | 1,208 | 115,722 | |
ZH-EN | 19,661 | 903 | 105,998 |
19,993 | 1,208 | 115,722 | |
′Abbreviated form. |
IV-B Baseline
To assess the performance of our model, we conducted comparisons against current leading entity alignment techniques. Broadly, these can be categorized as follows: Traditional Entity Alignment Methods. SEA proposes a semi-supervised entity alignment method that uses labeled entities and rich unlabeled entity information for alignment. BootEA proposes a Bootstrap-based entity alignment method. Relation-based Entity Alignment Methods. RAGA proposed a framework for capturing entity and relation interactions based on relation-aware graph attention networks. RANM[40] also demonstrates exceptional performance in leveraging relational assistance for entity alignment.
IV-C Evaluation setting
To ensure that the model remains optimal, five-fold cross-validation based on the training set (20%), validation set (10%), and test set (70%) was used, so the reported performance is the average of five independent training runs, and the training/validation/test datasets are shuffled in each round. The number of GAT layers is 2, the similarity threshold of entity embedding is 0.95, and the maximum number of neighbor nodes for alignment seeds is 982. The final training parameters are selected between the best performance period on the validation set (based on ”Hits@1”) and the last period (up to 1500 periods). The experiments were run on NVIDIA 4090 16 GB.
IV-D Ablation Studies
Our model exceeds baselines, as detailed in the table. Ablation studies adjusted parameters like GNN layers and learning rate, showing that the soft labels module enhanced DBP-15K accuracy by 0.3%-1%, especially in the ZH-EN subset with a 1% gain. Additionally, the Bidirectional Weighting module raised DBP-15K accuracy by 0.2%-0.9%, contributing to a consistent 0.7%-1.3% overall improvement across all subsets.
IV-E Robustness Analysis
In assessing our model’s robustness, we tested the impact of GNN layer count and learning rate. While more GNN layers slightly lowered entity alignment accuracy but kept overall performance stable, higher learning rates (0.01, 0.005, 0.001) significantly reduced performance due to training instability and convergence challenges.
V CONCLUSION
This paper presents the SEG framework for cross-lingual entity alignment in knowledge graphs, which enhances entity information through a multi-source fusion mechanism and utilizes semantic auxiliary information from relations to assist in entity alignment. Additionally, it employs a bidirectional weighting mechanism to improve the model’s differential processing of negative samples, complemented by a semi-supervised iterative mechanism that continually expands and strengthens alignment seeds. The robustness of this method has been validated on datasets, and future efforts will concentrate on refining relation representation and entity enhancement techniques to enhance alignment accuracy.
References
- [1] Y. Shou, T. Meng, W. Ai, S. Yang, and K. Li, “Conversational emotion recognition studies based on graph convolutional neural networks and a dependent syntactic analysis,” Neurocomputing, vol. 501, pp. 629–639, 2022.
- [2] Y. Shou, X. Cao, H. Liu, and D. Meng, “Masked contrastive graph representation learning for age estimation,” Pattern Recognition, vol. 158, p. 110974, 2025.
- [3] Y. Shou, T. Meng, W. Ai, C. Xie, H. Liu, and Y. Wang, “Object detection in medical images based on hierarchical transformer and mask mechanism,” Computational Intelligence and Neuroscience, vol. 2022, no. 1, p. 5863782, 2022.
- [4] Y. Shou, T. Meng, W. Ai, N. Yin, and K. Li, “A comprehensive survey on multi-modal conversational emotion recognition with deep learning,” arXiv preprint arXiv:2312.05735, 2023.
- [5] Y. Shou, T. Meng, W. Ai, F. Zhang, N. Yin, and K. Li, “Adversarial alignment and graph fusion via information bottleneck for multimodal emotion recognition in conversations,” Information Fusion, vol. 112, p. 102590, 2024.
- [6] Y. Shou, X. Cao, D. Meng, B. Dong, and Q. Zheng, “A low-rank matching attention based cross-modal feature fusion method for conversational emotion recognition,” arXiv preprint arXiv:2306.17799, 2023.
- [7] T. Meng, Y. Shou, W. Ai, N. Yin, and K. Li, “Deep imbalanced learning for multimodal emotion recognition in conversations,” IEEE Transactions on Artificial Intelligence, 2024.
- [8] Y. Shou, T. Meng, W. Ai, N. Yin, and K. Li, “Adversarial representation with intra-modal and inter-modal graph contrastive learning for multimodal emotion recognition,” arXiv preprint arXiv:2312.16778, 2023.
- [9] S. Auer, C. Bizer, G. Kobilarov, J. Lehmann, R. Cyganiak, and Z. Ives, “Dbpedia: A nucleus for a web of open data,” in International Semantic Web Conference. Springer, 2007, pp. 722–735.
- [10] F. M. Suchanek, G. Kasneci, and G. Weikum, “Yago: a core of semantic knowledge,” in Proceedings of the 16th International Conference on World Wide Web, 2007, pp. 697–706.
- [11] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, “Translating embeddings for modeling multi-relational data,” Advances in neural information processing systems, vol. 26, 2013.
- [12] W. Ai, Y. Shou, T. Meng, N. Yin, and K. Li, “Der-gcn: Dialogue and event relation-aware graph convolutional neural network for multimodal dialogue emotion recognition,” arXiv preprint arXiv:2312.10579, 2023.
- [13] W. Ai, Y. Shou, T. Meng, and K. Li, “Der-gcn: Dialog and event relation-aware graph convolutional neural network for multimodal dialog emotion recognition,” IEEE Transactions on Neural Networks and Learning Systems, 2024.
- [14] T. Meng, Y. Shou, W. Ai, J. Du, H. Liu, and K. Li, “A multi-message passing framework based on heterogeneous graphs in conversational emotion recognition,” Neurocomputing, vol. 569, p. 127109, 2024.
- [15] Y. Shou, H. Lan, and X. Cao, “Contrastive graph representation learning with adversarial cross-view reconstruction and information bottleneck,” arXiv preprint arXiv:2408.00295, 2024.
- [16] Y. Shou, X. Cao, and D. Meng, “Spegcl: Self-supervised graph spectrum contrastive learning without positive samples,” arXiv preprint arXiv:2410.10365, 2024.
- [17] Y. Shou, W. Ai, J. Du, T. Meng, and H. Liu, “Efficient long-distance latent relation-aware graph neural network for multi-modal emotion recognition in conversations,” arXiv preprint arXiv:2407.00119, 2024.
- [18] R. Ying, Y. Shou, and C. Liu, “Prediction model of dow jones index based on lstm-adaboost,” in 2021 International Conference on Communications, Information System and Computer Engineering (CISCE). IEEE, 2021, pp. 808–812.
- [19] Z. Wang, Q. Lv, X. Lan, and Y. Zhang, “Cross-lingual knowledge graph alignment via graph convolutional networks,” in Proceedings of the 2018 conference on empirical methods in natural language processing, 2018, pp. 349–357.
- [20] Y. Shou, W. Ai, T. Meng, and N. Yin, “Graph information bottleneck for remote sensing segmentation,” arXiv preprint arXiv:2312.02545, 2023.
- [21] Y. Shou, W. Ai, T. Meng, and K. Li, “Czl-ciae: Clip-driven zero-shot learning for correcting inverse age estimation,” arXiv preprint arXiv:2312.01758, 2023.
- [22] T. Meng, F. Zhang, Y. Shou, H. Shao, W. Ai, and K. Li, “Masked graph learning with recurrent alignment for multimodal emotion recognition in conversation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2024.
- [23] Y. Shou, T. Meng, F. Zhang, N. Yin, and K. Li, “Revisiting multi-modal emotion learning with broad state space models and probability-guidance fusion,” arXiv preprint arXiv:2404.17858, 2024.
- [24] W. Ai, Y. Wei, H. Shao, Y. Shou, T. Meng, and K. Li, “Edge-enhanced minimum-margin graph attention network for short text classification,” Expert Systems with Applications, vol. 251, p. 124069, 2024.
- [25] Y. Zhang, Y. Shou, T. Meng, W. Ai, and K. Li, “A multi-view mask contrastive learning graph convolutional neural network for age estimation,” Knowledge and Information Systems, pp. 1–26, 2024.
- [26] Y. Zhu, H. Liu, Z. Wu, and Y. Du, “Relation-aware neighborhood matching model for entity alignment,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 5, 2021, pp. 4749–4756.
- [27] W. Ai, F. Zhang, T. Meng, Y. Shou, H. Shao, and K. Li, “A two-stage multimodal emotion recognition model based on graph contrastive learning,” in 2023 IEEE 29th International Conference on Parallel and Distributed Systems (ICPADS). IEEE, 2023, pp. 397–404.
- [28] T. Meng, F. Zhang, Y. Shou, W. Ai, N. Yin, and K. Li, “Revisiting multimodal emotion recognition in conversation from the perspective of graph spectrum,” arXiv preprint arXiv:2404.17862, 2024.
- [29] W. Ai, W. Deng, H. Chen, J. Du, T. Meng, and Y. Shou, “Mcsff: Multi-modal consistency and specificity fusion framework for entity alignment,” arXiv preprint arXiv:2410.14584, 2024.
- [30] Y. Shou, W. Ai, T. Meng, F. Zhang, and K. Li, “Graphunet: Graph make strong encoders for remote sensing segmentation,” in 2023 IEEE 29th International Conference on Parallel and Distributed Systems (ICPADS). IEEE, 2023, pp. 2734–2737.
- [31] Z. Zhang, Y. Yang, and B. Chen, “Relation-aware heterogeneous graph neural network for entity alignment,” Neurocomputing, vol. 592, p. 127797, 2024.
- [32] W. Ai, J. Li, Z. Wang, J. Du, T. Meng, Y. Shou, and K. Li, “Graph contrastive learning via cluster-refined negative sampling for semi-supervised text classification,” arXiv preprint arXiv:2410.18130, 2024.
- [33] Z. Sun, W. Hu, Q. Zhang, and Y. Qu, “Bootstrapping entity alignment with knowledge graph embedding.” in IJCAI, vol. 18, no. 2018, 2018.
- [34] Y. Cao, Z. Liu, C. Li, J. Li, and T.-S. Chua, “Multi-channel graph neural network for entity alignment,” arXiv preprint arXiv:1908.09898, 2019.
- [35] C. Chen, K. Li, S. G. Teo, X. Zou, K. Wang, J. Wang, and Z. Zeng, “Gated residual recurrent graph neural networks for traffic prediction,” in Proceedings of the AAAI conference on artificial intelligence, vol. 33, no. 01, 2019, pp. 485–492.
- [36] C. Chen, K. Li, S. G. Teo, X. Zou, K. Li, and Z. Zeng, “Citywide traffic flow prediction based on multiple gated spatio-temporal convolutional neural networks,” ACM Transactions on Knowledge Discovery from Data (TKDD), vol. 14, no. 4, pp. 1–23, 2020.
- [37] M. Chen, Y. Tian, M. Yang, and C. Zaniolo, “Multilingual knowledge graph embeddings for cross-lingual knowledge alignment,” arXiv preprint arXiv:1611.03954, 2016.
- [38] H.-W. Yang, Y. Zou, P. Shi, W. Lu, J. Lin, and X. Sun, “Aligning cross-lingual entities with multi-aspect information,” arXiv preprint arXiv:1910.06575, 2019.
- [39] Y. Wu, X. Liu, Y. Feng, Z. Wang, R. Yan, and D. Zhao, “Relation-aware entity alignment for heterogeneous knowledge graphs,” arXiv preprint arXiv:1908.08210, 2019.
- [40] W. Cai, W. Ma, L. Wei, and Y. Jiang, “Semi-supervised entity alignment via relation-based adaptive neighborhood matching,” IEEE Transactions on Knowledge and Data Engineering, vol. 35, no. 8, pp. 8545–8558, 2022.