This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

SEG:Seeds-Enhanced Iterative Refinement Graph Neural Network for Entity Alignment

Wei Ai College of Computer and Mathematics
Central South University of Forestry and Technology
ChangSha, China
[email protected]
   Yinghui Gao College of Computer and Mathematics
Central South University of Forestry and Technology
ChangSha, China
[email protected]
                  Jianbin Li             College of Computer and Mathematics
            Central South University of Forestry and Technology
            ChangSha, China
            [email protected]
   Jiayi Du College of Computer and Mathematics
Central South University of Forestry and Technology
ChangSha, China
[email protected]
                  Tao Meng             College of Computer and Mathematics
            Central South University of Forestry and Technology
            ChangSha, China
            [email protected]
                  Yuntao Shou             College of Computer and Mathematics
            Central South University of Forestry and Technology
            ChangSha, China
            [email protected]
         Keqin Li This work is supported by National Natural Science Foundation of China (Grant No. 69189338, Grant No. 62372478), Excellent Young Scholars of Hunan Province of China (Grant No. 22B0275), Changsha Natural Science Foundation (Grant No. kq2202294), and program of Research on Local Community Structure Detection Algorithms in Complex Networks (Grant No. 2020YJ009) * is the corresponding author.          Department of Computer Science
         State University of New York
         New Paltz, New York 12561, USA
            [email protected]
Abstract

Entity alignment is crucial for merging knowledge across knowledge graphs, as it matches entities with identical semantics. The standard method matches these entities based on their embedding similarities using semi-supervised learning. However, diverse data sources lead to non-isomorphic neighborhood structures for aligned entities, complicating alignment, especially for less common and sparsely connected entities. This paper presents a soft label propagation framework that integrates multi-source data and iterative seed enhancement, addressing scalability challenges in handling extensive datasets where scale computing excels. The framework uses seeds for anchoring and selects optimal relationship pairs to create soft labels rich in neighborhood features and semantic relationship data. A bidirectional weighted joint loss function is implemented, which reduces the distance between positive samples and differentially processes negative samples, taking into account the non-isomorphic neighborhood structures. Our method outperforms existing semi-supervised approaches, as evidenced by superior results on multiple datasets, significantly improving the quality of entity alignment.

Index Terms:
Entity Alignment; GCN; Entity embedding;

I INTRODUCTION

Refer to caption
Figure 1: The difference in alignment for entities and relationships with non-isomorphic neighborhoods.

Knowledge graph is a large-scale structured semantic knowledge base used to describe concepts in the physical world and their interrelationships, aggregate information and data, and link relationships on the Web into knowledge [1, 2, 3, 4, 5, 6, 7, 8]. Knowledge graphs contain rich information such as entity attributes, semantic relationships, and entity types. There are many large knowledge graphs, such as DBpedia [9]and YAGO[10], which have been widely used in natural language processing, information retrieval, and intelligent question answering. Due to the frequent inclusion of many entities and relationships in knowledge graphs, there is also a close connection with scale computing. However, knowledge graphs are inherently incomplete. If downstream applications need support, integrating various KGs for knowledge fusion effectively solves this problem. Researchers have proposed many entity alignment methods. People have gradually shifted from traditional TransE [11, 12, 13, 14, 15, 16, 17, 18] methods to deep learning-based modeling of the graph structure of knowledge graphs. The methods based on deep learning [19, 20, 21, 22, 23, 24, 25] mainly follow the learning embedding-similarity calculation-greedy matching mode, which is to use seed entities to automatically extract implicit features of entities from different KGs using neural networks and then embed them into the same implicit space and calculate the similarity between different entity embeddings. However, in actual knowledge graphs, due to the heterogeneity of data sources, entities usually exhibit non-isomorphic neighborhood structures, and people have begun to realize that more is needed to consider entity embeddings for alignment. [26, 27, 28, 29, 30] considered the positive interaction between entity alignment and relationships in a semi-supervised context and mined relational information to assist in alignment. [31, 32] considers the simultaneous roles of entities and relationships. Although these scholars have attached varying degrees of importance to the importance of relationships in alignment, most methods still ignore two issues:
First, the difference between entities and relations is ignored. In the mainstream methods, embedding-similarity-matching, similarity methods can also be divided into two types: based on equivalent entities having similar attributes[33] and based on equivalent entities having similar adjacent entities. However, in these methods, the propagation of alignment information focuses more on the propagation from the similarity of entities to relations. Not only does it largely ignore the adequate auxiliary alignment information of the relational semantics itself, but there is also a problem of information loss. More importantly, it ignores the differences between entities and relations, which leads to inconsistent information. As shown in Fig 1, if only considering the propagation of similarity in entities, then Alan Turing and Alan Turing, Stephen Hawking, and Stephen Hawking can be fully aligned, but in relations, alma-mater, and graduate, student and worker cannot be aligned. GCN-based frameworks also use relations as weights[34] or information. However, noise is inevitably mixed in when the same model is used to train the representation of entities and relations jointly. The integration of relations into entity representation will cause the entity representation to be blurred and smooth.
Second, there is an obvious problem in current research: over-emphasis on reducing the distance between positive samples while ignoring the distance between negative samples, especially when facing pseudo-aligned entities with high similarity; there needs to be more differentiated treatment of these pseudo-entities. In cross-knowledge graph alignment, effectively dealing with the non-isomorphic neighborhood differences between positive and negative samples is a significant challenge. In order to overcome the limitations of existing methods, this paper proposes a soft label propagation framework based on multi-source fusion combined with an iterative process of seed enhancement. First, we conduct dual-angle modeling based on entities and relations and fuse the best-matching relationship pairs selected in both entity and relationship modes to generate soft labels. These soft labels contain rich entity neighbor features and integrate auxiliary information of relationship semantics, avoiding the problem of auxiliary information loss. Secondly, the soft labels fused by multi-source information propagate alignment information, and the topological structure and neighborhood differences are effectively captured. On this basis, a bidirectional adaptive weighted joint loss function is introduced, which not only shortens the distance between positive samples through commonalities but also widens the distance between high-similarity negative samples to varying degrees and more comprehensively considers the commonalities and differences in the alignment process. Some contributions are as follows:

  • Obtaining soft labels from the perspectives of entities and relations can reveal shared entity features and mine the rich information of the relationship semantics itself, thereby improving alignment accuracy.

  • The difference problem of non-isomorphic neighborhoods is solved by effectively shortening the distance between positive samples and using a weighted joint loss function to differentiate and distance negative samples.

  • Our method, tested on real datasets, significantly boosts model performance, proving highly feasible and valuable for entity alignment across knowledge graphs.

The remainder of this paper is organized as follows: Section II reviews related work, Section III provides a detailed description of the proposed method, Section IV presents the experimental setup and results, and finally, Section V offers a brief conclusion.

Refer to caption
Figure 2: The entity alignment approach of SEG, as outlined in the framework diagram, encompasses four principal components: Graph representation learning, Soft Labels, Pruning Based on Soft Labels, and Bidirectional Weighting Mechanism.

II RELATED WORK

In this part, we examine existing scholarly contributions about entity alignment, encompassing conventional translation-based tactics as well as contemporary graph representational techniques that leverage Graph Convolutional Networks (GCNs)[35][36].

II-A Conventional Approaches Utilizing Translation Techniques

Translation-based entity alignment. By continuously adjusting the vector representation, the entity vector + the relation vector = the entity vector. In order to solve the shortcomings of TransE. MTransE[37] embeds entities and relations in different knowledge graphs into different vector spaces and maps them to a unified space through distance calibration and linear transformation strategies to achieve cross-knowledge graph alignment. BootEA[33] is a Bootstrap-based entity alignment method that converts entity alignment into a classification problem. It completes the alignment task by iteratively marking possible entity alignments as training data and finding the maximum possibility of alignment.

II-B GCN-Driven Graph Representation Techniques

GNN-based entity alignment. GCN-align can embed entities from different KGs into a unified vector space. Embedding can be learned from the structure and attribute information of the entity. HMAN[38] embeds learning entity representation from three perspectives: structural information, relationship features, and attribute characteristics.RD-GCN[39] uses the graph attention mechanism to encourage dual relations between the interaction graph and the original graph and then feeds the vertex representation in the original graph into the GCN layer with highway neural network gating to capture the structural information of the neighbors.

II-C Semi-supervised entity alignment

III PROPOSED METHOD

We introduce the SEG methodology, whose structural outline is depicted in Fig 2. Entity encoding and alignment candidate generation.Soft label screening: The soft labels of the two modalities are fused, and the screening strategy is used to obtain the final accurate soft labels.Pruning when performing neighborhood aggregation based on soft labels: Propagate the alignment relationship in the soft labels to the neighborhood structure of the candidate alignment.Bidirectional weighted joint mechanism: Use a weighted method to differentiate negative samples to guide positive and negative sample learning between two knowledge graphs.Bidirectional iterative collaborative training.
In this study, we consider the scenario where two distinct knowledge graphs are provided as inputs. A knowledge graph is structured as KG=(E,R,T)KG=(E,R,T), where EE denotes the set of entities, RR represents the set of relations, and TT is a collection of triples defined as {(e1,r,e2)|e1,e2E,rR}\{(e_{1},r,e_{2})|e_{1},e_{2}\in E,r\in R\}. The source knowledge graph is denoted by kG1=(E1,R1,T1)kG_{1}=(E^{1},R^{1},T^{1}); similarly, the target knowledge graph is denoted by KG2=(E2,R2,T2)KG_{2}=(E^{2},R^{2},T^{2}). The set of seed alignment entities is given by Alignseed={(u,v)|uE1,vE2,uv}Align_{seed}=\{(u,v)|u\in E^{1},v\in E^{2},u\leftrightarrow v\}, indicating that entity uu from KG1KG_{1} and entity vv from KG2KG_{2} are identical, with \leftrightarrow signifying equivalence. To initiate the process, we employ a pre-trained model to assign initial embeddings to the entities.

III-A Graph representation learning

The Graph Attention Network serves as a feature extraction mechanism, introducing an innovative neural network framework for processing data structured as a graph. This architecture comprises an input layer, an intermediary hidden layer, and a final output layer. Within each hidden layer, the representation of a node is determined by its interaction with the surrounding nodes’ information, necessitating ongoing updates to the hidden layer’s state. The computation is based on the following formula:

Hei(l+1)=Hei(l)+d(l)σ(HejNiaij(l)Hej(l)),H_{e_{i}}^{(l+1)}=H_{e_{i}}^{(l)}+d^{(l)}\cdot\sigma(\sum_{H_{e_{j}}\in N_{i}}a_{ij}^{(l)}H_{e_{j}}^{(l)}), (1)

where Hei(l+1)H_{e_{i}}^{(l+1)} and Hei(l)H_{e_{i}}^{(l)} represent the embedding of the entity at the ll layer, ll represents the number of layers, and d(l)d^{(l)} represents the weight parameter of the ll layer,σ\sigma represents the ReLU activation function.

aij=exp(εSij)HekN(Hei)exp(εSik)a_{ij}=\frac{\exp(\varepsilon S_{ij})}{\sum_{H_{e_{k}}\in N(H_{e_{i}})}\exp(\varepsilon S_{ik})} (2)
Sij=(Hei,Rt,Hej)TijaT([Hei||Hej]Rt)S_{ij}=\sum_{(H_{e_{i}},R_{t},H_{e_{j}})\in T_{ij}}a^{T}([H_{e_{i}}||H_{e_{j}}]\odot R_{t}) (3)

where TijT_{ij} represents the set of relation triplets with eie_{i} as the head entity and eje_{j} as the tail entity, α\alpha is the attention parameter, |||| are vector concatenation, \odot element multiplication and LeakyReLU respectively function.

Rt=σ[HeiTtbTHei|Tt|HejWtbTHej|Wt|],\begin{split}R_{t}=\sigma\left[\frac{\sum_{H_{e_{i}}\in T_{t}}b^{T}H_{e_{i}}}{|T_{t}|}\bigg{\|}\frac{\sum_{H_{e_{j}}\in W_{t}}b^{T}H_{e_{j}}}{|W_{t}|}\right],\end{split} (4)

Where TtT_{t} and WtW_{t} represent the size of the head entity set and the tail entity set connected by RtR_{t}, respectively, bb is the attention parameter.

Hei=Wnn(Hei)+bn.H_{e_{i}}=W^{n}\cdot n(H_{e_{i}})+b^{n}. (5)

ww and bb are trainable weight matrices, and their output matrices are represented as HeiH_{e_{i}}.

Alignment candidates. After obtaining the embedding representation of the entity through GATGAT, the embedding representation of the two heterogeneous knowledge graphs is obtained respectively. E1={e11,e21,e31,ei1},E1KG1,E2={e12,e22e32,,ej2},E2KG2,E^{1}=\{e_{1}^{1},e_{2}^{1},e_{3}^{1}\ldots,e_{i}^{1}\},E^{1}\in KG_{1},E^{2}=\{e_{1}^{2},e_{2}^{2}e_{3}^{2},\ldots,e_{j}^{2}\},E^{2}\in KG_{2},, and then the candidate entity pairs need to be matched, that is, ei1ej2e_{i}^{1}\leftrightarrow e_{j}^{2}, cosine similarity is used to evaluate the similarity between entities. The similarity calculation results of all embeddings in E1E^{1} and E1E^{1} can be calculated according to the cosine similarity, and the calculation results are saved in the similarity matrix cosine_distances. The calculation formula is as follows:

Sim(ei,ej)=iKG1,jKG2neiejeiej,Sim_{(e_{i},e_{j})}=\sum_{i\in KG_{1},j\in KG_{2}}^{n}\frac{e_{i}\cdot e_{j}}{\|e_{i}\|\cdot\|e_{j}\|}, (6)

Sim[1,1]Sim\in[-1,1], Only the similarity in the matrix >=>= 0.95 is selected as the Candidate pre-selected set. However, these candidate sets only represent possible matching entity pairs, and there may even be one-to-many situations that cannot guarantee the accuracy of the final alignment.

III-B Soft Labels

Based on Entity propagate similarity. taking the aligned seeds of KG1KG_{1} and KG2KG_{2} as anchors, for any seed pair (ea1,eb2)Alignentity(e_{a}^{1},e_{b}^{2})\in Align_{entity}, we construct the neighbourhood node sets of ea1e_{a}^{1} and eb2e_{b}^{2}, n(e¯i1)|e¯i1N(ea1){n(\bar{e}_{i}^{1})|\bar{e}_{i}^{1}\in N(e_{a}^{1})} and n(e¯j2)|e¯j2N(eb2){n(\bar{e}_{j}^{2})|\bar{e}_{j}^{2}\in N(e_{b}^{2})}, (ea1,ra1,e¯11)(e_{a}^{1},r_{a1},\bar{e}_{1}^{1}) or (e¯11,ra1,ea1)(\bar{e}_{1}^{1},r_{a1},e_{a}^{1}). n(e¯i1)n(\bar{e}_{i}^{1}) represents the ii-th neighbor node of ea1e_{a}^{1}, and ra1r_{a1} represents the connection between the two. e¯j2\bar{e}_{j}^{2} represents the jj-th neighbor node of eb2e_{b}^{2}, (eb2,rb1,e¯12)(e_{b}^{2},r_{b1},\bar{e}_{1}^{2}), (e¯12,rb1,eb2)(\bar{e}_{1}^{2},r_{b1},e_{b}^{2}), rb1r_{b1} is the same. Here, the above cosine similarity is also used to calculate the similarity of the nodes (ea1,eb2)(e_{a}^{1},e_{b}^{2}) in the neighborhood of the seed pair. In order to ensure that the relationship pairs screened by the neighborhood candidate alignment set are highly accurate and convenient for subsequent training, two standards are set here: First, the neighborhood alignment nodes need to meet the set SimeSim_{e} similarity threshold and the MatcheMatch_{e} matching number threshold of the derived relationship. Second, we must ensure that the neighborhood nodes are aligned one-to-one. Only the neighborhood pairs with the highest similarity are selected when encountering a one-to-many or many-to-one situation.

Sum1={Sime>=0.98}+{Matche>=10},Sum_{1}=\{Sim_{e}>=0.98\}+\{Match_{e}>=10\}, (7)

Sum1Sum_{1} represents the soft label set screened based on the entity embedding model, SimeSim_{e} represents the cosine similarity of the neighborhood node pair, and match is the number of repetitions of the relationship between the entities. For example: (ea1,eb2)(ea1,ra1,e¯11),(eb2,rb1,e¯12)Sim(e¯11,e¯12)>=0.98(e_{a}^{1},e_{b}^{2})\to(e_{a}^{1},r_{a1},\bar{e}_{1}^{1}),\quad(e_{b}^{2},r_{b1},\bar{e}_{1}^{2})\to\text{Sim}(\bar{e}_{1}^{1},\bar{e}_{1}^{2})>=0.98 and Match(ra1,rb1)>=10\text{Match}(r_{a1},r_{b1})>=10,that is, keep this relationship pair in the set Sum1Sum_{1}.

Based on relation propagate similarity. In this section, the Bert model is introduced to encode the relations of the KG graph, and the semantic features of the relations are fully captured using the name and description of the relations.

Hr=Bert(RD),H_{r}=\text{Bert}(RD), (8)

RD: Relation Description. Similarly, setting the cosine similarity and matching threshold also requires one-to-one matching.

Sum2={Simr>=0.98}+{Matchr>=600},Sum_{2}=\{Sim_{r}>=0.98\}+\{Match_{r}>=600\}, (9)

Sum2Sum_{2} represents the number of soft labels selected based on the relation embedding model, SimrSim_{r} represents the cosine similarity of the relation pair, and MatchrMatch_{r} is the number of times this relation pair appears repeatedly. Sum=Sum1+Sum2,Order(Sum2)>Order(Sum1)\text{Sum}=\text{Sum}_{1}+\text{Sum}_{2},\quad\text{Order}(\text{Sum}_{2})>\text{Order}({Sum}_{1}), Sum2{Sum}_{2}has a higher priority than Sum2{Sum}_{2}.

III-C Pruning Based on Soft Labels

Soft labels enhance alignment precision by guiding neighborhood matching and are selectively retained based on match quality. GCN integrates this neighborhood data, while Highway networks combine entity names with neighborhood information to refine the final embeddings.

III-D Bidirectional Weighting Mechanism

Inspired by contrastive learning, we have engineered a bidirectional weighting mechanism based on similarity aimed at optimizing the weight distribution of samples during the entity alignment process. This mechanism adjusts the weights of negative samples differentially to achieve a precise approximation of positive samples and effectively optimize negative samples. Positive samples are actual alignments AlignentityAlign_{entity}, while negative samples are selected from the top 50 nodes with the highest similarity to the positive samples in either KG1KG_{1} or KG2KG_{2}. Notably, due to the most significant risk of misalignment, the node with the highest similarity is assigned the highest weight penalty.

Lossw=i=1Pj=1Nβexp(γj1K1)𝒟ij,Loss_{w}=\sum_{i=1}^{P}\sum_{j=1}^{N}\beta\exp\left(-\gamma\frac{j-1}{K-1}\right)\cdot\mathcal{D}_{ij}, (10)
TABLE I: Ablation Study and Comparative and Proprietary Method Performance on DBP15K Language Datasets using Hit@1, Hit@5, and MRR Metrics.
Method DBPJA-EN DBPFR-EN DBPZH-EN
Hit@1 Hit@5 MRR Hit@1 Hit@5 MRR Hit@1 Hit@5 MRR
SEA 29.12 55.64 0.416 32.64 60.28 0.455 34.56 59.54 0.464
BootEA 52.71 71.89 0.616 57.61 77.27 0.666 55.45 73.72 0.639
RAGA 79.29 89.12 0.838 85.27 93.17 0.889 68.72 82.55 0.750
RANM 90.56 94.30 0.923 90.96 94.74 0.927 77.64 85.66 0.813
SEG 91.56 95.2 0.933 92.26 95.44 0.937 78.55 86.53 0.822
SEG( w / 0 BWM) 91.0 94.07 0.924 92.18 95.47 0.937 78.2 86.16 0.819
SEG( w / 0 SL) 91.53 95.10 0.932 91.77 95.8 0.936 78.04 85.89 0.817

P represents the number of positive samples, while N denotes the number of negative samples corresponding to each positive sample (n_negatives_per_positive).DijD_{ij} represents the distance of features. The feature vector of the j-th negative sample corresponds to the i-th positive sample, where j{1,,K}j\in\{1,\ldots,K\},γ\gamma and β\beta are hyperparameters.

Alignment learning. In this paper, a margin-based loss function is used. Moreover, The entity alignment training optimization in this paper adopts a semi-supervised learning method and expands the training data through an iterative strategy to reduce the dependence on manually labelled seeds and the consumption of workforce and material resources.

Lossm\displaystyle Loss_{m} =(e1,e2)P,(e1,e2)N[D(e1,e2)+γD(e1,e2)]+\displaystyle=\sum_{(e^{1},e^{2})\in P,(e^{1^{\prime}},e^{2^{\prime}})\in N}\left[D(e^{1},e^{2})+\gamma-D\left(e^{1},e^{2^{\prime}}\right)\right]_{+}
+(e1,e2)P,(e1,e2)N[D(e1,e2)+γD(e1,e2)]+\displaystyle+\sum_{(e^{1},e^{2})\in P,(e^{1^{\prime}},e^{2^{\prime}})\in N}\left[D(e^{1},e^{2})+\gamma-D\left(e^{1^{\prime}},e^{2}\right)\right]_{+} (11)
Loss=Lossw+LossmLoss=Loss_{w}+Loss_{m} (12)

[x]+=max{0,x},(e1,e2)P[x]_{+}=\max\{0,x\},\quad(e^{1},e^{2})\in P represents the alignment seed positive sample set, and e1E1,E1KG1,e2e^{1}\in E^{1},\quad E^{1}\in KG_{1},\quad e^{2} is the same,(e1,e2)N\quad(e^{1^{\prime}},e^{2^{\prime}})\in N represents the alignment seed negative sample set,D(e1,e2)=e1e2L1\quad D(e^{1},e^{2})=\|e^{1}-e^{2}\|_{L_{1}}represents the L1L_{1} function between two vectors,γ>0\quad\gamma>0 is a marginal hyperparameter.

IV EXPERIMENTS

This section presents the experimental settings, including experimental datasets, evaluation metrics, model variants, and experimental implementation. The model in this paper is tested on the DBP15K datasets to study further the essential components of this model and their contributions.

IV-A Datasets

DBP-15K is built on DBpedia and includes three cross-language DBpedia datasets: ZH-EN (Chinese-English), JA-EN (Japanese-English), and FR-EN (French-English).

TABLE II: Statistic dates
Datasets Graph basic information statistics
Entity Relations Rel.triple
JA-EN 19,814 1,299 77,214
19,780 1,153 93,484
FR-EN 19,661 903 105,998
19,663 1,208 115,722
ZH-EN 19,661 903 105,998
19,993 1,208 115,722
Abbreviated form.

IV-B Baseline

To assess the performance of our model, we conducted comparisons against current leading entity alignment techniques. Broadly, these can be categorized as follows: Traditional Entity Alignment Methods. SEA proposes a semi-supervised entity alignment method that uses labeled entities and rich unlabeled entity information for alignment. BootEA proposes a Bootstrap-based entity alignment method. Relation-based Entity Alignment Methods. RAGA proposed a framework for capturing entity and relation interactions based on relation-aware graph attention networks. RANM[40] also demonstrates exceptional performance in leveraging relational assistance for entity alignment.

IV-C Evaluation setting

To ensure that the model remains optimal, five-fold cross-validation based on the training set (20%), validation set (10%), and test set (70%) was used, so the reported performance is the average of five independent training runs, and the training/validation/test datasets are shuffled in each round. The number of GAT layers is 2, the similarity threshold of entity embedding is 0.95, and the maximum number of neighbor nodes for alignment seeds is 982. The final training parameters are selected between the best performance period on the validation set (based on ”Hits@1”) and the last period (up to 1500 periods). The experiments were run on NVIDIA 4090 16 GB.

IV-D Ablation Studies

Our model exceeds baselines, as detailed in the table. Ablation studies adjusted parameters like GNN layers and learning rate, showing that the soft labels module enhanced DBP-15K accuracy by 0.3%-1%, especially in the ZH-EN subset with a 1% gain. Additionally, the Bidirectional Weighting module raised DBP-15K accuracy by 0.2%-0.9%, contributing to a consistent 0.7%-1.3% overall improvement across all subsets.

IV-E Robustness Analysis

In assessing our model’s robustness, we tested the impact of GNN layer count and learning rate. While more GNN layers slightly lowered entity alignment accuracy but kept overall performance stable, higher learning rates (0.01, 0.005, 0.001) significantly reduced performance due to training instability and convergence challenges.

V CONCLUSION

This paper presents the SEG framework for cross-lingual entity alignment in knowledge graphs, which enhances entity information through a multi-source fusion mechanism and utilizes semantic auxiliary information from relations to assist in entity alignment. Additionally, it employs a bidirectional weighting mechanism to improve the model’s differential processing of negative samples, complemented by a semi-supervised iterative mechanism that continually expands and strengthens alignment seeds. The robustness of this method has been validated on datasets, and future efforts will concentrate on refining relation representation and entity enhancement techniques to enhance alignment accuracy.

References

  • [1] Y. Shou, T. Meng, W. Ai, S. Yang, and K. Li, “Conversational emotion recognition studies based on graph convolutional neural networks and a dependent syntactic analysis,” Neurocomputing, vol. 501, pp. 629–639, 2022.
  • [2] Y. Shou, X. Cao, H. Liu, and D. Meng, “Masked contrastive graph representation learning for age estimation,” Pattern Recognition, vol. 158, p. 110974, 2025.
  • [3] Y. Shou, T. Meng, W. Ai, C. Xie, H. Liu, and Y. Wang, “Object detection in medical images based on hierarchical transformer and mask mechanism,” Computational Intelligence and Neuroscience, vol. 2022, no. 1, p. 5863782, 2022.
  • [4] Y. Shou, T. Meng, W. Ai, N. Yin, and K. Li, “A comprehensive survey on multi-modal conversational emotion recognition with deep learning,” arXiv preprint arXiv:2312.05735, 2023.
  • [5] Y. Shou, T. Meng, W. Ai, F. Zhang, N. Yin, and K. Li, “Adversarial alignment and graph fusion via information bottleneck for multimodal emotion recognition in conversations,” Information Fusion, vol. 112, p. 102590, 2024.
  • [6] Y. Shou, X. Cao, D. Meng, B. Dong, and Q. Zheng, “A low-rank matching attention based cross-modal feature fusion method for conversational emotion recognition,” arXiv preprint arXiv:2306.17799, 2023.
  • [7] T. Meng, Y. Shou, W. Ai, N. Yin, and K. Li, “Deep imbalanced learning for multimodal emotion recognition in conversations,” IEEE Transactions on Artificial Intelligence, 2024.
  • [8] Y. Shou, T. Meng, W. Ai, N. Yin, and K. Li, “Adversarial representation with intra-modal and inter-modal graph contrastive learning for multimodal emotion recognition,” arXiv preprint arXiv:2312.16778, 2023.
  • [9] S. Auer, C. Bizer, G. Kobilarov, J. Lehmann, R. Cyganiak, and Z. Ives, “Dbpedia: A nucleus for a web of open data,” in International Semantic Web Conference.   Springer, 2007, pp. 722–735.
  • [10] F. M. Suchanek, G. Kasneci, and G. Weikum, “Yago: a core of semantic knowledge,” in Proceedings of the 16th International Conference on World Wide Web, 2007, pp. 697–706.
  • [11] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, “Translating embeddings for modeling multi-relational data,” Advances in neural information processing systems, vol. 26, 2013.
  • [12] W. Ai, Y. Shou, T. Meng, N. Yin, and K. Li, “Der-gcn: Dialogue and event relation-aware graph convolutional neural network for multimodal dialogue emotion recognition,” arXiv preprint arXiv:2312.10579, 2023.
  • [13] W. Ai, Y. Shou, T. Meng, and K. Li, “Der-gcn: Dialog and event relation-aware graph convolutional neural network for multimodal dialog emotion recognition,” IEEE Transactions on Neural Networks and Learning Systems, 2024.
  • [14] T. Meng, Y. Shou, W. Ai, J. Du, H. Liu, and K. Li, “A multi-message passing framework based on heterogeneous graphs in conversational emotion recognition,” Neurocomputing, vol. 569, p. 127109, 2024.
  • [15] Y. Shou, H. Lan, and X. Cao, “Contrastive graph representation learning with adversarial cross-view reconstruction and information bottleneck,” arXiv preprint arXiv:2408.00295, 2024.
  • [16] Y. Shou, X. Cao, and D. Meng, “Spegcl: Self-supervised graph spectrum contrastive learning without positive samples,” arXiv preprint arXiv:2410.10365, 2024.
  • [17] Y. Shou, W. Ai, J. Du, T. Meng, and H. Liu, “Efficient long-distance latent relation-aware graph neural network for multi-modal emotion recognition in conversations,” arXiv preprint arXiv:2407.00119, 2024.
  • [18] R. Ying, Y. Shou, and C. Liu, “Prediction model of dow jones index based on lstm-adaboost,” in 2021 International Conference on Communications, Information System and Computer Engineering (CISCE).   IEEE, 2021, pp. 808–812.
  • [19] Z. Wang, Q. Lv, X. Lan, and Y. Zhang, “Cross-lingual knowledge graph alignment via graph convolutional networks,” in Proceedings of the 2018 conference on empirical methods in natural language processing, 2018, pp. 349–357.
  • [20] Y. Shou, W. Ai, T. Meng, and N. Yin, “Graph information bottleneck for remote sensing segmentation,” arXiv preprint arXiv:2312.02545, 2023.
  • [21] Y. Shou, W. Ai, T. Meng, and K. Li, “Czl-ciae: Clip-driven zero-shot learning for correcting inverse age estimation,” arXiv preprint arXiv:2312.01758, 2023.
  • [22] T. Meng, F. Zhang, Y. Shou, H. Shao, W. Ai, and K. Li, “Masked graph learning with recurrent alignment for multimodal emotion recognition in conversation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2024.
  • [23] Y. Shou, T. Meng, F. Zhang, N. Yin, and K. Li, “Revisiting multi-modal emotion learning with broad state space models and probability-guidance fusion,” arXiv preprint arXiv:2404.17858, 2024.
  • [24] W. Ai, Y. Wei, H. Shao, Y. Shou, T. Meng, and K. Li, “Edge-enhanced minimum-margin graph attention network for short text classification,” Expert Systems with Applications, vol. 251, p. 124069, 2024.
  • [25] Y. Zhang, Y. Shou, T. Meng, W. Ai, and K. Li, “A multi-view mask contrastive learning graph convolutional neural network for age estimation,” Knowledge and Information Systems, pp. 1–26, 2024.
  • [26] Y. Zhu, H. Liu, Z. Wu, and Y. Du, “Relation-aware neighborhood matching model for entity alignment,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 5, 2021, pp. 4749–4756.
  • [27] W. Ai, F. Zhang, T. Meng, Y. Shou, H. Shao, and K. Li, “A two-stage multimodal emotion recognition model based on graph contrastive learning,” in 2023 IEEE 29th International Conference on Parallel and Distributed Systems (ICPADS).   IEEE, 2023, pp. 397–404.
  • [28] T. Meng, F. Zhang, Y. Shou, W. Ai, N. Yin, and K. Li, “Revisiting multimodal emotion recognition in conversation from the perspective of graph spectrum,” arXiv preprint arXiv:2404.17862, 2024.
  • [29] W. Ai, W. Deng, H. Chen, J. Du, T. Meng, and Y. Shou, “Mcsff: Multi-modal consistency and specificity fusion framework for entity alignment,” arXiv preprint arXiv:2410.14584, 2024.
  • [30] Y. Shou, W. Ai, T. Meng, F. Zhang, and K. Li, “Graphunet: Graph make strong encoders for remote sensing segmentation,” in 2023 IEEE 29th International Conference on Parallel and Distributed Systems (ICPADS).   IEEE, 2023, pp. 2734–2737.
  • [31] Z. Zhang, Y. Yang, and B. Chen, “Relation-aware heterogeneous graph neural network for entity alignment,” Neurocomputing, vol. 592, p. 127797, 2024.
  • [32] W. Ai, J. Li, Z. Wang, J. Du, T. Meng, Y. Shou, and K. Li, “Graph contrastive learning via cluster-refined negative sampling for semi-supervised text classification,” arXiv preprint arXiv:2410.18130, 2024.
  • [33] Z. Sun, W. Hu, Q. Zhang, and Y. Qu, “Bootstrapping entity alignment with knowledge graph embedding.” in IJCAI, vol. 18, no. 2018, 2018.
  • [34] Y. Cao, Z. Liu, C. Li, J. Li, and T.-S. Chua, “Multi-channel graph neural network for entity alignment,” arXiv preprint arXiv:1908.09898, 2019.
  • [35] C. Chen, K. Li, S. G. Teo, X. Zou, K. Wang, J. Wang, and Z. Zeng, “Gated residual recurrent graph neural networks for traffic prediction,” in Proceedings of the AAAI conference on artificial intelligence, vol. 33, no. 01, 2019, pp. 485–492.
  • [36] C. Chen, K. Li, S. G. Teo, X. Zou, K. Li, and Z. Zeng, “Citywide traffic flow prediction based on multiple gated spatio-temporal convolutional neural networks,” ACM Transactions on Knowledge Discovery from Data (TKDD), vol. 14, no. 4, pp. 1–23, 2020.
  • [37] M. Chen, Y. Tian, M. Yang, and C. Zaniolo, “Multilingual knowledge graph embeddings for cross-lingual knowledge alignment,” arXiv preprint arXiv:1611.03954, 2016.
  • [38] H.-W. Yang, Y. Zou, P. Shi, W. Lu, J. Lin, and X. Sun, “Aligning cross-lingual entities with multi-aspect information,” arXiv preprint arXiv:1910.06575, 2019.
  • [39] Y. Wu, X. Liu, Y. Feng, Z. Wang, R. Yan, and D. Zhao, “Relation-aware entity alignment for heterogeneous knowledge graphs,” arXiv preprint arXiv:1908.08210, 2019.
  • [40] W. Cai, W. Ma, L. Wei, and Y. Jiang, “Semi-supervised entity alignment via relation-based adaptive neighborhood matching,” IEEE Transactions on Knowledge and Data Engineering, vol. 35, no. 8, pp. 8545–8558, 2022.