Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

An attribute-enhanced relationship-aware neighborhood matching model with dual attention

  • Junlin Gu,

    Roles Funding acquisition, Software

    Affiliation Department of Computer, Jiangsu Vocational College of Electronics and Information, Huai’an, China

  • Weiwei Liu,

    Roles Conceptualization, Resources, Supervision, Writing – original draft, Writing – review & editing

    Affiliation Department of Computer, Jiangsu Vocational College of Electronics and Information, Huai’an, China

  • Xiong Yang

    Roles Conceptualization, Writing – original draft, Writing – review & editing

    02116828@fdzcxy.edu.cn

    Affiliation Department of Computer Engineering, Fuzhou University Zhicheng College, Fuzhou, China

Abstract

The entity alignment task aims to match semantically corresponding entities in different knowledge graphs, which is important for knowledge fusion. Traditional graph-based methods often lose information due to insufficient use of attributes and imperfect relationship modeling, which makes it difficult to capture the deep semantic relationship between entities fully. To improve the effect of entity alignment, we propose a new model named ARNM-DAE2A, which strengthens the information aggregation capability of GCN by introducing a dual-attention mechanism to ensure a more balanced and comprehensive structural representation. The model contains the entity structure embedding module, the attribute structure embedding module, the joint alignment module and the relationship-aware neighborhood matching module. The entity structure embedding module optimizes the structure learning capability of GCN by introducing the pairwise attention mechanism. The attribute structural embedding module utilizes GCN to acquire entity attribute information. The joint alignment module weights and fuses the relationship structure information and attribute information as a comprehensive representation of entities. The relationship-aware neighborhood matching module then corrects the noise in the GCN aggregated information by comparing the neighborhood relationships of entity pairs. Experiments conducted on DBP15K and SRPRS datasets illustrate that the proposed ARNM-DAE2A outperforms baselines.

1 Introduction

Knowledge graph (KG) has shown great potential and value as a powerful form of knowledge representation and organization in various fields [1]. A KG is a semi-structured data model that graphically represents relationships among entities, including individuals, locations, occurrences, and various other elements [2]. It not only offers rich semantic information but also aids machines in comprehending and reasoning about the connections between entities. Despite presenting a wealth of semantic information, real-world hosts diverse KGs sourced from varying data sources, domains, and are described using different ontologies [3]. This heterogeneity poses substantial challenges for knowledge sharing and cross-graph applications.

Entity alignment refers to the process of discovering and matching mutually corresponding entities across different KGs or data sources [1,3]. Specifically, when there exist heterogeneous data describing similar concepts or entities in different graphs, the goal of entity alignment is to establish semantic correlations between these entities, enabling effective integration of these heterogeneous data on a semantic level. This implies that through entity alignment, entities with similar semantic meanings described in different data sources can be identified and mapped to consistent identifiers or representations [4]. Such mappings help eliminate semantic gaps between heterogeneous data, enhancing the quality and accuracy of data integration [5]. Through entity alignment, we can conduct knowledge queries, reasoning, and analysis in a cross-graph environment, maintaining semantic consistency while integrating heterogeneous data. This forms the basis for cross-domain, cross-data-source knowledge discovery, fostering deeper data collaboration and sharing. Entity alignment is not only a part of graph research but also a crucial tool driving cross-disciplinary knowledge fusion and application, laying the foundation for comprehensive, integrated intelligent information retrieval, and application [6].

In the context of entity alignment, the utilization of attribute information has become an indispensable aspect to capture semantic relationships and similarities between entities more accurately [7]. Entity attributes often encompass rich semantic features that can help differentiate entities to a certain extent. However, despite the wealth of semantic features in attribute information, existing entity alignment methods often overlook the correlations among attributes when utilizing them, leading to information loss and inaccuracy. Inherent connections and dependencies often exist among entity attributes, and these associations might be neglected in traditional approaches. Consequently, how to fully exploit attribute information and effectively capture the correlations between attributes has become a pressing challenge [8,9].

At the same time, to better model the relationships be-tween entities in entity alignment tasks, attention mechanisms have been incorporated into numerous entity alignment models. Traditional attention mechanisms often consider one-way information propagation and disregard the symmetry between entities, potentially resulting in incomplete information transfer, especially in entity pairs with symmetrical relationships. To comprehensively capture entity relationships, the dual attention mechanism has garnered significant attention in recent years [10,11]. The dual attention allows models to aggregate information from the perspectives of both entities simultaneously, aiding in a better understanding of their mutual influence. Through dual attention, the model can consider relationships between two entities simultaneously and map this relationship into a shared attention representation, thereby accurately capturing semantic connections between entities. However, the application of dual attention in the field of entity alignment remains relatively limited. While this mechanism has achieved some success in other domains, effectively integrating it into entity alignment models still necessitates further research and exploration [12,13].

To address these problems, we propose an Attribute-enhanced Relationship-Aware Neighborhood Matching Model with Dual Attention named ARNM-DAE2A for entity alignment. This model aims to fully exploit the correlations among entity attributes and accurately model entity relationships through the integration of attribute information and dual attention mechanism. The main contributions of this work are as below:

  • We introduce a relationship-aware neighborhood matching model that incorporates a dual attention mechanism, further enhancing the structural learning capability of GCN.
  • We introduce an innovative method for entity alignment that involves simultaneous learning of relationship structures and attribute details. These are embedded as preliminary matrix representations, followed by utilizing the relationship-aware neighborhood matching model for entity alignment.
  • Experimental results on DBP15K and SRPRS datasets demonstrate that the proposed model achieves improved alignment accuracy compared to baselines.

2 Related work

2.1 Entity alignment based on GCN

Wang et al. [3] first introduced graph convolutional networks (GCN) into the task of entity alignment, marking the initiation of using GCN-based methods as a benchmark for model expansion and attracting wide-spread attention from researchers. Since then, the focus of research has gradually shifted from translation-based models like TransE to models based on GCN, igniting a research trend with GCN models at its core. Although the GCN-Align model proposed by Wang et al. applied GCN to entity alignment for the first time, the alignment effect was limited by only using the basic GCN. Cao et al. [14] incorporated GCN into entity alignment using a multi-channel graph neural network (MuGNN), which complements multi-channel approaches involving attention-based KG completion and cross-lingual attention-based entity pruning. Wu et al. [15] designed a relationship-aware dual-graph convolutional network (RDGCN), introducing the concept of bidirectional dual graphs and enhancing discrimination between different entity network structures by constraining these dual graphs. However, they neglected the dependence between attribute information utilization and relationship, and therefore the alignment effect was not further improved. Zhu et al. [16] devised relationship-aware neighborhood matching (RNM), using a relationship-aware method that incorporates neighborhood matching to refine GCN-based entity alignment. Zhu et al. designed RNM that used relational perception to correct GCN alignment, but also did not fully consider the aspect of attributes. Wu et al. [17] introduced a neighborhood matching network (NMN) designed to handle the heterogeneous neighborhoods within KGs. This approach estimated entity similarity by considering both topological structure and neighborhood similarities. Although the NMN and HGCN reflected the idea of neighborhood matching, they did not deeply optimize the combination of attributes and relationships. Additionally, Wu et al. [18] introduced hyperbolic graph convolutional networks (HGCN), using a high-speed gate mechanism to regulate the spread of noise within GCN structures, and leveraging entity representations for approximating relationship representations to optimize the goal of relationship alignment. Although the NMN and the HGCN reflected the idea of neighborhood matching, they did not deeply optimize the combination of attributes and relationships. GCN, as a neural network, can effectively learn the dependencies and connection rules in structural relationships, and extract structural information to provide strong support. However, most GCN-based entity alignment methods mainly focus on learning structural information, ignoring the importance of attribute features. The interactive learning between different information is not deep enough, and the fusion effect of attribute and structural information is low. In addition, it is difficult to balance the proportion of structural and attribute information in the model, which directly affects the alignment effect.

2.2 Entity alignment based on attribute information and attention mechanism

Sun et al. [19] introduced the AliNet that harnesses the local structures of multi-hop entities and enforces equivalent entity pairs to possess identical hidden states across each layer of the graph attention network (GAT). However, AliNet relied too much on local structure and did not make good use of global context in-formation, which can easily lead to misalignment. Xin et al. [20] aggregated contextual information through a Transformer model and designed holistic reasoning based on embedding similarity, relationships, and entity functionality to evaluate alignment probabilities. Xin et al. used Transformer to collect context information, but ignored the correlation between structure and attributes, and could not express the complex features of entities well. Mao et al. [21] considered partitioning relationships into meta-relationships within a graph neural network framework, learning attention parameters among these meta-relationships and integrating them into entity representations, and ultimately training the model in a semi-supervised manner. Mao et al. learned relational attention under the GCN framework, but truncated the relationships into independent units, losing the order and continuity information. Furthermore, there were approaches relying on attention mechanisms, long short-term memory (LSTM), and the bidirectional encoder representations from transformers (BERT) model [22]. Attention mechanisms make up for the deficiencies of GCN in learning local relationships, and can effectively enhance the transfer of information related to the target. However, excessive reliance on local perspectives can also ignore the influence of global connectivity. When learning global contextual information, attention-based methods generally perform worse than GCN. In addition, how to master local and global information in the model and effectively combine them is a major problem that this method needs to address, and the current solution is not effective.

2.3 Entity alignment based on neighborhood matching and structural optimization

Wu et al. [15] utilized an attention mechanism to capture the interaction between the primary graph and the dual relation graph. Similar to Sun et al. [23] and Wang et al. [3], they utilized entity attribute information for neighborhood matching. However, the learning of entity relationship structures was not comprehensive enough, which could lead to the neglect of important entity information. Zhuang et al. [24] enhanced the modeling capability of different entity network structures by constraining the dual graph, indicating the potential effectiveness of this structural optimization approach in entity alignment tasks. Chami et al. [25] introduced a high-speed gate mechanism to control noise propagation in graph convolutional networks, optimizing approximate relationships. This structural optimization approach could be employed to improve relationship alignment performance. Zhu et al. [16] utilized neighborhood matching to enhance entity alignment, exploring useful information not only from neighboring nodes but also from connection relationships. Furthermore, iterative frameworks were designed to enhance structural optimization in a semi-supervised manner, fully harnessing the beneficial interactions between entity alignment and relationship alignment.

3 Problem definition

First, we define the KG as:

(1)

where E, R, V, A and T are distinct components within the KG [26], with E denoting entities, R denoting relationships, V denoting attributes, A denoting attribute values, and T denoting the set of knowledge triples, respectively.

Then, we define a relationship triple in the KG as:

(2)

where h, r and t denote the head, relationship, and tail of a relationship triple, respectively.

Finally, we define an attribute triple as:

(3)

where h, a and v denote the head, attribute, and attribute value of a attribute triple, respectively.

Given two KGs, denoted as and , and a pre-aligned entity pairs denoted as , the essential task of entity alignment is to identify the remaining equivalent entity pairs. Where, e1 denotes a specific instance in the entity set E1 of the KG G1, and similarly, e2 denotes another instance. Fig 1 shows an example of cross-language entity alignment, where the left nodes are from the Chinese KG (labels are presented in Chinese), and the right nodes are from the English KG (labels are presented in English). Each node on the left is a direct translation of its corresponding node on the right. In the mathematical description, the symbol E denotes the set of all entities in the two KGs, and this correspondence forms the basis of the interaction between structure and attribute information in the alignment process of this model. The objective of entity alignment is to identify relation pairs that have the same semantic meaning within the given two KGs.

thumbnail
Fig 1. An instance of cross-lingual entity alignment involving two KGs (Chinese KG and English KG).

https://doi.org/10.1371/journal.pone.0324290.g001

4 Model framework

We design a relationship-aware matching model that combines attribute information and dual attention mechanism named ARNM-DAE2A, as shown in Fig 2. The model consists four modules: entity structure embedding module, attribute structure embedding module, joint alignment module, and relationship-aware neighborhood matching module. The entity structure embedding module enhances the capability of RDGCN to learn entity structure information by introducing a dual attention mechanism. The attribute structure embedding module employs GCN to learn entity attributes, providing additional feature representations. The joint alignment module combines relationship structure information and attribute information through weighted fusion, obtaining comprehensive entity representations. The relationship-aware neighborhood matching module corrects noise in GCN aggregation information by comparing neighborhood relations of entity pairs, thereby improving alignment accuracy.

4.1 Entity structure embedding module

In the traditional GCN, information transfer is primarily based on unidirectional neighborhood aggregation, which is susceptible to neighborhood noise and information asymmetry problems, resulting in less robust node representations. If information can be aggregated from both source and target nodes at the same time, i.e., bi-directional information interaction is realized, it helps improve the symmetry and stability of information transfer. Therefore, this module introduces the dual-attention mechanism to theoretically achieving noise suppression and key information reinforcement by adaptively assigning neighborhood weights, thereby improving the overall entity representation quality. To start, this module establishes a dual relationship graph derived from the original graph. It amplifies the interaction between the dual relationship graph and the original graph by incorporating attention mechanisms. Next, the module processes the nodes from the original graph through a GCN layer, enhanced with a high-speed neural network gating mechanism, to capture the structural information of neighboring nodes. Finally, the resulting entity representations from this module are utilized to assess the alignment of two entities.

4.1.1 Constructing dual graph.

We denote the original graphs G1 and G2 as , where denotes the set of nodes in the graph, and denotes the set of relations in the graph. For the two original KGs , their dual relation graphs are denoted as and defined as: 1) For each relation r in , there exists a vertex , resulting in ; 2) If relations ri and rj share entities or tail entities, edges are created in to connect nodes ri and rj.

Based on the above definition, we establish the edge weights within the dual relationship graph as:

(4)(5)(6)

where Hi denotes the set of head entities, and Ti denotes the set of tail entities.

We use the graph attention network (GAT) [27] to iteratively obtain vertex representations of the dual relation graph and the original graph, where the attention mechanism helps facilitate interactions between the two graphs. The attention mechanism fosters interactions between these two graphs. Each bidirectional interaction comprises two layers: the dual attention layer and the original attention layer. This stacking of multiple interactions mutually reinforces both graphs.

(a) Dual Attention Layer

We represent the dual vertex representation matrix of the input KG as , where each row denotes to a vertex within the relation graph . We utilize the original node features generated by the original attention layer to calculate the dual attention scores:

(7)(8)

where denotes the output representation at vertex in the dual graph, denotes the dual representation of vertex , denotes the set of neighbor indices for dual vertex , denotes the dual attention score, denotes the fully connected layer, denotes the ReLU activation function, denotes the LeakyReLU activation function, denotes the concatenation operation, and ci denotes the relation representation from the previous original attention layer.

By concatenating the average representations of head and tail entities in , we obtain a relation representation ci obtained from the original attention layer:

(9)

where denotes the output representation of the k-th head node of relationship ri obtained from an original attention layer, denotes the output representation of the l-th tail node of ri.

(b) Original Attention Layer

In this layer, when utilizing GAT for the original graph, we calculate the attention scores for the original graph using the dual vertex representations obtained from , which are associated with the relations in the original KG .

We denote the original vertex representation matrix of the input KG as . For entity eq in the original KG , its representation is shown as:

(10)(11)

where denotes the dual representation of the relationship between entity eq and et obtained in , denotes the set of neighboring indices for entity eq in , denotes the original attention score, denotes a fully connected layer, and is an activation function.

In the model, the initial representation matrix for the original vertices can be initialized based on entity names, which offer crucial information for entity alignment [20]. Therefore, we retain the data by merging the output of the original attention layer with the initial representation:

(12)

where denotes the final representation of entity eq in the interactive module of the original KG, and denotes the weight parameters of the s-th layer of the original attention layer.

4.1.2 Merge structural information.

Following numerous iterations involving the dual relation graph and the original graph, we can accumulate entity representations that are sensitive to relationships from the original graph. Next, we employ a dual-layer GCN with high-speed neural network gates [28] to further merge the structural information of neighbors.

At each GCN layer l, we take the entity representations as input:

(13)

where denotes the adjacency matrix of the original graph , I denotes the identity matrix, denotes the activation function, denotes the weight for the l-th layer, and D denotes the degree matrix. To enable bidirectional information flow, when constructing matrix A, we treat as an undirected graph.

4.1.3 Relationship structure alignment.

Inspired by RDGCN[12], we gather the ultimate entity representations denoted as from the output of the GCN layers, and the distance between two entities serves as the alignment score:

(14)

where denotes the entity representation of entity e1 in G1, denotes the entity representation of entity e2 in G2, and denotes the distance between e1 and e2 in the same L1 norm-regularized space.

4.2 Attribute structure embedding module

In real KGs, it is often difficult to rely on structural information alone to distinguish entities with similar connection patterns but subtle semantic differences. The attribute data of entities, such as names, descriptions, and other semantic labels, can provide complementary information that enhances the discriminative power of entity representations. Based on this, this module uses an independent GCN to model the entity attributes and vectorize the extracted attribute features. Attribute information has a high degree of discrimination, which can distance structurally similar but semantically different entities in the embedding space, making it easier for the model to capture pairs of entities that are truly semantically related when aligned. With attribute-enhanced embedding, we hope to use this additional information to compensate for the limitations of single-structure embedding and further improve the overall discriminative power and alignment accuracy of the model. Therefore, in this module, we adopt the idea of graph convolutional alignment to separately analyze attributes. Unlike Wang et al. [3], we do not directly combine attribute information with GCN in joint training. Instead, we use a complete attribute matrix without considering its connection with the relationship structure matrix. We treat attributes as nodes, forming a KG where entities are connected by edges representing attributes. If an entity possesses a certain attribute, there is an edge connecting the entity to that attribute. Therefore, in the connection matrix of the GCN model, entities are only connected to attributes, and there are no connections between entities. The dimension of the connection matrix corresponds to the number of attributes. We initialize node vectors with random values as input for the first layer, and the convolutional process as:

(15)

where , Na denotes the number of attributes, a denotes the attribute structure embedding module, and denotes the activation function.

The loss function for this module is defined as:

(16)

where denotes the margin hyperparameter, , and denotes the entity-based embedding vector using attribute structure. We utilize stochastic gradient descent (SGD) [29] to minimize the aforementioned loss function.

4.3 Joint alignment module

After obtaining entity embedding based on both relationship structure and attribute structure, we compute the similarity between entities from these two aspects separately. Subsequently, we combine these similarities through weighted summation to derive the overall similarity between entities. The final entity similarity distance function is defined as:

(17)

where , , and denote the embedding of relationship structure and attributes, respectively. ds and da are the dimensions of the embedding for relationship structure and attributes, while and are hyperparameters that balance the importance of the two types of embedding.

4.4 Relationship-aware neighborhood matching module

4.4.1 Entity embedding.

We use the inputs of GCN as entity embedding and define the representation form of entities as:

(18)

where denotes the dimension of entity embedding, n denotes the number of entities. For an entity pair , denotes the aligned entity pairs, and denotes the candidate entity pairs for alignment. We define the distance as , where denotes the vector norm measure. The smaller the distance between entity pair , the higher the probability that entity pair is aligned.

To embed entities from two KGs into a shared latent space, we employ seed alignment as training data and devise an edge-based loss function for entity alignment:

(19)

where L denotes the set of aligned seed entity pairs, denotes a set of auxiliary alignments sampled from nearest neighbors, and is a margin hyperparameter that separates positive and negative entity alignment pairs. Our loss function is designed with the assumption that the distance between aligned entity pairs should approach zero, whereas the distance between negative samples should be maximized [30].

4.4.2 Relationships embedding.

We employ the information linking entities and the embedding of head and tail entities acquired through GCN to represent relationships within the KGs:

(20)

where denotes the embedding of relationship , denotes the concatenation operation, and respectively signify the average embedding of all distinct head and tail entities of relationship r.

Furthermore, to delve deeper into the translational insights stemming from ternary relationships, we employ a regularize similar to the TransE model [31]:

(21)

where T1 and T2 denote the triple sets of the given KGs G1 and G2, respectively. denotes the transformation matrix from the latent relationship space to the latent entity space, which is the model parameter to be learned.

To simultaneously acquire embedding for entities and relationships, we optimize the objective function following the pretraining of entity embedding:

(22)

where is a weighting coefficient used to balance the loss of entity alignment and the loss of regularize when considering the relation prior.

4.4.3 Relationship-aware neighbor matching for entity pairs.

For each candidate entity pair , in addition to comparing their one-hop neighbor entities in pairs, we also consider comparisons between connectivity relations [13]. Assuming that is the set of first-order neighboring entities of ei in G1 and is the set of first-order neighboring entities of in G2, for neighborhood matching between ei and , we compare . After that, we focus on matching neighbours with matching relations. In addition, the mapping property of the connectivity relation is also important for entity alignment. Therefore, for each matching case in the matching set , we compute the alignment probability based on r1, r2, n1 and n2:

(23)

where , , they denote the mapping probabilities of correspondence and neighboring entities, respectively. The distance between two entities is updated as:

(24)

where denotes a hyperparameter that regulates the balance between embedding distance and matching score. A higher matching score increases the likelihood of aligning candidate entity pairs.

4.4.4 Entity-aware relationship matching.

For relation r, we denote the set of associated entity pairs as , where T denotes the set of triples in the given KG. Given a candidate pair of relations , r1 from G1 and from G2. First, we form corresponding sets of entity pairs and . Then, we compare all entity pairs in and define the subset with the matching set . The distance between the updated relationship pairs is:

(25)

4.5 Complexity analysis

GCN is responsible for learning the structured representation of entities, and the main computational effort is from neighborhood feature aggregation and feature transformation. Assuming that a KG contains N entities and E edges with embedding dimension d and GCN adopts L layers, the complexity of neighborhood feature aggregation is O(Ed), the complexity of feature transformation is , and the overall computational complexity is .

The dual attention mechanism is used to optimize the GCN propagation process by introducing additional attention weight calculation and weighted neighbor aggregation. Assuming that the average number of neighbors of each entity is k, the complexity of attention computation is O(Nkd), the complexity of weighted aggregation is also O(Nkd), and the overall complexity is O(LNkd).

The joint alignment module calculates the entity similarity by fusing structural and attribute information, and the computational complexity is determined by embedding fusion and similarity calculation. Here, the computational complexity of the fusion process is O(Nd), the complexity of similarity computation is O(Md) based on the number of entity pairs M, and the overall complexity is .

The relationship-aware neighborhood matching module further adjusts the entity alignment results, and its computational complexity is primarily determined by the neighborhood matching computation. Here, the complexity of the neighborhood matching operation for each entity is O(Nkd).

In summary, the overall computational complexity of ARNM-DAE2A is O(L(Ed + Nd2)) + O(LNkd) + O(Nd + Md) + O(Nkd). The GCN computation grows linearly with the number of layers L, the neighborhood matching computation depends on the number of neighbors k, and the cost of the alignment computation is determined by the number of entity pairs M. The GCN computation is also a good example of GCN computation. It is worth mentioning that the computational cost may rise significantly in the case of dense graphs () or large-scale entity comparisons ().

5 Experimental Analysis

5.1 Datasets

We conducted simulation experiments using the DBP15K dataset [32], and comprehensive datasets details are presented in Table 1. DBP15K is a large-scale KG of a comprehensive encyclopaedia, containing rich semantics in different languages, with links across different languages, making it suitable for cross-lingual datasets. Since the DBP15K dataset is too dense and the degree distribution is very different from the real-world data, we also chose two cross-linguistic subsets from the relatively sparse SRPRS dataset [33], and comprehensive dataset details are presented in Table 2. For comparison with previous works, we used the same training/testing split with previous works [15], 30% for training and 70% for testing.

5.2 Experimental setup

5.2.1 Experimental platform.

We conducted experiments on a server with the specifications of Intel 4210R/2x Tesla V100-32G GPUs, an 8-core CPU, and 40GB of RAM. The model was implemented using the TensorFlow framework.

5.2.2 Evaluation metrics.

We use Hits@k and mean reciprocal rank (MRR), which are commonly used in KG entity alignment tasks, as evaluation metrics to evaluate all methods as:

(26)(27)

where N is the number of entities, ranki is the ranking of entities aligned with the i-th entity, and is an indicator function (the value of the function is 1 if the condition is true and 0 otherwise). For each source entity, entities in the other KGs are ranked in descending order based on their similarity to the source entity. measures the alignment accuracy, indicating the percentage of correctly aligned entities among those similar to the source entity. Meanwhile, MRR provides complementary insights into the entity alignment results. The higher the values of Hits@k and MRR, the better the performance of the model.

5.2.3 Parameter setting.

In the entity structure embedding module, according to [15], the hidden dimensions for both the dual attention layer and the original attention layer are set to 300. The hidden dimensions for all layers in the GCN are also set to 300. The learning rate is set to 0.001, and negative entity pairs are generated every 10 epochs of training, accumulating a total of 600 epochs.

For the attribute structure embedding module, we employ the SGD optimization algorithm for 2000 iterations of model updating. According to [15], the dimension da is set to 300, and the output, input, and final dimensions of the first and second layers of GCN are kept equal.

In the joint alignment module, according to [3], the parameters and in (17) are set to 0.9 and 0.1, respectively.

In the relationship-aware neighborhood matching module, according to [3], a 2-layer GCN is utilized to learn entity embedding. The hidden dimensions for both structure and attribute are set to 300, and the learning rate is set to 0.001. According to [34], seed alignment is set at a ratio of 30%. The boundary is set to 1, the learning rate is set to 0.001, is set to 10, and is set to 200. The nearest 100 entities and 20 relations are selected as candidates for matching. Each positive sample is paired with 125 negative samples, and a maximum iteration count T is set to 4. Optimization using (19) is performed for 50 epochs, followed by joint embedding training using (22) for an additional 10 epochs.

5.3 Experimental results and analysis

5.3.1 Comparative experiments.

For evaluating our proposed ARNM-DAE2A model, we compare it with the state-of-art methods: JAPE [23], GCN-Align [3], RDGCN [15], CTEA [35], HMAN [36], Dual-AMN [37], and MSNEA [38].

The comparative experimental results on DBP5K are shown in Table 3, where the results are presented as percentages rounded to two decimal places. Bold values indicate the best performance.

JAPE systematically introduces attribute information for the first time in a representation learning framework, incorporating both the structural and attribute information of entities into the embedding space to alleviate the heterogeneity problem in cross-lingual entity alignment. Attributes can provide additional discriminative features to help distinguish structurally similar but semantically different entities. However, early representation learning methods generally have poor accuracy benchmarks and are immature in fusing multiple modal information, resulting in poor performance of JAPE. For instance, on the ZH-EN dataset, the Hits@1 is only 41.18%.

GCN-Align introduces GCN into the entity alignment task for the first time and learns it jointly with attribute information. GCN can theoretically be regarded as a local filtering performed on the graph structure to obtain the contextual relationship of entities in the graph spectrum through neighborhood aggregation. Therefore, there are more GCN-based entity alignment methods in the current entity alignment tasks [22]. However, GCN-Align uses only the most basic GCN and is relatively inadequate in handling noise and heterogeneous structures, e.g., Hits@1 is only 41.25% on ZH-EN. This reflects that simply using GCNs for entity alignment cannot yet fully mitigate the complex heterogeneity in cross-lingual knowledge graphs.

CTEA and HMAN, on the other hand, take the GCN as the base model and utilize more auxiliary information to further improve the performance. CTEA adopts GCN and TransE to process entity alignment in parallel, which theoretically integrates two different representations of structural and translation models, and the Hits@10 on ZH-EN can reach 90.50%. However, due to the lack of deeper coupling or improvement of the GCN and TransE, the improvement of Hits@1 is still limited. HMAN attempts to mine richer semantic features from the context by introducing a BERT variant model, but the high time complexity and inconvenient use of BERT results in only 56.20% of Hits@1 on ZH-EN, which indicates that the task still needs to strike a balance when introducing large language models.

Dual-AMN further emphasizes the role of dual attention and incorporates difficult sample mining to accelerate the alignment process. Theoretically, dual-attention can focus on both the local neighborhood and the global structure to reduce the interference of noisy entities, while difficult sample mining can strengthen the model’s learning ability near the discriminative boundary. Experiments show that the method can significantly improve the alignment speed while maintaining high alignment accuracy, indicating that the combination of multi-angle attention and difficult sample mining has high utility in the alignment task.

MSNEA emphasizes characterizing entities from a Multi-Scale perspective, fusing different levels of graph structure and semantic information. According to the theory of multimodal feature fusion, information from multiple scales tends to provide a more comprehensive contextual semantics, thus helping models to better distinguish similar entities in the alignment task. The method also achieves excellent performance on cross-language datasets, suggesting that multi-scale representation can effectively alleviate the semantic divide across languages or modalities.

Consistently across multiple datasets, our proposed ARNM-DAE2A model outperforms existing baseline models on DBP15K, establishing itself as the state-of-the-art solution. Specifically, the ARNM-DAE2A model leverages an attribute information module to learn entity attribute features, optimizes structural learning with RDGCN, combines structural and attribute information for multi-dimensional entity representation, and utilizes a relationship-aware matching mechanism to correct noise introduced by GCN’s aggregation of neighbor information. Finally, the model employs a cyclic iterative learning framework that positively influences and mutually enhances entity alignment and relation alignment tasks. These innovations enrich entity representations, enhance structural learning capabilities, effectively integrate heterogeneous information sources, and improve the model’s resilience to disturbances.

To more intuitively demonstrate the performance of the ARNM-DAE2A model, we compared it with multiple benchmarks on the DBP15K dataset using Hits@1 to Hits@50 with a step size of 10. We selected JAPE, GCN-Align, and RD-GCN as the comparison models, as shown in Fig 3. It can be seen that the Hits@K values of our proposed ARNM-DAE2A model are higher than those of other models, achieving the highest score on both the JA-EN and FR-EN datasets. The Hits@K value of the GCN-Align model approaches our ARNM-DAE2A model after K is set to 20, indicating that alleviating the heterogeneity of the entity neighborhood structure is beneficial to entity alignment. However, the Hits@1 of the GCN-Align model is significantly lower than that of the ARNM-DAE2A model, indicating that the ARNM-DAE2A model has better alignment performance.

thumbnail
Fig 3. Comparison of Hits@K results.

(A) ZH-EN. (B) JA-EN. (C) FR-EN.

https://doi.org/10.1371/journal.pone.0324290.g003

The comparative experimental results on SRPRS are shown in Table 4, where the results are presented as percentages rounded to two decimal places. Bold values indicate the best performance. Compared with the results on DBP15K, the overall performance trend of the state-of-the-art methods on SRPRS remains largely the same, indicating that their strengths and weaknesses do not fundamentally change under different map sizes and densities. However, ARNM-DAE2A shows a more obvious relative advantage in this setting, even exceeding its achievement on DBP15K. First, the attribute structure embedding module effectively compensates for the lack of neighborhood information, and the complementary nature of the entity attributes becomes more prominent when the graph structure is sparse. Second, the dual-attention mechanism can perform bi-directional aggregation between the source and target entities, and is more resistant to noise and missing connections. Third, the relationship-aware neighborhood matching by iteratively comparing pairing and correcting alignment results is particularly suitable for reducing error propagation in sparse graphs. Taken together, all these mechanisms can play a more significant compensatory role in sparse environments, making the alignment effect of ARNM-DAE2A even exceed its achievement on denser datasets.

Finally, we compare the average training time of each method on the two datasets, as shown in Table 5, where the results are presented as hours rounded to two decimal places. JAPE employs only direct representation learning, characterized by smaller parameter sizes and a concise computational process, resulting in a shorter training time. GCN-Align introduces a GCN to enhance entity representations through neighborhood aggregation, which increases the computational effort. RDGCN further incorporates relational graph modeling, enabling the model to capture complex graph structure information while significantly increasing computational complexity. CTEA and HMAN utilize TransE and BERT as auxiliary modules, respectively, to enhance representations. However, these auxiliary modules inevitably add to the computational burden. In contrast, the proposed ARNM-DAE2A fuses both structural and attribute information within the RDGCN framework and employs a dual-attention mechanism combined with relation-aware matching to correct for noise. Although this design results in a higher computational complexity, we believe that the additional training cost is acceptable, as it significantly improves both the accuracy of entity alignment and the robustness of the model, thereby demonstrating a substantial advantage in practical applications.

5.3.2 Ablation experiments.

For further evaluation of our proposed ARNM-DAE2A model, we conducted ablation experiments on various modules within each model. We performed ablations by considering the model with only the matching module (MM), the matching module combined with structural embedding (MM+SE), and the matching module with both structural embedding and attribute embedding (MM+AE), as shown in Table 6 and Table 7.

Using only the MM for entity alignment yields poorer results compared to the baseline. This indicates that when relying solely on neighborhood matching, the entities themselves lack sufficient semantic representations, and the matching results are difficult to stabilize once the neighborhood information is limited or noisy. The effect of the matching module relies on the entity representation of the external input, which can constrain the upper limit of the effect of the matching module if the representation is limited. Meanwhile, the matching module only relies on the local neighbor information of the entities, which is prone to noise interference and lacks a global structural perspective, making it difficult to model symmetric relationships of entities and unable to play the role of relationship constraints in the KG. Comparing DBP15K and SRPRS, this trend is more obvious in both dense and sparse environments, indicating that the bottleneck of neighborhood matching mainly comes from the lack of semantic features of the entities. Lacking deep learning of the intrinsic structure or attributes of entities, neighborhood matching can only identify local relationships to a certain extent, but cannot globally grasp the semantic similarity between entities.

When the MM+SE for early-stage structural learning, some metrics show a relatively large improvement, suggesting that the structural representation brought by the GCN helps to make more accurate judgments in later matching. However, in some scenarios (e.g., ZH-EN of DBP15K), Hits@1 may show a small decrease, indicating that early GCN learning may instead bring some errors into later matching if it fails to converge sufficiently when there is more noise or higher sparsity. Nonetheless, overall, the SE module improves most of the metrics more than this cost, especially in sparse environments such as SRPRS, where the effect of complementing neighborhood gaps through structural information is more significant.

MM+AE yields impressive results, occasionally even surpassing the overall model’s performance. This is because that attribute feature can in some cases provide very different discriminative information from the graph structure, allowing the matching module to be more targeted in correcting initial misalignments. However, Hits@10, Hits@50, or MRR do not necessarily outperform the full model at the same time, suggesting that relying on attribute features alone may also be a bottleneck in a wider range of entity retrievals. Once the attributes are inadequate or noisy, the advantage of the AE module is diminished and cannot replace the value of structural embedding for global information capture.

5.3.3 Hyperparameter experiments.

To validate the impact of various hyperparameters in our proposed ARNM-DAE2A model on the experimental results, we conducted experiments with important model hyperparameters, taking Hits@1 as an example for visualization and analysis.

Firstly, we tested the entity dimension, as shown in Fig 4. When the dimension is between 0 and 1250, Hits@1 alignment results steadily increase. However, after reaching 1250 dimensions, the improvement in Hits@1 significantly diminishes. This could be due to the fact that as the dimensionality increases, the probability of similarity between two randomly generated high-dimensional vectors exponentially decreases. Therefore, increasing the dimensionality beyond 1250 has a diminishing impact on Hits@1, with a noticeable marginal effect.

Furthermore, we conducted tests on the number of GCN layers in the entity structure embedding module, as shown in Fig 5. When the number of layers is between 1 and 4, the Hits@1 results for alignment steadily improve. However, after increasing to 4 layers, Hits@1 not only stops improving but also declines. On one hand, this could be due to overfitting as multiple layers of neighborhood aggregation make entity nodes’ representations too similar, leading to the model’s inability to generalize well to the test data and resulting in decreased performance. On the other hand, the diminishing returns might be attributed to the cumulative errors introduced by the chained matrix operations in multi-layer GCN, which could degrade the model’s effectiveness.

Finally, we conducted tests on different numbers of dual attention heads, as illustrated in Fig 6. With an increase in the number of dual attention heads from 1 to 6, the model’s representation capacity gradually improved, allowing it to better capture the relational information among entities, thereby enhancing alignment accuracy. However, when the number exceeded 7 heads, the model might begin to overfit the training data, leading to a decrease in performance on the test data. This could be attributed to the model’s excessive sensitivity to the training data, making it challenging to generalize to other datasets or unseen entity pairs.

5.3.4 Robustness experiments.

As shown in Fig 7A, embedding dimension in the range of 200 to 700 , the five subsets show a gradual upward trend and stabilize at 500 to 600 dimensions. This suggests that as the embedding dimension increases, the model can learn richer entity features, thus improving the alignment accuracy. However, when the dimension exceeds a certain threshold, the performance gain gradually diminishes, indicating that the model is more robust to changes in the embedding dimension in the range of 500-600, and the performance degradation is limited even if the parameters slightly deviate from the optimal values. As shown in Fig 7B, with respect to the number of GCN layer, the result shows that as the number of GCN layers is increased from 1 to 4, the Hits@1 of the model is significantly improved in both datasets, which suggests that the increase in the number of layers can help to capture deeper graph structure information. However, the model performance slightly decreases when the number of layers exceeds 4, probably due to the effect of the over-smoothing phenomenon, which makes the node representations indistinguishable. This result shows that the model is more sensitive to the number of GCN layers, but within a reasonable range, the appropriate number of layers can effectively improve the model performance, while too many layers may lead to excessive aggregation of information, which may affect the alignment accuracy. As shown in Fig 7C, the result shows that in the range of 1 to 8, all the five subsets under perform best in the interval of 3 to 5 heads, followed by a slight decrease in performance as the number of dual attention heads increases. This suggests that an appropriate amount of dual attention heads can help the model capture multi-angle information more comprehensively and enhance entity matching. However, too few attention heads can limit the model’s learning ability, while too many heads may introduce additional noise or lead to a waste of computational resources. Therefore, although the model is robust to parameter variations in the range of 3 to 5 heads, it still needs to be tuned within a reasonable interval to ensure optimal performance.

Referring to [39], we test the robustness of the model by artificially introducing different degrees of textual noise, where textual noise includes entity errors, missing words, and additional characters. As shown in Fig 8, the alignment accuracy of each dataset tends to decrease as the proportion of noise increases, but the magnitude of the decrease is significantly different. In the DBP15K subsets, Hits@1 decreases more gently, which indicates that the model can rely on sufficient structural and attribute information to mitigate the negative impact of noise in the case of richer information. In the SRPRS subsets, the decrease in Hits@1 is more pronounced due to the sparser graphs and less redundant information. This suggests that textual perturbations interfere more with entity alignment in noisy scenarios with sparse data. Overall, the experimental results illustrate that the proposed ARNM-DAE2A is robust under low to moderate noise levels, but there is still room for improvement under high noise and sparse data conditions, and further optimization of the noise processing and information fusion strategies is needed.

5.3.5 Case study.

We take Fig 1 as an illustrative example of a case study of the workflow of the dual attention mechanism. While the traditional one-way attention mechanism only passes information from entities to their neighbors, our dual attention mechanism calculates two sets of attention weights simultaneously. On the one hand, in the Chinese KG, by calculating the source attention from a Chinese-side entity to its neighboring nodes, the importance of each neighbor to the representation of that entity can be quantified; on the other hand, by calculating the target attention from the neighboring nodes back to that Chinese-side entity, the feedback contribution of the neighbors to the representation of the entity is reflected. Similarly, in the English KG, the exact bidirectional computation is performed for the corresponding English-side entity and its neighbor nodes. Subsequently, these two attention weights are fused to form a more comprehensive and robust entity representation. Based on this fused representation, the alignment process shown in Fig 1 can accurately capture the semantic consistency between the two sides of the KG in terms of languages, population, capital, and central bank, significantly improving the accuracy of entity alignment and the model’s noise immunity.

6 Potential applications

KG entity alignment plays a crucial role in various real-world applications where integrating heterogeneous data sources is necessary. The proposed ARNM-DAE2A model, with its ability to effectively align entities across different KGs, has potential applications in the following fields:

Intelligent search and question answering: Search engines and question-answering systems rely on knowledge graphs to provide accurate and context-aware responses. However, inconsistencies in entity representations across different sources can lead to suboptimal results. ARNM-DAE2A improves search relevance and enhances answer accuracy by aligning equivalent entities, ensuring consistent information retrieval. In biomedical databases, for instance, the model helps unify different names for the same disease, facilitating more comprehensive access to medical knowledge.

Enterprise knowledge management and data integration: Organizations maintain multiple databases with overlapping but inconsistently labeled information. Effective entity alignment enables seamless integration of disparate data sources, improving data consistency and decision-making. ARNM-DAE2A can assist in unifying customer records, product catalogs, or financial transactions, making it particularly valuable in industries such as healthcare, finance, and supply chain management.

Recommendation systems: Knowledge graphs enhance recommendation models by representing relationships between users, items, and contextual information. However, inconsistencies in entity representations across platforms can reduce recommendation accuracy. By aligning entities from multiple sources, ARNM-DAE2A enables more precise associations, improving personalization in e-commerce, media streaming, and online advertising.

Multilingual knowledge graph integration: Many large-scale KGs, such as DBpedia and Wikidata, contain equivalent entities in different languages without explicit mappings. ARNM-DAE2A facilitates automatic cross-lingual entity alignment, reducing manual efforts and improving interoperability between multilingual knowledge bases. This is particularly beneficial for cross-lingual information retrieval, machine translation, and global-scale semantic search.

7 Conclusion

This paper introduces an entity alignment neighborhood matching model called ARNM-DAE2A that combines attribute information and dual attention. This model incorporates attribute information, which was not considered in the relationship-aware neighborhood matching model for entity alignment, and introduces dual attention mechanisms to enhance the learning capabilities of GCN structures. Simulations on three cross-lingual datasets validate the effectiveness of dual attention and attribute information in entity alignment methods. In future work, we plan to explore more proactive fusion mechanisms, such as cross-modal attention or dynamic feature selection-based methods, to enable attribute embedding to adaptively adjust the weights between different entity pairs and improve the robustness of alignment. At the same time, we consider introducing dyadic attention to other association tasks, such as cross-language information retrieval or KG complementation, to expand its application areas.

References

  1. 1. Zhang R, Trisedya BD, Li M, Jiang Y, Qi J. A benchmark and comprehensive survey on knowledge graph entity alignment via representation learning. VLDB J. 2022;31(5):1143–68.
  2. 2. Cai H, Zheng VW, Chang KCC. A comprehensive survey of graph embedding: problems, techniques, and applications. IEEE Trans Knowl Data Eng. 2018;30(9):1616–37.
  3. 3. Wang Z, Lv Q, Lan X, Zhang Y. Cross-lingual knowledge graph alignment via graph convolutional networks. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 349–57.
  4. 4. Zhu L, Li N, Bai L. Embedding-based entity alignment between multi-source temporal knowledge graphs. Eng Appl Artif Intell. 2024;133:108451.
  5. 5. Masmoudi M, Ben Abdallah Ben Lamine S, Karray MH, Archimede B, Baazaoui Zghal H. Semantic data integration and querying: a survey and challenges. ACM Comput Surv. 2024;56(8):1–35.
  6. 6. Zhang D, Yuan Z, Liu H, lin X, Xiong H. Learning to walk with dual agents for knowledge graph reasoning. In: Proceedings of the AAAI Conference on Artificial Intelligence, 2022, pp. 5932–41.
  7. 7. Zhu B, Wang R, Wang J, Shao F, Wang K. A survey: knowledge graph entity alignment research based on graph embedding. Artif Intell Rev. 2024;57(9):229.
  8. 8. Tam NT, Trung HT, Yin H, Van Vinh T, Sakong D, Zheng B, et al. Entity alignment for knowledge graphs with multi-order convolutional networks (extended abstract). In: 2021 IEEE 37th International Conference on Data Engineering (ICDE), 2021, pp. 2323–4.
  9. 9. Wang Q, Mao Z, Wang B, Guo L. Knowledge graph embedding: a survey of approaches and applications. IEEE Trans Knowl Data Eng. 2017;29(12):2724–43.
  10. 10. Chawla K, Yang D. Semi-supervised formality style transfer using language model discriminator and mutual information maximization. arXiv, preprint, 2020.
  11. 11. Shen L, He R, Huang S. Entity alignment with adaptive margin learning knowledge graph embedding. Data Knowl Eng. 2022;139:101987.
  12. 12. Wu Z, Pan S, Chen F, Long G, Zhang C, Yu PS. A comprehensive survey on graph neural networks. IEEE Trans Neural Netw Learn Syst. 2021;32(1):4–24. pmid:32217482
  13. 13. Wu F, Souza A, Zhang T, Fifty C, Yu T, Weinberger K. Simplifying graph convolutional networks. In: Proceedings of the 36th International Conference on Machine Learning. vol. 97. PMLR; 2019, pp. 6861–71.
  14. 14. Cao Y, Liu Z, Li C, Liu Z, Li J, Chua TS. Multi-channel graph neural Network for entity alignment. 2019.
  15. 15. Wu Y, Liu X, Feng Y, Wang Z, Yan R, Zhao D. Relation-aware entity alignment for heterogeneous knowledge graphs. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence. International Joint Conferences on Artificial Intelligence Organization; 2019, pp. 5278–84.
  16. 16. Zhu Y, Liu H, Wu Z, Du Y. Relation-aware neighborhood matching model for entity alignment. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 35. 2021. pp. 4749–56.
  17. 17. Wu Y, Liu X, Feng Y, Wang Z, Zhao D. Neighborhood matching network for entity alignment. 2020.
  18. 18. Wu Y, Liu X, Feng Y, Wang Z, Zhao D. Jointly learning entity and relation representations for entity alignment. 2019.
  19. 19. Sun Z, Wang C, Hu W, Chen M, Dai J, Zhang W, et al. Knowledge graph alignment network with gated multi-hop neighborhood aggregation. 2020;34:222–9.
  20. 20. Xin K, Sun Z, Hua W, Hu W, Zhou X. Informed multi-context entity alignment. In: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining. New York, NY: Association for Computing Machinery; 2022, pp. 1197–205.
  21. 21. Mao X, Wang W, Xu H, Lan M, Wu Y. MRAEA: an efficient and robust entity alignment approach for cross-lingual knowledge graph. In: Proceedings of the 13th International Conference on Web Search and Data Mining. New York, NY: Association for Computing Machinery; 2020, pp. 420–8.
  22. 22. Zeng K, Li C, Hou L, Li J, Feng L. A comprehensive survey of entity alignment for knowledge graphs. AI Open. 2021;2:1–13.
  23. 23. Sun Z, Hu W, Li C. Cross-lingual entity alignment via joint attribute-preserving embedding. In: d'Amato, C, et al. The Semantic Web – ISWC 2017. ISWC 2017. Lecture Notes in Computer Science, vol 10587. Cham: Springer; 2017. https://doi.org/10.1007/978-3-319-68288-4_37
  24. 24. Zhuang C, Ma Q. Dual graph convolutional networks for graph-based semi-supervised classification. In: Proceedings of the 2018 World Wide Web Conference. Republic and Canton of Geneva, CHE: International World Wide Web Conferences Steering Committee; 2018, pp. 499–508.
  25. 25. Chami I, Ying Z, Ré C, Leskovec J. Hyperbolic graph convolutional neural networks. In: Advances in Neural Information Processing Systems 32. 2019.
  26. 26. Asif NA, Sarker Y, Chakrabortty RK, Ryan MJ, Ahamed MH, Saha DK, et al. Graph neural network: a comprehensive review on non-Euclidean space. IEEE Access. 2021;9:60588–606.
  27. 27. Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y. Graph attention networks. arXiv, preprint, 2018.
  28. 28. Kipf TN, Welling M. Semi-supervised classification with graph convolutional networks. arXiv, preprint, 2017.
  29. 29. Ketkar N, Santana E. Deep learning with Python. vol. 1. Springer; 2017.
  30. 30. HaoChen JZ, Wei C, Gaidon A, Ma T. Provable guarantees for self-supervised deep learning with spectral contrastive loss. In: Advances in Neural Information Processing Systems 34. Red Hook, NY: Curran Associates Inc.; 2021, pp. 5000–11.
  31. 31. Bordes A, Usunier N, Garcia-Duran A, Weston J, Yakhnenko O. Translating embeddings for modeling multi-relational data. In: Advances in neural information processing systems 26. Red Hook, NY: Curran Associates Inc.; 2013.
  32. 32. Sun Z, Hu W, Li C. Cross-lingual entity alignment via joint attribute-preserving embedding. In: The Semantic Web–ISWC 2017: 16th International Semantic Web Conference, Vienna, Austria, October 21–25, 2017, Proceedings, Part I 16. Springer; 2017, pp. 628–44.
  33. 33. Guo L, Sun Z, Hu W. Learning to exploit long-term relational dependencies in knowledge graphs. In: Chaudhuri K, Salakhutdinov R, editors. Proceedings of the 36th International Conference on Machine Learning, vol. 97. 2019, pp. 2505–14.
  34. 34. Sun Z, Hu W, Zhang Q, Qu Y. Bootstrapping entity alignment with knowledge graph embedding. In: IJCAI'18: Proceedings of the 27th International Joint Conference on Artificial Intelligence, vol. 18. 2018.
  35. 35. Yan Z, Peng R, Wang Y, Li W. CTEA: context and topic enhanced entity alignment for knowledge graphs. Neurocomputing. 2020;410:419–31.
  36. 36. Yang HW, Zou Y, Shi P, Lu W, Lin J, Sun X. Aligning cross-lingual entities with multi-aspect information. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics; 2019.
  37. 37. Mao X, Wang W, Wu Y, Lan M. Boosting the speed of entity alignment 10: Dual attention matching network with normalized hard sample mining. In: Proceedings of the web conference 2021. 2021, pp. 821–32.
  38. 38. Chen L, Li Z, Xu T, Wu H, Wang Z, Yuan NJ, et al. Multi-modal siamese network for entity alignment. In: Proceedings of the 28th ACM SIGKDD conference on knowledge discovery and data mining. 2022, pp. 118–26.
  39. 39. Yoon S, Ko S, Kim T, Kang S, Yeo J, Lee D. Unsupervised robust cross-lingual entity alignment via neighbor triple matching with entity and relation texts. 2025. Available from: https://arxiv.org/abs/2407.15588