Next Article in Journal
Special Issue on Intelligent Diagnostic and Prognostic Methods for Electronic Systems and Mechanical Systems
Next Article in Special Issue
Survey of Text Mining Techniques Applied to Judicial Decisions Prediction
Previous Article in Journal
Applying Natural Language Processing and TRIZ Evolutionary Trends to Patent Recommendations for Product Design
Previous Article in Special Issue
Automatic Classification of Eyewitness Messages for Disaster Events Using Linguistic Rules and ML/AI Approaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Leveraging Multi-Modal Information for Cross-Lingual Entity Matching across Knowledge Graphs

1
School of Computer Science and Engineering, Southeast University, Nanjing 211189, China
2
School of Computer and Software, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Submission received: 5 September 2022 / Revised: 30 September 2022 / Accepted: 7 October 2022 / Published: 8 October 2022
(This article belongs to the Special Issue Natural Language Processing (NLP) and Applications)

Abstract

:
In recent years, the scale of knowledge graphs and the number of entities have grown rapidly. Entity matching across different knowledge graphs has become an urgent problem to be solved for knowledge fusion. With the importance of entity matching being increasingly evident, the use of representation learning technologies to find matched entities has attracted extensive attention due to the computability of vector representations. However, existing studies on representation learning technologies cannot make full use of knowledge graph relevant multi-modal information. In this paper, we propose a new cross-lingual entity matching method (called CLEM) with knowledge graph representation learning on rich multi-modal information. The core is the multi-view intact space learning method to integrate embeddings of multi-modal information for matching entities. Experimental results on cross-lingual datasets show the superiority and competitiveness of our proposed method.

1. Introduction

Knowledge graphs [1] are attracting attention from both academics and industries due to their power to model structured information and professional knowledge. In recent years, many high-quality knowledge graphs have been built, such as Microsoft Concept Graph [2], NELL [3], Zhishi.me [4], etc. Knowledge graphs have been widely used in many vertical fields [5], e.g., finance, medical and e-commerce. Knowledge graphs and their related technologies have gradually become indispensable basic technologies in the era of artificial intelligence.
However, a knowledge graph can be freely constructed by any organization or individual according to their own needs and languages. Therefore, the data in a knowledge graph can be multilingual and diverse, and there is a large amount of overlapping knowledge and complementary information across different knowledge graphs. To better utilize such various knowledge, knowledge graph matching has attracted the attention of more and more researchers. The purpose of knowledge graph matching is to integrate different knowledge graphs to form a global knowledge base and establish interoperability between the applications based on different knowledge graphs, especially in the fields of information retrieval, machine reading and knowledge base question answering [5].
The research of knowledge graph matching starts with traditional ontology matching [6], namely matching classes and properties in the schema level of knowledge graphs. More recently, with the fast development of knowledge graphs, the number of entities has increased rapidly. As a result, the entity matching between different knowledge graphs is becoming more and more important to overcome the heterogeneity problem. The process of entity matching decides whether two entities from different knowledge graphs are the same object or not. Owing to the growing number of multilingual knowledge graphs due to information globalization, entity matching across multilingual knowledge graphs (namely, cross-lingual entity matching) is becoming an urgent sub-problem to be solved for multilingual intelligent applications.
As part of the technologies on knowledge graph representation learning, cross-lingual entity matching has become much easier since we only need to train vector representations of entities from the knowledge graphs of different languages and then decide whether two entities are matched by computing vector similarities. Most of existing studies on cross-lingual entity matching with knowledge graph representation learning only use the structural information of the knowledge graph, i.e, the relation triples in the form of entity, relation, entity to learn the structural embedding of entities. For example, MTransE [7] uses TransE to map two knowledge graphs into their respective vector spaces and then learns the mapping relationship between knowledge graphs. IPTransE [8] and BootEA [9] use iterative methods to continuously discover new matched entities with updated embeddings. In addition to relation triples, some works also consider attribute triples, such as JAPE [10], etc. In addition, KDCoE [11] and HMAN [12] use the text descriptions of entities as the supplement of relation triples to realize entity matching. EVA [13] incorporates images as the complement to align entities across different knowledge graphs.
Although the existing entity matching methods based on representation learning have made remarkable achievements, they still have the following two problems. Firstly, the existing entity matching methods cannot make full use of knowledge graph relevant muti-modal information. Most of them only use the relation triples in knowledge graphs but seldom use other modal information, such as attributes, texts, images, etc. Such information has its own characteristics and can provide useful information for entity representations from different perspectives. For example, text description can provide rich textual context information for entities, and images can provide concrete visual information for entities. Secondly, it is difficult for the existing entity matching methods to integrate the representation learning methods from different modalities. Due to the heterogeneity of cross-lingual knowledge graphs, the representations of entities, relationships, attributes and images with the same meaning in different knowledge graphs may be completely different. That means the entity representation learning from each single modality is insufficient and complementary. Therefore, an effective entity matching method requires dealing with this insufficiency and integrating the information from different modalities.
To address the above problems, we propose a new cross-lingual entity matching method (called CLEM) using knowledge graph representation learning which integrates the rich multi-modal information. The motivation is that the representation of one entity from each single modality (called view) can only capture the partial entity information, while different views describe different aspects of the entity and may share some common redundant information. Therefore, the entity representations can be learned from each particular view and jointly optimized to improve the representation learning (i.e., entity embedding learning) performance. We first attempt to perform representation learning on relation triples, attribute triples, entity text descriptions and images. Then we apply multi-view intact space learning (MISL) [14] to combine multi-modal information of entities and conduct entity matching on cross-lingual datasets. The main contributions are summarized as follows:
  • We leverage four views including relations, attributes, text descriptions and images to learn entity representations, and each view corresponds to an independent learning model;
  • We apply multi-view intact space learning to solve the insufficiency in each individual view and integrate the multi-modal view information to obtain the entity representation in the intact space;
  • We perform experiments on cross-lingual datasets and evaluate our CLEM with different evaluation metrics. The experimental results show that the proposed method outperforms the state-of-the-art methods in most evaluation criteria.

2. Related Work

We discuss two lines of works that are relevant to this paper.
Knowledge Graph Representation Learning. In recent years, knowledge graph representation learning has attracted the attention of researchers. The current representation learning models can be roughly divided into three categories, which are translation-based models, semantic matching models and neural-network-based models. The typical methods of translation-based models are TransE [15] and its improved models TransH [16] and TransR [17]. The core idea of TransE is to regard the relationship as the translation from the head entity to the tail entity. For any relation triples in the knowledge graph, TransE expects that the head entity vector plus the relation vector equals the tail entity vector. The semantic matching models use similarity functions to infer relational facts, e.g., the Hadamard product in ComplEx [18] and the circular correlation in HolE [19]. The neural-network-based models exploit deep learning techniques for knowledge graph embedding. For example, ConvE [20] is a multi-layer convolution neural network which learns the representation through deep network structure and convolution operations; R-GCN [21] is a relational graph convolution neural network which generates the representation by convolving semantic information on the local graph structure. All the above models focus on relational facts and are mostly evaluated by the task of link prediction in a single knowledge graph.
Entity Matching. With the emergence of various knowledge graphs in different domains, entity matching, especially cross-lingual entity matching, is becoming more and more important to solve the problem of heterogeneity. Entity matching decides whether two entities of different knowledge graphs refer to the same object. Ref. [22] transforms the entity matching problem based on attribute similarity scores into a multi-classification problem, which is divided into matching, possible matching and mismatching, and establishes the probability model of the entity matching problem. Although the above method is simple and intuitive, it requires sufficient labeled matching entity pairs and the corresponding annotation costs. Additionally, the calculation of feature similarities is often interfered with by semantic heterogeneity among different knowledge graphs. Recently, the use of the embedding models to match the entities across cross-lingual knowledge graphs has attracted extensive attention. JE [23] uses the model TransE to embed two independent knowledge graphs into the same vector space; MTransE [7] jointly trains a translation embedding model to encode language-specific knowledge graphs in separate embedding spaces and align counterpart entities across embeddings. Some studies also consider other aspects of the entity for the matching task. Some of them (e.g., MultiKE [24] and DAEA [25]) consider the name and structure of entities; others (e.g., JAPE [10], GCN-Align [26], AttrE [27] and RAKA [28]) consider entity attributes; some works (e.g., KDCoE [11] and HMAN [12]) consider the text descriptions of entities. Moreover, EVA [13] incorporates images along with structures, relations and attributes to align entities across different knowledge graphs.

3. Method

3.1. Problem Definition

In this paper, we study the cross-lingual entity matching using knowledge graph representation learning, which aims to learn the multi-view entity embeddings based on different modal perspectives (i.e., views). We consider four views for the entity embedding, including relation view, attribute view, description view and image view. Based on these views, we formalize a knowledge graph as a 6-tuple G = { E , R , A , L , X , I } , where E, R, A, L, X and I are respectively employed to represent entities, relations, attributes, attribute values, entity text descriptions and images of knowledge graph G.
The problem of entity matching based on knowledge graph representation learning can be defined as follows: Assuming G 1 and G 2 are, respectively, employed to represent two different knowledge graphs.
G 1 = { E 1 , R 1 , A 1 , L 1 , X 1 , P 1 }
G 2 = { E 2 , R 2 , A 2 , L 2 , X 2 , P 2 }
S denotes the set of known equivalent entity pairs from G 1 and G 2 , and ≡ denotes the equivalent relation.
S = e 1 , e 2 | e 1 E 1 , e 2 E 2 , e 1 e 2
We obtain the embedded representation of each entity from G 1 and G 2 by using the representation learning model of different views and integrate the multi-modal information from each view to generate the final entity representation. Assuming E 1 and E 2 respectively represent the embedding matrix for the set of all entities E 1 and E 2 , then, the function computes the similarities between the entity representations to find the equivalent entity set M.
M = ( E 1 , E 2 )
The purpose is to find a set M that contains all matched entity pairs which do not belong to the known set S.
M = e 1 , e 2 | e 1 E 1 , e 2 E 2 , e 1 e 2 , ( e 1 , e 2 ) S

3.2. Overall Framework

Our framework is shown in Figure 1 and consists of two parts: entity representation learning in different views and multi-view entity matching.
Entity Representation Learning in Different Views. We make full use of the knowledge graph, extracting the entity information from the relation view, the attribute view, the description view and the image view.
Multi-view Entity Matching. Since the embedded representations of entities from a single view only capture partial entity information, we employ multi-view intact space learning to integrate the complementary information from each view, which obtains the multi-view embedded representations. We utilize the K-nearest neighbor algorithm [29] to find out all matched entity pairs.

3.3. Entity Representation Learning in Different Views

This part consists of the entity embedded representations from four views. We study the embedding models for different views, respectively.

3.3.1. Relation View

Relations are an important part of a knowledge graph, since the relation view describes the structures among different entities. The translation-based knowledge graph embedding model shows its power in characterizing such relational structures [30]. Therefore, we employ the TransE [15] model, which has good generalization capability, to embed knowledge graphs from the relation view. Assuming G 1 and G 2 denote the two knowledge graphs to match, the translation score could be obtained as below:
f r e l ( h , r , t ) = | | h + r t | | 2 2
where T = ( h , r , t ) denotes a relation triple in G { G 1 , G 2 } such that h means the head entity, t means the tail entity, and r means the relation. To learn the common embeddings of entities and relations, we embed two different knowledge graphs into the same vector space. The learning objective of the model can be achieved by minimizing the following margin-based [31] loss function:
L o s s r e l = G G 1 , G 2 ( h , r , t ) G | | f r e l ( h , r , t ) f r e l ( h ˜ , r , t ˜ ) + γ | | + ( e 1 , e 2 ) S | | e 1 e 2 | |
where ( h ˜ , r , t ˜ ) is the negative example obtained by randomly replacing the head entity or tail entity in ( h , r , t ) with another entity, S means the known equivalent entity set, which is defined in Section 3.1, and  γ is the margin parameter describing the boundary between positive and negative examples.

3.3.2. Attribute View

The attribute view characterizes attribute and attribute value information of entities. There are word abbreviations in attributes and cross-lingual differences in attribute values, which make entity matching based on the attribute view quite difficult. To leverage the information of attributes and attribute values to help match entities, we use a graph convolutional network (GCN) [32] to encode the neighbour attributes and attribute values into the the low-dimensional representations of the given entities. GCN can perform feature extraction on arbitrary graphs and will not be affected by the number of neighbour nodes associated with entities.
Firstly, we construct the entity attribute graph. All attribute triples are classified based on the head entity, i.e., attribute triples of the same head entity are grouped together. Then, for an entity e, we collect all its attributes and attribute values. We combine each collected attribute and its corresponding attribute value as a neighbour node of the entity node in the entity attribute graph. Figure 2 shows an example of the entity attribute graph for the entity “Donald Trump”.
Secondly, we adopt GCN to encode attributes and attribute values into the representations of entities. We initialize the vector representations of entities, attributes and attribute values using the pre-training model BERT [33]. BERT is a multi-layer bidirectional Transformer encoder, and we use its original implementation, the BERT-base model. To obtain the representation of each neighbour node for the entity node in the built entity attribute graph, we perform element-wise multiplication between the vectors of attributes and attribute values, respectively. Afterwards, GCN is utilized to conduct convolution operations as follows:
H ( l + 1 ) = R e L U ( D ^ 1 2 A ^ D ^ 1 2 H ( l ) W ( l + 1 ) )
where H denotes the feature matrix of all nodes, each row in H is the vector representation of a node, A is an adjacency matrix showing the connectivity between nodes, A ^ = A + I , I is the unit matrix, D ^ is the diagonal matrix of A ^ , W is the weight matrix, and l means the l-layer. We choose R e L U = max ( 0 , x ) for the activation function.
Owing to the single-layer structure of the entity attribute graph, we assign a single-layer GCN for each knowledge graph to encode attributes and attribute values into the representations of entities. Specifically, given two knowledge graphs G 1 and G 2 , the objective can be achieved by minimizing the following margin-based loss function:
L o s s a t t r = ( e 1 , e 2 ) S [ f ( e 1 , e 2 ) γ 1 ] + + ( e 1 , e 2 ) S [ γ 2 f ( e 1 , e 2 ) ] +
where [ x ] + means m a x { 0 , x } , f ( x , y ) means the L 2 distance of vector x and vector y , ( e 1 , e 2 ) is a negative example obtained by randomly replacing one of the two entities in ( e 1 , e 2 ) with another entity, and  γ 1 / 2 is the margin parameter describing the boundary between positive and negative examples.

3.3.3. Description View

The description view contains rich textual semantic information, which is difficult to obtain from other views. To extract such semantic information from the description view, we encode the entity text descriptions through two steps. Firstly, the embedded representation of words in the entity text descriptions is obtained by using a pre-training model. After pre-training, the description of each entity will be converted into a vector sequence, which will be input to the description encoder, and then the encoder will output the embedded representation of the entity text description. The details are as follows:
For the pre-training language model, we also use BERT which is proposed by Devlin et al. [33]. As for the encoder, we adopt gated recurrent unit (GRU) [34] since it can model the sequential data well, while it is computationally more efficient than LSTM [35] or other models. GRU is a kind of recurrent neural network (RNN), which is often used to encode natural language text. It consists of two types of gating units, namely the reset gate and the update gate, which are used to track the sequence status without using a separate storage unit. In this paper, two stacked GRU layers are used to model entity descriptions in different knowledge graphs. The input is the embeddings of an entity and its descriptions obtained by BERT. We follow the standard method where the tokens are surrounded by [CLS] and [SEP] on the left and right, respectively. For example, if the input entity and description are “Trump” and “President Trump has insisted on a full-scale convention.”, the input to our encoder would be: [CLS]+ “Trump” +[SEP]+ “President Trump has insisted on a full-scale convention.” + [SEP]. The output of the second GRU layer can be obtained as follows:
t e = R e L U ( W o [ s 1 ; s 2 ; ; s l ] + b o )
where W o denotes the mapping matrix, b o is the bias, and l is the length of the entity and description sequence. We also choose ReLU as the activation function. Then, each t e is normalized to | | t e | | 2 2 = 1 . Finally, we can maximize the similarity (i.e., the dot product) between the embeddings of descriptions for matched entities and minimize the dot product between the embeddings of descriptions for irrelevant entities. This objective can be achieved by minimizing the following loss function:
L o s s t e x t = ( e 1 , e 2 ) S { l o g ( R e L U ( t e 1 T t e 2 ) ) + i = 1 k E e i U ( e i E 2 ) [ l o g ( R e L U ( t e 1 T t e i ) ) ] }
where S means the known equivalent entity set, k is the number of negative entity examples, and U is the distribution of the entities in the entity set E 2 .

3.3.4. Image View

The image view provides concrete images of entities, characterizing the visual information of entities more intuitively and vividly. We can directly distinguish the president “Donald Trump” from the physicist “John G. Trump” because the facial features are totally different.
To extract image features, we obtain the embeddings of the entity images according to the VGG16 [36] model. The VGG16 model was pre-trained on the ILSVRC 2012 dataset which is derived from ImageNet [37]. The model we introduce has thirteen convolutional layers, which are followed by three fully connected layers. To obtain the embeddings of the entity images, we remove the softmax layer. Owing to the fact that the image vectors generated from the VGG16 model do not share the same vector space with entity embeddings, we use a m a p function to make them in the same space. Specifically, given a pair of an entity and its corresponding image { ( e , i ) | e E , i I } , where E and I denote the entity and image set in the given two knowledge graphs, respectively, we extract image features as follows:
f i m ( e , i ) = | | e R e L U ( m a p ( i ) ) | | 2 2
where i denotes the embedding of the entity image i, and each i is normalized to | | i | | 2 2 = 1 . ReLU is the activation function. Given the known equivalent entity set S, we can maximize the similarity (i.e., the dot product) between the embeddings of images for matched entities and minimize the dot product between the embeddings of images for irrelevant entities. Thus, we define the following loss function:
L o s s i m = ( e , i ) { ( e , i ) | e E , i I } l o g ( 1 + e x p ( f i m ( e , i ) ) ) ( e 1 , e 2 ) S { l o g ( R e L U ( i e 1 T i e 2 ) ) + i = 1 k E e i U ( e i E 2 ) [ l o g ( R e L U ( i e 1 T i e i ) ) ] }
where k is the number of negative entity examples, and U is the distribution of the entities in the entity set E 2 .

3.4. Multi-View Entity Matching

In Section 3.3, we introduce four representation learning models for modeling different views. Each model can be trained to obtain the embedded representations of entities, and the entity matching task can be completed by calculating similarities between entity embeddings. However, the performance of entity matching cannot be maximized by only using one of the models [14]. Therefore, it is crucial to combine the four representations for each entity, i.e., integrate the multi-modal information from each view to generate the final entity representations.

3.4.1. Multi-View Intact Space Learning

The information from different views can provide different aspects of one entity. For example, the attribute view can provide abstract attribute information of entities themselves, and the relation view can provide the relational structures among different entities. However, each view only captures the entity information from one aspect, and all views may share the common redundant information. We adopt the multi-view intact space learning (MISL) [14] to integrate multi-view entity representations to solve the insufficiency problem when using each individual view.
We assume a multi-view training data set D = { z i v | 1 i n , 1 v m } , where z i v denotes the v-th view of the i-th entity embedding, n is the number of the entity, and m is the number of the entity embedding views. Suppose that x i represents the embedding of an entity in the intact space, so the loss function of the m-view space being reconstructed by the intact space can be defined as follows:
L o s s M I S L = 1 m n i = 1 n v = 1 m l o g ( 1 + | | z i v W v x i | | 2 c 2 ) + C 1 v = 1 m | | W v | | F 2 + C 2 i = 1 n | | x i | | 2 2
where c is a constant scale parameter, C 1 and C 2 are non-negative constants, and  W v is the v-th view generation matrix. The optimization problem can be decomposed into two sub-problems over the view generation function W and the variable on entity embedding x with the alternating optimization method. Given fixed view generation functions { W v } v = 1 m , the loss function can be reduced as follows:
m i n x ( L o s s M I S L ) = 1 m v = 1 m l o g ( 1 + | | z v W v x | | 2 c 2 ) + C 2 | | x | | 2 2
Setting the gradient of L o s s M I S L with respect to x to 0, we can obtain the following equations:
Q v = 1 | | z v W v x | | 2 + c 2
x = ( v = 1 m W v T Q v W v + m C 2 ) 1 v = 1 m W v T Q v z v
Given fixed entity embeddings { x i } i = 1 n , the loss function can be reduced as follows:
m i n W ( L o s s M I S L ) = 1 n i = 1 n l o g ( 1 + | | z i W x i | | 2 c 2 ) + C 1 | | W | | 2 2
We can set the gradient of L o s s M I S L with respect to W to 0 and obtain the following equations:
Q i = 1 | | z i Wx i | | 2 + c 2
W = i = 1 n z i Q i x i T ( i = 1 n z i Q i x i T + n C 1 ) 1
By constantly iterating the above formulas until the entity embedding x converges, we can obtain the final integrated embedding for each entity (see Algorithm 1).

3.4.2. Entity Matching Algorithm

This section introduces the whole entity matching process in Algorithm 1. All random initialization for the embeddings from the relation view is provided by the X a v i e r initializer [38], and the pre-training initialization for the attribute and description view is provided by the BERT. The input of the algorithm is two knowledge graphs G 1 and G 2 to be matched, the known equivalent entity set S and the number of iterations Q. We first separately train the entity embeddings from the relation view, attribute view, description view and image view until Q is reached. Then, we obtain the final multi-view entity representations by MISL. Finally, we find matched entities by the K-nearest neighbor algorithm [39] on the final entity representations. The output of the algorithm is the predicted pairs of matched entities M across two knowledge graphs in different languages.
Algorithm 1: The Entity Matching Algorithm in CLEM.
Input: 
G 1 , G 2 , the known equivalent entity set S, and the number of iterations Q.
Output: 
The predicted matched entity pairs M.
1:
for q = 1, 2, 3,...,Q do
2:
    Minimize L o s s r e l under the relation view;
3:
    Minimize L o s s a t t r under the attribute view;
4:
    Minimize L o s s t e x t under the description view;
5:
    Minimize L o s s i m under the image view;
6:
end for
7:
repeat
8:
    Iterate the Equations (16), (17), (19) and (20);
9:
until The entity embedding x in the intact space converges
10:
Find matched entities by the K-Nearest Neighbor algorithm on the final entity representations;
11:
return The predicted matched entity pairs M

4. Experiments

To verify the effectiveness of our proposed method, we used Python to implement our approach with the aid of PyTorch (https://pytorch.org/) All reported experiments were performed on one Linux server with Xeon CPU (2.10GHz) processor with 64 GB RAM and one NVIDIA TITAN Xp GPU (8 GB).

4.1. Dataset

The experiments were performed on DBP15K [10], which is a subset of DBpedia and a classic benchmark dataset for entity matching. DBpedia is a large multilingual knowledge graph which extracts structured content from Wikipedia. We obtained entity descriptions and entity images, respectively, from the “dbpedia-owl:abstract” and “dbpedia-owl:thumbnail”. As for the missing entity images, we obtained the supplement from Google Images (https://images.google.com) by querying entity names. DBP15K contains three sub-datasets, i.e., Chinese–English (zh-en), Japanese–English (ja-en), and French–English (fr-en) subsets, which contain language links between different language versions. Table 1 shows the details of DBP15K. Each subset contains the knowledge in two languages and 15,000 pairs of matched entities from DBpedia. In the experiment, the known matched entity pairs were used for model training and testing.

4.2. Evaluation Measures

Link prediction is used to predict possible links in knowledge graphs or compute link losses caused by incomplete data. Similar to MTransE [7] and subsequent related works, we used link prediction as the evaluation method for entity matching. We applied mean reciprocal rank (MRR) and Hits@n as the evaluation criteria. Given query samples Q, the MRR and Hits@n are defined as follows:
M R R = 1 | Q | i = 1 | Q | 1 r a n k i
where r a n k i is the rank of the correct answer in the response list for the i-th query.
H i t s @ n = 1 | Q | i = 1 | Q | H i t ( q i , L i , n )
where H i t ( q i , L i , n ) means that if the correct answer of the query q i is at the first n items for the ranked list L i , then its value is one, otherwise, the value is zero.
Moreover, we also tried to record traditional evaluation metrics for entity matching including precision, recall and F1-score. However, since the embedding-based methods always return a list of candidates for each input entity, recall and F1-score are equal to precision, and the precision actually equals Hits@1 [24].

4.3. Comparison Methods

We compared our method with the following baselines:
  • MTransE [7]: MTransE is one representative work of multilingual knowledge graph embedding for entity matching. MTransE combines monolingual models with a jointly trained alignment model and achieves good results on the single relation view.
  • JAPE [10]: JAPE is a joint embedding model on relation triples and attribute triples for entity matching between knowledge graphs.
  • KDCoE [11]: KDCoE is a semi-supervised entity matching method based on collaborative training which enhances multilingual knowledge graph embedding. It iteratively trains the two parts of the embedding model on the relation triples and entity descriptions in different languages, respectively.
  • MultiKE [24]: Multi-KE is an entity matching framework based on multi-view knowledge graph embedding. The underlying idea is to divide the various features of knowledge graphs into multiple subsets, which are complementary to each other.
  • EVA [13]: This research proposes the idea of entity alignment using visual information, which incorporates images along with relations and attributes to align entities in different knowledge graphs.
  • DAEA [25]: DAEA utilizes graph convolutional networks (GCNs) to integrate the information of entities, relations, attributes and entity name embeddings to learn a unified latent representation to perform cross-lingual entity alignment.
Moreover, we performed the ablation experiment and the multi-view integration experiment to evaluate the effectiveness of each view and our MIST method. For the ablation experiment, we eliminated one of the four views to generate the CLEM variant each time and denoted the CLEM variants without the image view, relation view, the attribute view and the description view by CLEM-RAD, CLEM-ADI, CLEM-RDI and CLEM-RAI, respectively. For the the multi-view integration experiment, we followed three multi-view integration methods of the MultiKE model (i.e., weighted view averaging, shared space learning, and in-training combination) to generate the CLEM variants which are denoted by CLEM-WVA, CLEM-SSL and CLEM-ITC, respectively. In addition, the the multi-view integration experiment included the CLEM variant which employs the vector concatenation strategy to combine the view-specific embeddings. This CLEM variant is denoted by CLEM-CAT.

4.4. Experiment Settings

The following hyper-parameters were used in the experiments. Each training took Q = 3000 epochs, and the proposed model used in our experiments was trained with the Adam as optimizer, the batch size among { 32 , 64 , 128 , 256 , 512 } , the embedding dimension among { 50 , 75 , 100 , 125 , 150 } , the margin among among { 0.5 , 1.0 , 1.5 , 2.0 , 2.5 , 3.0 } and the learning rate among { 0.0001 , 0.0005 , 0.001 , 0.005 , 0.01 } . The best hyperparameters for “(batch size, embedding dimension, margin, learning rate)” vary in different sub-datasets, and the details are given in Table 2. For the baseline models, we used the reported results in their papers or ran their source codes on DBP15K. Specifically, for MTransE and JAPE, we used the results given in [10] (page 639); for EVA and DAEA, we used the results given in [13] (page 4261) and [25] (page 6), respectively; for MultiKE (https://github.com/nju-websoft/MultiKE) and KDCoE (https://github.com/muhaochen/MTransE-tf), we applied their published source codes to DBP15K to obtain evaluation results.

4.5. Results and Discussions

Table 3 shows the comparison results between CLEM and other entity matching methods. We found that CLEM achieved the best performance on the three sub-data sets. For example, on DBP15K zh-en, CLEM achieved the Hits@1 score 85.41% with 2.6% improvement compared to the second best method. The effectiveness of CLEM partly comes from the integration of the multi-modal entity information. CLEM fuses the multi-modal information from the relation view, attribute view, description view and the image view. In contrast to that, MTransE, JAPE and KDCoE only use a part of the above four views. Moreover, although MultiKE uses multi-view integration models to deal with each kind of entity information equally, but the performance is not as good as ours (details are presented in Figure 3). Instead, CLEM adopts multi-view intact space learning (MISL) to solve the insufficiency problem of each individual view and integrate the multi-modal information from all given views. As for DAEA and EVA, they perform relatively well. EVA uses the additional visual embedding to characterise each entity in the embedding space, while EVA ignores the importance of the text description. Although DAEA integrates multi-aspect information to achieve good results, it neglects the use of entity images to eliminate the ambiguity caused by text relevant modal information.
The Ablation Experiment.Table 4 shows the results of the ablation experiment. As expected, the performance of different variants without the corresponding modal information decreases obviously. The results show that each modal information is indispensable and plays an important role in the task of entity matching. In addition, we noticed that CLEM-RAD performed the worst among these variants. This phenomenon indicates that CLEM-RAD lacks the most effective view of the four views, i.e., the image view. The effectiveness of the image view is attributed to the fact that we extracted visual features to eliminate the ambiguity in the relation view, attribute view and text view, which is common in heterogeneous knowledge graphs.
Moreover, although we found that the image view is the most effective view, the entity matching results rely on image quality to a certain extent according to our error analysis. Since some images are missing in the DBpedia dataset, we collected images from Google Images which also introduces some noise due to the color, background and brightness differences. Such noisy images do cause a number of errors in the entity matching results.
The Multi-View Integration Experiment. To verify the effectiveness of our multi-view representation learning method, we conducted the multi-view integration experiment in terms of Hits@1. Figure 3 shows the results. Compared with other variants, we observed that the Hits@1 of our proposed method increased at least 10.22%. This is because compared with the multi-view integration methods of MultiKE, our applied MISL handled the relevance and complementarity between different modal entity information in the intact space better. The use of the MISL method successfully integrates the multi-modal information from each view to generate the final entity representation, which effectively improves the results of the entity matching. In addition, we found that the CLEM-CAT is inferior to the other variants. This may be due to the fact that the concatenation method does not have the training process, and the integration after training cannot fully consider the association between different modalities.

5. Conclusions

In this paper, we proposed a cross-lingual knowledge graph embedding method CLEM for entity matching, which extracts and integrates the rich multi-modal information from different views, including the relation view, attribute view, description view and image view. The multi-view intact space learning was adopted to generate the final multi-view entity representations, which effectively solved the insufficiency problem of the single modal information in cross-lingual entity matching. Our experiments on three real-world datasets show the effectiveness of CLEM and our multi-view learning strategy.
As for future work, we plan to study the few-shot cross-lingual entity matching with knowledge graph embedding when the known matched entities are quite few. Moreover, we will explore a global matching method for align at least three knowledge graphs in different languages simultaneously.

Author Contributions

T.W. and C.G. proposed the idea of this work and designed the method; L.L. finished a part of the experiments; T.W. and C.G. wrote the paper; Y.W. revised this paper and gave a lot of suggestions. All authors read and approved the final manuscript.

Funding

This work is supported by the NSFC (Grant No. 62006040, 62072149), the Project for the Doctor of Entrepreneurship and Innovation in Jiangsu Province (Grant No. JSSCBS20210126), the Fundamental Research Funds for the Central Universities, and ZhiShan Young Scholar Program of Southeast University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, T.; Qi, G.; Li, C.; Wang, M. A survey of techniques for constructing Chinese knowledge graphs and their applications. Sustainability 2018, 10, 3245. [Google Scholar] [CrossRef] [Green Version]
  2. Wu, W.; Li, H.; Wang, H.; Zhu, K.Q. Probase: A Probabilistic Taxonomy for Text Understanding. In Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data, Scottsdale, AZ, USA, 20–24 May 2012; pp. 481–492. [Google Scholar]
  3. Carlson, A.; Betteridge, J.; Kisiel, B.; Settles, B.; Hruschka, E.R.; Mitchell, T.M. Toward an Architecture for Never-Ending Language Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Atlanta, GA, USA, 11–15 July 2010. [Google Scholar]
  4. Wu, T.; Wang, H.; Li, C.; Qi, G.; Niu, X.; Wang, M.; Li, L.; Shi, C. Knowledge graph construction from multiple online encyclopedias. World Wide Web 2020, 23, 2671–2698. [Google Scholar] [CrossRef]
  5. Wang, J.; Wang, X.; Ma, C.; Kou, L. A survey on the development status and application prospects of knowledge graph in smart grids. IET Gener. Transm. Distrib. 2021, 15, 383–407. [Google Scholar] [CrossRef]
  6. Liu, X.; Tong, Q.; Liu, X.; Qin, Z. Ontology Matching: State of the Art, Future Challenges, and Thinking Based on Utilized Information. IEEE Access 2021, 9, 91235–91243. [Google Scholar] [CrossRef]
  7. Chen, M.; Tian, Y.; Yang, M.; Zaniolo, C. Multilingual knowledge graph embeddings for cross-lingual knowledge alignment. arXiv 2016, arXiv:1611.03954. [Google Scholar]
  8. Zhu, H.; Xie, R.; Liu, Z.; Sun, M. Iterative Entity Alignment via Joint Knowledge Embeddings. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Melbourne, Australia, 19–25 August 2017; pp. 4258–4264. [Google Scholar]
  9. Sun, Z.; Hu, W.; Zhang, Q.; Qu, Y. Bootstrapping Entity Alignment with Knowledge Graph Embedding. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 13–19 July 2018; pp. 4396–4402. [Google Scholar]
  10. Sun, Z.; Hu, W.; Li, C. Cross-Lingual Entity Alignment via Joint Attribute-Preserving Embedding. In Proceedings of the International Semantic Web Conference, PART I, Vienna, Austria, 21–25 October 2017; Springer: Cham, Switzerland; pp. 628–644. [Google Scholar]
  11. Chen, M.; Tian, Y.; Chang, K.W.; Skiena, S.; Zaniolo, C. Co-training embeddings of knowledge graphs and entity descriptions for cross-lingual entity alignment. arXiv 2018, arXiv:1806.06478. [Google Scholar]
  12. Yang, H.W.; Zou, Y.; Shi, P.; Lu, W.; Lin, J.; Sun, X. Aligning cross-lingual entities with multi-aspect information. arXiv 2019, arXiv:1910.06575. [Google Scholar]
  13. Liu, F.; Chen, M.; Roth, D.; Collier, N. Visual Pivoting for (Unsupervised) Entity Alignment. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 2–9 February 2021; pp. 4257–4266. [Google Scholar]
  14. Xu, C.; Tao, D.; Xu, C. Multi-view intact space learning. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 2531–2544. [Google Scholar] [CrossRef] [Green Version]
  15. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 2013, 26, 1–8. [Google Scholar]
  16. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge Graph Embedding by Translating on Hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; pp. 1112–1119. [Google Scholar]
  17. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning Entity and Relation Embeddings for Knowledge Graph Completion. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 2181–2187. [Google Scholar]
  18. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex embeddings for simple link prediction. In Proceedings of the International Conference on Machine Learning, New York City, NY, USA, 19–24 June 2016; pp. 2071–2080. [Google Scholar]
  19. Nickel, M.; Rosasco, L.; Poggio, T. Holographic Embeddings of Knowledge Graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1955–1961. [Google Scholar]
  20. Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2D Knowledge Graph Embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 1811–1818. [Google Scholar]
  21. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Berg, R.v.d.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In Proceedings of the European Semantic Web Conference, Heraklion, Crete, Greece, 3–7 June 2018; Springer: Berlin/Heidelberg, Germany; pp. 593–607. [Google Scholar]
  22. Fellegi, I.P.; Sunter, A.B. A theory for record linkage. J. Am. Stat. Assoc. 1969, 64, 1183–1210. [Google Scholar] [CrossRef]
  23. Hao, Y.; Zhang, Y.; He, S.; Liu, K.; Zhao, J. A Joint Embedding Method for Entity Alignment of Knowledge Bases. In Proceedings of the China Conference on Knowledge Graph and Semantic Computing, Beijing, China, 19–22 September 2016; Springer: Berlin/Heidelberg, Germany; pp. 3–14. [Google Scholar]
  24. Zhang, Q.; Sun, Z.; Hu, W.; Chen, M.; Guo, L.; Qu, Y. Multi-view knowledge graph embedding for entity alignment. arXiv 2019, arXiv:1906.02390. [Google Scholar]
  25. Zhang, G.; Zhou, Y.; Wu, S.; Zhang, Z.; Dou, D. Cross-lingual entity alignment with adversarial kernel embedding and adversarial knowledge translation. arXiv 2021, arXiv:2104.07837. [Google Scholar]
  26. Wang, Z.; Lv, Q.; Lan, X.; Zhang, Y. Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 349–357. [Google Scholar]
  27. Trisedya, B.D.; Qi, J.; Zhang, R. Entity Alignment between Knowledge Graphs Using Attribute Embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 297–304. [Google Scholar]
  28. Chen, B.; Zhang, J.; Tang, X.; Chen, H.; Li, C. RAKA: Co-Training of Relationships and Attributes for Cross-lingual Knowledge Alignment. arXiv 2019, arXiv:1910.13105. [Google Scholar]
  29. Fix, E.; Hodges, J.L. Discriminatory analysis. Nonparametric discrimination: Consistency properties. Int. Stat. Rev. Int. Stat. 1989, 57, 238–247. [Google Scholar] [CrossRef]
  30. Ji, S.; Pan, S.; Cambria, E.; Marttinen, P.; Philip, S.Y. A survey on knowledge graphs: Representation, acquisition, and applications. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 494–514. [Google Scholar] [CrossRef]
  31. Wu, C.Y.; Manmatha, R.; Smola, A.J.; Krahenbuhl, P. Sampling matters in deep embedding learning. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2840–2848. [Google Scholar]
  32. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2017, arXiv:1609.02907. [Google Scholar]
  33. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  34. Jozefowicz, R.; Zaremba, W.; Sutskever, I. An Empirical Exploration of Recurrent Network Architectures. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 2342–2350. [Google Scholar]
  35. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef]
  36. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  37. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  38. Kumar, S.K. On weight initialization in deep neural networks. arXiv 2017, arXiv:1704.08863. [Google Scholar]
  39. Peterson, L.E. K-nearest neighbor. Scholarpedia 2009, 4, 1883. [Google Scholar] [CrossRef]
Figure 1. The framework of entity matching. G 1 and G 2 denote different knowledge graphs, M I S L means multi-view intact space learning, and we use M I S L to integrate entity information from different modalities.
Figure 1. The framework of entity matching. G 1 and G 2 denote different knowledge graphs, M I S L means multi-view intact space learning, and we use M I S L to integrate entity information from different modalities.
Applsci 12 10107 g001
Figure 2. The example of the entity attribute graph for the entity “Donald Trump”.
Figure 2. The example of the entity attribute graph for the entity “Donald Trump”.
Applsci 12 10107 g002
Figure 3. The result of the multi-view integration experiment in terms of Hits@1.
Figure 3. The result of the multi-view integration experiment in terms of Hits@1.
Applsci 12 10107 g003
Table 1. The details of the DBP15K dataset.
Table 1. The details of the DBP15K dataset.
Sub-Dataset#Entity#Relation#Relation Triple#Attribute Triple
Chinese66,4692830153,929379,684
English98,1252317237,674567,755
Japanese65,7442043164,373354,619
English95,6802096233,319497,230
French66,8581379192,191528,665
English105,8892209278,590576,543
Table 2. The best hyper-parameters for different sub-datasets of DBP15K.
Table 2. The best hyper-parameters for different sub-datasets of DBP15K.
DatasetBatch SizeEmbedding
Dimension
MarginLearning Rate
zh-en0.0051251256
ja-en0.0051251.5256
fr-en0.0051251.5128
Table 3. The comparison results between CLEM and other entity matching methods.
Table 3. The comparison results between CLEM and other entity matching methods.
DBP15K zh-en ja-en fr-en
MethodsHits@1
(Prec.)
Hits@10MRRHits@1
(Prec.)
Hits@10MRRHits@1
(Prec.)
Hits@10MRR
MTransE30.8361.4136.4027.8657.4534.9024.4155.5533.50
JAPE41.1874.4649.0036.2568.5047.6032.3966.6843.00
KDCoE43.4275.7752.7039.4870.7550.5033.5969.5244.70
MultiKE50.8757.6153.2039.3048.8542.6063.9471.1966.50
EVA76.1090.7081.4076.2091.3081.7079.3094.2084.70
DAEA82.8092.4084.3087.0095.1088.2093.6097.1094.80
CLEM85.4193.4587.9088.4895.8190.4093.6097.6495.20
Table 4. The comparison results of CLEM and its variants in the ablation experiments.
Table 4. The comparison results of CLEM and its variants in the ablation experiments.
DBP15K zh-en ja-en fr-en
ModelsHits@1Hits@10MRRHits@1Hits@10MRRHits@1Hits@10MRR
CLEM-RAD53.4580.0256.5051.7084.1955.0757.1371.6160.40
CLEM-ADI64.4583.4168.2074.3780.3877.9074.2580.4076.20
CLEM-RDI56.4678.3560.8062.4573.6265.1067.8572.2469.30
CLEM-RAI61.8780.2964.7069.0982.6474.3071.9978.2074.50
CLEM85.4193.4587.9088.4895.8190.4093.6097.6495.20
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, T.; Gao, C.; Li, L.; Wang, Y. Leveraging Multi-Modal Information for Cross-Lingual Entity Matching across Knowledge Graphs. Appl. Sci. 2022, 12, 10107. https://0-doi-org.brum.beds.ac.uk/10.3390/app121910107

AMA Style

Wu T, Gao C, Li L, Wang Y. Leveraging Multi-Modal Information for Cross-Lingual Entity Matching across Knowledge Graphs. Applied Sciences. 2022; 12(19):10107. https://0-doi-org.brum.beds.ac.uk/10.3390/app121910107

Chicago/Turabian Style

Wu, Tianxing, Chaoyu Gao, Lin Li, and Yuxiang Wang. 2022. "Leveraging Multi-Modal Information for Cross-Lingual Entity Matching across Knowledge Graphs" Applied Sciences 12, no. 19: 10107. https://0-doi-org.brum.beds.ac.uk/10.3390/app121910107

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop