Next Article in Journal
Natural Frequencies and Modes of Electrostatically Actuated Curved Bell-Shaped Microplates
Next Article in Special Issue
Machine Learning Approach for Personality Recognition in Spanish Texts
Previous Article in Journal
Inline Optical Coherence Tomography for Multidirectional Process Monitoring in a Coaxial LMD-w Process
Previous Article in Special Issue
Conversational AI over Military Scenarios Using Intent Detection and Response Generation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Entity Linking by Introducing Knowledge Graph Structure Information

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
2
Key Laboratory of Network Information System Technology (NIST), Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100190, China
3
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100190, China
4
QILU Research Institute, Aerospace Information Research Institute, Chinese Academy of Sciences, Jinan 250000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(5), 2702; https://doi.org/10.3390/app12052702
Submission received: 27 January 2022 / Revised: 25 February 2022 / Accepted: 3 March 2022 / Published: 5 March 2022
(This article belongs to the Special Issue Current Approaches and Applications in Natural Language Processing)

Abstract

:
Entity linking involves mapping ambiguous mentions in documents to the correct entities in a given knowledge base. Most of the current methods are a combination of local and global models. The local model uses the local context information around the entity mention to independently resolve the ambiguity of each entity mention. The global model encourages thematic consistency across the target entities of all mentions in the document. However, the known global models calculate the correlation between entities from a semantic perspective, ignoring the correlation information between entities in nature. In this paper, we introduce knowledge graphs to enrich the correlation information between entities and propose an entity linking model that introduces the structural information of the knowledge graph (KGEL). The model can fully consider the relations between entities. To prove the importance of the knowledge graph structure, extensive experiments are conducted on multiple public datasets. Results illustrate that our model outperforms the baseline and achieves superior performance.

1. Introduction

The named entity linking (NEL) task refers to correctly linking entity mentions in text to entities in a structured knowledge base (such as Wikipedia, Freebase [1], or YAGO [2]), which can solve the ambiguity of mentions in natural language processing. In Figure 1, for example, a mention of “Michael Jordan” may correspond to entity entries in the knowledge base (KB) such as “Michael Jordan”, “Michael I. Jordan”, “Michael Jordan (footballer)”, “Michael B. Jordan”, etc. The entity linking (EL) involves linking the mention “Michael Jordan” to the correct entity “Michael I. Jordan” in the KB. Entity linking is also the basis of many other natural language processing tasks, such as knowledge base question and answer [3], information retrieval [4], and content analysis [5].
Given a document, the named entity mentions are recognized in advance by a named entity recognition (NER) method. Generally speaking, a typical entity linking system consists of two steps: (1) candidate entity generation, in which a model retrieves a set of candidate entities, which contains the entities that the mention may refer to; and (2) candidate entity ranking, in which a model ranks the entities in the candidate set and selects the entity that the mention is most likely to link to. Recently, some methods such as techniques based on a named dictionary and techniques based on surface form expansion have achieved high candidate recalls, and thus most work focuses on methods for downstream candidate entity ranking, as described in this paper.
In early work, prior distribution and local contexts played important roles in disambiguating different candidate entities. However, in many cases, local features alone cannot provide sufficient information for disambiguation. Therefore, many global models have emerged to solve the task of entity linking. For example, Ganea and Hofmann [6] combine local and global information. First, the word-entity co-occurrence counts are used to train the entity embeddings, then the local scores between contexts of mentions and the entity embeddings are calculated in the local model, and the scores between candidate entities of all mentions in the document are calculated in the global model. On the basis of [6], Le and Titov [7] model the latent relations between mentions. Based on [7], Hou et al. [8] inject fine-grained semantic information into entity embeddings. In addition, Yang et al. [9] propose the dynamic context augmentation method, which uses the entity embedding in [6].
However, the above methods still have some shortcomings. They essentially calculate the similarity between entity embeddings when obtaining global scores, which only consider the semantic proximity between entities. While there are real relations between some entity mentions in a document, these relations are contained in some knowledge graphs, and comprise the so-called knowledge graph structural information. As shown in Figure 1, there is an association relation of “colleague” between entity “Michael I. Jordan” and entity “Xiandong Jing” in the knowledge base. In addition, although there are also some works [10,11,12,13] that involve knowledge graphs, this is because their target knowledge base is a knowledge graph, and our method is different from them essentially. For example, Cetoli et al. [12] use bi-directional long short-term memory (Bi-LSTM) to encode graph triplets. Mulang et al. [13] develop a context-aware attentive neural network approach on Wikidata. Instead, on the basis of Wikipedia, we introduce the structural information of other knowledge graphs to complement the semantic information of Wikipedia, which is somewhat similar to the fusion of information from different knowledge bases.
To address the limitations of existing methods, we propose an entity linking model that introduces knowledge graph structural information (KGEL). First, under the premise that the target knowledge base is Wikipedia, we obtain the entities and triples in the knowledge graph Wikidata corresponding to the candidate entities. Then, the knowledge graph embedding method is used to train entity embeddings and relation embeddings. Finally, according to the different characteristics of local and global models, we use the previously trained entity embeddings and relation embeddings only for the global model of entity linking; that is, the global scores are computed from the perspective of the graph structure and fused with the Ment–Norm [7] model. Existing methods have been able to achieve more than 90% F1 on the standard AIDA-CoNLL dataset; for example, Ment–Norm achieves 93.07% F1. Our KGEL method achieves an improvement of 0.4% F1 on the basis of Ment–Norm, and the average result of KGEL on the five out-of-domain datasets is also 0.2% higher than Ment–Norm, which indicates that our model also has better generalization. Our method can also further improve the performance when using a more superior baseline.
The main contributions of our paper can be summarized as follows. (1) We propose to introduce knowledge graph structure information into the entity linking model, so as to complement the semantic information. (2) We obtain the Wikipedia–Wikidata mappings of entities and the required triples, and then obtain the entity and relation embeddings containing the graph structure through the knowledge graph embedding method. This provides a new idea for information fusion between different knowledge bases (graphs). (3) Extensive experiments on multiple datasets show the excellent performance of our method and demonstrate the effectiveness of the knowledge graph structure for entity linking.

2. Background and Related Work

2.1. Problem Definition

Given a knowledge base containing a set of entities E s = { e 1 , , e t } and a set of entity mentions M = { m 1 , , m n } in corpus D , the goal of entity linking is to map each entity mention m i M in the text to its corresponding entity e i * E s . Because a KB may contain a large number of entities, in order to reduce complexity, we usually use a heuristic to choose potential candidates, thus obtaining candidate set C i = ( e i 1 , , e i l i ) , which is the candidate entity generation we mentioned earlier. Then, we select gold entities on the candidate set in the candidate entity ranking stage.

2.2. Entity Linking

As it is an important task in natural language processing, there is a lot of work in the field of entity linking. Most of the early work comprises methods based on manually designed features and rule-based methods, which are not enough to capture the potential dependence and interaction in the data. With the rapid development of deep learning, a large number of deep-learning-based methods have appeared in the field of entity linking, and they have achieved better results than previous methods. Topics related to the work of this article are as follows.
Local model. The local model uses the local text context information around the entity mention to independently resolve the ambiguity of each entity mention. He et al. [14] were early adopters of deep learning for entity linking. They learned distributed representations of entities to measure similarity, avoiding manually designed features, so that words and entities could be in the joint semantic space, and then candidate entities could be sorted based on vector similarity. Subsequently, Sun et al. [15] used neural networks to encode mentions, contexts of mentions, and entities. Among them, contexts of mentions are encoded by convolutional neural networks (CNN), which are combined with representations of the mention titles to obtain the final mention representations. The entity representations are obtained from the entity titles and entity categories. Finally, the similarities between the mention representations and the entity representations are calculated to obtain local scores. Based on [15], Francis-Landau et al. [16] used CNN and stacked denoising auto-encoders to encode different granular information of mentions and entities to enhance the representation. In addition, Gupta et al. [17] cascaded the output of two long short-term memory (LSTM) [18] networks. The two LSTM networks independently encode the left and right context of the entity mention, including the entity mention itself. Kolitsas et al. [19] expressed entity mention as a combination of LSTM hidden states contained in the span of entity mention. Eshel et al. [20] used a variant of LSTM-GRU [21]. Ganea and Hofmann [6] introduced an attention mechanism in the local model. They assumed that a context word was important if it was strongly related to at least one candidate entity, and the context words were hard pruned. The local model in this paper is based on Ganea and Hofmann [6].
Global model. The global model links all the mentions in a document at the same time and considers that the target entities of all the mentions are consistent on the subject. The previous global methods usually executed RandomWalk [22] or PageRank [23] algorithms on the graph containing candidate entities. Another solution is to maximize the conditional random field [24], but the problem is NP-hard. Ganea and Hofmann [6] used loopy belief propagation (LBP) [25] to iteratively propagate entity scores to reduce complexity. Based on [6], Le and Titov [7] modeled the latent relations between mentions and added them to the global model in the form of features, achieving better results. Some recent studies have defined the global entity linking problem as a sequential decision task, where the linking of the new entity is based on the already linked entity. Fang et al. [26] used LSTM to maintain long-term memory for previous decisions; Yang et al. [9] proposed a dynamic context integration method that uses previous decisions as dynamic context to improve subsequent decisions; Yamada et al. [27] calculated the confidence scores based on the previous decisions. In addition, graph neural networks (GNNs) can also be used for the global model of entity linking. Wu et al. [28] proposed a dynamic graph convolutional network model, in which the graph structure is dynamically calculated and changed during training, and fusion of knowledge through dynamically linked nodes can effectively obtain the theme consistency in the document. Fang et al. [29] proposed a sequential graph attention network to synthesize the advantages of the graph model and the sequence model, which dynamically encodes the preceding and following entity mentions, and assigns different weights to these entity mentions. The global model of this article refers to the work of [7].
Entity embedding. Entity embedding is a key component in entity linking to avoid manual features and enhance model effects. There is also a lot of work for entity embedding. Yamada et al. [30] proposed to map words and entities to the same continuous vector space. They used two models to extend the skip-gram model. The KB graph model uses the link structure in the KB to learn the relevance of entities. The anchor context model aims to use KB anchor text and context words to align vectors so that similar words and entities are close in the vector space. Yamada et al. [31] further proposed to jointly learn distributed representations of text and entities. Given a piece of text in the knowledge base, a model is trained to predict entities related to the text; that is, using a large amount of text extracted from Wikipedia and their entity annotations to train the model. Ganea and Hofmann [6] used pre-trained word embeddings and word-entity co-occurrence counts to obtain entity embeddings so that words and entities were represented in the same low-dimensional vector space. Ling et al. [32] proposed a fill-in-the-blank task to learn context-independent entity representations from the text context. Hou et al. [8] proposed incorporating fine-grained semantic information into entity embedding to reduce uniqueness and promote the learning of contextual commonality. Yamada et al. [27] used the pre-trained model BERT [33] to generate the representation of words and entities, and the results were greatly improved compared to the previous method. This paper also uses the entity embeddings of [6].

2.3. Knowledge Graph Embedding

The knowledge graph is a multi-relational graph composed of entities (nodes) and relations (edges), and each edge is in the form of a triple (head entity, relation, tail entity). The existing knowledge graphs include Freebase [1], DBpedia [34], Wikidata, etc. Knowledge graph embedding [35] involves embedding the entities and relations in the knowledge graph into a continuous vector space. In general, knowledge graph embedding methods can be divided into two groups: translational distance models and semantic matching models [36,37,38]. The former use distance-based scoring functions, and the latter similarity-based ones. Among translational distance models, TransE [39] is the most representative. The main idea is to give a triple h , r , t , the goal is h + r t , where h , r , t are the head entity, relation, and tail entity, respectively, and h , r , t are, respectively, vector representations. To solve the limitations of the TransE model in dealing with 1-to-N, N-to-1, and N-to-N complex relations, TransH [40] introduces relation-specific hyperplanes that allow an entity to have different representations under different relations. In order to further improve the representation ability, TransR [41] introduces relation-specific spaces, rather than hyperplanes. TransD [42] simplifies TransR by further decomposing the projection matrix into a product of two vectors. TransM [43] assigns specific relation weight to each triple h , r , t .
There are also recent knowledge graph embedding methods with better performance. Zhang et al. [44] proposed the hierarchy-aware knowledge graph embedding model (HAKE), which maps entities into a polar coordinate system. PairRE [45] has paired vectors for each relation representation, which can adaptively adjust the margin in a loss function to fit for complex relations. Additionally, PairRE can encode three relation patterns: symmetry/antisymmetry, inverse, and composition. DualE [46] introduces dual quaternions into knowledge graph embedding, where a dual quaternion is similar to a “complex quaternion” with its real and imaginary part all being quaternar. DualE universally models relations as the combination of a series of translation and rotation operations. EIGAT [47] allows correct incorporation of global information into the graph attention network (GAT) family of models by using scaled entity importance, which is computed by an attention-based global random walk algorithm. In order to focus on the importance of the knowledge graph structure for the entity linking task, the knowledge graph embedding method used in this article is the most basic TransE model.

3. Learning Entity Embeddings KGEmbs

3.1. Wikipedia–Wikidata Mappings

Since the target knowledge base of the dataset we use is Wikipedia, and we want to introduce the structural information of other knowledge graphs, for the Wikipedia entities used, we need to obtain their corresponding Wikidata entities, i.e., obtain the Wikipedia–Wikidata mappings. In the entity’s Wikipedia page, there is a corresponding Wikidata hyperlink, as shown in Figure 2. Therefore, we can obtain the Wikidata ID of the Wikipedia entity through the crawler. Examples of the Wikipedia–Wikidata mappings are shown on the left side of Table 1.

3.2. Triple Knowledge

We can obtain the triple knowledge of Wikidata from OpenKE: http://139.129.163.161/index/toolkits (accessed on 1 March 2022), including 20,982,733 entities, 594 relations, and 68,904,773 triples. According to the work of [7], we obtain 274,474 entities in the candidate entity generation stage to filter relations and triples, and finally obtain 486 relations and 807,587 triples. The triple format is shown on the right side of Table 1. For example, (Q1, P2670, Q523) is a triple, where Q1 is the head entity and its corresponding entity is “universe”, Q523 is the tail entity and its corresponding entity is “star”, and P2670 is the relation between entities Q1 and Q523; that is, “instance has part(s) of the class”. Therefore, the triple can be represented as (universe, instance has part(s) of the class, star).

3.3. Entity and Relation Embeddings

In order to demonstrate more intuitively the effectiveness of the knowledge graph structure for entity linking, and also considering the speed differences of each model, we use the TransE model to train entity and relation embeddings on triples, where h , t E (the set of entities) and r R (the set of relations). The main idea is that the functional relation obtained from the edges labeled by r corresponds to the translation of the embedding; that is, we hope that h + r t when h , r , t holds, while h + r should be far away from t otherwise.
In order to learn entity and relation embeddings, we minimize the following loss:
L = ( h , r , t ) S ( h , r , t ) S ( h , r , t ) γ 1 + d h + r , t d h + r , t +
where x + denotes the positive part of x, γ > 0 is a margin hyperparameter, and d h + r , t is an indicator to measure similarity. Here we use the L 1 -norm, and
S ( h , r , t ) = h , r , t | h E h , r , t | t E
The optimization is performed by stochastic gradient descent, and an additional constraint is that the L 2 -norm of the embeddings of the entities is 1.

4. Model

The entity linking model in this paper integrates local and global features and is a conditional random field model in form. Figure 3 provides an overview of our model. Specifically, a scoring function g is defined to evaluate the mappings from entity mentions m 1 , , m n to the entities e 1 , , e n in a document D:
g ( e 1 , . . . , e n ) = i = 1 n Ψ ( e i ) + i j Φ ( e i , e j , D )
where n represents the number of entity mentions in the document. The first part of Equation (3) is the local score, which is the matching score between the local context of the entity mention and the candidate entity, and the second part is the global score, which is the score between entities in the document. The local model and the global model are described below.

4.1. Local Model

According to Ganea and Hofmann [6], this paper takes the local model as an attention model based on entity embedding. For an entity mention m, if a word in the context is strongly related to at least one candidate entity, the word is considered important.
In the candidate generation stage, we can obtain the candidate entity set C i = ( e i 1 , , e i l i ) . Then we calculate the score of each candidate entity e C i according to the P-word window local context c = { w 1 , . . . , w p } around m. First, we calculate the unnormalized support score of each word in the context; that is, the weight of each word
u w = max e C i e T A w
where A is a parameterized diagonal matrix, w is the word embedding (we use the pre-trained word2vec word embedding), and e is the candidate entity embedding, which is trained based on the co-occurrence counts of the word-entity in Wikipedia [6]. If the word is strongly related to at least one candidate entity, its weight score is relatively high. In addition, it is observed that some words with insufficient information will introduce noise to the local model, so the hard pruning method is used to select Q P words with the highest weight scores:
c = w c | u w t o p Q u
Therefore, the final attention weight is:
β w = exp u w v c exp u v i f w c 0 o t h e r w i s e
Finally, we can obtain the local scores of the candidate entities:
Ψ e = w c β w e T B w
where B is another diagonal matrix that can be trained.

4.2. Global Model

Ganea and Hofmann [6] mainly considered the consistency between entities. However, Le and Titov [7] proposed that there is not only consistency between entities, but there are also some latent relations that can support the constraints on entities. Assuming that there are K latent relations, each relation k corresponds to a pair ( m i , m j ) , so the second term of Equation (3) can be written as:
Φ e i , e j , D = k = 1 K α i j k Φ k e i , e j , D
That is, the paired score ( m i , m j ) is the weighted sum of the corresponding scores of each relation, and α i j k is the weight corresponding to the relation k. Here, each relation k is a diagonal matrix R k R d × d , and
Φ k e i , e j , D = e i T R k e j
The weight α i j k is the normalized score:
α i j k = 1 Z i j k exp f T ( m i , c i ) D k f ( m j , c j ) d
where Z i j k is the normalization factor, D k R d × d is a diagonal matrix, and f ( m i , c i ) is a single-layer neural network, which is used to obtain the local context representation of the mention m i . For c i , we first obtain the average c l of the word embeddings of the context words on the left of the mention m i , then obtain the average c r of the word embeddings of the context words on the right, and finally take the concatenation of c l and c r . In addition, Le and Titov [7] proposed two normalization methods of Z i j k : normalization over relations and normalization over mentions. We adopt the method of normalization over mentions, then
Z i j k = j = 1 j i n exp f T ( m i , c i ) D k f ( m j , c j ) d
Now j = 1 , j i n α i j k = 1 , which means that for each relation k and mention m i , we want to find another mention that has a relation k with the mention m i . The entity embeddings e i , e j here are obtained by training using word-entity co-occurrence counts in Wikipedia, so the global model is called the WikiEmbs model, and there is Φ w i k i e i , e j , D = Φ e i , e j , D . The WikiEmbs model essentially only uses the semantic information of the entities; that is, the more semantically related entities have a greater probability of appearing in the same document. However, the structural information in the knowledge graph is ignored, so we propose the KGEmbs model, which explicitly uses the knowledge graph structure information in the global model. Our motivation is that the knowledge graph structure should be maintained when the entity mentions in a document are mapped to the knowledge base. Assuming that there are R n relations (Section 3.2), the second term in Equation (3) can be written as:
Φ K G e i , e j , D = max r R n f K G e i , e j , r
where f K G e i , e j , r is the scoring function of the knowledge graph embedding method; that is, for all relations R, the score of ( e i , e j ) must be calculated, and then the maximum value is taken. The TransE [39] model is used here, and because the head entity and tail entity in ( e i , e j ) cannot be distinguished, there is:
f K G e i , e j , r = max γ 1 d e i + r , e j , γ 1 d e j + r , e i
where γ 1 is consistent with γ 1 in Equation (1), and
d h + r , t = h + r t 1
Among them, the smaller d ( h + r , t ) , the greater the probability that the entities h and t have the relation r. In addition, h , t are the entity embeddings obtained by the TransE model, and r is the relation embedding. Finally, we combine the two global scores obtained above:
Φ e i , e j , D = f g l o b a l Φ w i k i e i , e j , D , Φ K G e i , e j , D
where f g l o b a l is a two-layer neural network.

4.3. Model Training

The solution of Equation (3) is NP-hard. Following Le and Titov [7], we also adopt max-product loopy belief propagation (LBP) to estimate the max-marginal probability:
g ^ i e | D max e 1 , , e i 1 e i + 1 , , e n g e 1 , , e n
Then we obtain the final score of mention m i
ρ i e = f f i n a l g ^ i e | D , p ^ e | m i
The one with the highest score is the candidate entity to be linked to, f f i n a l is another two-layer neural network, and p ^ e | m is the mention-entity prior. We optimize the parameters in the model by minimizing the ranking loss as follows:
L θ = D D m i D e C i h m i , e
h m i , e = max 0 , γ 2 ρ i e i * + ρ i e
where θ denotes the model parameters, D is the training corpus, D is a document, and e i * is the gold entity.

5. Experiments

5.1. Datasets

To prove the effectiveness of our method, we conducted experiments on six popular open-source datasets, including an in-domain dataset and five out-domain datasets. For the in-domain dataset, we used the AIDA-CoNLL dataset [48], which contains AIDA-train, AIDA-A, and AIDA-B, which were used for training, verification, and testing, respectively. For out-domain datasets, we used MSNBC (MSB), AQUAINT (AQ), and ACE2004 (ACE), which are cleaned and updated by Guo and Barbosa [22]; and WNED-WIKI (WW) and WNED-CWEB (CWEB), which are automatically extracted from ClueWeb and Wikipedia corpora by Guo and Barbosa [22]. Among them, the latter two datasets are larger in scale and noisier, making linking of entities more difficult. Statistics of these datasets are summarized in Table 2. The target knowledge base is Wikipedia. Based on previous work [6,7], we do not consider mentions that have no corresponding entities in the KB.

5.2. Candidate Entity Generation

To ensure fairness and comparable results, we use the candidate generation method of Le and Titov [7]. First, we select the top 30 candidate entities for each mention m i based on the prior p ^ e | m i , and then select 7 from them. Among them, the top 4 entities are selected based on p ^ e | m i , and the top 3 entities are selected based on the score e T w d i w , where e , w R d are entity and word embeddings, respectively, and d i is the 50-word local context surrounding m i . The quality of the candidate set obtained by the above method is shown in Table 2.

5.3. Hyper-Parameter Setting

Our models are implemented in the Pytorch framework. For the Local model, according to Ganea and Hofmann [6], we use the following hyper-parameters: P = 100 , Q = 25 (Equation (5)). We set the dimensions of word embedding and entity embedding to 300, where word embedding and entity embedding are from [6]. For the WikiEmbs Global model, when calculating f (Equation (10)), we use the word embedding in Le and Titov [7] and the entity embedding in [6], both of which have a dimension of 300. In addition, according to [7], the number of LBP loops is set to 10, the dropout rate for f is set to 0.3, the window size c i of the local context used when calculating pairwise score functions is 6, and the number of relations in Ment-norm is 3. For the KGEmbs Global model, we use the TransE model to train entity embeddings and relation embeddings, where learning rate λ = 0.0001 , margin γ 1 = 24 (Equation (1)), batch size is 1024, hidden size is 300, and the dimensions of entity embedding and relation embedding are 300. When training the model, we set γ 2 = 0.01 (Equation (19)). When the F1 score of the model on the validation set reaches 91%, we adjust the learning rate from 1 × 10 4 to 1 × 10 5 , and we stop learning if the F1 on the validation set does not improve after 20 epochs.

5.4. Main Results

The following methods are selected as baselines.
  • AIDA [48] combines the previous methods into a comprehensive framework that contains three measures: the prior probability of an entity being mentioned, the similarity between the context of mention and the candidate entity, and the consistency among candidate entities for all mentions. It constructs a weighted graph whose nodes are mentions and candidate entities and calculates a dense subgraph to obtain an approximately optimal mention-entity mapping.
  • GLOW is a global entity disambiguation system proposed by [49], which formulates the entity disambiguation task as an optimization problem with local and global variants.
  • RI [50] combines statistical methods to perform richer relational analysis on the text. It proposes a modular formulation that includes the entity-relation inference problem. It also proves that the recognition of relations in the text is not only helpful for candidate entities, but also the subsequent ranking stage.
  • PBoH [51] uses a graphical model to perform global entity disambiguation. It simultaneously disambiguates mentions in a document by using the co-occurrence probability between entities in the document and the local context information of the mentions. It uses LBP to perform approximate inference.
  • Deep-ED [6] introduces an attention mechanism into the local model, and the context words of mentions are hard pruned. Its global model is a fully-connected pairwise conditional random field. Because the problem is NP-hard, it uses LBP to iteratively propagate entity scores to reduce complexity.
  • Ment-Norm [7] models the latent relations between mentions and adds them to the global model in the form of features. There are two options for normalization, where it is normalization over mentions.
  • DCA-SL [9] regards entity linking as a sequence decision task and uses the previous decision as dynamic contexts to improve the later decisions. It explores supervised learning strategies for learning the DCA model.
  • DCA-RL [9] involves the use of reinforcement-learning strategies to learn the DCA model.
Table 3 shows micro F1 scores on AIDA-B and five out-domain test sets. Compared with Deep-ED [6], our method achieves a substantial improvement on both the in-domain dataset AIDA-B and the average result on five out-domain datasets. Moreover, KGEL’s F1 score is still 0.4% higher than Ment–Norm on the AIDA-B dataset, and for the average result on the five out-domain datasets, KGEL also has an improvement of 0.2% F1 on Ment–Norm. It should be noted that although the DCA-SL model has good results on the datasets AIDA-B and MSNBC, it has poor results on the dataset CWEB, so its average result on the out-domain datasets is not good. The same is true for DCA-RL. This indicates that our method has better generalization. Therefore, overall, our method achieves very competitive results on the AIDA-B dataset. Moreover, KGEL achieves higher F1 scores than previous methods on the ACE2004 dataset as well as on the average of out-domain datasets. This fully demonstrates the effectiveness of our method, i.e., the importance of knowledge graph structure for entity linking.

5.5. Ablation Study

In order to study the role of each module of the model, an ablation study was also performed in this research, and the experimental results are shown in Table 4. We utilize the following variants:
  • KGEL is our proposed method, which includes three modules: Local model, WikiEmbs Global model, and KGEmbs Global model.
  • -KGEmbs represents the results on each dataset after removing the KGEmbs global model.
  • -WikiEmbs represents the experimental results after removing the WikiEmbs global model.
  • -local-WikiEmbs is the result of removing the Local model and WikiEmbs Global model at the same time.
As can be seen in Table 4, when the KGEmbs Global model is removed, the results on four datasets and the average result on the out-domain datasets drop dramatically. This proves the validity of the KGEmbs Global model, i.e., the necessity of introducing knowledge graph structural information. Similarly, we can find that the results on each dataset drop more significantly when the WikiEmbs Global model is removed, indicating that using only the structural information in the knowledge graph is insufficient because there is a certain sparsity in the knowledge graph, i.e., not every pair of entities has a clear relationship with each other, so the structural information of the knowledge graph has a certain guiding effect on the linking of entities, but cannot be used independently. After removing the Local model based on -WikiEmbs, we find that the results on each dataset have further decreased, which illustrates the necessity of the local model. Thus, the entire ablation experiment shows that all modules of the model are valid.

5.6. Other Ways of Using KG Structure

In addition to using knowledge graph embedding methods such as TransE on triples, we also try to use triples directly. We consider two entities to be related if there is a relation between them, i.e., two entities that can form a triple are related. Therefore, for entity e 1 , we obtain the entity set E r related to it from the triples. For example, in Table 1, the related entity set of entity Q1 is {Q523, Q136407, Q323}. To incorporate information about its related entities in the representation of entity e 1 , we perform the following operations:
e r = 1 a i = 1 a e i
e = α e 1 + 1 α e r
where e i E r is the entity associated with entity e 1 , a is the size of the entity set E r , e r is the average embedding of entities associated with entity e 1 , e 1 is the original embedding of entity e 1 , e is the embedding of entity e 1 after fusing information, and α is a hyperparameter. This operation is equivalent to using 1-hop information of the knowledge graph.
In order to determine the optimal value of α , we performed a lot of experiments for different α ; that is, directly replacing the original entity embedding with the entity embedding after fusion, and the model structure is consistent with Le and Titov [7]. The experimental results are shown in Figure 4.
From the figure, it is clear that the best results are obtained when α = 0.9 . In addition, we also tried some other variants:
  • Ment-Norm is the model of Le and Titov [7] and also our basic model.
  • KGEL is our main model; that is, the entity and relation embeddings obtained by the knowledge graph embedding method are used in the global model of entity linking.
  • Related-Fixed refers to the method of using related entities mentioned in this section, in which the parameter α is fixed at 0.9.
  • Related-Vari means that the parameter α is variable; that is, it changes during training.
  • Based on Related-Vari, Related-Vari-diff makes the α in the global model and the local model different.
  • Related-nn indicates the use of a neural network to fuse e 1 and e r .
From the Table 5, it can be seen that the parameter α fixed to 0.9 is the optimal result when using related entities. The result of Related-Fixed is slightly better than that of Ment–Norm, indicating that the knowledge graph structure is beneficial for the effect of entity linking. However, the result of Related-Fixed is worse than that of KGEL, which shows that how the knowledge graph structure is used is also very important. Obviously, it is better for us to use the entity embedding obtained by the knowledge graph embedding for the characteristics of the global model considering the correlations between entities.

5.7. Better Baseline

To further prove the importance of the knowledge graph structure to the entity linking, we used the KGEmbs module for a better baseline. FGS2EE [8] is an improvement of Ment–Norm [7], which introduces fine-grained semantic information into the original entity embedding to improve the model performance. KGEL-FGS2EE adds the KGEmbs module on the basis of FGS2EE. The experimental results are shown in Figure 5. We can find that for the average F1 score, KGEL-FGS2EE can further improve the performance based on FGS2EE. This shows that the KGEmbs module we proposed is effective. Similarly, the KGEmbs module can also be used in other methods. In other words, it should be useful to introduce knowledge graph structure based on other methods.

5.8. Case Study

Table 6 shows the mentions and their real entities, as well as the results predicted by the model. Examples of incorrect model predictions are shown in red, e.g., “Scotland” is predicted to be “Scotland_national_cricket_team”. This shows that in some cases, only semantic information cannot complete the link to the entity. We note that a document contains a knowledge graph structure. As shown in Figure 6, there is a certain connection between the entities “Scotland” and “England”. When calculating the global score, the score between “Scotland” and “England” will be higher than the scores between other entities, indicating that mentions “English” and “Scotland” are more likely to refer to entities “England” and “Scotland”, respectively. Therefore, we can guide the prediction of mention “Scotland” based on this connection. Similarly, we can use the knowledge graph structure between “Edgbaston” and “Birmingham” to guide the prediction of “Edgbaston”. In summary, the introduction of the knowledge graph structure solves the problem of incorrect prediction of some mentions.

5.9. Execution Times of the Models

To investigate the complexity of the method, we conducted experiments on the training and inference time of the model. Among them, the model was trained on the AIDA-train dataset and inference was performed on AIDA-B and five out-of-domain datasets. The results are shown in Table 7, where the second column indicates the time spent for one epoch during model training, and the third column indicates the total time spent by the model for inference on several datasets. As can be seen from the table, under the same experimental conditions, our proposed model KGEL is close to the model Ment–Norm [7] in both training and inference time, because we calculated the scores between entities in the KGEmbs Global model offline. In addition, the epochs required for KGEL and Ment–Norm to converge are similar, so the introduced knowledge graph structure does not have much impact on the execution times.

6. Conclusions

In this work, we proposed a simple but effective method, KGEL, to introduce knowledge graph structure information into entity linking. In addition to considering the relevance of entities at the semantic level, the relations between entities were also considered from the perspective of structure. We first obtained the triples and then trained them using the knowledge graph embedding method to obtain the entity embeddings and relation embeddings that contained the graph structure. Finally, the entity embeddings and relation embeddings obtained above were used in the calculation of the global score. Extensive experiments on multiple datasets prove the effectiveness of our method; that is, the knowledge graph structure is useful for entity linking tasks. In addition, KGEmbs can be used as a module to enhance the effects of other baseline models.
In future work, we will solve the sparsity problem of the knowledge graph. Not every entity has a corresponding triple, nor is there a relation between every pair of entities. In addition, we will try to use better methods to utilize the knowledge graph structure, such as other knowledge graph embedding methods. As introduced in Section 2.3, some recent knowledge graph embedding methods such as HAKE [44], PairRE [45], DualE [46], and EIGAT [47] can better encode entities and relations in knowledge graphs, and theoretically they should further improve the performance of entity linking.

Author Contributions

Conceptualization, Q.L. (Qijia Li) and F.L.; methodology, Q.L. (Qijia Li) and S.L.; validation, Q.L. (Qijia Li), X.L. and K.L.; formal analysis, Q.L. (Qijia Li) and S.L.; investigation, Q.L. (Qijia Li); resources, Q.L. (Qing Liu); data curation, P.D.; writing—original draft preparation, Q.L. (Qijia Li); writing—review and editing, F.L. and S.L.; visualization, X.L.; funding acquisition, F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Strategic Priority Research Program of the Chinese Academy of Sciences (grant number Y835120378).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

https://github.com/lephong/mulrel-nel (accessed on 1 March 2022).

Conflicts of Interest

The authors declare that they do not have any conflict of interest. This research does not involve any human or animal participation. All authors have checked and agreed with the submission.

References

  1. Bollacker, K.; Evans, C.; Paritosh, P.; Sturge, T.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, Vancouver, BC, Canada, 9–12 June 2008; pp. 1247–1250. [Google Scholar]
  2. Fabian, M.; Gjergji, K.; Gerhard, W. Yago: A core of semantic knowledge unifying wordnet and wikipedia. In Proceedings of the 16th International World Wide Web Conference, WWW, Banff, AB, Canada, 8–12 May 2007; pp. 697–706. [Google Scholar]
  3. Yih, S.W.t.; Chang, M.W.; He, X.; Gao, J. Semantic parsing via staged query graph generation: Question answering with knowledge base. In Proceedings of the Joint Conference of the 53rd Annual Meeting of the ACL and the 7th International Joint Conference on Natural Language Processing of the AFNLP, Beijing, China, 26–31 July 2015. [Google Scholar]
  4. Hoffmann, R.; Zhang, C.; Ling, X.; Zettlemoyer, L.; Weld, D.S. Knowledge-based weak supervision for information extraction of overlapping relations. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; pp. 541–550. [Google Scholar]
  5. Michelson, M.; Macskassy, S.A. Discovering users’ topics of interest on twitter: A first look. In Proceedings of the Fourth Workshop on Analytics for Noisy Unstructured Text Data, Toronto, ON, Canada, 26 October 2010; pp. 73–80. [Google Scholar]
  6. Ganea, O.E.; Hofmann, T. Deep joint entity disambiguation with local neural attention. arXiv 2017, arXiv:1704.04920. [Google Scholar]
  7. Le, P.; Titov, I. Improving entity linking by modeling latent relations between mentions. arXiv 2018, arXiv:1804.10637. [Google Scholar]
  8. Hou, F.; Wang, R.; He, J.; Zhou, Y. Improving entity linking through semantic reinforced entity embeddings. arXiv 2021, arXiv:2106.08495. [Google Scholar]
  9. Yang, X.; Gu, X.; Lin, S.; Tang, S.; Zhuang, Y.; Wu, F.; Chen, Z.; Hu, G.; Ren, X. Learning dynamic context augmentation for global entity linking. arXiv 2019, arXiv:1909.02117. [Google Scholar]
  10. Fang, W.; Zhang, J.; Wang, D.; Chen, Z.; Li, M. Entity disambiguation by knowledge and text jointly embedding. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, Berlin, Germany, 11–12 August 2016; pp. 260–269. [Google Scholar]
  11. Luo, A.; Gao, S.; Xu, Y. Deep semantic match model for entity linking using knowledge graph and text. Procedia Comput. Sci. 2018, 129, 110–114. [Google Scholar] [CrossRef]
  12. Cetoli, A.; Akbari, M.; Bragaglia, S.; O’Harney, A.D.; Sloan, M. Named entity disambiguation using deep learning on graphs. arXiv 2018, arXiv:1810.09164. [Google Scholar]
  13. Mulang, I.O.; Singh, K.; Vyas, A.; Shekarpour, S.; Sakor, A.; Vidal, M.E.; Auer, S.; Lehmann, J. Context-aware entity linking with attentive neural networks on wikidata knowledge graph. arXiv 2019, arXiv:1912.06214. [Google Scholar]
  14. He, Z.; Liu, S.; Mu, L.; Ming, Z.; Wang, H. Learning Entity Representation for Entity Disambiguation. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Sofia, Bulgaria, 4–9 August 2013. [Google Scholar]
  15. Sun, Y.; Lin, L.; Tang, D.; Yang, N.; Ji, Z.; Wang, X. Modeling mention, context and entity with neural networks for entity disambiguation. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
  16. Francis-Landau, M.; Durrett, G.; Klein, D. Capturing semantic similarity for entity linking with convolutional neural networks. arXiv 2016, arXiv:1604.00734. [Google Scholar]
  17. Gupta, N.; Singh, S.; Roth, D. Entity linking via joint encoding of types, descriptions, and context. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2681–2690. [Google Scholar]
  18. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  19. Kolitsas, N.; Ganea, O.E.; Hofmann, T. End-to-end neural entity linking. arXiv 2018, arXiv:1808.07699. [Google Scholar]
  20. Eshel, Y.; Cohen, N.; Radinsky, K.; Markovitch, S.; Yamada, I.; Levy, O. Named entity disambiguation for noisy text. arXiv 2017, arXiv:1706.09147. [Google Scholar]
  21. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  22. Guo, Z.; Barbosa, D. Robust named entity disambiguation with random walks. Semant. Web 2018, 9, 459–479. [Google Scholar] [CrossRef]
  23. Pershina, M.; He, Y.; Grishman, R. Personalized page rank for named entity disambiguation. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Denver, CO, USA, 31 May–5 June 2015; 2015; pp. 238–243. [Google Scholar]
  24. Lafferty, J.; McCallum, A.; Pereira, F.C. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data. 2001. Available online: https://repository.upenn.edu/cis_papers/159/?ref=https://githubhelp.com (accessed on 1 March 2022).
  25. Murphy, K.; Weiss, Y.; Jordan, M.I. Loopy belief propagation for approximate inference: An empirical study. arXiv 2013, arXiv:1301.6725. [Google Scholar]
  26. Fang, Z.; Cao, Y.; Li, Q.; Zhang, D.; Zhang, Z.; Liu, Y. Joint entity linking with deep reinforcement learning. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 438–447. [Google Scholar]
  27. Yamada, I.; Washio, K.; Shindo, H.; Matsumoto, Y. Global entity disambiguation with pretrained contextualized embeddings of words and entities. arXiv 2019, arXiv:1909.00426. [Google Scholar]
  28. Wu, J.; Zhang, R.; Mao, Y.; Guo, H.; Soflaei, M.; Huai, J. Dynamic graph convolutional networks for entity linking. In Proceedings of the Web Conference 2020, Ljubljana, Slovenia, 19–23 April 2020; pp. 1149–1159. [Google Scholar]
  29. Fang, Z.; Cao, Y.; Li, R.; Zhang, Z.; Liu, Y.; Wang, S. High quality candidate generation and sequential graph attention network for entity linking. In Proceedings of the Web Conference 2020, Ljubljana, Slovenia, 19–23 April 2020; pp. 640–650. [Google Scholar]
  30. Yamada, I.; Shindo, H.; Takeda, H.; Takefuji, Y. Joint learning of the embedding of words and entities for named entity disambiguation. arXiv 2016, arXiv:1601.01343. [Google Scholar]
  31. Yamada, I.; Shindo, H.; Takeda, H.; Takefuji, Y. Learning distributed representations of texts and entities from knowledge base. Trans. Assoc. Comput. Linguist. 2017, 5, 397–411. [Google Scholar] [CrossRef] [Green Version]
  32. Ling, J.; FitzGerald, N.; Shan, Z.; Soares, L.B.; Févry, T.; Weiss, D.; Kwiatkowski, T. Learning cross-context entity representations from text. arXiv 2020, arXiv:2001.03765. [Google Scholar]
  33. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  34. Lehmann, J.; Isele, R.; Jakob, M.; Jentzsch, A.; Kontokostas, D.; Mendes, P.N.; Hellmann, S.; Morsey, M.; Van Kleef, P.; Auer, S.; et al. Dbpedia–a large-scale, multilingual knowledge base extracted from wikipedia. Semant. Web 2015, 6, 167–195. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, Q.; Mao, Z.; Wang, B.; Guo, L. Knowledge graph embedding: A survey of approaches and applications. IEEE Trans. Knowl. Data Eng. 2017, 29, 2724–2743. [Google Scholar] [CrossRef]
  36. Nickel, M.; Tresp, V.; Kriegel, H.P. A Three-Way Model for Collective Learning on Multi-Relational Data. 2011. Available online: https://openreview.net/forum?id=H14QEiZ_WS (accessed on 1 March 2022).
  37. Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex embeddings for simple link prediction. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 2071–2080. [Google Scholar]
  38. Bordes, A.; Glorot, X.; Weston, J.; Bengio, Y. A semantic matching energy function for learning with multi-relational data. Mach. Learn. 2014, 94, 233–259. [Google Scholar] [CrossRef] [Green Version]
  39. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 2013, 26, 9. [Google Scholar]
  40. Wang, Z.; Zhang, J.; Feng, J.; Chen, Z. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, Quebec City, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  41. Lin, Y.; Liu, Z.; Sun, M.; Liu, Y.; Zhu, X. Learning entity and relation embeddings for knowledge graph completion. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  42. Ji, G.; He, S.; Xu, L.; Liu, K.; Zhao, J. Knowledge graph embedding via dynamic mapping matrix. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015; pp. 687–696. [Google Scholar]
  43. Fan, M.; Zhou, Q.; Chang, E.; Zheng, F. Transition-based knowledge graph embedding with relational mapping properties. In Proceedings of the 28th Pacific Asia Conference on Language, Information and Computing, Phuket, Thailand, 12–14 December 2014; pp. 328–337. [Google Scholar]
  44. Zhang, Z.; Cai, J.; Zhang, Y.; Wang, J. Learning hierarchy-aware knowledge graph embeddings for link prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3065–3072. [Google Scholar]
  45. Chao, L.; He, J.; Wang, T.; Chu, W. PairRE: Knowledge graph embeddings via paired relation vectors. arXiv 2020, arXiv:2011.03798. [Google Scholar]
  46. Cao, Z.; Xu, Q.; Yang, Z.; Cao, X.; Huang, Q. Dual quaternion knowledge graph embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 6894–6902. [Google Scholar]
  47. Zhao, Y.; Zhou, H.; Xie, R.; Zhuang, F.; Li, Q.; Liu, J. Incorporating Global Information in Local Attention for Knowledge Representation Learning. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online Event, 1–6 August 2021; pp. 1341–1351. [Google Scholar]
  48. Hoffart, J.; Yosef, M.A.; Bordino, I.; Fürstenau, H.; Pinkal, M.; Spaniol, M.; Taneva, B.; Thater, S.; Weikum, G. Robust disambiguation of named entities in text. In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, Scotland, UK, 27–31 July 2011; pp. 782–792. [Google Scholar]
  49. Ratinov, L.; Roth, D.; Downey, D.; Anderson, M. Local and global algorithms for disambiguation to wikipedia. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Portland, OR, USA, 19–24 June 2011; pp. 1375–1384. [Google Scholar]
  50. Cheng, X.; Roth, D. Relational inference for wikification. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, DC, USA, 18–21 October 2013; pp. 1787–1796. [Google Scholar]
  51. Ganea, O.E.; Ganea, M.; Lucchi, A.; Eickhoff, C.; Hofmann, T. Probabilistic bag-of-hyperlinks model for entity linking. In Proceedings of the 25th International Conference on World Wide Web, Montreal, QC, Canada, 11–15 April 2016; pp. 927–938. [Google Scholar]
Figure 1. An example of NEL whose goal is to link each mention to an entity in the KB (e.g., “Michael Jordan” is linked to Michael I. Jordan; “Artificial intelligence” is linked to Artificial intelligence). Note that there are various relations between entities in the KB.
Figure 1. An example of NEL whose goal is to link each mention to an entity in the KB (e.g., “Michael Jordan” is linked to Michael I. Jordan; “Artificial intelligence” is linked to Artificial intelligence). Note that there are various relations between entities in the KB.
Applsci 12 02702 g001
Figure 2. Example for the Wikipedia–Wikidata mapping. We can obtain the corresponding Wikidata ID through the entity’s Wikipedia page.
Figure 2. Example for the Wikipedia–Wikidata mapping. We can obtain the corresponding Wikidata ID through the entity’s Wikipedia page.
Applsci 12 02702 g002
Figure 3. The architecture of the proposed KGEL model. It contains three parts: Local model, WikiEmbs Global model, and KGEmbs Global model. Specifically, in the Local model, the similarity calculated by context embedding and entity embedding is used as the local score. In the Global model, the scores between the candidate entities of all mentions in the document are taken as the global score. Among them, in the WikiEmbs Global model, entity embedding is obtained through word-entity co-occurrence counts, which consider the semantic information. In the KGEmbs Global model, entity embedding is obtained through triples, considering the structural information of the knowledge graph.
Figure 3. The architecture of the proposed KGEL model. It contains three parts: Local model, WikiEmbs Global model, and KGEmbs Global model. Specifically, in the Local model, the similarity calculated by context embedding and entity embedding is used as the local score. In the Global model, the scores between the candidate entities of all mentions in the document are taken as the global score. Among them, in the WikiEmbs Global model, entity embedding is obtained through word-entity co-occurrence counts, which consider the semantic information. In the KGEmbs Global model, entity embedding is obtained through triples, considering the structural information of the knowledge graph.
Applsci 12 02702 g003
Figure 4. F1 scores for different α , where F1 is the average result on five out-domain datasets.
Figure 4. F1 scores for different α , where F1 is the average result on five out-domain datasets.
Applsci 12 02702 g004
Figure 5. F1 scores of different baselines on out-domain datasets.
Figure 5. F1 scores of different baselines on out-domain datasets.
Applsci 12 02702 g005
Figure 6. The knowledge graph structure contained in the example.
Figure 6. The knowledge graph structure contained in the example.
Applsci 12 02702 g006
Table 1. Examples of Wikipedia–Wikidata mappings and triples.
Table 1. Examples of Wikipedia–Wikidata mappings and triples.
Wikipedia–Wikidata MappingsTriples
en.wikipedia.org/wiki/UniverseQ1Q1P2670Q523
en.wikipedia.org/wiki/StarQ523Q1P2184Q136407
en.wikipedia.org/wiki/Big_BangQ323Q1P793Q323
en.wikipedia.org/wiki/HappinessQ8Q8P31Q331769
en.wikipedia.org/wiki/Mood_(psychology)Q331769Q8P31Q9415
en.wikipedia.org/wiki/Toledo, _Minas_GeraisQ22065023Q22065023P131Q39109
en.wikipedia.org/wiki/Minas_GeraisQ39109Q22065023P17Q155
Table 2. Statistics of experiment datasets. Gold recall is the probability that the candidate sets of mentions contain the ground truth entities.
Table 2. Statistics of experiment datasets. Gold recall is the probability that the candidate sets of mentions contain the ground truth entities.
DatasetNumber MentionsNumber DocsMentions per DocGold Recall
AIDA-train18,44894619.5-
AIDA-A479121622.197.3
AIDA-B448523119.498.3
MSNBC6562032.898.5
AQUAINT7275014.594.2
ACE2004257367.190.6
CWEB11,15432034.891.1
WIKI682132021.392.4
Table 3. F1 scores on AIDA-B and five out-domain test sets. The last column is the average of F1 scores on the five out-domain datasets. The best results are in bold.
Table 3. F1 scores on AIDA-B and five out-domain test sets. The last column is the average of F1 scores on the five out-domain datasets. The best results are in bold.
ModelAIDA-BMSNBCAQUAINTACE2004CWEBWIKIAvg
AIDA-79568058.66367.32
GLOW-75838256.267.272.68
RI-90908667.573.481.38
PBoH87.69189.288.7---
Deep-ED92.2293.788.588.577.977.585.22
Ment–Norm93.0793.988.389.977.57885.5
DCA-SL94.6494.5787.3889.4473.4778.1684.6
DCA-RL93.7393.8088.2590.1475.5978.8485.32
KGEL(ours)93.4794.2688.1190.5477.2178.4085.7
Table 4. F1 scores of the ablation experiments.
Table 4. F1 scores of the ablation experiments.
ModelAIDA-BMSNBCAQUAINTACE2004CWEBWIKIAvg
KGEL93.4794.2688.1190.5477.2178.4085.7
- KGEmbs93.0793.988.389.977.578.085.5
- WikiEmbs87.1692.1281.5487.7372.8468.9680.64
- local84.8691.0579.1686.927064.4678.32
- WikiEmbs
Table 5. F1 scores of different variants on out-domain datasets.
Table 5. F1 scores of different variants on out-domain datasets.
ModelMSNBCAQUAINTACE2004CWEBWIKIAvg
Ment–Norm93.988.389.977.578.085.5
KGEL94.2688.1190.5477.2178.4085.7
Related-Fixed94.2688.3989.7477.4178.0685.57
Related-Vari93.6588.2588.1377.0777.8284.98
Related-Vari-diff93.887.4187.7377.0177.3984.67
Related-nn92.5886.4388.1374.8771.6782.74
Table 6. The examples predicted by the baseline model. The bold font in the first column denotes the mention, the second column is the entity predicted by the model, and the last column is the real entity corresponding to the mention.
Table 6. The examples predicted by the baseline model. The bold font in the first column denotes the mention, the second column is the entity predicted by the model, and the last column is the real entity corresponding to the mention.
MentionPredGold
…Arrive in London May 14…LondonLondon
…matches against English county sides…EnglandEngland
…Counties and Scotland Tour itinerary…Scotland_national_cricket_teamScotland
…match (at Edgbaston,Birmingham)…Edgbaston_Cricket_GroundEdgbaston
…Edgbaston, Birmingham) June…BirminghamBirmingham
…international (at The Oval, London)…The_OvalThe_Oval
…Sussex or Surrey (three days)…Surrey_County_Cricket_ClubSurrey_County_Cricket_Club
Table 7. The execution times of the models.
Table 7. The execution times of the models.
ModelTrain Time/EpochInference Time
Ment-Norm23 s9 s
KGEL25 s10 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Q.; Li, F.; Li, S.; Li, X.; Liu, K.; Liu, Q.; Dong, P. Improving Entity Linking by Introducing Knowledge Graph Structure Information. Appl. Sci. 2022, 12, 2702. https://doi.org/10.3390/app12052702

AMA Style

Li Q, Li F, Li S, Li X, Liu K, Liu Q, Dong P. Improving Entity Linking by Introducing Knowledge Graph Structure Information. Applied Sciences. 2022; 12(5):2702. https://doi.org/10.3390/app12052702

Chicago/Turabian Style

Li, Qijia, Feng Li, Shuchao Li, Xiaoyu Li, Kang Liu, Qing Liu, and Pengcheng Dong. 2022. "Improving Entity Linking by Introducing Knowledge Graph Structure Information" Applied Sciences 12, no. 5: 2702. https://doi.org/10.3390/app12052702

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop