Next Article in Journal
State Prediction Method for A-Class Insulation Board Production Line Based on Transfer Learning
Next Article in Special Issue
Biomedical Interaction Prediction with Adaptive Line Graph Contrastive Learning
Previous Article in Journal
On Convergence of Support Operator Method Schemes for Differential Rotational Operations on Tetrahedral Meshes Applied to Magnetohydrodynamic Problems
Previous Article in Special Issue
Identification of Nonlinear State-Space Systems via Sparse Bayesian and Stein Approximation Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperbolic Directed Hypergraph-Based Reasoning for Multi-Hop KBQA

1
Science and Technology on Information Systems Engineering Laboratory, National University of Defense Technology, Changsha 410073, China
2
People’s Liberation Army, Guangzhou 510600, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(20), 3905; https://doi.org/10.3390/math10203905
Submission received: 6 September 2022 / Revised: 27 September 2022 / Accepted: 11 October 2022 / Published: 21 October 2022
(This article belongs to the Special Issue Mathematics-Based Methods in Graph Machine Learning)

Abstract

:
The target of the multi-hop knowledge base question-answering task is to find answers of some factoid questions by reasoning across multiple knowledge triples in the knowledge base. Most of the existing methods for multi-hop knowledge base question answering based on a general knowledge graph ignore the semantic relationship between each hop. However, modeling the knowledge base as a directed hypergraph has the problems of sparse incidence matrices and asymmetric Laplacian matrices. To make up for the deficiency, we propose a directed hypergraph convolutional network modeled on hyperbolic space, which can better deal with the sparse structure, and effectively adapt to the problem of an asymmetric incidence matrix of directed hypergraphs modeled on a knowledge base. We propose an interpretable KBQA model based on the hyperbolic directed hypergraph convolutional neural network named HDH-GCN which can update relation semantic information hop-by-hop and pays attention to different relations at different hops. The model can improve the accuracy of the multi-hop knowledge base question-answering task, and has application value in text question answering, human–computer interactions and other fields. Extensive experiments on benchmarks—PQL, MetaQA—demonstrate the effectiveness and universality of our HDH-GCN model, leading to state-of-the-art performance.

1. Introduction

Knowledge base question answering (KBQA) has been a hot task in the field of natural language processing and is very challenging [1]. Several different QA datasets have been proposed, such as the Stanford Question Answering Dataset (SQuAD) [2,3], NarrativeQA [4] and CoQA [5], and the kind of reasoning based on these datasets is termed single-hop reasoning, since it requires reasoning only over a single piece of evidence [6]. For a QA task of single-hop reasoning, the performance of previous work [1,7] has been improved a lot over the last years.
However, in real-world QA tasks, obtaining answers often requires multi-hop reasoning [8], that is, to find a knowledge path consisting of multiple pieces of knowledge in the knowledge base to deduce the answer. Figure 1 shows a two-hop KBQA example. The reasoning path starts from the entity mentioned in the query and consists of the relations at each hop and the intermediate entities. The methods mentioned above that focus only on single-hop reasoning lack the ability to deal with multi-hop reasoning QA tasks. To solve a multi-hop QA task, some work has been proposed recently which can be mainly divided into two categories. One is the neural network-based methods such as models in [9,10] and the other is the graph neural network-based methods such as [11]; this has achieved desirable performance [12].
When we deal with multi-hop tasks, generally, entities extracted from the query begin to be retrieved in the knowledge base, go down to the next hop according to the specific relationship of different hops, repeat this step to form a reasoning path, and finally find the final answer. In this process, we argue that semantic relational information is crucial for multi-hop reasoning, while previous studies have not fully exploited semantic relational information. The work in [13] computes relation-specific transformations by separating different relations, which does not consider semantic relation information. Ref. [14] does not require updating relational information during multi-hop reasoning, but exploits relational information to obtain attention for static graphs.
In addition, the pairwise connections between nodes based on a graph network (GNN) are insufficient with respect to fully representing the higher-order relationships between relationships and entities in the knowledge graph. Recently, some main works on hypergraph convolutional networks (HGNN) have been proposed. HGNN uses hyperedges to connect more than two nodes at the same time, which is conducive to imitating human reasoning and accurately locating a group of entities connected by the same relationship, rather than reasoning entity by entity. The disadvantage is that HGNN is aimed at undirected hypergraphs, while knowledge graphs are directed, and each triplet has a specific directional meaning. HGNN collects potential learning relationships from connected entities, but does not reveal and further utilize them.
Based on the study in the hypergraph neural network introduced above, a directed hypergraph convolutional network-based model for multi-hop KBQA (2HR-DR) was proposed [15]. 2HR-DR models the entities extracted from questions and their related relationships and entities in the knowledge base into directed hypergraphs, and then uses Directed Hypergraph Convolutional Networks (DHGCN) [15] to predict relations hop-by-hop and form a sequential relation path to make the reasoning interpretable. 2HR-DR can explicitly learn and update relation information and dynamically concentrate on different relations at different hops.
Although 2HR-DR can better solve some of the challenges mentioned above, it still has some disadvantages. First, using directed hypergraphs to model entities and relationships may lead to a situation in which an entity can be related to many entities based on some relations while it may be related to a few entities based on some other relations. For example, as shown in Figure 1, when constructing a hypergraph of the query, the number of entities (actors) related to “Child_of_Deaf_Adults” can be very small based on some relations such as “Staring_in” while a large number of entities (actors) do “Act_in” this movie. This results in a large difference in the number of nodes contained by each hyperedge in the modeled directed hypergraph. In that case, the incidence matrices of constructed hypergraphs will be much sparser, and that will have a bad effect on the training efficiency and accuracy of the model. Second, 2HR-DR used the directed hypergraph convolution network, which needs the eigenvalue decomposition of Laplacion matrices when calculating the spectrum convolution of hypergraphs, and that requires that the Laplacian matrices are real symmetric matrices (we are not able to ensure that non-symmetric matrices can certainly perform eigenvalue decomposition). However, as for directed hypergraph convolution networks, since each hyperedge has a direction, the degree of each node is supposed to be divided into in degree and out degree, which are different in most cases; this leads to the fact that Laplacian matrices are often asymmetric matrices.
To solve the problems mentioned above, we firstly propose a Hyperbolic Directed Hypergraph Convolutional Network (HDH-GCN) for a directed hypergraph to take the direction of information transmission into account. We investigate hyperbolic embedding spaces [16] and manage to map the sparse data points and hypergraph to the hyperboloid manifold directly. The rationale is that hyperbolic space has a stronger ability than Euclidean space to accommodate networks with long-tailed distributions and sparse structures [17], which is also verified in our experiments. On that basis, we propose a Hyperbolic Directed Hypergraph Convolutional Network (HDH-GCN)-based framework for multi-hop QA. This framework explicitly updates the relation information and dynamically focuses on specific relations at every hop of the query. In addition, we record the semantic representation of the relationship in each hop, and the representation of the relationship in every hop is influenced by the representation of the relationship in the previous hops, which makes the QA task interpretable to a large extent.
In summary, we make the following contributions:
  • For solving the problem of sparse incidence matrices of directed hypergraphs modeled on a knowledge base, we design a method of modeling a directed hypergraph in hyperbolic space.
  • Based on the hyperbolic directed hypergraph, we propose a Hyperbolic Directed Hypergraph Convolutional Network (HDH-GCN) for a directed hypergraph and design a framework on this basis that can handle multi-hop knowledge base question-answering tasks well.
  • The modules constitute a new model, namely, HDH-GCN for handling the multi-hop knowledge base question-answering task. Through the experiments on several real-world datasets, we confirm the superiority of HDH-GCN over state-of-the-art models.

2. Related Work

2.1. Multi-Hop Question Answering

The multi-hop KBQA model can be basically divided into two types. The first is to apply the neural network mentioned earlier. These models use the previous single-hop question-answering method [18,19,20] for multi-hop question-answering tasks. Xu et al. improved KVMemNet to achieve better results across multiple triples [10]. Zhong et al. used coarse-grained modules and fine-grained modules [21]. Part of the method introduces an end-to-end framework, which is explicitly designed to simulate the step-by-step reasoning process involved in multi-hop QA and MRC. Kundu et al.’s [22] model constructed a path connecting questions and candidate answers, and then scored them through the neural architecture. Jiang et al. [23] also constructed a proposer to propose an answer from each root to leaf path in the reasoning tree, extract a key sentence containing the proposed answer from each path and finally combine them to predict the final answer. However, these methods lack consideration of graph structure information.
The other kind of method is based on graph neural networks. Sun et al. learnt what to retrieve from the KB and corpus and then reasoned over the built graph [24]. Tu et al. employed GCN to reason over heterogeneous graphs [25]. Xiong et al. achieved better performance by applying graph attention networks [14]. Cao et al. proposed a bi-directional attention entity graph convolutional network [26]. These models use r-GCN [13], which does not consider the semantic relation information, or use graph attention networks to assign static weights. Different from these models, ref. [15] proposes a dynamic relation strategy, which dynamically updates relation states during the reasoning process. Documents unrelated to the complex query may affect the accuracy of the model. In the “select, answer, and explain” (SAE) model proposed by Tu et al. [27], BERT [28] acts as the encoder in the selection module. Then a sentence extractor is applied to the output of BERT to obtain the sequential output of each sentence with precalculated sentence start and end indices, to filter out answer-unrelated documents and thus reduce the amount of distraction information. The selected answer-related documents are then input to a model, which jointly predicts the answer and supporting sentences. Concurrently to the SAE model, Bhargav et al. [29] used a two-stage BERT-based architecture to first select the supporting sentence and then used the filtered supporting sentence to predict the answer. The upstream side of Jiang et al.’s [23] proposed model is the Document Explorer to iteratively address relevant documents. Han et al. [15] proposed two-phase hypergraph-based reasoning with dynamic relations which explicitly learns and updates relation information and dynamically concentrates on different relations at different hops.

2.2. Hypergraph Convolutional Networks

Feng et al. [30] proposed a hypergraph neural network, which replaces the general graph with a hypergraph structure, effectively encoding the higher-order data correlation. Bai et al. [31] further enhanced the representational learning ability by using attention modules. Yadati, N. et al. [32] proposed a new method of training a GCN on a hypergraph using tools from the spectral theory of hypergraphs and applying the method to the problems of SSL (hypergraph-based semi-supervised learning) and combinatorial optimization on real-world hypergraphs. Zhang et al. [33] developed a new self-attention-based graph neural network applicable to homogeneous and heterogeneous hypergraphs with variable hyperlink sizes. Han et al. [15] proposed a directed hypergraph convolutional network that incorporates direction information into HGNN to deal with a directed knowledge graph.

2.3. Hyperbolic Neural Networks

Hyperbolic space has always been a popular research domain in mathematics. Some works have been conducted to explore the treelike structure of graphs [34,35] and the relations between hyperbolic space and hierarchical data such as languages and complex networks [36,37]. Such works have demonstrated the consistency between real-world scale-free and hierarchical data and the hyperbolic space, providing a theoretical basis for recent works which apply hyperbolic space to various tasks including link prediction, node classification and recommendation. Some researchers apply hyperbolic space to traditional metric learning approaches such as HyperBPR [38] and HyperML [39]. Some try to adopt hyperbolic space to neural networks and define hyperbolic neural network operations, producing powerful models such as hyperbolic neural networks [40], hyperbolic graph neural networks [41] and hyperbolic convolutional neural networks [17]. Meanwhile, ref. [42] provides a scalable hyperbolic recommender system for industry use. Ref. [43] applies hyperbolic space to heterogeneous networks for link prediction tasks. Ref. [44] applies hyperbolic space to next-POI recommendation. Ref. [45] proposes a path-based recommendation approach with hyperbolic embeddings, etc.

3. Methods

3.1. Hyperbolic Directed Hypergraph Convolutional Networks

In this section, we are going to introduce the directed hypergraph convolutional network constructed on hyperbolic space. Definitions of notations used in the text are shown in the Table 1.

3.1.1. Undirected Hypergraph Convolutional Network

We first introduce undirected hypergraph convolutional neural networks. Different from simple graphs, hyperedges in a hypergraph may contain two or more vertices. A hypergraph can be defined as ϱ = ( V , E , W ) , which includes a vertex set V, a hyperedge set E and each hyperedge is assigned with a weight by W which is a diagonal matrix whose diagonal lines are the weights of each hyperedge. We use a | V | × | E | incidence matrix H to denote a hypergraph ϱ , and H can be concretely expressed as:
h ( v , e ) = 1 v e 0 v e
For every vertex v V , the degree can be defined as d ( v ) = Σ e E ω ( e ) h ( v , e ) , For every hyperedge e E , the degree can be defined as d ( e ) = Σ v V ω ( v ) h ( v , e ) , D e and D v denote the diagonal matrices of the hyperedge degrees and the vertex degrees. We let Θ = D v 1 2 H W D e 1 H T D v 1 2 and Δ = I Θ is defined as the hypergraph Laplacian [46], according to the Laplacian expression form, it can be known that it is a symmetric semidefinite matrix. It can be obtained by eigenvalue decomposition of symmetric positive semidefinite matrices that Δ = Φ Λ Φ T ; [30] use the eigen vectors as the Fourier bases and the eigenvalues as frequencies to express the spectral convolution as:
g * X = Φ ( ( Φ T g ) ( Φ T X ) ) = Φ g ( Δ ) Φ T X
Ref. [30] then approximate the above equation by Chebyshev polynomials, modify inside parameters appropriately and finally formulate hyperedge convolution as:
X ( l + 1 ) = D v 1 H W D e 1 H T X ( l ) P
where X ( l ) is the node feature of the hyperedge at layer l, and P is the learnable parameter.

3.1.2. Hyperbolic Directed Hypergraph Convolutional Neural Network

As shown in Section 3.1.1, the derivation process of hyperedge convolution is based on the eigenvalue decomposition of the Laplacian matrix of the hypergraph. In an undirected hypergraph, the Laplacian matrix is a symmetric positive semidefinite matrix because the degree matrices of its vertices and hyperedges are unique, so the eigenvalue decomposition always works. However, for directed hypergraphs, due to hyperedges having direction, the degree matrices of vertices and hyperedges should be divided into out-degree matrix and in-degree matrix, and according to the random walk explanation of spectral hypergraph partitioning [46], the specific representation of the Laplacian matrix of a directed hypergraph should be as follows:
Δ = I Θ = I D v t a i l 1 H t a i l W D e h e a d 1 H h e a d T
where D v t a i l and D e h e a d are the diagonal matrices of tail degrees of nodes and head degrees of hyperedges, H t a i l and H h e a d stand for the tail and head incidence matrices.
H t a i l ( i , j ) = 1 v i e j t a i l 0 v i e j t a i l
H h e a d ( i , j ) = 1 v i e j h e a d 0 v i e j h e a d
Since the two incidence matrices are generally different in a directed hypergraph, the Laplacian matrix is often not a symmetric matrix; as a result, the directed hypergraph modeled on the knowledge base may not be able to derive the hyperedge convolution as the undirected hypergraph shown in Section 3.1.1, and this will produce calculation error to some extent.
Aiming at solving the problem above and the sparsity issue in hypergraphs modeled on a knowledge base, we apply the variant forms of GCN on the hyperbolic space [17] to the directed hypergraph and obtain the matrix form of the directed hypergraph convolution network on the hyperbolic space. The directed hypergraph convolution operations on hyperbolic spaces aggregate each representation vector individually in vector dimensions, without the aforementioned problem of symmetry of the Laplacian matrix of directed hypergraphs. The specific process is as follows:
We first transform the initial item features from Euclidean space to hyperbolic space H K , and then we feed the initial hyperbolic item embeddings to learn item embeddings. For the hyperbolic space, we set α : = { K , 0 , 0 , , 0 } H K as the north pole in H K , and the negative curvature of the hyperboloid manifold is 1 K . Then the initial item features in hyperbolic space can be deduced from Euclidean space as follows:
x ( 0 , H ) = e x p α K ( ( 0 , x ( 0 , E ) ) ) = ( K c o s h ( x ( 0 , E ) 2 K ) , K s i n h ( x ( 0 , E ) 2 K ) x ( 0 , E ) x ( 0 , E ) 2 )
where x ( 0 , H ) and x ( 0 , E ) are the initial hyperbolic embedding and the initial Euclidean embedding, respectively. For the directed hypergraph convolutional network, when updating entity representation through the convolutional layer, we first aggregate the head entities in directed hyperedges to obtain the representation of relations, and then accumulate relation representation containing the same tail entity to obtain the representation of the tail entity so as to continuously update the representation of entities. The representation can be aggregated via the following convolutional operation in hyperbolic space:
r i L , H = e x p α K ( j e i h e a d M i j l o g α K ( x j L , H ) )
where r i L , H is the hyperbolic hidden embedding of relation e i in the L-th layer after aggregation, the node j’s hyperbolic embedding is transformed to Euclidean embedding via l o g α K , so the Euclidean-based sum and add operations are available. e x p α K aims to transform the Euclidean-based embedding to hyperbolic embedding. M i j is the projecting weight defined as follows:
M i j = s o f t m a x j e i h e a d H ( M L P ( l o g α K ( x i L , H ) | | l o g α K ( x j L , H ) ) )
Accordingly, we can write an expression that evaluates the tail entity representation:
x t L + 1 , H = e x p α K ( s e t t a i l M t s l o g α K ( r s L , H ) )
where x t L , H is the hyperbolic embedding of tail entity x t in the L-th layer, e t t a i l stands for the directed hyperedge containing the tail entity x t .
Because e x p α K and l o g α K are the inverse of each other, the total convolution is as follows:
x t L + 1 , H = e x p α K ( s e t t a i l M t s w r L ( j e i h e a d M i j l o g α K ( x j L , H ) ) )
The formula cancels out the adjacent inverse operation e x p α K and l o g α K , and applies max pooling to obtain weights for each relation,
w r L = s o f t m a x ( M a x P o o l i n g ( R i L , H ) )
To facilitate the formulation of the model, we rewrite it in matrix form:
X L + 1 , H = e x p α K ( M t a i l W M h e a d l o g α K ( X L , H ) )
where X L + 1 , H and X ( L , H ) are the entity representation matrix in the L + 1 -th and L-th layer, respectively. W stands for a diagonal matrix of hyperedge weights, M h e a d and M t a i l are the aggregate matrices of the head part and tail part of directed hyperedges.

3.2. Model

In this section, we are going to introduce the concrete model for the multi-hop knowledge base QA task. The overview of how the model works is shwon in Figure 2.
Table 1. Descriptions of notations used in the following parts.
Table 1. Descriptions of notations used in the following parts.
SymbolDefinition
Ha | V | × | E | incidence matrix
D e a diagonal matrices of the hyperedge degrees
D v a diagonal matrices of the vertex degrees
Δ the hypergraph Laplacian
Φ the eigenvectors of the hypergraph Laplacian
the element-wise Hadamard product
H K a hyperboloid manifold which negative curvature is 1 K
α the north pole in H K
e x p α K the exponential map of the hyperboloid model
l o g α the logarithmic map of the hyperboloid model
c o s h hyperbolic cosine function
s i n h hyperbolic sine function
M h e a d the aggregate matrix of the head-part of directed hyperedges
M t a i l the aggregate matrix of the tail-part of directed hyperedges

3.2.1. Query-Aware Entity Encoder

The query-aware entity encoder encodes entities and relations in questions and their potential related entities in knowledge base to vector representation. Let L q , L e and L r respectively denote the embedding matrices of question, entities and relations; we begin by encoding each question using bidirectional Gated Recurrent Units (GRUs) [47].
E q = B i G R U ( t a n h ( W q L q + b q ) )
We then follow the work of [21]; we employ co-attention to learn query-aware entity representation,
A e q = L e ( H q T )
C e = s o f t m a x ( A e q ) E q
C q = s o f t m a x ( A e q T ) L e
D e = B i G R U ( s o f t m a x ( A e q ) C q )
E a t t n = f c ( [ C e ; D e ] )
where s o f t m a x stands for column-wise normalization, f c is a linear network which converts 2h dimension to h dimension.

3.2.2. Reasoning over Hypergraph

According to the property that relation embedding can be obtained in the intermediate processes of hyperbolic directed hypergraph convolutional networks, we separate the hyperbolic directed hypergraph convolutional networks into two steps. Specifically, the model first collects the features of nodes onto the connected hyperedges and explicitly represents the learning relationship. Then, it dynamically assigns the weight of the relationship according to the similarity between the problem and the relationship, and predicts the current relation. Finally, the node status is updated through the connection relationship information. The specific process is as follows.
Firstly, assuming that the current hop is l, we use a linear network to concatenate the node status obtained by the previous l 1 layer and the input entity representation of the current hop, and then map it onto the hyperbolic space.
U e l , H = e x p α K ( f i n ( [ E ( l 1 ) ; E a t t n ] ) )
where operator [ ; ] is column-wise concatenation. Then the model learns the relation representation R l , H by aggregating the connected head entity feature.
U r l , H = M h e a d l o g α K ( U e l , H ) P r
where P r is the relation-specific learnable parameter, we then use a linear network to concatenate the relation representation obtained by the previous l 1 layer and U r l , H to obtain a representation of relations at hop l.
R l , H = f r ( [ R ( l 1 ) , H ; U r l , H ] )
After that, we apply w r L in (12) to obtain weights for each relation and the diagonal matrix W of edge weights is W = d i a g ( w r L ) . The weight of the dynamic allocation relation depends on the updated relation representation hop-by-hop. This model predicts the current relation based on the relation weight.
Finally, the model adaptively updates entity states by accumulating connected relation feature
E l = e x p α K ( M t a i l W R l , H P e )
where P e is the entity-specific learnable parameter.

3.3. Training

For an L hop question, we sum the entity representation of each layer to obtain the final representation and use a liner layer f a n s to predict the answer distribution.
P = σ ( f a n s ( l = 1 L E ( l ) ) )
where σ is the sigmoid function.
Since the model needs to predict both the answer to the question and the reasoning path, the loss function consists of two parts, one is binary cross-entropy loss of the final answer prediction, the other is the negative log likelihood of the intermediate prediction of the reasoning path. The specific expression of the loss function is as follows:
L = i = 1 n ( y i l o g ( p i ) + ( 1 y i ) l o g ( 1 p i ) ) + λ × l = 1 L + 1 ( l o g ( w r l ( r l * ) ) )
where y i is the golden distribution over entities. r l * is the golden relation index at hop L. λ is a hyper parameter to balance the two terms.

4. Experiment

This section reports the experiments.

4.1. Experiment Setup

We detail the adopted datasets, evaluation metrics, parameters and baselines.

4.1.1. Datasets

We use two single answer KBQA datasets and two large-scale multi-answer KBQA datasets for the multi-hop KBQA task. We briefly outline these datasets in Table 2.
  • PQL-2H [48]: PQL-2H is a single answer KBQA dataset, which includes a knowledge base containing 5035 entities and 364 relationships, and a two-hop question set containing 1594 two-hop questions. These questions can be answered by following the reasoning path consisting of several relations and intermediate entities. The path has been given.
  • PQL-3H [48]: PQL-3H is a single-answer KBQA dataset, which includes the same knowledge base with 5035 entities and 364 relations as PQL-2H, and a three-hop question set with 1031 three-hop questions. The characteristics of questions and the reasoning path are the same as PQL-2H.
  • MetaQA-1H [49]: MetaQA-1H contains 116,045 questions for single-hop reasoning QA and the knowledge base in the dataset contains 40,128 entities and nine relations. To test QA systems in more realistic (and more difficult) scenarios, MetaQA-1H also provides neural-translation-model-paraphrased datasets, and text-to-speech-based audio datasets.
  • MetaQA-2H [49]: MetaQA-2H contains 148,724 questions for two-hop reasoning and the knowledge base in the dataset contains 40,128 entities and nine relations. MetaQA-2H provides neural-translation-model-paraphrased datasets, and text-to-speech-based audio datasets just like MetaQA-1H.

4.1.2. Metrics and Parameters

We test the effectiveness of the model in four datasets. The total questions in datasets are divided into three parts: 70% for training, 10% for validation and 20% for testing. We evaluate the experiment results via two standard metrics: F1 and Hits@1. The F1 value is an overall evaluation of the precision and recall, which can evaluate the performance of the model well, while Hits@1 measures the proportion of top 1 rankings. The aim of the training is to achieve high F1 and Hits@1.
The reported results are given for the best set of hyper-parameters evaluated on the validation set for each model after grid search on the following values: embedding size ∈ {100, 200, 300, 400, 500}, learning rate ∈ {1, 0.1, 0.01, 0.001}, λ and dropout are set to 1 and 0.4.

4.1.3. Baselines

We compare HDH-GCN with the following baselines:
  • KVMemNet [50]: This is an end-to-end memory network which divides the memory into two parts, the key memory stores the head entity and relation, and the value memory stores the tail entity.
  • IRN [48]: This is an interpretable reasoning network, which uses a hop-by-hop reasoning process and answers questions based on knowledge maps.
  • VRN [49]: An end-to-end variational learning algorithm is proposed, which can effectively solve the multi hop reasoning problem and simultaneously deal with the noise in the problem
  • GraftNet [12]: Text information and entities are introduced to construct a graph, and GCN is applied to reasoning.
  • SGReader [14]: This also combines the unstructured text and knowledge graph to figure out the incompleteness of the knowledge graph. The model employs graph attention to reason effectively.
  • 2HR-DR [15]: This models the entities extracted from questions and their related relationships and entities in the knowledge base into directed hypergraphs, then uses Directed Hypergraph Convolutional Networks to predict relations hop-by-hop and form a sequential relation path to make the reasoning interpretable.

4.2. Results of Main Experiment

Table 3 and Table 4 show the main experiment results for two kinds of multi-hop KBQA datasets. The highest scores are in bold. As shown in Table 3, we can find out that our proposed HDH-GCN can achieve optimal results under Hit@1 measurement standards (there is only one answer for each question in PQL datasets, so we only adopt Hits@1 for evaluation). For the rest of the datasets, except for the F1 value of HDH-GCN on the MetaQA-1H dataset, which does not exceed the baseline model, other evaluation indexes have been improved to some extent, as shown in Table 4. Specifically, HDH-GCN achieves an improvement for PQL-2H which is 0.9% higher than the second best model. It also obtains good result on PQL-3H, 1.2% higher than the second best one. Table 4 demonstrates the performance of the baseline methods and HDH-GCN on the MetaQA-1H dataset; our model improves Hits@1 by 1.8% and obtains competitive F1. For MetaQA2-Hop, we improve Hits@1 and F1 by 0.3% and 0.8%, respectively.
First of all, compared with models based on a knowledge base modeled on a simple graph, our model reconstructs the knowledge graph to hypergraph structure which fully considers the high-order data correlation. Meanwhile, we dynamically concentrate on relation information at different hops by performing loop operations for each hop of inference to guide the model to follow the golden relation path and select the final answers, so we can introduce the information of the intermediate reasoning path into our model to supervise the model focus on the dynamic relations at different hops.
When comparing with the directed hypergraph-based model 2HR-DR, the improvement of both evaluation values on the two PQL datasets is more obvious than on the other datasets. The reasons why our method performs better include: (1) The directed hypergraph is modeled on hyperbolic space, which effectively reduces the sparsity of the incidence matrix of the directed hypergraph; this will reduce the scale of matrix calculation during training and reduce the inadequacy of training. This also explains why the results on the two PQL datasets are better than on MetaQA-1H and MetaQA-2H, the number of relationships in PQL is much higher than in MetaQA, which leads to more obvious matrix sparsity problems (the number of relations in MetaQA is small, and each relation can relate to many entities). (2) In the convolutional network of a directed hypergraph, there is a computational process of eigenvalue decomposition of the asymmetric Laplacian matrix, but the asymmetric matrix may not be able to carry out the eigenvalue decomposition, which leads to the forced eigenvalue decomposition of the matrix that cannot be diagonalized similarly in the training process, and the training error is always caused. The problem can be solved effectively by deforming the Laplace matrix in hyperbolic space. This problem is evident in the single-answer QA task, while the multi-answer QA task will dilute the influence of this problem to a certain extent when calculating the F1 value, which also explains why the F1 value of HDH-GCN is not improved in MetaQA-1H.

4.3. Parameter Analysis

Embedding size is a significant factor in KBQA models, determining the performance of the model to a large extent. Hence, we will analyze the results obtained by the model on PQL-2H in different embedding sizes to investigate its impact. First, according to the Figure 3a, HDH-GCN outperforms other methods when the dimension is {100, 200, 300, 400}. The Hit@1 of HDH-GCN increases sharply with the early stage of increasing the embedding size and becomes smooth after the embedding size increases to 400. The Hit@1 of 2HR-DR is almost identical to TF-DHP’s from the start; however, because the sparsity issue becomes intense as the dimension increases, it cannot remain smooth like HDH-GCN when the embedding size increases. After the embedding size increases to a certain extent, 2HR-DR’s Hit@1 will decrease slightly. For other methods, since the knowledge base is not modeled on the hypergraph, the sparsity issue has no obvious effect on the training results of higher dimensions; however, due to the reasons mentioned above, its results in each dimension are not as superior as HDH-GCN. We also count the Hit@1 results on PQL-2H of each training session. In the Figure 3b, we compare the Hit@1 between HDH-GCN and 2HR-DR on model training. 2HR-DR is stopped early around 35 epochs because of not updating Hit@1 for 10 epochs, so the line is not complete. HDH-GCN always achieves better performance, and is still updating until around 34 epochs.

4.4. Approximate Training Time Comparison

On the two kinds of datasets PQL and MetaQA, HDH-GCN takes around 75 min and 3 h, respectively, of training time, while 2HR-DR takes around 2 h and 3 h, respectively. All were run on a GeForce GTX 1080 super GPU machine with Python 3.

4.5. Case Study

As Figure 4 shows, we give an exemplar question from PQL-2H and its corresponding reasoning path and triples in KB. It is clear that HDH-GCN has the ability to predict relations hop-by-hop and stop reasoning automatically. For question “Who is the singer of the theme song of the movie “Titanic”?”, the model firstly detects the relation “theme song”, then “singer” successively and finally meets <STOP> to end the reasoning process. From the “path” in Figure 4, we can observe our model’s predicted relation path (Theme song → Singer → <STOP>).

5. Conclusions and Future Work

In this paper, we introduce HDH-GCN, a novel model for multi-hop KBQA tasks. We model the directed hypergraph convolutional network in hyperbolic space, which effectively reduces the influence of the sparsity issue on the model effect. Our model can improve the accuracy of the multi-hop knowledge base question-answering task, and has application value in text question answering, human–computer interactions and other fields. The experimental results verify the advantages of HDH-GCN in both single-answer questions and multi-answer question datasets.
In the future, we will study the multi-hop knowledge base question-answering task in multi-modal data, study the possibility of modeling a multi-modal knowledge base by directed hypergraph and explore the possible application prospect.

Author Contributions

Conceptualization, G.X. and J.L.; methodology, G.X.; software, G.X. and J.L.; validation, G.X., J.L. and Z.T.; formal analysis, G.X.; investigation, G.X., J.L., and Y.Y.; data curation, G.X.; writing—original draft preparation, G.X.; writing—review and editing, G.X., J.L., Z.T. and B.G.; visualization, G.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by NSFC under grants Nos.61902417 and 71971212.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, X.; Zhang, J.; Li, D.; Li, P. Knowledge Graph Embedding Based Question Answering. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, WSDM 2019, Melbourne, VIC, Australia, 11–15 February 2019; Culpepper, J.S., Moffat, A., Bennett, P.N., Lerman, K., Eds.; ACM: New York, NY, USA, 2019; pp. 105–113. [Google Scholar]
  2. Rajpurkar, P.; Zhang, J.; Lopyrev, K.; Liang, P. SQuAD: 100,000+ Questions for Machine Comprehension of Text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, Austin, TX, USA, 1–4 November 2016; Su, J., Carreras, X., Duh, K., Eds.; The Association for Computational Linguistics: Stroudsburg, PA, USA, 2016; pp. 2383–2392. [Google Scholar]
  3. Rajpurkar, P.; Jia, R.; Liang, P. Know What You Don’t Know: Unanswerable Questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, 15–20 July 2018; Gurevych, I., Miyao, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; Volume 2: Short Papers, pp. 784–789. [Google Scholar]
  4. Kociský, T.; Schwarz, J.; Blunsom, P.; Dyer, C.; Hermann, K.M.; Melis, G.; Grefenstette, E. The NarrativeQA Reading Comprehension Challenge. Trans. Assoc. Comput. Linguist. 2018, 6, 317–328. [Google Scholar] [CrossRef] [Green Version]
  5. Reddy, S.; Chen, D.; Manning, C.D. CoQA: A Conversational Question Answering Challenge. Trans. Assoc. Comput. Linguist. 2019, 7, 249–266. [Google Scholar] [CrossRef]
  6. Cao, X.; Liu, Y. Coarse-grained decomposition and fine-grained interaction for multi-hop question answering. J. Intell. Inf. Syst. 2022, 58, 21–41. [Google Scholar] [CrossRef]
  7. Xu, K.; Reddy, S.; Feng, Y.; Huang, S.; Zhao, D. Question Answering on Freebase via Relation Extraction and Textual Evidence. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016, Berlin, Germany, 7–12 August 2016; The Association for Computer Linguistics: Stroudsburg, PA, USA, 2016; Volume 1: Long Papers. [Google Scholar]
  8. Lin, X.V.; Socher, R.; Xiong, C. Multi-Hop Knowledge Graph Reasoning with Reward Shaping. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 3243–3253. [Google Scholar]
  9. Zhou, M.; Huang, M.; Zhu, X. An Interpretable Reasoning Network for Multi-Relation Question Answering. In Proceedings of the 27th International Conference on Computational Linguistics, COLING 2018, Santa Fe, NM, USA, 20–26 August 2018; Bender, E.M., Derczynski, L., Isabelle, P., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 2010–2022. [Google Scholar]
  10. Xu, K.; Lai, Y.; Feng, Y.; Wang, Z. Enhancing Key-Value Memory Neural Networks for Knowledge Based Question Answering. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019; Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1 (Long and Short Papers), pp. 2937–2947. [Google Scholar]
  11. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Sun, H.; Dhingra, B.; Zaheer, M.; Mazaitis, K.; Salakhutdinov, R.; Cohen, W.W. Open Domain Question Answering Using Early Fusion of Knowledge Bases and Text. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Riloff, E., Chiang, D., Hockenmaier, J., Tsujii, J., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 4231–4242. [Google Scholar]
  13. Schlichtkrull, M.S.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In Proceedings of the Semantic Web—15th International Conference, ESWC 2018, Heraklion, Greece, 3–7 June 2018; Gangemi, A., Navigli, R., Vidal, M., Hitzler, P., Troncy, R., Hollink, L., Tordai, A., Alam, M., Eds.; Lecture Notes in Computer Science; Springer: Berlin, Germany, 2018; Volume 10843, pp. 593–607. [Google Scholar]
  14. Xiong, W.; Yu, M.; Chang, S.; Guo, X.; Wang, W.Y. Improving Question Answering over Incomplete KBs with Knowledge-Aware Reader. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, 28 July–2 August 2019; Korhonen, A., Traum, D.R., Màrquez, L., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1: Long Papers, pp. 4258–4264. [Google Scholar]
  15. Han, J.; Cheng, B.; Wang, X. Two-Phase Hypergraph Based Reasoning with Dynamic Relations for Multi-Hop KBQA. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, Yokohama, Japan, 11–17 July 2020; Bessiere, C., Ed.; pp. 3615–3621. [Google Scholar]
  16. Zhang, S.; Chen, H.; Ming, X.; Cui, L.; Yin, H.; Xu, G. Where are we in embedding spaces? A Comprehensive Analysis on Network Embedding Approaches for Recommender Systems. arXiv 2021, arXiv:2105.08908. [Google Scholar]
  17. Chami, I.; Ying, Z.; Ré, C.; Leskovec, J. Hyperbolic Graph Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; 2019; pp. 4869–4880. [Google Scholar]
  18. Bordes, A.; Weston, J.; Usunier, N. Open Question Answering with Weakly Supervised Embedding Models. In Proceedings of the Machine Learning and Knowledge Discovery in Databases—European Conference, ECML PKDD 2014, Nancy, France, 15–19 September 2014; Part I; Lecture Notes in Computer Science. Calders, T., Esposito, F., Hüllermeier, E., Meo, R., Eds.; Springer: Berlin, Germany, 2014; Volume 8724, pp. 165–180. [Google Scholar]
  19. Hao, Y.; Zhang, Y.; Liu, K.; He, S.; Liu, Z.; Wu, H.; Zhao, J. An End-to-End Model for Question Answering over Knowledge Base with Cross-Attention Combining Global Knowledge. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, BC, Canada, 30 July–4 August 2017; Barzilay, R., Kan, M., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; Volume 1: Long Papers, pp. 221–231. [Google Scholar]
  20. Min, S.; Zhong, V.; Socher, R.; Xiong, C. Efficient and Robust Question Answering from Minimal Context over Documents. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, 15–20 July 2018; Gurevych, I., Miyao, Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; Volume 1: Long Papers, pp. 1725–1735. [Google Scholar]
  21. Zhong, V.; Xiong, C.; Keskar, N.S.; Socher, R. Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  22. Kundu, S.; Khot, T.; Sabharwal, A.; Clark, P. Exploiting Explicit Paths for Multi-hop Reading Comprehension. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, 28 July–2 August 2019; Korhonen, A., Traum, D.R., Màrquez, L., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1: Long Papers, pp. 2737–2747. [Google Scholar]
  23. Jiang, Y.; Joshi, N.; Chen, Y.; Bansal, M. Explore, Propose, and Assemble: An Interpretable Model for Multi-Hop Reading Comprehension. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, 28 July–2 August 2019; Korhonen, A., Traum, D.R., Màrquez, L., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1: Long Papers, pp. 2714–2725. [Google Scholar]
  24. Sun, H.; Bedrax-Weiss, T.; Cohen, W.W. PullNet: Open Domain Question Answering with Iterative Retrieval on Knowledge Bases and Text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, 3–7 November 2019; Inui, K., Jiang, J., Ng, V., Wan, X., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 2380–2390. [Google Scholar]
  25. Tu, M.; Wang, G.; Huang, J.; Tang, Y.; He, X.; Zhou, B. Multi-hop Reading Comprehension across Multiple Documents by Reasoning over Heterogeneous Graphs. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, 28 July–2 August 2019; Korhonen, A., Traum, D.R., Màrquez, L., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1: Long Papers, pp. 2704–2713. [Google Scholar]
  26. Cao, Y.; Fang, M.; Tao, D. BAG: Bi-directional Attention Entity Graph Convolutional Network for Multi-hop Reasoning Question Answering. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019; Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1 (Long and Short Papers), pp. 357–362. [Google Scholar]
  27. Tu, M.; Huang, K.; Wang, G.; Huang, J.; He, X.; Zhou, B. Select, Answer and Explain: Interpretable Multi-Hop Reading Comprehension over Multiple Documents. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; AAAI Press: Menlo Park, CA, USA, 2020; pp. 9073–9080. [Google Scholar]
  28. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, 2–7 June 2019; Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1 (Long and Short Papers), pp. 4171–4186. [Google Scholar]
  29. Bhargav, G.P.S.; Glass, M.R.; Garg, D.; Shevade, S.K.; Dana, S.; Khandelwal, D.; Subramaniam, L.V.; Gliozzo, A. Translucent Answer Predictions in Multi-Hop Reading Comprehension. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; AAAI Press: Menlo Park, CA, USA, 2020; pp. 7700–7707. [Google Scholar]
  30. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph Neural Networks. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, HI, USA, 27 January–1 February 2019; AAAI Press: Menlo Park, CA, USA, 2019; pp. 3558–3565. [Google Scholar]
  31. Bai, S.; Zhang, F.; Torr, P.H.S. Hypergraph convolution and hypergraph attention. Pattern Recognit. 2021, 110, 107637. [Google Scholar] [CrossRef]
  32. Yadati, N.; Nimishakavi, M.; Yadav, P.; Nitin, V.; Louis, A.; Talukdar, P.P. HyperGCN: A New Method For Training Graph Convolutional Networks on Hypergraphs. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; 2019; pp. 1509–1520. [Google Scholar]
  33. Zhang, R.; Zou, Y.; Ma, J. Hyper-SAGNN: A self-attention based graph neural network for hypergraphs. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  34. Adcock, A.B.; Sullivan, B.D.; Mahoney, M.W. Tree-Like Structure in Large Social and Information Networks. In Proceedings of the 2013 IEEE 13th International Conference on Data Mining, Dallas, TX, USA, 7–10 December 2013; Xiong, H., Karypis, G., Thuraisingham, B., Cook, D.J., Wu, X., Eds.; IEEE Computer Society: Washington, DC, USA, 2013; pp. 1–10. [Google Scholar]
  35. Chen, W.; Fang, W.; Hu, G.; Mahoney, M.W. On the Hyperbolicity of Small-World and Treelike Random Graphs. Internet Math. 2013, 9, 434–491. [Google Scholar] [CrossRef] [Green Version]
  36. Krioukov, D.V.; Papadopoulos, F.; Kitsak, M.; Vahdat, A.; Boguñá, M. Hyperbolic Geometry of Complex Networks. arXiv 2010, arXiv:1006.5169. [Google Scholar] [CrossRef] [PubMed]
  37. Nickel, M.; Kiela, D. Poincaré Embeddings for Learning Hierarchical Representations. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R., Eds.; 2017; pp. 6338–6347. [Google Scholar]
  38. Vinh, T.D.Q.; Tay, Y.; Zhang, S.; Cong, G.; Li, X. Hyperbolic Recommender Systems. arXiv 2018, arXiv:1809.01703. [Google Scholar]
  39. Tran, L.V.; Tay, Y.; Zhang, S.; Cong, G.; Li, X. HyperML: A Boosting Metric Learning Approach in Hyperbolic Space for Recommender Systems. In Proceedings of the WSDM ’20: The Thirteenth ACM International Conference on Web Search and Data Mining, Houston, TX, USA, 3–7 February 2020; Caverlee, J., Hu, X.B., Lalmas, M., Wang, W., Eds.; ACM: New York, NY, USA, 2020; pp. 609–617. [Google Scholar]
  40. Ganea, O.; Bécigneul, G.; Hofmann, T. Hyperbolic Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, QC, Canada, 3–8 December 2018; Bengio, S., Wallach, H.M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; 2018; pp. 5350–5360. [Google Scholar]
  41. Liu, Q.; Nickel, M.; Kiela, D. Hyperbolic Graph Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H.M., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E.B., Garnett, R., Eds.; 2019; pp. 8228–8239. [Google Scholar]
  42. Chamberlain, B.P.; Hardwick, S.R.; Wardrope, D.R.; Dzogang, F.; Daolio, F.; Vargas, S. Scalable Hyperbolic Recommender Systems. arXiv 2019, arXiv:1902.08648. [Google Scholar]
  43. Wang, X.; Zhang, Y.; Shi, C. Hyperbolic Heterogeneous Information Network Embedding. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, HI, USA, 27 January–1 February 2019; AAAI Press: Menlo Park, CA, USA, 2019; pp. 5337–5344. [Google Scholar]
  44. Feng, S.; Tran, L.V.; Cong, G.; Chen, L.; Li, J.; Li, F. HME: A Hyperbolic Metric Embedding Approach for Next-POI Recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, 25–30 July 2020; Huang, J.X., Chang, Y., Cheng, X., Kamps, J., Murdock, V., Wen, J., Liu, Y., Eds.; ACM: New York, NY, USA, 2020; pp. 1429–1438. [Google Scholar]
  45. Papadis, N.; Stai, E.; Karyotis, V. A path-based recommendations approach for online systems via hyperbolic network embedding. In Proceedings of the 2017 IEEE Symposium on Computers and Communications, ISCC 2017, Heraklion, Greece, 3–6 July 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 973–980. [Google Scholar]
  46. Zhou, D.; Huang, J.; Schölkopf, B. Learning with Hypergraphs: Clustering, Classification, and Embedding. In Proceedings of the Advances in Neural Information Processing Systems 19, Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; Schölkopf, B., Platt, J.C., Hofmann, T., Eds.; MIT Press: Cambridge, MA, USA, 2006; pp. 1601–1608. [Google Scholar]
  47. Cho, K.; van Merrienboer, B.; Gülçehre, Ç.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, Doha, Qatar, 25–29 October 2014; A meeting of SIGDAT, a Special Interest Group of the ACL. Moschitti, A., Pang, B., Daelemans, W., Eds.; ACL, 2014; pp. 1724–1734. [Google Scholar]
  48. Zhou, M.; Huang, M.; Zhu, X. An Interpretable Reasoning Network for Multi-Relation Question Answering. In Proceedings of the 27th International Conference on Computational Linguistics (COLING), Santa Fe, NM, USA, 20–26 August 2018; pp. 2010–2022. [Google Scholar]
  49. Zhang, Y.; Dai, H.; Kozareva, Z.; Smola, A.J.; Song, L. Variational Reasoning for Question Answering With Knowledge Graph. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th Innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, LA, USA, 2–7 February 2018; McIlraith, S.A., Weinberger, K.Q., Eds.; AAAI Press: Menlo Park, CA, USA, 2018; pp. 6069–6076. [Google Scholar]
  50. Miller, A.H.; Fisch, A.; Dodge, J.; Karimi, A.; Bordes, A.; Weston, J. Key-Value Memory Networks for Directly Reading Documents. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, EMNLP 2016, Austin, TX, USA, 1–4 November 2016; Su, J., Carreras, X., Duh, K., Eds.; The Association for Computational Linguistics: Stroudsburg, PA, US, 2016; pp. 1400–1409. [Google Scholar]
Figure 1. Sketch of an exemplar of multi-hop question answering based on knowledge graph. In the figure, arrows with different colors represent the relations in two hops that need to be extracted in order to obtain the reasoning path to the query, and colored dots represent the entities related to the question in knowledge base.
Figure 1. Sketch of an exemplar of multi-hop question answering based on knowledge graph. In the figure, arrows with different colors represent the relations in two hops that need to be extracted in order to obtain the reasoning path to the query, and colored dots represent the entities related to the question in knowledge base.
Mathematics 10 03905 g001
Figure 2. An overview of how the model works.
Figure 2. An overview of how the model works.
Mathematics 10 03905 g002
Figure 3. (a) Hit@1 over different embedding sizes of knowledge base QA models, evaluated on PQL-2H. (b) Hit@1 over different training epochs, evaluated on PQL-2H.
Figure 3. (a) Hit@1 over different embedding sizes of knowledge base QA models, evaluated on PQL-2H. (b) Hit@1 over different training epochs, evaluated on PQL-2H.
Mathematics 10 03905 g003
Figure 4. An exemplar of a two-hop query in a KBQA dataset; the figure shows the reasoning path and triples related to entities in the query, and graphically shows how the model reaches the answer to the question through the reasoning path.
Figure 4. An exemplar of a two-hop query in a KBQA dataset; the figure shows the reasoning path and triples related to entities in the query, and graphically shows how the model reaches the answer to the question through the reasoning path.
Mathematics 10 03905 g004
Table 2. Statistics of the hypergraph datasets used in the experiments.
Table 2. Statistics of the hypergraph datasets used in the experiments.
DatasetsPQL-2HPQL-3HMetaQA-1HMetaQA-2H
Number of
Questions
15941031116,049148,724
Number of
Entities in
knowledge base
5034503440,12840,128
Number of
Relations in
knowledge base
36436499
Table 3. Results of Hit@1 on PQL-2H and PQL-3H.
Table 3. Results of Hit@1 on PQL-2H and PQL-3H.
ModelPQL-2HPQL-3H
KVMemNet0.6900.617
IRN0.7250.710
GraftNet0.7070.910
SGReader0.7190.893
2HR-DR0.7550.921
HDH-GCN0.7640.933
Table 4. Results of Hit@1 and F1 on MetaQA-1H and MetaQA-2H.
Table 4. Results of Hit@1 and F1 on MetaQA-1H and MetaQA-2H.
ModelMetaQA-1HMetaQA-2H
Hit@1F1Hit@1F1
KVMemNet0.958-0.251-
VRN0.975-0.898-
GraftNet0.9700.9100.9480.727
SGReader0.9670.9600.8070.798
2HR-DR0.9880.9730.9370.814
HDH-GCN0.9900.9680.9510.822
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiao, G.; Liao, J.; Tan, Z.; Yu, Y.; Ge, B. Hyperbolic Directed Hypergraph-Based Reasoning for Multi-Hop KBQA. Mathematics 2022, 10, 3905. https://doi.org/10.3390/math10203905

AMA Style

Xiao G, Liao J, Tan Z, Yu Y, Ge B. Hyperbolic Directed Hypergraph-Based Reasoning for Multi-Hop KBQA. Mathematics. 2022; 10(20):3905. https://doi.org/10.3390/math10203905

Chicago/Turabian Style

Xiao, Guanchen, Jinzhi Liao, Zhen Tan, Yiqi Yu, and Bin Ge. 2022. "Hyperbolic Directed Hypergraph-Based Reasoning for Multi-Hop KBQA" Mathematics 10, no. 20: 3905. https://doi.org/10.3390/math10203905

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop