Next Article in Journal
Some Basic Inequalities on (ϵ)-Para Sasakian Manifold
Next Article in Special Issue
Big Data Clustering Using Chemical Reaction Optimization Technique: A Computational Symmetry Paradigm for Location-Aware Decision Support in Geospatial Query Processing
Previous Article in Journal
A Numerical Confirmation of a Fractional-Order COVID-19 Model’s Efficiency
Previous Article in Special Issue
Interactive Image Segmentation Based on Feature-Aware Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hypernetwork Representation Learning Based on Hyperedge Modeling

1
Department of Computer Technology and Applications, Qinghai University, Xining 810000, China
2
State Key Laboratory of Tibetan Intelligent Information Processing and Application, Qinghai Normal University, Xining 810000, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(12), 2584; https://doi.org/10.3390/sym14122584
Submission received: 9 November 2022 / Revised: 2 December 2022 / Accepted: 4 December 2022 / Published: 7 December 2022

Abstract

:
Most network representation learning approaches only consider the pairwise relationships between the nodes in ordinary networks but do not consider the tuple relationships, namely the hyperedges, among the nodes in the hypernetworks. Therefore, to solve the above issue, a hypernetwork representation learning approach based on hyperedge modeling, abbreviated as HRHM, is proposed, which fully considers the hyperedges to obtain ideal node representation vectors that are applied to downstream machine learning tasks such as node classification, link prediction, community detection, and so on. Experimental results on the hypernetwork datasets show that with regard to the node classification task, the mean node classification accuracy of HRHM approach goes beyond other best baseline approach by about 1% on the MovieLens and wordnet, and with regard to the link prediction task, except for HPHG approach, the mean AUC value of HRHM approach surpasses that of other baseline approaches by about 17%, 18%, and 6%, respectively, on the GPS, drug, and wordnet. The mean AUC value of HRHM approach is very close to that of other best baseline approach on the MovieLens.

1. Introduction

With the rapid development of artificial intelligence, hypernetwork representation learning has gradually become a research hotspot in the field of machine learning. Hypernetwork representation learning maps the nodes in the hypernetwork to a low-dimensional vector representation space. The learned node representation vectors can be applied to node classification [1], link prediction [2], community detection [3], and so on.
According to the types of the hypernetwork, the hypernetwork representation learning can be divided into homogeneous hypernetwork representation learning and heterogeneous hypernetwork representation learning. Homogeneous hypernetwork representation learning aims to map the nodes of a single type in the homogeneous hypernetwork to a low-dimensional vector representation space. For example, Zhou [4] proposed the hypergraph embedding approach on the basis of the spectral hypergraph clustering [5], but the high computational complexity of this approach limits the wide application. HGNN [6] extended the convolution operation to hypergraph embedding, but the datasets used in this approach are not really hypernetwork datasets. In a word, although the above homogeneous hypernetwork representation learning approaches have good representation learning abilities, they do not consider the heterogeneity of the hypernetwork. Therefore, the researchers propose heterogeneous hypernetwork representation learning approaches, which aim to learn distinguishing representation vectors for different types of nodes in the heterogeneous hypernetwork. For example, HHNE [7] is designed as a fully connected graph convolution layer to project different types of nodes into a common low-dimensional vector representation space, but the computational complexity of this approach is too high to be suitable for large-scale heterogeneous networks. DHNE [8] realizes the local and global proximity of nonlinear tuple similar functions in the embedding space, but because the multi-layer perceptron is used in the DHNE approach, the approach is limited to heterogeneous hyperedges with fixed size, and the relationships among multi-type instances with unfixed size cannot be considered.
To sum up, although the above hypernetwork representation learning methods can obtain nice node representation vectors, there are various issues, especially high computational complexity and the limitation of the hyperedges with a fixed size. Therefore, to solve the above issues, a hypernetwork representation learning approach based on hyperedge modeling to effectively capture complex tuple relationships (i.e., hyperedges) among the nodes is proposed, which is suitable for the hypernetwork with the hyperedges with unfixed size and improves the computational efficiency.
The following two aspects are the features of this paper:
  • A hypernetwork representation learning approach based on hyperedge modeling is proposed to map the nodes in the hypernetwork to a low-dimensional vector representation space, where the main components of the learned node representation vectors are the hypernetwork structure and the hyperedges;
  • The advantage of HRHM approach is that the hyperedges with unfixed size are encoded in the learned node representation vectors. The disadvantage of HRHM approach is that the partial information of the hypernetwork structure is lost because the hypernetwork abstracted as the hypergraph is transformed into the ordinary network abstracted as two-section graph.

2. Related Works

Nowadays, researchers have proposed some hypernetwork representation learning approaches to obtain node representation vectors that are rich in hypernetwork structure information. The existing hypernetwork representation learning approaches can be divided into homogeneous hypernetwork representation learning approaches and heterogeneous hypernetwork representation learning approaches. With regard to homogeneous hypernetwork representation learning approaches, Zhou [4] proposed the hypergraph embedding approach on the basis of the spectral hypergraph clustering. HyperGCN [9] approximates each hyperedge of a hypergraph by a set of pairwise edges connecting the nodes in the hyperedge and treats the hypergraph learning as graph learning. HGNN [6] extends the convolution operation to hypergraph embedding, which is convolved by the hypergraph Laplacian function in the spectral domain and further approximated by truncated Chebyshev polynomials. LHCN [10] maps the hypergraph to a weighted attributed line graph and learns graph convolutions on this line graph. With regard to heterogeneous hypernetwork representation learning approaches, HHNE [7] is designed as a fully connected graph convolutional layer to project different types of nodes into a common low-dimensional space and uses a tuple similarity function to protect the network structure, and a rank-based loss function is used to improve the similarity scores of hyperedges in the embedding space. DHNE [8] is a new deep model to realize the local and global proximity of the nonlinear tuple similarity function in the embedding space. HPHG [11] is a deep model called Hyper-gram to capture pairwise and tuple relationships in the node sequences. Hyper-SAGNN [12] utilizes a self-attention mechanism [13] to aggregate hypergraph information.

3. Problem Definition of HRHM Approach

The hypernetwork H = ( V , E ) abstracted as the hypergraph consists of the node set V = { v i } i = 1 | V | and the hyperedge set E = { e i = { v 1 , v 2 , , v τ } } i = 1 | E |   ( τ 2 ) . The HRHM approach aims to learn a low-dimensional representation vector r n R k for each node n in the hypernetwork, where k is much smaller than | V | .
In order to understand the process of hypernetwork representation learning well, take the drug hypernetwork as an example, where the triplet < user, drug, adverse reaction > is the hyperedge. Since each drug has several specific adverse reactions, there is a semantic relevance between the drug and the adverse reaction. How to assess the above semantic relevance is the real-life problem, which is solved by hypernetwork representation learning, which obtains a node representation vector to calculate the similarity between the drug and the adverse reaction to assess the semantic relevance.

4. Preliminaries

4.1. Two-Section Graph Transformed from Hypergraph

According to the literature [14], the two-section graph structure is closer to the hypergraph structure than the line graph and the incidence graph. Therefore, the two-section graph transformed from the hypergraph is used in this paper to carry out the research of hypernetwork representation learning. The hypergraph and its corresponding two-section graph are shown in Figure 1.
The two-section graph S = ( V , E ) transformed from the hypergraph H = ( V , E ) is an ordinary graph to meet the following conditions:
  • The node set V′ of two-section graph S is identical with the node set V of the hypergraph H;
  • Any two different nodes are associated with one edge if and only if these two nodes belong to at least one hyperedge simultaneously.

4.2. TransE Model

TransE [15] is a knowledge representation model with the translation mechanism, which thinks that if the head entity h and the tail entity t are with the relationship r , the triplet ( h , r , t ) holds. Moreover, the head entity vector h plus the relationship vector r is almost identical with the tail entity vector t , namely h + r t , when the triplet ( h , r , t ) holds; otherwise, h + r t . The TransE framework is shown in Figure 2.

5. HRHM Approach

This section introduces hypernetwork representation learning based on hyperedge modeling in more detail. Firstly, the cognitive topology model is introduced in Section 5.1. Secondly, the cognitive hyperedge model is introduced in Section 5.2. Finally, the joint optimization of the above two models is introduced in more detail in Section 5.3.

5.1. Cognitive Topology Model

In order to improve the computational efficiency, a cognitive topology model with the negative sampling [16] is introduced to capture the hypernetwork structure. Under the condition of the node sequences C , we try to maximize the following target function of the cognitive topology model to obtain the representation vectors rich in the hypernetwork structure.
D 1 = n C u { { n } N E G 1 ( n ) } p ( u | c o n t e x t ( n ) )
where N E G 1 ( n ) is the subset of negative samples of the center node n , regarded as the positive sample. c o n t e x t ( n ) is the contextual nodes of the center node n . p ( u | c o n t e x t ( n ) ) is defined as follows.
p ( u | c o n t e x t ( n ) ) = { σ ( X n T θ u ) , L n ( u ) = 1 1 σ ( X n T θ u ) , L n ( u ) = 0
where σ ( X n T θ u ) = 1 / ( 1 + e X n T θ u ) is a sigmoid function. X n is the sum vector of all nodes representation vectors corresponding to c o n t e x t ( n ) . θ n is the parameter vector. For   u V , the node label L n ( u ) is defined as follows.
L n ( u ) = { 1 , u { n } 0 , u N E G 1 ( n )
By means of Equation (3), Equation (2) can also be written as an integral expression.
p ( u | c o n t e x t ( n ) ) = [ σ ( X n T θ u ) ] L n ( u ) [ 1 σ ( X n T θ u ) ] 1 L n ( u )
By substituting Equation (4) into Equation (1), Equation (1) can be rewritten as follows:
D 1 = n C u { { n } N E G 1 ( n ) } { [ σ ( X n T θ u ) ] L n ( u ) [ 1 σ ( X n T θ u ) ] 1 L n ( u ) }
Formally, maximizing the target function D 1 makes the learned node representation vectors rich in hypernetwork structure.

5.2. Cognitive Hyperedge Model

Because the qualities of the node representation vectors from the above cognitive topology model do not consider that the hyperedges are not high, a novel cognitive hyperedge model with the negative sampling is proposed to consider the hyperedges to learn node representation vectors of high quality, where the hyperedges are deemed as the interaction relationships among the nodes and regarded as the translation operations in the representation vector space in the TransE model.
Under the condition of the hyperedge constraint, we try to maximize the following target function of the cognitive hyperedge model to obtain the representation vectors rich in the hyperedges.
D 2 = n C r R n h H r ξ { { n } N E G 2 ( n ) } p ( ξ | h + r ) = n C r R n h H r ξ { { n } N E G 2 ( n ) } { σ ( e h + r T θ ξ ) δ n ( ξ ) [ 1 σ ( e h + r T θ ξ ) ] 1 δ n ( ξ ) }
where R n is the set of the hyperedges, namely relationships associated with the center node n ; H r is the set of the nodes with the relation r with the center node n ; N E G 2 ( n ) is the subset of negative samples of the center node n ; θ ξ is the parameter vector; and the parameter vector e h + r is the sum vector of the parameter vectors e h and e r , namely e h + r = e h + e r .
For   ξ V , the node label δ n ( ξ ) in the Equation (6) is defined as follows:
δ n ( ξ ) = { 1 , ξ { n } 0 , ξ N E G 2 ( n )
Formally, maximizing the target function D 2 makes the learned node representation vectors rich in hyperedges.

5.3. Joint Optimization

In this subsection, the hypernetwork representation learning approach based on hyperedge modeling, abbreviated as HRHM, is proposed. The HRHM approach can jointly optimize the cognitive topology model and cognitive hyperedge model. Figure 3 shows the HRHM framework.
From Figure 3, the network topology representation and hyperedge representation from the cognitive topology model and cognitive hyperedge model, respectively, share the same representation.
For ease of calculation, take the logarithm of D 1 and D 2 to maximize the following joint optimization target function to make the hyperedges fully incorporated into the node representation vectors.
L = n C { u { { n } N E G 1 ( n ) } { L n ( u ) l o g [ σ ( X n T θ u ) ] + [ 1 L n ( u ) ] l o g [ 1 σ ( X n T θ u ) ] } + β r R n h H r ξ { { n } N E G 2 ( n ) } { δ n ( ξ ) l o g [ σ ( e h + r T θ ξ ) ] + [ 1 δ n ( ξ ) ] l o g [ 1 σ ( e h + r T θ ξ ) ] } } = n C { u { { n } N E G 1 ( n ) } { L n ( u ) l o g [ σ ( X n T θ u ) ] + [ 1 L n ( u ) ] l o g [ 1 σ ( X n T θ u ) ] } + r R n h H r ξ { { n } N E G 2 ( n ) } β { δ n ( ξ ) l o g [ σ ( e h + r T θ ξ ) ] + [ 1 δ n ( ξ ) ] l o g [ 1 σ ( e h + r T θ ξ ) ] } }
where the harmonic factor β is used to equilibrate the contribution rate between the cognitive topology model and cognitive hyperedge model.
For ease of derivation, L ( n , u , r , h , ξ ) is denoted as follows:
L ( n , u , r , h , ξ ) = { L n ( u ) l o g [ σ ( X n T θ u ) ] + [ 1 L n ( u ) ] l o g [ 1 σ ( X n T θ u ) ] } + β { δ n ( ξ ) l o g [ σ ( e h + r T θ ξ ) ] + [ 1 δ n ( ξ ) ] l o g [ 1 σ ( e h + r T θ ξ ) ] }
The stochastic gradient ascent approach is used to optimize the target function L . The acquisition of four kinds of gradients of L is the key to this paper.
Firstly, the gradient on the parameter vector θ u is calculated as follows:
L ( n , u , r , h , ξ ) θ u = L n ( u ) [ 1 σ ( X n T θ u ) ] X n [ 1 L n ( u ) ] σ ( X n T θ u ) X n = { L n ( u ) [ 1 σ ( X n T θ u ) ] [ 1 L n ( u ) ] σ ( X n T θ u ) } X n = [ L n ( u ) σ ( X n T θ u ) ] X n
Therefore, the parameter vector θ u is updated as follows:
θ u = θ u + α [ L n ( u ) σ ( X n T θ u ) ] X n
where α is the learning rate.
Secondly, the gradient on the sum vector X n is calculated as follows via the symmetry property between θ u and X n :
L ( n , u , r , h , ξ ) X n = [ L n ( u ) σ ( X n T θ u ) ] θ u
Therefore, the representation vector v v is updated as follows, where v c o n t e x t ( n ) :
v v = v v + α u { { n } N E G 1 ( n ) } L ( n , u , r , h , ξ ) X n   =   v v + α u { { n } N E G 1 ( n ) } [ L n ( u ) σ ( X n T θ u ) ] θ u
Thirdly, the gradient on the parameter vector θ ξ is calculated as follows:
L ( n , u , r , h , ξ ) θ ξ = β { θ ξ { δ n ( ξ ) l o g [ σ ( e h + r T θ ξ ) ] + [ 1 δ n ( ξ ) ] l o g [ 1 σ ( e h + r T θ ξ ) ] } } = β { δ n ( ξ ) [ 1 σ ( e h + r T θ ξ ) ] e h + r [ 1 δ n ( ξ ) ] σ ( e h + r T θ ξ ) e h + r } = β { { δ n ( ξ ) [ 1 σ ( e h + r T θ ξ ) ] [ 1 δ n ( ξ ) ] σ ( e h + r T θ ξ ) } e h + r } = β [ δ n ( ξ ) σ ( e h + r T θ ξ ) ] e h + r
Therefore, the parameter vector θ ξ is updated as follows:
θ ξ = θ ξ + α β [ δ n ( ξ ) σ ( e h + r T θ ξ ) ] e h + r
Finally, the gradient on the parameter vector e h + r is calculated as follows via the symmetry property between θ ξ and e h + r :
L ( n , u , r , h , ξ ) e h + r = β [ δ n ( ξ ) σ ( e h + r T θ ξ ) ] θ ξ
Specially, e h + r = e h + e r , and the gradient L ( n , u , r , h , ξ ) e h + r is used to update the parameter vectors e h and e r , respectively, as follows:
e h = e h + α β ξ { { n } N E G 2 ( n ) } [ δ n ( ξ ) σ ( e h + r T θ ξ ) ] θ ξ
e r = e r + α β ξ { { n } N E G 2 ( n ) } [ δ n ( ξ ) σ ( e h + r T θ ξ ) ] θ ξ
The stochastic gradient ascent approach is used for optimization. More details are shown in Algorithm 1.
Algorithm 1: HRHM
1 Input:
2    hypernetwork H = ( V , E )
3    vector dimension size d
4 Output:
5    node representation matrix Y R | V | × d
6 for  node n in V do
7    initializing the representation vector v n R 1 × d
8    initializing the parameter vector θ n R 1 × d
9    for  hyperedge r in R n do
10       for  node h in H r do
11         initializing the parameter vector e h + r R 1 × d
12        end for
13      end for
14    end for
15    node sequences C = r a n d o m w a l k ( )
16    for ( n , c o n t e x t ( n ) ) in C do
17    updating the parameter vector according to the Equation (11)
18    updating the representation vector according to the Equation (13)
19    updating the parameter vector according to the Equation (15)
20    for  hyperedge r in R n do
21      for  node h in H r do
22       updating the parameter vector according to the Equation (17)
23       updating the parameter vector according to the Equation (18)
24       end for
25     end for
26 end for
27 for   i = 0 ;   i < | V | ;   i + + do
28      Y i = v n
29 end for
30 return   Y

6. Experiments and Result Analysis

6.1. Datasets

Four hypernetwork datasets are used to evaluate HRHM approach. Detailed dataset statistics are shown in Table 1.
Four datasets are shown as follows:
  • GPS [17] describes a situation where a user takes part in an activity in a location. The triplet < user, location, activity> is utilized to construct the hypernetwork;
  • MovieLens [18] describes personal tag activities from MovieLens. The triplet < user, movie, tag> is utilized to construct the hypernetwork, where each movie has at least one genre;
  • Drug (http://www.fda.gov/Drugs/, accessed on 27 January 2020) describes a situation where the user who takes some drugs has certain reactions that lead to adverse events. The triplet < user, drug, reaction> is utilized to construct the hypernetwork;
  • Wordnet [15] is made up of a set of triplets <head, relation, tail> extracted from Wordnet3.0. The triplet < head, relation, tail> is utilized to construct the hypernetwork.

6.2. Baseline Approaches

DeepWalk: DeepWalk [19] is a popular approach for learning node representation vectors to encode social relationships.
Node2vec: Node2vec [20] maps the nodes in the network to a low-dimensional representation space to preserve network neighborhoods of the nodes.
LINE: LINE [21] embeds huge networks into low-dimensional vector spaces to preserve both local and global network structures.
GraRep: GraRep [22] integrates global graph structural information into the process of representation learning.
HOPE: HOPE [23] is a learning approach that preserves higher-order proximities of large scale graphs and captures the asymmetric transitivity.
SDNE: SDNE [24] is a semi-supervised learning approach with multiple layers of non-linear functions to capture the highly non-linear network structure.
HPGH: HPHG [11] proposes a deep model called Hyper-gram to capture pairwise and tuple relationships in the node sequences.
HRHM: HRHM regards the interaction relationships among the nodes as the translation operation in the representation space and incorporates the relationships among the nodes into node representation vectors.

6.3. Node Classification

Because the labels are only on the MovieLens and wordnet, our approach is assessed via node classification [1] on the two datasets. The node classification accuracies are calculated by means of SVM [25].
The observations from Table 2 and Table 3 are shown as follows:
  • With regard to the two datasets, the mean node classification accuracy of HRHM approach surpasses that of other baseline approaches, and in terms of the node classification accuracies with different training ratios, HRHM approach surpasses other baseline approaches. Furthermore, the node classification accuracies of HRHM approach is directly proportional to the training ratios, which shows that a large amount of training data is helpful for node classification. It is worth noting that the node classification accuracies of HRHM approach are not high. The reason is that the categorical attributes on the two datasets are less prominent;
  • The mean node classification accuracy of DeepWalk ranks only second to that of HRHM approach because DeepWalk captures the hypernetwork structure to a certain extent in the node sequences generated by random walks.
In short, it is found that the node representation vectors from the HRHM approach are better than other baseline approaches, which shows that HRHM approach is effective.

6.4. Link Prediction

The link prediction is assessed by the AUC [26]. The observations from Table 4, Table 5, Table 6 and Table 7 are shown as follows.
  • The HRHM approach performs worse than HPHG approach. To be specific, the mean AUC value of HPHG approach goes beyond that of HRHM approach by about 29%, 9%, 24%, and 4%, respectively, for the GPS, MovieLens, drug, and wordnet. The reason is that HRHM approach transforms the hypergraph into a two-section graph, which leads to partial loss of the hypernetwork structure information, but the HPHG approach does not decompose the hyperedges, which leads to almost complete preservation of hypernetwork structure information;
  • Except for HPHG approach, the mean AUC value of HRHM approach surpasses that of other baseline approaches on the GPS, drug, and wordnet. The mean AUC value of HRHM approach is very close to that of the other best baseline approach DeepWalk on the MovieLens. To sum up, HRHM approach surpasses most baseline approaches, which shows that the HRHM approach is effective;
  • The HRHM approach performs consistently at different training ratios compared to other baseline approaches, which shows its feasibility and robustness;
  • The HRHM approach almost surpasses other baseline approaches that do not consider the hyperedges, which verifies the assumption that the hyperedges are useful for link prediction.
In short, the above observations show that the HRHM approach can learn node representation vectors of high quality.

6.5. Parameter Sensitivity

The contribution rate between the cognitive topology model and the cognitive hyperedge model is equilibrated by the harmonic factor β . We set the raining ratio to 50% and calculate node classification accuracies with different β within the ranges from 0.1 to 0.9 on MovieLens and wordnet. Figure 4 shows the comparisons of node classification accuracies.
From Figure 4, it is found that the variation ranges of node classification accuracies on the two datasets are both within 2%, which indicates that the HRHM approach is not sensitive to the parameter β and shows the robustness of the HRHM approach.
In short, the best node classification results on both MovieLens and wordnet datasets are achieved at β = 0.5 .

7. Conclusions

This hypernetwork representation learning approach based on hyperedge modeling consists of the cognitive topology model and the cognitive hyperedge model, which incorporate the hypernetwork topology structure and the hyperedges into node representation vectors, respectively, where the learning process of node representation vectors is regarded as a joint optimization problem, which is resolved via the stochastic gradient ascend approach. The advantage of the HRHM approach is that the hyperedges with unfixed size are encoded in the learned node representation vectors. The experimental results show that the performance of the HRHM approach is almost all better than that of other baseline approaches, more or less, except for the HPHG approach. In future research, we will not transform the hypergraph into the ordinary graph but regard the hyperedges as a whole to carry out the research of the hypernetwork representation learning.

Author Contributions

Conceptualization, Y.Z. and H.Z.; methodology, Y.Z. and H.Z.; software, Y.Z.; validation, Y.Z.; formal analysis, Y.Z.; investigation, Y.Z. and H.Z.; resources, Y.Z.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z. and H.Z.; visualization, Y.Z.; supervision, Y.Z. and H.Z.; project administration, Y.Z. and H.Z.; funding acquisition, Y.Z., X.W. and J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by National Natural Science Foundation of China, grant number 62166032, grant number 62162053, and grant number 62062059; by Natural Science Foundation of Qinghai Province, grant number 2022-ZJ-961Q; by the Project from Tsinghua University, grant number SKL-IOW-2020TC2004-01; and by the Open Project of State Key Laboratory of Plateau Ecology and Agriculture, Qinghai University, grant number 2020-ZZ-03.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pedroche, F.; Tortosa, L.; Vicent, J.F. An eigenvector centrality for multiplex networks with data. Symmetry 2019, 11, 763. [Google Scholar] [CrossRef] [Green Version]
  2. Papageorgiou, I.; Bittner, D.; Psychogios, M.N.; Hadjidemetriou, S. Brain immunoinformatics: A symmetrical link between informatics, wet lab and the clinic. Symmetry 2021, 13, 2168. [Google Scholar] [CrossRef]
  3. Guerrero, M.; Banos, R.; Gil, C.; Montoya, F.G.; Alcayde, A. Evolutionary algorithms for community detection in continental-scale high-voltage transmission grids. Symmetry 2019, 11, 1472. [Google Scholar] [CrossRef]
  4. Zhou, D.Y.; Huang, J.Y.; Schölkopf, B. Learning with hypergraphs: Clustering, classification and embedding. In Proceedings of the 19th International Conference on Neural Information Processing Systems, Vancouver, Canada, 4–7 December 2006; pp. 1601–1608. [Google Scholar]
  5. Sharma, K.K.; Seal, A.; Herrera-Viedma, E.; Krejcar, O. An enhanced spectral clustering algorithm with s-distance. Symmetry 2021, 13, 596. [Google Scholar] [CrossRef]
  6. Feng, Y.F.; You, H.X.; Zhang, Z.Z.; Ji, R.R.; Gao, Y. Hypergraph neural networks. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 3558–3565. [Google Scholar]
  7. Baytas, I.M.; Xiao, C.; Wang, F.; Jain, A.K.; Zhou, J.Y. Heterogeneous hyper-network embedding. In Proceedings of the 18th IEEE International Conference on Data Mining, Singapore, 17–20 November 2018; pp. 875–880. [Google Scholar]
  8. Tu, K.; Cui, P.; Wang, X.; Wang, F.; Zhu, W.W. Structural deep embedding for hyper-networks. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 426–433. [Google Scholar]
  9. Yadati, N.; Nimishakavi, M.; Yadav, P.; Nitin, V.; Louis, A.; Talukdar, P. HyperGCN: A new approach of training graph convolutional networks on hypergraphs. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  10. Bandyopadhyay, S.; Das, K.; Murty, M.N. Line hypergraph convolution network: Applying graph convolution for hypergraphs. arXiv 2020, arXiv:2002.03392. [Google Scholar]
  11. Huang, J.; Liu, X.; Song, Y.Q. Hyper-path-based representation learning for hyper-networks. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 449–458. [Google Scholar]
  12. Zhang, R.C.; Zou, Y.S.; Ma, J. Hyper-SAGNN: A self-attention based graph neural network for hypergraphs. arXiv 2019, arXiv:1911.02613. [Google Scholar]
  13. Song, G.; Li, J.W.; Wang, Z. Occluded offline handwritten chinese character inpainting via generative adversarial network and self-attention mechanism. Neurocomputing 2020, 415, 146–156. [Google Scholar] [CrossRef]
  14. Zhu, Y.; Zhao, H.X. Hypernetwork representation learning with the set constraint. Appl. Sci. 2022, 12, 2650. [Google Scholar] [CrossRef]
  15. Bordes, A.; Usunier, N.; Garcia-Duran, A.; Weston, J.; Yakhnenko, O. Translating embeddings for modeling multi-relational data. In Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 2787–2795. [Google Scholar]
  16. Zhu, Y.; Ye, Z.L.; Zhao, H.X.; Zhang, K. Text-enhanced network representation learning. Front. Comput. Sci. 2020, 14, 146322. [Google Scholar] [CrossRef]
  17. Zheng, V.W.; Cao, B.; Zheng, Y.; Xie, X.; Yang, Q. Collaborative filtering meets mobile recommendation: A user-centered approach. In Proceedings of the 24th AAAI Conference on Artificial Intelligence, Atlanta, GA, USA, 11–15 July 2010; pp. 236–241. [Google Scholar]
  18. Harper, F.M.; Konstan, J.A. The movielens datasets: History and context. ACM Trans. Interact. Intell. Syst. 2015, 5, 19. [Google Scholar] [CrossRef]
  19. Perozzi, B.; Al-Rfou, R.; Skiena, S. DeepWalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD Internatonal Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  20. Grover, A.; Leskovec, J. Node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  21. Tang, J.; Qu, M.; Wang, M.Z.; Zhang, M.; Yan, J.; Mei, Q.Z. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  22. Cao, S.S.; Lu, W.; Xu, Q.K. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM International Conference on Information and Knowledge Management, Melbourne, Australia, 19–23 October 2015; pp. 891–900. [Google Scholar]
  23. Ou, M.D.; Cui, P.; Pei, J.; Zhang, Z.W.; Zhu, W.W. Asymmetric transitivity preserving graph embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1105–1114. [Google Scholar]
  24. Wang, D.X.; Cui, P.; Zhu, W.W. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1225–1234. [Google Scholar]
  25. Gholamnia, K.; Nachappa, T.G.; Ghorbanzadeh, O.; Blaschke, T. Comparisons of diverse machine learning approaches for wildfire susceptibility mapping. Symmetry 2020, 12, 604. [Google Scholar] [CrossRef] [Green Version]
  26. Almardeny, Y.; Boujnah, N.; Cleary, F. A novel outlier detection method for multivariate data. IEEE Trans. Knowl. Data Eng. 2020, 34, 4052–4062. [Google Scholar] [CrossRef]
Figure 1. Hypergraph and two-section graph. (a) Hypergraph; (b) two-section graph.
Figure 1. Hypergraph and two-section graph. (a) Hypergraph; (b) two-section graph.
Symmetry 14 02584 g001
Figure 2. TransE framework, where h + r t .
Figure 2. TransE framework, where h + r t .
Symmetry 14 02584 g002
Figure 3. HRHM framework, where v i is the center node, and other nodes v i s , v i s + 1 , v i + s 1 , v i + s , etc. are contextual nodes of the center node v i , namely c o n t e x t ( v i ) , where the vectors corresponding to v i s , v i s + 1 , v i , v i + s 1 and v i + s are denoted as the first red dot from the left, the second red dot from the left, the red dot in the middle, the second red dot from the right, the first red dot from the right. r is the interaction relation namely the hyperedge; h is a node with the relationship r with the center node v i ; and R v i is the hyperedge set associated with the center node v i , where the vectors corresponding to r and h are denoted as the first red dot from the right and the second red dot from the left. The left-hand rectangle denoted as all red is the projection layer representation derived from r and h ; The right-hand rectangle denoted as all red is the projection layer representation derived from v i s , v i s + 1 , v i + s 1 and v i + s .
Figure 3. HRHM framework, where v i is the center node, and other nodes v i s , v i s + 1 , v i + s 1 , v i + s , etc. are contextual nodes of the center node v i , namely c o n t e x t ( v i ) , where the vectors corresponding to v i s , v i s + 1 , v i , v i + s 1 and v i + s are denoted as the first red dot from the left, the second red dot from the left, the red dot in the middle, the second red dot from the right, the first red dot from the right. r is the interaction relation namely the hyperedge; h is a node with the relationship r with the center node v i ; and R v i is the hyperedge set associated with the center node v i , where the vectors corresponding to r and h are denoted as the first red dot from the right and the second red dot from the left. The left-hand rectangle denoted as all red is the projection layer representation derived from r and h ; The right-hand rectangle denoted as all red is the projection layer representation derived from v i s , v i s + 1 , v i + s 1 and v i + s .
Symmetry 14 02584 g003
Figure 4. Parameter sensitivity. (a) Sensitivity on MovieLens. (b) Sensitivity on wordnet.
Figure 4. Parameter sensitivity. (a) Sensitivity on MovieLens. (b) Sensitivity on wordnet.
Symmetry 14 02584 g004
Table 1. Dataset statistics.
Table 1. Dataset statistics.
DatasetNode Type#(V)#(E)
GPSuserlocationactivity1467051436
MovieLensusermovietag457168815305965
druguserdrugreaction41322211195
wordnetheadrelationtail1754715492174
Table 2. Node classification results on MovieLens (%).
Table 2. Node classification results on MovieLens (%).
ApproachesTraining Ratios
10%20%30%40%50%60%70%80%90%MeanRank
DeepWalk48.0150.3551.4152.6052.5953.4753.5754.2354.0952.262
node2vec46.9349.2850.7751.5152.6252.5853.0453.4452.7151.434
LINE43.9345.4646.5247.2947.7048.1648.0249.0948.3447.176
GraRep47.7550.1151.1652.0152.1053.1553.3453.4353.2451.813
HOPE46.3348.5749.9550.6951.0651.0451.2952.5351.7950.365
SDNE41.7441.7942.3442.7343.3643.2743.8943.4342.8142.827
HRHM48.7351.2652.7153.6254.3854.5755.0355.8256.3053.601
Table 3. Node classification results on wordnet (%).
Table 3. Node classification results on wordnet (%).
ApproachesTraining Ratios
10%20%30%40%50%60%70%80%90%MeanRank
DeepWalk29.9133.4434.5335.0535.7036.8037.9336.7139.0035.452
node2vec29.2732.2333.7134.5236.1736.0537.5337.6637.3034.944
LINE22.7724.1125.1124.9425.2325.5925.8726.6025.4425.076
GraRep32.5934.7434.6335.2135.3836.0535.1036.6337.7935.353
HOPE30.5333.6135.0235.9734.9035.1136.2136.2034.8434.715
SDNE21.9621.5722.0522.3723.2622.5923.6323.6025.3122.937
HRHM31.3033.7935.5936.1836.9537.9438.1438.6340.0736.511
Table 4. Link prediction results on GPS.
Table 4. Link prediction results on GPS.
ApproachesTraining Ratios
60%65%70%75%80%85%90%MeanRank
DeepWalk0.43080.42780.42050.45830.44180.49140.48310.45054
node2vec0.36600.36140.38080.39390.38340.39580.36490.37807
LINE0.45750.48290.47610.45620.44290.46630.45740.46283
GraRep0.38730.38050.38820.37650.38200.38570.38740.38396
HOPE0.38050.36760.34160.29710.27940.25180.23340.30738
SDNE0.32620.43710.43190.31570.43790.35270.45400.39365
HPHG0.90260.91580.91420.92690.93470.93260.93150.92261
HRHM0.68450.64280.64830.64030.62160.60050.58560.63192
Table 5. Link prediction results on MovieLens.
Table 5. Link prediction results on MovieLens.
ApproachesTraining Ratios
60%65%70%75%80%85%90%MeanRank
DeepWalk0.78450.81290.83010.84400.87290.88000.90250.84672
node2vec0.70780.73900.74180.76960.79390.80360.82960.76936
LINE0.82820.82420.82530.83200.83650.81720.82310.82664
GraRep0.72900.78330.79070.81210.82770.84810.85440.80655
HOPE0.68950.73330.72030.75220.77870.79860.80490.75397
SDNE0.40040.35110.34940.34060.34330.35980.41710.36608
HPHG0.93560.93670.93880.92760.91380.92450.93010.92961
HRHM0.84950.84970.83510.83250.83870.82810.83200.83793
Table 6. Link prediction results on drug.
Table 6. Link prediction results on drug.
ApproachesTraining Ratios
60%65%70%75%80%85%90%MeanRank
DeepWalk0.48520.49540.49340.45800.49010.46380.47130.47965
node2vec0.45000.45250.44900.45250.43290.47120.43450.44897
LINE0.47500.46720.46360.46250.47410.45230.47680.46746
GraRep0.50250.50890.48670.50510.55570.58350.53620.52553
HOPE0.50550.52690.49330.46900.49410.46680.42710.48324
SDNE0.29480.43100.44540.50500.51960.35360.38360.41908
HPHG0.94510.94580.94670.95520.95830.95480.94890.95071
HRHM0.71530.70710.71340.71340.68680.71080.72400.71012
Table 7. Link prediction results on wordnet.
Table 7. Link prediction results on wordnet.
ApproachesTraining Ratios
60%65%70%75%80%85%90%MeanRank
DeepWalk0.77800.81810.83050.83410.87080.87650.88800.84233
node2vec0.78070.82420.83090.82850.85190.85030.85950.83234
LINE0.80630.81840.80560.80910.80000.79380.79260.80375
GraRep0.76850.77420.78880.78060.79580.79720.77560.78306
HOPE0.69020.73140.74170.74030.76490.77630.77000.74507
SDNE0.37120.53480.47840.48240.42540.61590.48500.48478
HPHG0.92170.93780.93860.94880.95560.94950.95010.94321
HRHM0.90300.91150.90500.90160.90980.90270.89120.90352
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, Y.; Zhao, H.; Wang, X.; Huang, J. Hypernetwork Representation Learning Based on Hyperedge Modeling. Symmetry 2022, 14, 2584. https://doi.org/10.3390/sym14122584

AMA Style

Zhu Y, Zhao H, Wang X, Huang J. Hypernetwork Representation Learning Based on Hyperedge Modeling. Symmetry. 2022; 14(12):2584. https://doi.org/10.3390/sym14122584

Chicago/Turabian Style

Zhu, Yu, Haixing Zhao, Xiaoying Wang, and Jianqiang Huang. 2022. "Hypernetwork Representation Learning Based on Hyperedge Modeling" Symmetry 14, no. 12: 2584. https://doi.org/10.3390/sym14122584

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop