Next Article in Journal
DAG: Dual Attention Graph Representation Learning for Node Classification
Previous Article in Journal
Cervical Precancerous Lesion Image Enhancement Based on Retinex and Histogram Equalization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Leveraging Minimum Nodes for Optimum Key Player Identification in Complex Networks: A Deep Reinforcement Learning Strategy with Structured Reward Shaping

1
School of International Business and Management, Sichuan International Studies University, Chongqing 400031, China
2
College of Systems Engineering, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(17), 3690; https://doi.org/10.3390/math11173690
Submission received: 20 July 2023 / Revised: 15 August 2023 / Accepted: 24 August 2023 / Published: 28 August 2023
(This article belongs to the Topic Complex Systems and Network Science)

Abstract

:
The problem of finding key players in a graph, also known as network dismantling, or network disintegration, aims to find an optimal removal sequence of nodes (edges, substructures) through a certain algorithm, ultimately causing functional indicators such as the largest connected component (GCC) or network pair connectivity in the graph to rapidly decline. As a typical NP-hard problem on graphs, recent methods based on reinforcement learning and graph representation learning have effectively solved such problems. However, existing reinforcement-learning-based key-player-identification algorithms often need to remove too many nodes in order to achieve the optimal effect when removing the remaining network until no connected edges remain. The use of a minimum number of nodes while maintaining or surpassing the performance of existing methods is a worthwhile research problem. To this end, a novel algorithm called MiniKey was proposed to tackle such challenges, which employs a specific deep Q-network architecture for reinforcement learning, a novel reward-shaping mechanism based on network functional indicators, and the graph-embedding technique GraphSage to transform network nodes into latent representations. Additionally, a technique dubbed ‘virtual node technology’ is integrated to grasp the overarching feature representation of the whole network. This innovative algorithm can be effectively trained on small-scale simulated graphs while also being scalable to large-scale real-world networks. Importantly, experiments from both six simulated datasets and six real-world datasets demonstrates that MiniKey can achieve optimal performance, striking a perfect balance between the effectiveness of key node identification and the minimization of the number of nodes that is utilized, which holds potential for real-world applications such as curbing misinformation spread in social networks, optimizing traffic in transportation systems, and identifying key targets in biological networks for targeted interventions.

1. Introduction

Complex networks hold substantial significance given their extensive reach and impact on diverse aspects of our lives. At the heart of complex networks lie the key players, also known as influential players [1], vital players [2], or critical players [3]. These players represent certain nodes, edges, or substructures that, when removed, can substantially degrade a network’s specific functionality [4]. The importance of identifying these key players has profound implications in a variety of domains, such as epidemic control [5], drug design [6], viral marketing [7], criminal networks analysis [8] and combat network disintegration [9].
In the quest to identify key players in complex networks, many techniques and methods have been developed over the years, each with its own strengths and weaknesses. At one end of the spectrum, mathematical-programming-based methods represent one of the earliest and most fundamental approaches to this problem [10,11]. They tackle the problem by formulating it as an optimization challenge, using techniques such as integer programming, or mixed-integer programming to seek optimal solutions. However, as network size increases, these methods can quickly become computationally intensive, restricting their practical applications to larger-scale networks.
Another well-studied approach focuses on using network centrality measures such as degree centrality [12], k-core centrality [13], betweenness centrality [14], pagerank centrality [15], and collective influence (CI) [16]. These techniques identify key players based on their importance within the network, as determined by their centrality measure. Typically, before removing the next node, these algorithms adaptively recalculate the network’s centrality indicators. The method that removes the node with the highest degree is termed HDA. Similarly, methods that remove nodes with the highest betweenness, pagerank, and collective influence are referred to as HBA, HPRA, and HCI, respectively. This process is repeated iteratively until there are no connected edges left in the network. However, while intuitively appealing and computationally efficient, these approaches can sometimes oversimplify the complex dynamics of the network. Additionally, they typically overlook node attributes, which can offer insights into a node’s role or importance, leading to suboptimal results as they fail to account for the intricate interconnections among nodes.
Moving beyond these techniques, a wide range of heuristic-based methods have been developed, including greedy algorithms [17], acquaintance immunization [18], belief propagation dismantling [19,20], the reverse-reinsertion algorithm [21,22], and spectral partitioning [23]. These methods are designed around heuristics to select nodes for removal, with the aim of maximizing the impact on network functionality. Despite their innovative approaches, the performance of these methods can be heavily reliant on the specific heuristic that is used, and they may not always yield optimal results.
Evolutionary-algorithm-based strategies, on the other hand, have been inspired by processes found in nature. Techniques such as Tabu Search [24], genetic algorithms [25], simulated annealing [26], and artificial bee colonies [27] use these naturally inspired processes to guide the search for optimal node removal sequences. Despite their ability to handle complex problems and search a larger solution space, they are typically computationally intensive, and their probabilistic nature can impact the quality of the solutions they provide.
Continuing the exploration of these methodologies, the landscape of key player identification has been recently reshaped with the emergence of deep reinforcement learning (DRL) [28,29]. Different from traditional approaches, DRL learns from interactions within the environment and makes decisions based on the observed states, offering a more dynamic way to understand and manipulate complex networks. A representative work is the network key node identification algorithm framework called FINDER, based on deep reinforcement learning [30]. FINDER can be trained on small graphs and extended to millions of nodes, and it has a good identification performance. Since the FINDER algorithm operates on a single-objective deep reinforcement learning framework, it usually requires a large number of nodes to be removed to achieve optimal disintegration effects, even when there are no remaining edges in the network. How to achieve the best results with as few nodes as possible is an important issue that is worth studying.
In light of these issues, we introduce MiniKey (Leveraging Minimum Nodes for Optimum Key Players Identification in Complex Networks), a novel algorithm that integrates DRL with reward-shaping. MiniKey extends the original FINDER framework into a multi-objective optimization field, which can navigate the balance between functionality optimization and minimal node usage. By harnessing the power of DRL and reward shaping, it ensures a more efficient and robust learning process, resulting in the identification of fewer key players while still maintaining the network’s optimal functionality.
To provide a vivid illustration of the effectiveness of MiniKey, we showcase its application in the domain of crime networks in Figure 1. This figure depicts the efficient performance of MiniKey, clearly demonstrating how it outperforms other methods, maintaining network connectivity while minimizing the number of utilized nodes. In a direct comparison using the 9/11 terrorist network as an example, MiniKey outperforms the original FINDER model, preserving network connectivity by using fewer nodes, which represents a significant step forward in this research area. This demonstration of MiniKey’s superior performance emphasizes its potential in solving real-world problems, particularly in maintaining optimal network functionality while minimizing node usage.

2. Preliminaries

2.1. Accumulated Normalize Connectivity (ANC)

Given a graph G ( V , E ) with vertices set V , edge set E and a connectivity metric σ , where each edge E is a pair of distinct vertices ( u , v ) V × V , the learning target in key players’ identification is to find a optimal node removal sequence ( v 1 , v 2 , v n ) which minimizes the following accumulated normalize connectivity (ANC):
R ( v 1 , v 2 , v n ) = 1 N k = 1 n σ ( G \ { v 1 , v 2 , v n } ) σ ( G )
In this paper, connectivity metric σ represents the size of the largest connected component remaining in the network, in which case the problem is referred to as the network dismantling problem (NC) [31]. This problem involves the identification of a sequence of node removals that minimizes the size of the largest connected component remaining in the network. By addressing this issue, we aim to offer new insights into the process of network disintegration and propose more efficient strategies for identifying key players in complex networks.
It is worth noting that the connectivity metric σ can also represent various other network characteristics. For instance, it could represent the sum of the connectivity between nodes, referred to as the critical node problem (CN) [32]. It could also indicate the average geodesic distance [33] in the network among other possible measures.

2.2. Solution Length Ratio (SLR)

Given a graph G ( V , E ) , we define the solution length ratio (SLR) under a certain strategy as the ratio of the total number of nodes needed to disintegrate the network until no edges remain to the total number of nodes in the original network. This can be mathematically expressed as follows:
SLR = | V S | | V |
Here, V S is the number of nodes required to dismantle the network until no edges remain in the network; | V | is the number of nodes in the original network G ( V , E ) . The importance of the SLR lies in its ability to measure the resource efficiency of a network-disintegration strategy. In practical scenarios, it is often preferable to achieve the network disintegration with as few node removals as possible due to resource constraints. Hence, a strategy with a lower SLR would be more desirable as it indicates less resource consumption in terms of node removals. When optimizing based solely on this metric, the problem can be transformed into the minimum vertex cover problem (MVC) [34].

2.3. Pareto Frontier

In the field of multi-objective optimization, the pareto frontier is characterized as the assortment of solutions (or, correspondingly, points within the objective function space) for which no other viable solution can be found that would decrease any given criterion without concurrently instigating an increase in at least one additional criterion.
Formally, for a problem with objectives to minimize, a solution X is said to dominate another solution Y if, and only if:
(1)
For all objectives i , 1 i k , the score of X on i is less than or equal to the score of Y on i ;
(2)
There exists at least one objective j , 1 j k , such that the score of X on j is strictly less than the score of Y on j .

3. Model of MiniKey

Here, we employ a typical encoder–decoder architecture to model the key node identification problem on graphs. Simultaneously, we utilize a reinforcement learning algorithm with structured reward-shaping to train the entire MiniKey model. The design of each part of the algorithm is as follows.

3.1. Encoding Process of MiniKey

The identification of key players in complex networks heavily depends on the feature representation of elements within the network. Therefore, an effective feature-learning model can enhance the performance of the algorithm. In this work, we employ GraphSage [35] to encode the nodes within the network into a latent representation. Furthermore, we utilize a technique known as virtual node technology to capture the overall feature representation of the entire network. The detailed algorithmic framework of our graph encoding process is outlined as follows.
Algorithm 1 is the encoding process of MiniKey; given a graph G ( V , E ) , node features X v , depth K , and learnable weight parameters W 1 R c × p , W 2 R p × ( p / 2 ) , W 3 R p × ( p / 2 ) , the purpose of this algorithm is to produce node embeddings z v and graph embedding z s .
Algorithm 1: Encoding process of MiniKey
Input: Graph G ( V , E ) , node features { X v R 1 × c , v V } , depth K , learnable weight parameters W 1 R c × p , W 2 R p × ( p / 2 ) , W 3 R p × ( p / 2 )
Output: Node embedding z v , v V { s }
1. Add a virtual node s , which connects all nodes in G ( V , E )
2. Initialize h v ( 0 ) ReLU ( X v W 1 ) , h v ( 0 ) h v ( 0 ) / h v ( 0 ) 2 , v V { s }
3. FOR l = 1 to K
4.    FOR v V { s }
5.       h N ( v ) ( l 1 ) j N ( v ) h j ( l 1 )
6.       h v ( l ) ReLU ( [ W 2 h v ( l 1 ) , W 3 h N ( v ) ( l 1 ) ] )
7.    END FOR
8.     h v ( l ) h v ( l ) / h v ( l ) 2 , v V { s }
9. END FOR
10.  z v h v ( K ) , v V { s }
The algorithm starts by adding a virtual node s that connects to all nodes in the graph. This node is referred to as the graph state. The embeddings h v ( 0 ) for the nodes and the virtual node are then initialized. The function Rectified Linear Unit (ReLU) is applied to the dot product of the node features X v and the weight parameters W 1 , and the result is then normalized using the L2 norm. The algorithm then enters a loop that runs K times. This loop corresponds to the K layers of the graph neural network. In each iteration, it calculates the new embeddings for each node and the virtual node. Inside this loop, there is a nested loop that runs for every node v in the graph and the virtual node. For each node v , the embedding h N ( v ) ( l 1 ) is calculated by summing the embeddings h j ( l 1 ) of all nodes j in its neighborhood N ( v ) . Then, the new embedding h v ( l ) for each node v is calculated by applying the ReLU function to the concatenation of the dot product of W 2 , and the previous embedding h v ( l 1 ) of the node and the dot product of W 3 , and the newly calculated neighbor nodes embedding h N ( v ) ( l 1 ) . After all nodes have been processed, the new embeddings h v ( l ) are normalized using the L2 norm. Finally, after K layers, the final embeddings z v are calculated for each node and the virtual node. These are simply the final h v ( K ) values. These embeddings can then be used for subsequent tasks, such as the identification of key players in the network.

3.2. Decoding Process of MiniKey

We use a two-layered MLPs to decode a state–action pair ( s , a ) to a scalar value Q ( s , a ) that predicts the maximal rewards after taking action a in a given state s , which is defined as shown below:
Q ( s , a ) = W 5 ReLU   ( z a z s W 4 )
Here, W 4 R p × 1 , W 5 R p × 1 , are learnable weight parameters, z s and z a R 1 × p are the state-embedding (Graph) and action-embedding, respectively, which are produced by Algorithm 1.

3.3. Training Algorithm of MiniKey

To leverage a minimum number nodes to find the optimum key players sets in complex networks, we formulate this problem as a Markov Decision Process (MDP) in a graph. The reinforcement learning components of MiniKey are outlined below:
States: The state, denoted as S in MiniKey, is the current configuration of the network graph G . It represents the partial solution to the key player identification problem. Each state is represented as a vector in a p -dimensional space, which is defined in Algorithm 1 as z s .
Transitions: The transition function in MiniKey is deterministic and corresponds to the selection of a node v G that is not yet part of the state S .
Actions: An action in MiniKey corresponds to the selection of a node v that is not currently part of the state S . The node is also represented by its p -dimensional embedding as z v in Algorithm 1.
Rewards: The reward r ( S , v ) in MiniKey can be be defined as ANC, i.e., the increase in ANC after selecting node v k as the action and transitioning to the new state S = ( S , v k ) .Therefore, it can be expressed as:
r ( v k ) = 1 N σ ( G \ { v 1 , v 2 , v k } ) σ ( G )
The cumulative reward of a terminal state S ^ aligns with the cost function value of S ^ :
R ( S ^ ) = i = 1 | S ^ | r ( S i , v i ) = R ( v 1 , v 2 , v n ) = 1 N k = 1 n σ ( G \ { v 1 , v 2 , v n } ) σ ( G )
In addition to considering the original ANC as the objective of reinforcement learning, we also incorporate an additional structural penalty as a constraint for the agent. The learning objective of MiniKey is to simultaneously optimize ANC and SLR, that is, to identify the optimal key nodes with as few nodes as possible, which could lead to multi-objective reinforcement learning [36,37].
Inspired by the network dismantling algorithm CoreHD [38], we observed that, for the network dismantling problem, star-shaped networks are more optimal for disintegration. In such cases, removing only the central node can disconnect many edges. Conversely, if there are too many nodes with degrees 1 or 2 in the largest connected component, a large number of nodes will need to be consumed in order to ensure that there are no remaining edges in the network. Nodes with a degree of 1 in a network are usually leaf nodes, and nodes with a degree of 2 are typically located in chains and cycles; the more of these two types of nodes there are, the more nodes need to be removed to ensure no edges exist in the network. In the extreme case of a chain of length N, to ensure that there are no edges, it would be necessary to remove N-1 nodes.
Based on these observations, we modified our reward function to include a penalty term, specifically, the number of nodes in the largest connected component of the remaining network with degrees 1 or 2. This design intends to minimize the presence of nodes with degrees 1 or 2 in the largest connected component while optimizing network dismantling. Experiments demonstrate that this crucial reward function design enables our MiniKey framework to identify the key nodes in the network with fewer nodes.
R t o t ( v 1 , v 2 , v n ) = R ( v 1 , v 2 , v n ) + R p e n a l t y ( v 1 , v 2 , v n ) = 1 N k = 1 n ( σ ( G \ { v 1 , v 2 , v k } ) σ ( G ) ) + 1 N k = 1 n ( N 1 , 2 L C C ( G \ { v 1 , v 2 , , v k } ) σ ( G ) ) = 1 N k = 1 n ( σ ( G \ { v 1 , v 2 , v k } ) σ ( G ) + N 1 , 2 L C C ( G \ { v 1 , v 2 , , v k } ) σ ( G ) )
As illustrated in Equation (6), R t o t ( v 1 , v 2 , v n ) is the total reward of MiniKey, R ( v 1 , v 2 , v n ) is the original reward of NC, and R p e n a l t y ( v 1 , v 2 , v n ) is the penalty term, where N 1 , 2 L C C ( G \ { v 1 , v 2 , , v k } ) is the number of nodes with degrees of 1 or 2 in the largest connected component (LCC) of the left graph.
It is worth noting that the learning objective of MiniKey is to minimize the reward function R t o t ( v 1 , v 2 , v n ) , and the original reward function R ( v 1 , v 2 , v n ) will inevitably monotonically decrease as the network scale decreases. At this time, the optimal effect of the reward function R t o t ( v 1 , v 2 , v n ) after reward-shaping can only be achieved when the structural penalty reward R p e n a l t y ( v 1 , v 2 , v n ) also decreases. That is, while reducing the number of nodes with a degree of 1 or 2 in the largest connected component, the size of the largest connected component in the remaining network decreases the fastest.
Policy: The policy in MiniKey is based on the approximated Q-function, Q ^ . A deterministic greedy policy π ( v | S ) = argmax v S ¯ Q ^ ( z s , v ) is applied. When action v is taken, a node from G is added to the current partial solution, leading to a reward r ( S , v ) = Q ( s , a ) which is defined in Equation (3).
Based on the aforementioned modeling, we use the DQN [39] algorithm to train MiniKey using simulated Barabási–Albert (BA) graphs as training samples. DQN is a variant of reinforcement learning where Q-Learning is combined with deep neural networks. This objective can be mathematically defined by the following cost function:
J ( θ ) = E ( s , a , r , s ) ~ D [ ( r + γ max a Q ( s , a ; θ ) Q ( s , a ; θ ) ) 2 ]
In Equation (7), s represents the current state, a is the action taken, r is the reward received, and s is the new state after taking action a . D is the experience replay memory, Q ( s , a ; θ ) is the Q-value function approximated by the network with parameters θ , and γ is the discount factor.

4. Results

4.1. Results on Synthetic Graphs

During the training phase, we employed the Barabási Albert model (BA) with a default parameter setting: m = 4 (the number of edges attached from a new node to existing ones), and node number is uniformly chosen from the range [30, 50] (indicating that the node count varies between 30 and 50). All experiments were conducted on a platform equipped from Huawei Cloud with a Nvidia GeForce Tesla V100-32GB GPU.
In the testing phase, synthetic graphs were generated, maintaining m = 4, but with varying node numbers divided into scales: 30–50, 50–100, 100–200, 200–300, 300–400, and 400–500. For each node size category, we created 100 test graphs. Subsequently, we gauged the performance of several algorithms, namely, HDA [12], HBA [14], HCA [40], HPRA [15], FINDER [30], and MiniKey, on these generated datasets. This methodology ensured a comprehensive evaluation, permitting a detailed comparison of the methods across different graph sizes.
Figure 2 provides a comprehensive overview of the average performance (ANC and SLR) of various methods on these synthetic datasets. Importantly, we highlighted the error bars in the figure to depict the standard deviations of each method across the 100 test graphs. From Figure 2a, which shows the ANC results, we can observe that the MiniKey method generally outperforms the other techniques across all graph sizes. It achieves the lowest ANC in all categories except for the 30–50 and 50–100 range, where it is narrowly beaten by HBA. This indicates that MiniKey is more effective at identifying key nodes within the networks, as a lower ANC indicates a more effective disintegration of the largest connected component. The performance of MiniKey is especially noticeable in larger graph sizes (200–500), where it consistently outperforms the other methods. Figure 2b presents the SLR results across different graph sizes. The SLR measures the ratio of the nodes used by an algorithm to break the network completely. Here, MiniKey demonstrates an impressive performance again, consistently using fewer nodes compared to the majority of other methods across all graph sizes. This performance indicates a higher efficiency in utilizing network nodes for MiniKey. However, it is noteworthy that HDA method also showcases a competitive performance, with SLR closely following MiniKey’s.
Moveover, according to Figure 3, which illustrates the Pareto Front comparison of different methods on six simulated datasets, it is evident that the competitive performance of MiniKey in both identifying critical nodes within the network and efficiently using nodes to break down the network.

4.2. Results on Real-World Networks

During the testing phase on real-world graphs, we used the best model, trained on simulated BA graphs (with node range 30–50 and m = 4), as the default parameters for MiniKey. We selected six real datasets from SNAP Datasets [41]: Crime, HI-II-14, Digg, Enron, Gnutella31, and Facebook. These datasets cover a wide range of fields, including criminal networks, biological networks, social networks, and communication networks. Table 1 presents the details of the network structures of these datasets, including node number, edge number, maximum degree, average degree, diameter, clustering coefficient and assortativity.
Furthermore, we compared MiniKey with network critical node identification algorithms that can run on large-scale networks, such as CI [16], MinSum [31], CoreHD [38], GND [23], and FINDER [30]. Table 2 and Figure 4, respectively, present the ANC score and the ANC curves of different methods on six real-world datasets. From the results, we can observe that FINDER and MiniKey consistently outperform the other methods across all datasets. Specifically, on Crime, HI-II-14, and Digg datasets, FINDER and MiniKey excel by providing the most accurate identification of key nodes. For the Gnutella31 and Facebook datasets, although BPD and MiniKey perform slightly better, FINDER still delivers results that are quite close to the top performers, indicating its effectiveness in key node identification.
Table 3 provides the solution length ratio (SLR), which measures the number of nodes each method uses to successfully decompose the network, a critical factor when evaluating efficiency. MiniKey exhibits an excellent performance in this respect, requiring the fewest nodes across all datasets to accomplish this task. It is particularly remarkable in the Crime and HI-II-14 datasets, significantly outperforming all other methods. FINDER also shows impressive results, especially on the Digg and Enron datasets, demonstrating high efficiency by maintaining a low SLR.
Additionally, as depicted in Figure 5, showing the pareto frontiers comparison of various methods across six real-world datasets, MiniKey evidently has a superior performance. In summary, both FINDER and MiniKey display an exceptional performance in critical node identification and the efficiency of network decomposition. MiniKey stands out due to its efficiency in maintaining a low SLR across all test datasets. This efficiency results in resource savings and an enhanced performance, marking it as a highly effective solution for network-dismantling problems.

5. Discussion

The MiniKey algorithm proposed in this paper highlights a significant step forward in identifying key players in complex networks, emphasizing the use of minimal nodes. Evaluations across different datasets showcase its strength and flexibility. However, upon closer examination, clear challenges and details emerge, which are discussed as follows.
Scalability: while MiniKey exhibits notable proficiency with both small simulated graphs and extensive real-world datasets, care should be taken when navigating exceptionally large-scale networks. Networks characterized by complex structures or abundant node interactions might introduce challenges, potentially impacting the algorithm’s reliability.
Dependency on Graph Representation: central to MiniKey’s efficacy is its reliance on the graph representation technique, GraphSage. Nevertheless, as the landscape of graph neural networks (GNNs) continues to evolve, considering newer architectures might be beneficial. For instance, Graph Attention Network (GAT) [42], Message Passing Neural Networks (MPNN) [43], and Geometric Graph Convolutional Networks (Geom-GCN) [44] and et al. Incorporating or adapting components from these latest developments could potentially offer a more comprehensive embedding for Minikey, enhancing its capacity to tackle more intricate and diverse graph structures.
Structured Reward Shaping: our approach to reward-shaping is certainly innovative. However, it strongly relies on preset network functional markers. The absence of comprehensive markers or overlooking of pivotal network dynamics might steer the reinforcement learning agent away from the most fruitful actions.
Adaptability to Dynamic Networks: at present, MiniKey is fine-tuned to cater to static graphs. The ever-evolving landscape of dynamic networks, where node relationships fluctuate over time, presents a distinct challenge. Addressing this issue would require major changes to our current approach.
In conclusion, MiniKey presents a promising approach to optimum key players’ identification with minimum nodes in complex networks; however, a comprehensive understanding and careful consideration of its limitations and intricacies are pivotal to harnessing its full potential and paving the way for future refinements in complex network analyses.

6. Conclusions

This paper presents MiniKey, an innovative approach to leveraging minimum nodes for optimum key players’ identification in complex networks. Experiments conducted on a range of simulated and real-world datasets attest to the superior performance of MiniKey compared to other leading methods in terms of the Accumulated Normalized Connectivity (ANC) score and Solution Length Ratio (SLR). MiniKey outperforms existing strategies in its ability to identify and eliminate network edges, thereby proving its practical utility in various fields, including crime networks, social networks, communication networks, and bio-networks. Notably, MiniKey’s unique strength lies in its efficiency, as it consistently uses fewer nodes to break network connectivity, providing resource savings and an enhanced performance. MiniKey reshapes our understanding of how to maintain or break network connectivity with minimal movements. Such insights can redefine current network analysis techniques, benefiting stakeholders in domains such as social media, urban planning, or epidemiology. Despite these promising results, the potential for future improvements and applications of MiniKey remains vast. Future research might focus on extending the approach to handling larger, more complex network structures or integrating it with other network analysis tools for more comprehensive network solutions. The adaptability and efficacy of MiniKey make it an exciting and promising frontier for network analysis.

Author Contributions

Conceptualization, methodology, formal analysis, investigation, resources, data curation, writing, L.Z. and C.F.; original draft preparation and writing, review and editing, L.Z., C.F. and C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the School-Level Scientific Research Project of Sichuan International Studies University, with the project number sisu202216.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Morone, F.; Makse, H.A. Influence Maximization in Complex Networks through Optimal Percolation. Nature 2015, 524, 65–68. [Google Scholar] [CrossRef]
  2. Lü, L.; Chen, D.; Ren, X.-L.; Zhang, Q.-M.; Zhang, Y.-C.; Zhou, T. Vital Nodes Identification in Complex Networks. Phys. Rep. 2016, 650, 1–63. [Google Scholar] [CrossRef]
  3. Lalou, M.; Tahraoui, M.A.; Kheddouci, H. The Critical Node Detection Problem in Networks: A Survey. Comput. Sci. Rev. 2018, 28, 92–117. [Google Scholar] [CrossRef]
  4. Borgatti, S.P. Identifying Sets of Key Players in a Social Network. Comput. Math. Organ. Theory 2006, 12, 21–34. [Google Scholar] [CrossRef]
  5. Pastor-Satorras, R.; Vespignanj, A. Epidemic Spreading in Scale-Free Networks. In The Structure and Dynamics of Networks; Princeton University Press: Princeton, NJ, USA, 2011; pp. 493–496. [Google Scholar]
  6. Kuntz, I.D. Structure-Based Strategies for Drug Design and Discovery. Science 1992, 257, 1078–1082. [Google Scholar] [CrossRef] [PubMed]
  7. Leskovec, J.; Adamic, L.A.; Huberman, B.A. The Dynamics of Viral Marketing. In Proceedings of the 7th ACM Conference on Electronic Commerce, Ann Arbor, MI, USA, 11–15 June 2006; ACM: New York, NY, USA, 2006. [Google Scholar]
  8. Bright, D.; Greenhill, C.; Britz, T.; Ritter, A.; Morselli, C. Criminal Network Vulnerabilities and Adaptations. Glob. Crime 2017, 18, 424–441. [Google Scholar] [CrossRef]
  9. Chen, L.; Wang, C.; Zeng, C.; Wang, L.; Liu, H.; Chen, J. A Novel Method of Heterogeneous Combat Network Disintegration Based on Deep Reinforcement Learning. Front. Phys. 2022, 10, 1021245. [Google Scholar] [CrossRef]
  10. Walteros, J.L.; Veremyev, A.; Pardalos, P.M.; Pasiliao, E.L. Detecting Critical Node Structures on Graphs: A Mathematical Programming Approach. Networks 2018, 73, 48–88. [Google Scholar] [CrossRef]
  11. Ventresca, M.; Aleman, D. A Derandomized Approximation Algorithm for the Critical Node Detection Problem. Comput. Oper. Res. 2014, 43, 261–270. [Google Scholar] [CrossRef]
  12. Hooshmand, F.; Mirarabrazi, F.; MirHassani, S.A. Efficient Benders Decomposition for Distance-Based Critical Node Detection Problem. Omega 2020, 93, 102037. [Google Scholar] [CrossRef]
  13. Albert, R.; Jeong, H.; Barabási, A.-L. Error and Attack Tolerance of Complex Networks. Nature 2000, 406, 378–382. [Google Scholar] [CrossRef]
  14. Carmi, S.; Havlin, S.; Kirkpatrick, S.; Shavitt, Y.; Shir, E. A Model of Internet Topology Using K-Shell Decomposition. Proc. Natl. Acad. Sci. USA 2007, 104, 11150–11154. [Google Scholar] [CrossRef]
  15. Wandelt, S.; Sun, X.; Feng, D.; Zanin, M.; Havlin, S. A Comparative Analysis of Approaches to Network-Dismantling. Sci. Rep. 2018, 8, 13513. [Google Scholar] [CrossRef]
  16. Qin, J.; Xu, J.J.; Hu, D.; Sageman, M.; Chen, H. Analyzing Terrorist Networks: A Case Study of the Global Salafi Jihad Network. In Intelligence and Security Informatics; Springer: Berlin/Heidelberg, Germany, 2005; pp. 287–304. [Google Scholar]
  17. Aringhieri, R.; Grosso, A.; Hosteins, P.; Scatamacchia, R. VNS Solutions for the Critical Node Problem. Electron. Notes Discret. Math. 2015, 47, 37–44. [Google Scholar] [CrossRef]
  18. Mahdavi Pajouh, F.; Boginski, V.; Pasiliao, E.L. Minimum Vertex Blocker Clique Problem. Networks 2014, 64, 48–64. [Google Scholar] [CrossRef]
  19. Zhou, H.-J. Spin Glass Approach to the Feedback Vertex Set Problem. Eur. Phys. J. B 2013, 86, 455. [Google Scholar] [CrossRef]
  20. Qin, S.-M.; Ren, X.-L.; Lü, L.-Y. Efficient Network Dismantling via Node Explosive Percolation. Commun. Theor. Phys. 2019, 71, 764. [Google Scholar] [CrossRef]
  21. Šimon, M.; Dirgová Luptáková, I.; Huraj, L.; Hosťovecký, M.; Pospíchal, J. Combined Heuristic Attack Strategy on Complex Networks. Math. Probl. Eng. 2017, 2017, 6108563. [Google Scholar] [CrossRef]
  22. Fan, C.; Zeng, L.; Feng, Y.; Xiu, B.; Huang, J.; Liu, Z. Revisiting the Power of Reinsertion for Optimal Targets of Network Attack. J. Cloud Comput. 2020, 9, 24. [Google Scholar] [CrossRef]
  23. Ren, X.-L.; Gleinig, N.; Helbing, D.; Antulov-Fantulin, N. Generalized Network Dismantling. Proc. Natl. Acad. Sci. USA 2019, 116, 6554–6559. [Google Scholar] [CrossRef]
  24. Deng, Y.; Wu, J.; Tan, Y. Optimal Attack Strategy of Complex Networks Based on Tabu Search. Phys. A Stat. Mech. Its Appl. 2016, 442, 74–81. [Google Scholar] [CrossRef]
  25. Zhou, Y.; Hao, J.-K.; Fu, Z.-H.; Wang, Z.; Lai, X. Variable Population Memetic Search: A Case Study on the Critical Node Problem. IEEE Trans. Evol. Comput. 2021, 25, 187–200. [Google Scholar] [CrossRef]
  26. Arulselvan, A.; Commander, C.W.; Elefteriadou, L.; Pardalos, P.M. Detecting Critical Nodes in Sparse Graphs. Comput. Oper. Res. 2009, 36, 2193–2200. [Google Scholar] [CrossRef]
  27. Lozano, M.; García-Martínez, C.; Rodríguez, F.J.; Trujillo, H.M. Optimizing Network Attacks by Artificial Bee Colony. Inf. Sci. 2017, 377, 30–50. [Google Scholar] [CrossRef]
  28. Khalil, E.; Dai, H.; Zhang, Y.; Dilkina, B.; Song, L. Learning Combinatorial Optimization Algorithms over Graphs. In Advances in Neural Information Processing Systems 30 (NIPS 2017); NIPS: Grenada, Spain, 2017. [Google Scholar]
  29. Fan, C.; Shen, M.; Nussinov, Z.; Liu, Z.; Sun, Y.; Liu, Y.-Y. Searching for Spin Glass Ground States through Deep Reinforcement Learning. Nat. Commun. 2023, 14, 725. [Google Scholar] [CrossRef]
  30. Fan, C.; Zeng, L.; Sun, Y.; Liu, Y.-Y. Finding Key Players in Complex Networks through Deep Reinforcement Learning. Nat. Mach. Intell. 2020, 2, 317–324. [Google Scholar] [CrossRef] [PubMed]
  31. Braunstein, A.; Dall’Asta, L.; Semerjian, G.; Zdeborová, L. Network Dismantling. Proc. Natl. Acad. Sci. USA 2016, 113, 12368–12373. [Google Scholar] [CrossRef]
  32. Sarker, S.; Veremyev, A.; Boginski, V.; Arvind, S. Critical Nodes in River Networks. Sci Rep. 2019, 9, 11178. [Google Scholar] [CrossRef]
  33. Holme, P.; Kim, B.J.; Yoon, C.N.; Han, S.K. Attack Vulnerability of Complex Networks. Phys. Rev. E 2002, 65, 056109. [Google Scholar] [CrossRef]
  34. Dinur, I.; Safra, S. On the Hardness of Approximating Vertex Cover. Ann. Math. 2005, 162, 439–485. [Google Scholar] [CrossRef]
  35. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive Representation Learning on Large Graphs. In Advances in Neural Information Processing Systems 30 (NIPS 2017); NIPS: Grenada, Spain, 2017. [Google Scholar]
  36. Liu, C.; Xu, X.; Hu, D. Multiobjective Reinforcement Learning: A Comprehensive Overview. IEEE Trans. Syst. Man Cybern. Syst. 2015, 45, 385–398. [Google Scholar] [CrossRef]
  37. Van Moffaert, K.; Nowé, A. Multi-objective reinforcement learning using sets of pareto dominating policies. J. Mach. Learn. Res. 2014, 15, 3483–3512. [Google Scholar]
  38. Zdeborová, L.; Zhang, P.; Zhou, H.-J. Fast and Simple Decycling and Dismantling of Networks. Sci. Rep. 2016, 6, 37954. [Google Scholar] [CrossRef]
  39. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.A.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  40. Bavelas, A. Communication Patterns in Task-Oriented Groups. J. Acoust. Soc. Am. 1950, 22, 725–730. [Google Scholar] [CrossRef]
  41. Leskovec, J.; Krevl, A. SNAP Datasets: Stanford Large Network Dataset Collection; SNAP: Santa Monica, CA, USA, 2014. [Google Scholar]
  42. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  43. Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural Message Passing for Quantum Chemistry. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  44. Pei, H.; Wei, B.; Chang, K.; Lei, Y.; Yang, B. Geom-GCN: Geometric Graph Convolutional Networks. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
Figure 1. Leveraging minimum nodes for optimum key players’ identification in complex net-works. (a) The 9/11 terrorist network with 62 terrorists (nodes) and 159 relations (edges). The size of a node corresponds to its degree. (b): HDA algorithm adaptively removes 28 nodes (grey); only 3 edges remain in the network, but this algorithm has a poor ANC score. (c): FINDER algorithm removes the same number of nodes as HDA, and 6 edges remain in the network. (d): MiniKey algorithm removes the same number of nodes as HDA; only 3 edges remain in the network and an optimal ANC score is achieved.
Figure 1. Leveraging minimum nodes for optimum key players’ identification in complex net-works. (a) The 9/11 terrorist network with 62 terrorists (nodes) and 159 relations (edges). The size of a node corresponds to its degree. (b): HDA algorithm adaptively removes 28 nodes (grey); only 3 edges remain in the network, but this algorithm has a poor ANC score. (c): FINDER algorithm removes the same number of nodes as HDA, and 6 edges remain in the network. (d): MiniKey algorithm removes the same number of nodes as HDA; only 3 edges remain in the network and an optimal ANC score is achieved.
Mathematics 11 03690 g001
Figure 2. Accumulated Normalized Connectivity and Solution Length Ratio of different methods on six synthetic graphs. (a) Accumulated Normalized Connectivity of different methods on synthetic graphs; a lower value is preferable for this metric. (b) Solution Length Ratio of of different methods on synthetic graphs; a lower value is preferable for this metric.
Figure 2. Accumulated Normalized Connectivity and Solution Length Ratio of different methods on six synthetic graphs. (a) Accumulated Normalized Connectivity of different methods on synthetic graphs; a lower value is preferable for this metric. (b) Solution Length Ratio of of different methods on synthetic graphs; a lower value is preferable for this metric.
Mathematics 11 03690 g002
Figure 3. Comparison of pareto frontiers across different methods on six simulated datasets (af). MiniKey has the best performance when considering both ANC and SLR metrics simultaneously.
Figure 3. Comparison of pareto frontiers across different methods on six simulated datasets (af). MiniKey has the best performance when considering both ANC and SLR metrics simultaneously.
Mathematics 11 03690 g003
Figure 4. ANC curves for different methods on six real world datasets (af).
Figure 4. ANC curves for different methods on six real world datasets (af).
Mathematics 11 03690 g004
Figure 5. Comparison of pareto frontiers of different methods on six real-world datasets (af).
Figure 5. Comparison of pareto frontiers of different methods on six real-world datasets (af).
Mathematics 11 03690 g005
Table 1. Statistical analysis of network structures for six real-world datasets examined in this study.
Table 1. Statistical analysis of network structures for six real-world datasets examined in this study.
Statistics/DatasetCrimeHI-II-14DiggEnronGnutella31Facebook
Node Number829416529,65233,69662,56163,392
Edge Number147313,08784,781180,811147,878816,831
Maximum Degree252863101383951098
Average Degree3.556.285.7210.734.7325.77
Diameter101112111115
Clustering Coefficient0.00580.04440.00540.50920.00550.2218
Assortativity−0.1645−0.20160.0027−0.1165−0.09270.1768
Table 2. The ANC score of different methods on six real world dataset.
Table 2. The ANC score of different methods on six real world dataset.
Method/DatasetCrimeHI-II-14DiggEnronGnutella31Facebook
CI0.12430.06160.08660.04450.11740.2695
MinSum0.13830.06520.09520.04770.11730.2725
CoreHD0.11330.06060.08680.05140.11970.2747
GND0.13810.06940.10660.04560.12570.2742
FINDER0.10990.05540.08660.04300.11100.2684
MiniKey0.10850.05660.08580.05090.10450.2810
Table 3. The SLR score of different methods on six real-world datasets.
Table 3. The SLR score of different methods on six real-world datasets.
Method/DatasetCrimeHI-II-14DiggEnronGnutella31Facebook
CI0.67550.96760.99851.00000.99010.9980
MinSum0.68520.96760.99841.00000.99010.9976
CoreHD0.68280.96760.99850.96190.99010.9776
GND0.66830.96760.99851.00000.99010.9980
FINDER0.50180.35410.42630.49010.27480.6547
MiniKey0.42700.29360.38360.46260.26240.6157
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, L.; Fan, C.; Chen, C. Leveraging Minimum Nodes for Optimum Key Player Identification in Complex Networks: A Deep Reinforcement Learning Strategy with Structured Reward Shaping. Mathematics 2023, 11, 3690. https://doi.org/10.3390/math11173690

AMA Style

Zeng L, Fan C, Chen C. Leveraging Minimum Nodes for Optimum Key Player Identification in Complex Networks: A Deep Reinforcement Learning Strategy with Structured Reward Shaping. Mathematics. 2023; 11(17):3690. https://doi.org/10.3390/math11173690

Chicago/Turabian Style

Zeng, Li, Changjun Fan, and Chao Chen. 2023. "Leveraging Minimum Nodes for Optimum Key Player Identification in Complex Networks: A Deep Reinforcement Learning Strategy with Structured Reward Shaping" Mathematics 11, no. 17: 3690. https://doi.org/10.3390/math11173690

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop