Next Article in Journal
Novel Hybrid Intelligent Secure Cloud Internet of Things Based Disease Prediction and Diagnosis
Next Article in Special Issue
An Approach to the Construction of a Recursive Argument of Polynomial Evaluation in the Discrete Log Setting
Previous Article in Journal
Exploiting the S-Iteration Process for Solving Power Flow Problems: Novel Algorithms and Comprehensive Analysis
Previous Article in Special Issue
A Novel Key Generation Method for Group-Based Physically Unclonable Function Designs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SOD2G: A Study on a Social-Engineering Organizational Defensive Deception Game Framework through Optimization of Spatiotemporal MTD and Decoy Conflict

Department of Computer Science, Kyonggi University, Suwon-si 16227, Korea
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(23), 3012; https://doi.org/10.3390/electronics10233012
Submission received: 13 November 2021 / Revised: 28 November 2021 / Accepted: 30 November 2021 / Published: 2 December 2021
(This article belongs to the Special Issue Advanced Security Protection Mechanism in Networks)

Abstract

:
Existing moving target defense (MTD) and decoy systems are conceptually limited in avoiding and preventing attackers’ social-engineering real-time attacks by organization through either structural mutations or induction and isolation only using static traps. To overcome the practical limitations of existing MTD and decoy and to conduct a multi-stage deception decision-making in a real-time attack-defense competition, the current work presents a social-engineering organizational defensive deception game (SOD2G) as a framework, consi dering hierarchical topologies and fingerprint characteristics by organization. The present work proposed and applied deception concepts and zero-sum-based two-player game models as well as attacker and defender decision-making process based on deceivable organizational environments and vulnerability information. They were designed in consideration of limited organizational resources so that they could converge in the positive direction to secure organizational defender dominant share and optimal values of the defender deception formulated by both scenario and attribute. This framework could handle incomplete private information better than existing models and non-sequentially stratified, and also contributed to the configuration of the optimal defender deception strategy. As the experimental results, they could increase the deception efficiency within an organization by about 40% compared to existing models. Also, in the sensitivity analysis, the proposed MTD and decoy yielded improvements of at least 60% and 30% in deception efficiency, respectively, compared to the existing works.

1. Introduction

Defensive cyber deception [1,2] is a type of non-cooperative decision-making contaminating technology that involves manipulating the cognitive perspective of potential attackers and deceiving them so that they continuously construct and maintain erroneous post-action strategies. This is carried out by leveraging defender-dominant information asymmetry. Defensive cyber deception is a new concept [3,4] with unique characteristics that distinguish it from other security elements, such as induction, isolation, backtracking, and mutation, while separately having dedicated kill chain processes by application environment and scenario. By introducing defensive deception as such, the attacker-dominant spatial and temporal asymmetry remaining in conventional security (such as access control and intrusion blocking) can be drastically alleviated in existing system architectures. Related deceptive security systems can be also flexibly applied and distributed in multiple operating layers more easily without having to make more substantial changes than in conventional security. It can also serve as a major source of technology, driving the development of the concept of active defense in an organizational operating environment. Its expected benefits include dynamically detecting and collecting artifacts such as attackers’ attack vectors or intrusion paths to trace them back, formulating all immediate detection, proactive evasion, induction, and isolation schemes by advanced persistent threat (APT) type, and maximizing attackers’ total attack cost while reducing defenders’ entropic time spent to capture infiltration attempts [5].

1.1. Background

Defensive cyber deception technology includes MTD [6], Honey-X [7], and Decoy [8,9]. It can be classified according to its operational goals and defensive purposes. For example, based on its mobility properties such as shuffling, diversity, and redundancy, MTD can limit the effectiveness of defender intelligences collected in advance by attackers either periodically or aperiodically, mutate observable attacks and exploration surfaces [10] according to the defender’s intention to increase the attacker’s uncertainty and confusion, thus blocking serial compositions of attack chains and ultimately resolving the attacker dominant information asymmetry to realize a proactive defense. Meanwhile, Decoy and Honey-X are dynamic sandboxing elements. They can perform induction, isolation, and decoy schemes known to cause attackers to attack false targets that are not protected. These can interact with each other to mislead attackers based on the network or system level depending on the defender’s intention. MTD has been combined with decision-making strategies among decision-making entities in areas such as game theory [11,12,13,14], MDP (Markov decision process) [15,16,17,18], and machine learning [19,20,21,22]. It has also been combined with the learning theory for benefit optimization between attacks and defenses to achieve diverse optimized mutation strategies and deception thresholds while attenuating effects of intrusion by cyber kill chain (CKC) [23] stage and surface spaces. And this MTD has also been combined with other elements such as Honey-X and Decoy.

1.2. Limitation of Existing Defensive Deception Based on Non-Common and Common Games

However, the existing MTD and Decoy design presented in previous studies has a few limitations. Limitations of the non-game theory-based defensive deception studies are as follows:
  • Organizational deception management strategies are not considered: To improve the reliability and practicality of defensive cyber deception in an organization and to optimize the common goal of maximizing security while minimizing the impact on performance, deception modeling dedicated to MTD and Decoy should be constructed and deception management strategies tailored to unique characteristics of each organization should be formulated. This is also related to the seamless connection of legal services enabled by the application of MTD [24,25]. However, the actual operability of this setup is low due to the fact that previous studies have responded by opening separate encrypted closed communication channels and mapping random virtual network information with real information, thus partially mutating the information. In addition, since the generated virtual information is also randomly composed, the attacker’s judgment cannot be adjusted according to the intention of the defender. In addition, the artificial cognitive bias is insignificant.
  • Having a low flexibility and scalability: The existing naive MTD is only focused on securing proactive defensibility such as through obstructing the formation of attack chains by evading attackers’ initial reconnaissance through structural mutations or judging those users who have requested illegal characteristic information (e.g., IP address, port number, service, etc.) at random points in time as illegal and then inducing and isolating them into static traps. In addition, since post-responses by threat were not considered as major actions in the MTD concept, there was also a complete lack of countermeasures against deviations of sophisticated attackers who had identified MTD mutation patterns or fatal vulnerability attack points in the organization’s operational architecture.
  • Lacking deception strategies against social engineering APT attacks: Quantitative deception management strategies against various social engineering APT attack vectors [26], such as malvertising [27], spear phishing [28], and watering hole [29], should be presented along with optimization plans. That is, the skewness and outliers in the detection and blocking system of an organization with limited available resources should be alleviated. Mutations of MTD related to open-source intelligence (OSINT) [30,31], HUMINT, Decoy management-distribution, dynamics diversification of attack, exploration surfaces, and so on should all be considered.
  • Defender deception supplementation and intelligence composition are not systematized: Although many approaches can be used to reduce the skewness or errors in threat intelligences and improve response quality using research models and platforms integrating existing MTD and decoys, they are limited to the test bed configuration based on certain domains. In addition, the conceptualization of active defender deception actions appropriate to the actual organizational environment has also been limited to studies involving a small number of vendors based on the concept of disinformation and adaptive pollution of external OSINT [32].
Limitations of game theory-based deception studies can be summarized as follows:
  • Having decision-making conflicts and making it impossible to optimize deceptions: Game theories are used to model dynamic decision-making in uncertain situations under the assumption that each player has a consistent view. However, when decision-making is based only on asymmetric information that can be observed by the player at random points in the same game, an unoptimized deceptive value can be calculated depending on the game strategy and the uncertainty of the equilibrium state [33]. Therefore, a separate concept of view and share considering the presence of uncertainty that can lead to different optimal judgments each time should be added.
  • Concept of high dynamics and real-time quality are not considered: Although many different types of disturbances and contingent factors in the network can affect the security state in an organization with rapid changes of the attack-defense environment, previous studies did not accurately characterized real-time changes in the performance of the actual network attack-defense process or consider the noise that can potentially affect the rewards by episode since they sequentially modeled the relationship between defenders and attackers mainly based on simple leaders and followers. In addition, when players are taking actions intended to achieve their goals, it is impossible to dynamically calculate suboptimal deception strategies against incompletely perceived enemies due to the sequential consumption of episodes for equilibrium [34].
  • Lacking organizational unique characteristics and operating principle: Since previous game theory studies have defined attack-defense targets, equilibrium states, weights, and state-transition probabilities in either a static manner or a weak dynamic manner based on CVE (common vulnerability-ties and exploits) and CVSS (common vulnerability scoring systems) in NVD (national vulnerability database) when constructing scenarios and sequences, they are efficient in comparing and verifying the efficiency and utility of defender deception based on vulnerable points of contact and attack vectors. They can be used to calculate the effectiveness of a dedicated system reflecting some characteristics of a simulated domain (e.g., multi-tenant cloud network [35], vehicle-type internal/external communication networks [36,37]) depending on the definition of the topology structure. However, measures for applying the concept of defensive cyber deception to the actual organizational environment have not been officially reported yet. They do not have the dedication of the concept of deception according to organizational unique characteristics, operating principles, or possible attack-defense patterns, etc.
This study of social engineering defensive organizational deception games aims to alleviate practical limitations and theoretical scalability issues of non-game theory- and game theory-based defensive cyber deception studies and to practically model the decision-making process between attackers and defenders in an organizational environment.

1.3. Research Gap and Key Contributions

To optimize existing MTD and Decoy deception management strategies by organization and overcome their limitation of low practical scalability, the objective of this study was to present PBNE (perfect Bayesian Nash equilibrium) and BSSG (Bayesian stochastic Stackelberg game)-based zero sum game foregrounds considering characteristics of real organizations and existing SOD2Gs and systematic deception game frameworks based on the POMDP (partially observable Markov decision process) state-transition background by attack model and defense model. This study will also simulate concepts of limited asymmetric rewards, views, and share manipulation according to defender’s disinformation-based false information push. Adaptive deception decision-making was also characterized to decoy attackers’ cognitive bias according to the defender’s goal. This study also calculated a general-purpose-defined organizational scenario based on the proposed architecture, conducted simulations based on the CKC sequence as well as deception metrics and parameters for comparison, and calculated the optimal value of deception according to evaluation and verification results.
Major contributions of this study are as follows:
  • Game theory was used to formulate an MTD that newly defined defender dominant deceiving behaviors according to the selection of a mutation cycle by accommodating attacks with loose adaptive mutations instead of strong prior avoidances as well as OSINT-based decoys by organization that could maintain decoying, isolation, and continuous deception resulting from the attackers’ initial cognitive bias.
  • By applying social engineering characteristics by organization to customized MTD (LPC-MTD) and Decoy (HS-Decoy), measures to strengthen prior preventiveness and post responsiveness within the organization were conceptually secured and deception efficiency was calculated based on the open organizational OSINT.
  • PBNE and BSSG-based zero-sum game models universally composed by scenario were used to simulate real-time attack-defense competition within an organization with limited resources and model a multi-staged spatiotemporal deception decision-making process. In addition, by combining POMDP state-transition models, limited views by attacker and defender along with related attack surface concepts were simulated. The defender can also push disinformation-based deception signals for any single host initially intruded upon or finally occupied to the attacker as well as inverse transfer false host occupation information to the attacker to make the attack chain biased to be defender dominant or make it stay in the sandbox for a long period of time.
  • When responding to in-depth attackers by organization, it is possible to establish an optimal defensive deception that minimizes operational performance degradation and maximizes security while operating independently without separately applying additional procedures or dedicated protocols based on scenarios and metrics.
The composition of this paper is as follows. First, Section 2 presents preceding game theory studies on MTD and Decoy concepts of defensive deception and compares them with the newly proposed concept of organizational deception games. In Section 3, with the goal to address limitations of existing MTD and Decoy, to evaluate and optimize real-time decision-making strategies between attackers and defenders by actually defined organizational environment, PBNE- and BSSG-based zero-sum game foregrounds and SOD2Gs (which are POMDP state-transition background-based organizational deception game frameworks) in practical application are described with an internal attack-defense model. Related game metrics, formulas, social engineering knowledge, and deceptive parameters are then detailed. In Section 4, multiple organizational scenarios in SOD2Gs are specified together with CVEs and CVSSs by host to configure topology test beds, a social engineering attack-defense competition is simulated, and topology test beds are compared and analyzed along with an analysis of sensitivity by metric. In Section 5, inherent limitations of this study are discussed in addition to improvement measures. Finally, conclusions are drawn in Section 6.

2. Related Studies and Taxonomy of Proposed Deceptive Game Framework

Although MTD partially belongs to the scope of defensive cyber deception, unlike Honey-X or Decoy, it neither projects false information to actively mislead the attacker nor conceptually involves any scheme that actively induces the attacker’s cognitive bias according to the defender’s intention. In other words, their show crucial differences in that MTD mainly efficiently changes and proactively avoids configuration of internal networks and system hosts to be protected while maintaining availability of major service functions for legitimate users, whereas Honey-X and Decoy manipulate attackers’ perceptions while carrying out reactive induction and isolation [1].
Based on these differences and goals of the present study, designs in previous studies are categorized as ‘Game-theoretic Deception with MTD’ or ‘Game-theoretic Deception with Honey-X and Decoy’ to upgrade the proposed game framework and attack-defense model into concepts that can be applied in practice to improve their performances and optimize their equilibrium state.

2.1. Game-Theoretic Defensive Deception with MTD

The core of game theory studies using MTD involves competitively modeling the attacker’s CKC tactics and the defender’s proactive evasion tactics as well as optimizing MTD variables to achieve goals held by each player. The main purposes of such studies are to minimize performance degradation and maximize security (thereby maximizing the defenders’ gains) and to advance deception strategies in order to minimize attackers’ gains such as through lateral movements and occupation of the target. Such studies are divided into ‘general game theoretic-based studies’, ‘Bayesian Stackelberg game theoretic-based studies’, and ‘stochastic game theoretic-based studies’ [1,6,7].
First, among general game theoretic-based studies, Zhu et al. [38] have quantified the trade-off based on the defender’s security enhanced by MTD and degraded operational performance by applying the general game theory-based sequential attack-defense competition formula and related metrics to the MTD mutation concept. This allows for the concept of MTD games related to attacker reconnaissance evasion and intelligence collection efficiency minimization to be established for the first time. Ge et al. [39] have proposed incentive compatible MTD games based on communication mapping between normal users during server migration as a methodology that could be used to ensure high network service visibility and throughput for legitimate users as well as improve cyber agility [40] in terms of availability. This methodology could compose the concept of MTD operation for securing service availability and cyber agility in an architecture defined in detail for the first time. Carter et al. [41,42] have proposed a dedicated game to secure a migration optimization strategy to maximize seamless connection to legitimate users’ services while minimizing suspicion among attackers that have been induced and isolated into the sandbox. Results of their study have verified that platform diversity based on overlapping cloning for the application of MTD at the lower stage is effective for some exploit attacks and more efficient than pure random mutations. However, proposed methodology is inefficient for rapid local attacks and also has proven that ensuring high hierarchical diversity by dummy platform can yield higher proactive benefits than having a large number of dummy platforms horizontally.
Second, among Bayesian Stackelberg game theoretic-based studies, Hasan [43] has proposed co-resident attack mitigation and prevention (CAMP), which is a Nash equilibrium game model for detecting co-resident attacks in a virtual environment where the same spatiotemporal resources are shared while minimizing the impact of internal and external threats in a compromised joint virtual environment. The proposed CAMP has been proven to be able to immediately provide active attacker induction and active isolation tactics related to the optimal MTD defense strategy in the VM made by cloning the normal server. Feng et al. [44] have proposed a Bayesian Stackelberg-based artificial information disclosure model. Based on the Stackelberg game for applying the leader and follower scheme and the signaling game, which is an interactive decision strategy, they have proven that intentional disclosure of false information by the defender is a rather promising methodology that can disturb and bias the initial decision of the attacker to attenuate the continuity of the attack while simultaneously enhancing the agility of the defender. In a follow-up study, Zhu et al. [45] have proposed a Stackelberg game framework that generates and publishes false packets characterized in the reconnaissance stage of internal and external attackers. The efficiency of the indeterminate attacker induction and isolation mechanism in the multi-routing network topology was improved in their study. Among related studies, their study secured the directionality of application of the deception concept according to the construction of a scenario model close to the practical working environment for the first time. Sengupta et al. [35,46] have proposed a Bayesian Stackelberg game to optimize the deployment of security systems in the web and cloud. With that game, it is possible to formulate an MTD strategy that maximizes the proactive security using the system configuration set candidate while minimizing the mutation cost and performance degradation rate of the defender with limited resources. In a follow-up study, Li et al. [47] have proposed a Markov Stackelberg game to calculate the spatiotemporal MTD mutation decision-making of the defender against advanced attackers and composed an optimization formula based on the average-cost semi-Markov decision process (SMDP) and the discrete time Markov decision process (DTMDP). With their game and formula, it is possible to empirically evaluate and verify the improvement of the MTD decision-making performance of the defender who proactively responds to the short-sighted adaptive professional attacker with better performance than previous related studies.
Third, among stochastic game theoretic-based studies, Manadhata et al. [48] have diversified the dynamics between attacks and defenses based on the concept of probabilistic transition according to the decision-making flow and proposed a game model that reflects the foregoing in the strategies by MTD mutation state as an extension of previous studies regarding the diversification of attack surfaces and exploration surfaces. Their study showed that it was possible to formalize the MTD equilibrium concept and related trade-offs for modeling the optimal pro-active defense strategies based on dynamic surface compositions by scenario environment.

2.2. Game-Theoretic Defensive Deception with Non-MTD

The core of game theory studies using Honey-X and Decoy involves modeling attackers’ detour tactics and defender’s reactive defense tactics and conceptualizing decoying and sandboxing to realize disturbance, induction, and isolation based on signals corresponding to certain requests. The main goal of such studies is to present a strategy that can minimize performance degradation and maximize the efficiency of deception. In this subsection, ‘game theoretic with MDP and POMDP studies’, ‘signaling game theoretic studies’, and ‘static game theoretic studies’ are described briefly.
First, among studies on game theory with MDP and POMDP, Bilinski et al. [49] have described solutions based on masking games and Stackelberg games that can hide the essence of the defender node through processes such as inquiries between attackers and defenders, payment of costs, and selection of the final compromise target. These solutions make it possible to probabilistically analyze the attacker’s potential behaviors before and after the exploit. They can also make adaptive configurations of the defender’s deception strategies by episode. Durkota et al. [50] have proposed an MDP and sibling-class pruning-based Stackelberg game to optimize deception strategies in the form of searches by time point of the time to the defender who deploys a honeypot to induce an attacker who then complies with the CKC attack graph. Anwar et al. [51] have quantified a static game that combines POMDP and a stochastic game as a honeypot and decoy allocation methodology considering probabilistic uncertainty through attack-defense interaction and model attack graph-based decision making. In that game, related to ramifications of the honeypot and decoy that perform attacker cognitive bias and induction, the researchers constructed a deceptive container management plan that yielded the highest defendability and the lowest cost for defenders with limited utilities. They also presented a heuristic optimization for alleviating the complexity of deceptive container allocation according to the potential decision-making interaction between attack and defense.
Second, among signaling game theory studies, Pawlick [52] has investigated and analyzed multiple signaling game and modeled them to devise a related solution while considering conditions and situations in which an attacker detects and recognizes the defender’s honeypot. That study proved that the attacker detection ability formulated based on the present situation of the defender’s deception identified in advance or evidence collected according to the interacted signals by episode did not necessarily diminish the deception efficiency and security of the defender. It also verified that the concept of artificial exposure could be expanded and applied to other non-signal game-based deception models. In a follow-up study [53], the concept of artificial exposure has been improved as a signaling game framework that could probabilistically induce and make judgments to estimate evidence based on the sender type and transmitted message, thereby proving again that the concept is vulnerable to attackers who can analyze the signals. Mohammadi et al. [54] have proposed a signaling game for identifying and deceiving external attackers using false defender avatars to optimize decoy strategies and alert thresholds. In that study, they defined the uncertainty of the judgment of the identity of the defender by the attacker who received signals from a real defender or a false defender avatar and considered the concept of insider authorization based on access control in the system architecture. Casey et al. [55] have proposed a basic compliance signaling game with which to consider the interaction between an organization’s defender and a hostile insider, who can acquire the surface of the organization at the beginning and attempt to infringe on all hosts. The researchers have established the concept of a honey surface that perturbs and isolates the cognitive bias of malicious insiders. A follow-up study [56] has proposed an advanced compliance signaling game to simulate the interaction between an attacker that adaptively attempts to compromise and a defender by learning the defender intelligence that changes as the insider threat within the organization deepens.
Third, among static game theoretic studies, Nan et al. [57,58] have proposed a sandbox-based mixed-type perfect Nash equilibrium game framework in which the defender intelligently selects the most likely false node to formulate a deception strategy that biases the attacker to attack the deceptive node while wasting resources. Using that framework, they presented the most efficient deception node deployment strategy to the defender with limited resources, as well as configure a reactive management plan to achieve isolation, access blocking, and exit the induced attacker along with a primary deception action. Dimitriadis et al. [59] have proposed 3GHNET as a non-cooperative zero-sum architecture that can prevent DDoS and node breaches and enhance the security of 3G core networks within an organization. Their study showed that it was possible to formulate a payoff matrix to assists the defender in constructing an optimal deception strategy through two gateways that could control and capture the data flow between the sending and receiving nodes as well as a Nash equilibrium calculation focusing on the network attacker. Huang et al. [60] have proposed multi-frameworks reflecting the static Bayesian game and the asymmetric information one-shot game to consider the covertness and social engineering deceitfulness of professional attackers. The proposed multi-frameworks were verified within the Tennessee Eastman-based APT scenario.

2.3. Analysis by Previous Studies for Proposed Deceptive Game Concepts

To alleviate and solve the above-mentioned limitations, SOD2G, a systemic deception game framework based on PBNE and BSSG-based zero-sum game foreground and POMDP state-transition background, together with internal attacker and defender models, and the analysis of their classification are shown in Table 1.

3. Proposed Organizational Deceptive Game Framework

The PBNE and BSS-based zero-sum game foreground and POMDP state-transition background-based systematic defensive deception game framework in this study is a measure that can alleviate all operational limitations, unrealistic decision-making issues, practical scalability issues, and so on in an organization environment of MTD and Decoy concepts as well as normalize the deception optimality in each zero-sum-based competition scenario. Therefore, this section details the novel formulation by module and by the process of the characterized MTD and decoy concepts based on OSINTs by organization in addition to the attack-defense player model in the proposed SOD2G framework and related components. PBNE, BSSG, and POMDP-related deception-game metrics and equations are also defined with related tuples.

3.1. Concept of Social-Engineering Organizational Defensive Deception Game Framework

The major architecture of the proposed SOD2G is illustrated in Figure 1. First, the OSINT elements by organization collected and formulated with node-link-based information graphs in advance are applied to LPC-MTD and HS-Decoy (which are MTD- and Decoy-applied concepts, respectively) by major property and by function, leading to advancements in social engineering deception knowledge. Social engineering deception knowledge on which unique characteristics by organization are reflected by hierarchical element is then used as an interface between the game-based foreground and the state-transition-based background. First, in the game-based foreground, it is used as a preprocessed parameter for generating and updating deception strategies of the defender model in the PBNE- and BSSG-based zero-sum game components and signaling-based distribution behavior, as a variable to be referenced when plotting by metric in the optimizer module, and as a major template concept for the formation of the defender’s countermeasure tactics in the attack and defense graph component. Further, in the state-transition-based background, it is applied as a deceptive signaling and decoying concept for the defender’s microscopic view and occupancy rate, surface, and partial reward manipulation concepts in the POMDP. It is also used as an adaptive probability change factor in the transition probability matrix based on the vulnerability table in the transition component. Finally, it fosters all related concept of deception, competition between players, payoff tactics, and game equilibrium.

3.2. Organizational Deception Knowledge with OSINT

Open-source intelligence (OSINT) is reliable information from open sources, government agencies, and so on. It goes beyond detailed information and characteristics to generally refer to even the collection process from open sources. It is mainly used in various cyber and information security fields such as military, security, cyber security, infringement responses, disinformation, and deception-based intelligence. It is mainly classified as a domain-based category consisting of information flow on the Internet, public datasets maintained by governments, and private open datasets. Organizational OSINT [61] is characterized as the use of an information group consisting of unique OSINTs and rough fingerprints of foreign government military organizations collected and composed based on open sub-party services. It involves performing all necessary stages to proactively establish an OSINT strategy based on security requirements related to an organization’s legacy operating environment mainly by identifying information sources, including a stage in which open OSINTs are collected based on identified organizational information sources and strategies, a stage in which acquired data are normalized based on correlation graphs and then supplemented based on parent-child nodes followed by a reinforcing step based on parent-child nodes, and a stage in which the reinforced OSINT information group is refined into social engineering deceptive knowledge and concretized as cyber space intelligence (CYBINT) in sequence. The composition of the related organizational OSINT information group is reduced and recomposed based on the network, host, and service layers as presented in Table A1 based on the concept of simulation of the view and surface of the attack-defense model in SOD2G.

3.3. LPC-MTD and HS-Decoy for Organizational Defensive Deception Strategy

First, loosely proactive control-based MTD (LPC-MTD) [61] is an organizational MTD concept designed to alleviate both conceptual limitations and social engineering scalability problems of the existing MTD. It artificially adjusts the mutation strength based on the intention of the defender to intentionally expose false information or present disinformation whereby the defender dominantly forces the attacker to make a hasty judgment as if he/she has successfully bypassed the defender’s MTD interface. It also induces early cognitive biases of attackers, such as leading attackers to the defender’s isolated sandbox by perturbing the defender’s intelligence gathered by the attacker that has performed reconnaissance. The concept of LPC-MTD as such is defined with a focus on the Bellman value iteration-based batch sampling optimizer to select entropy-based adaptive MTD mutation strength according to changes in attack behaviors along with the perturbation [7] to asymmetrically impart noises to the attacker’s cognitive directivity based on a limited view and related attack surface information. Based on that concept, along with additional advancement of the defender’s deception strategies using the MTD cycle selection schemes by organization, MTD-based deception operation can also be formulated for network separation and closed network.
Due to the mobility-based MTD principle [6,11], the LPC-MTD is constructed as shown in Equations (1) and (2) for perturbation (P) and value iteration sampling (VS), respectively:
P = P r [ L = l | O S = μ ( o s ) ] ,
where L is a set of virtualized IP, port, service, and host fingerprint elements that are similar to the actual information of the target to be protected in an organization, but are hierarchically composed so that only the defender and legitimate user can identify them. which are dynamically signaled under the leadership of the defender in order to bias the normalized gradient of the defender’s intelligence possessed by the attacker toward the defender’s advantage. OS is a set of OSINT information groups by organization. It is intended to improve the reliability of the virtual information generated for deception while minimizing the attacker’s suspicion of disinformation and artificial information exposure through L. In Equation (1), μ(os) calculates the attacker’s predictability for the secret being protected by the deceptive defender based on both observable arbitrary OSINT-based exposure factors os and the defender intelligence possessed by the attacker at present.
V S n ( i ) = min τ i , m i M [ c i , j + m ˜ i , j V S n 1 ( j ) ] ,
where c i , j is the mutation cost to shift or shuffle the surface index j of the defender of the next episode with private information through an adaptive deception strategy of signaling a limited view and a false occupancy rate at the surface index i of the defender of the current episode in order to optimize the trade-off of the legacy topology operating environment in an organization that has limited resources. m ˜ i , j is the possibility to distinguish i and j following changes in i and j. It means the possibility of minimizing the suspicion of the APT attacker who continuously observes the defender’s environment. τ i is the temporospatial resource consumed in relation to the intervention of the defender until the completion of the mutation of the mutation time slot length of the surface information related to i, and mi is the surface information sampled and optimized based on i in the deceptive surface set M of the defender, which can be created through mutation.
HS-Decoy (hierarchical social engineering decoy) [61] is also an OSINT-based organizational decoy concept that either statically induces attackers to continue to attack a false target that is not a target of protection to further improve the deception efficiency within an organization of the existing decoy or forces successive attack chains of attackers that have been isolated to remain in the sandbox while deceiving the attackers to stop them from attempting to escape. Since it mainly appears similar to normal service or important information while being more enticing based on OSINTs by organization, it looks valuable to attackers. However, it has a complex multitenant-based architecture composed of a large number of false data in reality, thus enabling legitimate users within the organization to distinguish the foregoing from real information while operating such that attackers cannot distinguish it from real information without prior knowledge of the target system. It also increases hierarchies and number of entities to statically deceive attackers using layered OSINT elements while independently improving redundancy and diversity concepts based on detailed deception elements in the decoy to minimize suspicion of the attacker who invades in the early stage in relation to the parallel defender environment recognition and verification behavior of the attacker. The concept of HS-Decoy is therefore upgraded by combining hierarchical OSINT information groups by organization together with detailed complex properties such as ‘reliability and distinguishability’, ‘attack inducibility’, ‘permanent exposability’, ‘diversity’, ‘redundancy and non-coherence’, ‘distinguishability’, ‘detectability’, ‘role-based access control’, and so on to induce continuous attacker inferiority in relation to access event count, intrusion tolerance and restoration, induction of static attackers’ cognitive bias in the early stage, isolation, and so on. Through the foregoing, for dynamic attackers’ cognitive bias in the early stage and intentional permitting of invasion, which is primarily performed at the network and host level, HS-Decoy, which has been secondarily combined, can continuously force defender predominance in the isolation environment using independent distributed multi-tenant based multiplexed OSINT-based decoys with cloning and mimicking.
As such, the HS-Decoy will focus on the property-based decoy principle [9] based on Equations (3)–(9):
P r [ E x p A , H , O b e l i e v e = 1 ] 1 2 ,
First, Equation (3) is Believability (B). It calculates the attacker A ’s trustworthiness of HS-Decoy, who does not have information on the internal distinction criteria related to the protection target. and for the hierarchically formulated HS-Decoy set H based on both O, which is a set of organizational OSINT elements reconstructed for social engineering decoy operation, and information on the unique characteristics of the actual protection target. To elaborate, when A composes an arbitrary defender’s surface information, it is judged to be an active defender deception, and cases where H, which is defender dominant, is excluded as a stochastic noise are guaranteed to be 0.5 or less.
P r [ o O | o P F ] = P r [ h O | h H ] ,
Equation (4) is Enticingness (E). It normalizes the attacker’s hasty judgment such as IP, port, service, host fingerprint, and so on from o, which is an arbitrary element of organizational OSINT for decoy operation in O, to PF, which is an index for the degree to which attackers are quickly forced by the defender dominantly to calculate the possibility of static signaling to intentionally induce an infringement action on the o-based h. It also verifies that characteristics of PF in h, which are created, distributed, and managed by the deceptive defender, do not deviate from the uniqueness of o, which minimizes the attacker’s suspicion based on gradual mutation.
i = 0 n P r [ V i ] > δ ,
Equation (5) is Conspicuousness (C). It calculates the possibility of static dazzling for the limited view Vi related to the i-based surface information of the attacker and defender formulated in the current episode based on mutual signaling through a comparison with δ, which is an index standardized for the degree of salience of factors causing extreme bias in the formation of continuous CKC chains such as CVE vulnerability among o. That is, beyond passively advertising only CVE vulnerability information to the attacker, it also guarantees the execution of Vi-based projection to additionally prove the false reliability of the relevant vulnerability information to the attacker.
P r [ h O : C D A , h = 1 ] ϵ ,
Equation (6) is Detectability (DE). It calculates the detectability for C D A , h , which is an index for the degree to which the act of attacker A infringes on h through comparison with the detection threshold ϵ in h for adaptive management and for updating of h according to changes in the external entropy. It also ensures the possibility of false positive and false negative to improve the defender’s deceivability according to the decoy beacon and emulator.
P r [ h H : E x p A , H , O , h b e l i e v e = 1 ] 1 2 ,
Equation (7), which is based on (3), is Variability (V). It calculates the trustworthiness of h, a single HS-Decoy element, in a certain H. That is, by inducing the attacker’s cognitive bias and securing different distinguishability for all h in H where the decision gradient will be defender dominantly perturbed, it ensures the enhancement of hierarchical deception diversity for each h.
P r [ C T D , o , h = 1 ] = P r [ C T D , o , h = 1 | H ] ,
Equation (8) is Non-Interference (NI). It calculates the possibility of reactive non-interference for the contiguity of defender D (a legitimate user in the organization for o-based h) to be clearly distinguished as a sender for deceptive signaling while immediately being treated as an exception so that it would not be selected as a decoying target. That is, it guarantees role-based access control schemes that are divided such that attackers can access only h where legitimate users and D can only access the real protection target.
P r [ E x p D , H , O b e l i e v e = 1 ] = 1 ,
Finally, Equation (9) is Differentiability (DI), which is based on Equations (3), (7) and (8). It calculates both the distinguishability of defender D and the indistinguishability of attacker A for O-based H and the real protection target. That is, the protection target cannot be distinguished from H with a secondary deception policy if the attacker understands the abstracted organizational protection target in the isolation environment, which the attacker cannot reconnoiter or explore, through primary deceptive signaling such as MTD.
Finally, utilizing all organizational OSINT information groups, LPC-MTD, and HS-Decoy, they are normalized with the social engineering deceptive knowledge in SOD2G. They are then applied as major atomic input parameters in each of PBNE- and BSSG-based foreground and POMDP-based background components.

3.4. Construction of Organizational Dynamic Deceptive Game Framework

As shown in Figure 1, the game-based foreground in SOD2G consists of a zero-sum-based dynamic game component based on PBNE and BSSG, clustered elements for performing the next action based on transition probabilities by player of the transition branches by state in the POMDP, and attack and defense graph components, which represent an event information group.

3.4.1. Design Principles with Deceptive Multi-Player Game Architecture

Attack and defense graph components are configured as a threat model for the attack flow to achieve occupation of the final intrusion or suboptimal intrusion target point based on the temporospatial search and CKC stages formulated by the organizational scenario predefined in the scenario template component with POMDP-based background. The dynamic game components adopt the PBNE game strategy to optimize the judgment that can maximize the payoffs by episode for private asymmetry based on incomplete information by player. They also adopt BSSG strategies based on the leader and responsive followers. Ripple effects of signaling by player are also added to apply a quantitative sequential relationship for partial signaling between the attacker and the defender stochastically based on the concept of compensation.
In this case, dynamic game components, which are configurable based on Figure 1 and Figure 2, are composed in detail while centering on the following 9-tuples.
  • N = ( N A , N D ) is a set of players where N A is an attacker and N D is a defender. In this case, depending on the direction of the payoff, signaling, and feedback by player in a random episode, N is formulated with the attacker as an active APT leader and sender/defender as a passive reactive follower/receiver, or conversely, with the defender as a dynamic deceptive leader and sender/attacker as a passive naïve follower/receiver.
  • T S = ( T S N A , T S N D ) , T S N D = ( t s i | i = 1 , 2 , , n ) , and T S N A = ( ρ ) are sets of player types. T S N D is defined as an element of the hierarchical deceptive and defense graph-based private information of defender N D . T S N A is defined as an element of the attack-graph-based private information of attacker N A . Player types are divided or combined with abstraction based on abilities to subtract and add rewards by player. And the attacker to the defender who has non-deterministic elements decisively composes elements with the defender intelligence validity, which is termed ρ .
  • G S = ( G S N A , G S N D ) , G S D = ( g s N d i | i = 1 , 2 , ) , and G S A = ( g s N a j | j = 1 , 2 , ) are sets of game strategies related to mutual zero-sum competition between individual attackers N A and defenders N D . They are composed according to the sender and receiver signaling relationship. G S D is the deceptive and defense graph-based strategy possessed by the defender. G S A is defined as the attack-graph and intelligence-based strategy possessed by the attacker as valid defender surface information.
  • S S = ( S S N A , S S N D ) , S S N D = ( s s N d i | i = 1 , 2 , ) , and S S N A = ( S S N a j | j = 1 , 2 , ) are sets of signals of attacker N A and the defender N D . They are selected or released according to signaling mechanisms, which vary between active and passive from player to player. The attacker has S S N A as a leading attack signal set to achieve the final invasion target. The defender has S S N D as the leading deceptive and defense signal set for complete blockage and exit of the attacker.
  • S = ( s i | i = 0 , 1 , k ) is a finite set of states based on GS and SS in game components. It defines multi-level and transitivity in the attack-defense competitive game environment along with actions.
  • A = ( A N A , A N D ) , A N D = ( a N d i i | i = 1 , 2 , x ) , and A N A = ( a N a i j | j = 1 , 2 , y ) are finite sets of actions of attacker N A and defender N D for S . A N D defines the defender’s deceiving, defending, or false negative actions for s i as a transitive relationship. A N A defines attacker’s actions for s i , such as reconnaissance, search, vulnerability and fingerprint-based exploits, early occupation, lateral movement, and final invasion through target point privilege elevation and takeover.
  • θ ( S k , a x , a y , S k ) is a probability distribution function used to calculate the probability of reaching S k in the case where attacker N A performs the action termed a x and defender N D performs the action termed a y in the current episode S k .
  • R ( S k , a x , a y ) is a function used to calculate the reward that can be obtained from the player’s judgment in the episode when attacker N A and defender N D perform actions termed a x and a y , respectively, in S k . In response, players compete in the direction to maximize R ’s reward.
  • U = ( U A , U D ) is a signaling-based discount factor function. It cuts off the judgment ranges by player within [0, 1] to attenuate effects of signaling actions by player. It also simulates pre- and post-competitive strategy judgments by player in the leader-follower relationship, such as limited views and the concept of true-false shares.
As specified in the nine-tuples above, the signaling performed by each player in SOD2G is defined based on the signal game strategy to perform multistage-layer-based information transmission and BSSG. Causality of transmission and reception is determined according to the optimization directionality of payoff by player and the signaling initiative. It is terminated as a PBNE-based optimized equilibrium state.

3.4.2. Configuration and Optimization of Zero-Sum-Based Attack-Defense Competition

First, the attacker, who is a sender and leader from the viewpoint of attack actions by episode, delivers signals related to concepts of early reconnaissance and search, exploitation, privilege escalation, and host occupation to the victim host in the defender model or the defender dominant deception environment. In response to signals, the passive defender, who is a receiver and follower, dynamically performs feedback according to the concept of a vulnerable point of contact or a pre-configured deceptive sandbox that does not consider defense in advance. Next, the attacker verifies the validity and reliability of the defender intelligence and surface, view, and occupancy information possessed at the present time through the naive or deceptive response of the defender to the signals. The attacker then accordingly reconstructs its invasion strategies, attack vectors, and so on. In this case, the attacker’s judgment range is cut-off spatiotemporally according to a predefined discount factor to induce reduction of solution space and derivation of approximate values to calculate the optimal value of the total reward related to the invasion target. In addition, based on the maximum number of intrusion attempts, which is a private information-based conceptual factor to achieve the attack goal, the payoff that is inversely added by the defender dominantly due to the achievement of attacker dominant asymmetry or the attacker’ wrong judgment is calculated in terms of utility and cost or reward.
Next, the adaptive defender, who is an active deceptive sender and leader from the viewpoint of defensive actions by episode, forces defender dominant early cognitive bias based on disinformation operation to the attacker before or during the process of carrying out reconnaissance and search. As well as delivering deceptive signals to deceive the attacker so that the attacker is induced to a false environment as invasion strategy reconstruction input values in the attacker model. In response to signals, the passive attacker, who is the receiver and follower, dynamically performs the feedback to update the defender intelligence possessed by the defender dominantly without doubt. Through the response of the attacker who accepts the signals, the defender then evaluates the efficiency and scalability of the proactive MTD as well as the reactive decoy-based defensive deception strategy being followed at the present time. It additionally updates layered deception strategies and host defense elements according to results. At this time, the defender’s judgment range is also a cut-off according to a predefined discount coefficient, thus yielding a payoff approximate value that is not completely optimized by episode.
The maximization of reward through reasoning between leading players in the leader-follower relationship based on signaling by episode is organized as a Q-value as shown in Equation (10). In this case, U and TS are defined as players with signaling initiative in the current episode:
Q ( S k , a x , a y ) = R ( S k , a x , a y ) + U S k θ ( S k , a x , a y , S k ) · T S · O P T ( S k ) ,
That is, O P T ( S k ) , from the viewpoint of the attacker who is actively signaling, is expressed as Equation (11) through SS, which is the signaling action that can be performed in S k . This produces the optimized maximum reward with judgement based on incomplete and private information:
O P T ( S k ) = max S S   min a x a y Q ( S k , a x , a y ) · ( s s N d i | i = 1 , 2 , ) ,
Next, the partial signal game-based PBNE in the game environment based on OD and OA is generally summarized as shown in Equations (12)–(15) based on Equations (10) and (11). In this case, PD in Equation (12), which is similarly constructed based on O P T 2, is the prior probability-based defender’s judgment probability for S S N A -related T S N D . P D   in Equation (13) is the defender’s inference probability based on the posterior probability related to S S N D reconstructed based on the updated internal deception-defense strategy after the feedback-based signaling of the defender for S S N A . In addition, when calculating PBNE through PD and P D as such, it is affected by U, a discount factor. The PBNE state entry in game is also controlled according to the configuration of UA or UD based on whether or not the signaling leader is selected:
P D = ( p D · ( T S N D i ) | i = 1 , 2 , n ) ,
P D = P D ( ( T S N D i | i = 1 , 2 , n ) | S S N A ) ,
O D ( S S N A j ) = a r g max S S N D k S S N D T S N D i T S N D P D · F ( T S N D i , S S N A j , S S N D k ) ,
O A ( T S N A i ) = a r g max S S N A j S S N A F ( T S N A i , S S N A j , O D ( S S N A j ) ) ,
Finally, to reflect the zero-sum based attack-defense competitive game presented in SOD2G above, it is structured as an interface as shown in Figure 3 based on the state, transition, and arrival concepts in POMDP components specified in Figure 1.
Transition probability and reward setting in the POMDP are also detailed in Table 2, although they are configured differently according to pre-defined scenarios.

4. Experiments and Sensitivity Analysis

In this section, the attack-defense scenarios in the SOD2G framework are formulated according to the concept of organizational operation due to the normalization scheme of conflict of game caused by the independent Monte Carlo based on true random generation as well as the fluid intrusion probability distribution between the attacker and the defender. Along with a comparison of the efficiency of zero-sum-based attack and defensive deception according to successful intrusion and detection-induction-isolation-blocking success by node in the configured scenario, related sensitivity analysis is also performed in a general manner to ultimately calculate the relevant deception optimal value.

4.1. Definition of Experimental Bayesian Stochastic Game-Based Scenarios and Related Metrics

To experimentally verify the efficiency of defensive deception, complex metrics, attack-defense scenarios, and topologies related to the host-switch-security solution-based organizational legacy network operation are calculated as shown in Figure 4, Figure 5 and Figure 6. Each scenario is formulated in detail with operational equipment and topology structure diagrams, spatiotemporal performances and costs by participant, invasion-defense discrimination criteria, CKC, attacker’s and defender’s final goals, attack-defense and deception sequence, other assumptions such as CVE and CVSS, and so on. Variables such as surfaces of attackers and defenders with limited views, counter-measure strategies, limited resources between service availability and security, real-time share competition related to successful invasions and defenses by node, disturbances due to deceptive signals, and so on are also implicitly applied by scenario and subdivided as listed in Table A1, Table A2 and Table A3.
Accordingly, the formulated organizational operation scenarios have the following performance premises in common. A universal concept of attack-defense competition interface is established based on these premises.
  • Attack Behavior Standard: The attacker performs an attack on the organization’s legacy internal network from the outside and selects either single or multiple final invasion target points to proceed with the invasion according to the episode. The selection of target points by episode differs depending on the intelligence of the defender possessed. The formation of attack chains continues in the direction that yields the highest partial gain.
  • Attacker’s execution of professional APT attacks: The attacker gives the top priority to the final occupation of the defined single or multiple target points and maximizes the attack gain while optimizing all possible attack-based branches without any fixed attack path. The attacker also has the flexibility to select another host as the next best attack target such that invasion origins can change by episode immediately before the failure. Lateral movement can also be activated along with a change in the direction of movement in cases where the attacker fails to achieve the first priority target. However, the attack efficiency may decrease due to the deceptive false surface information of the defender. It may be impossible to judge Decoy and Honey-X based sandbox based on biased view selection for a long time.
  • Defender behavior standard: The defender carries out monitoring, reactive response, blocking, and exit of all hosts and terminals. However, resources for the operation of the defense solution and the execution of the adaptive active deception sequence are limited. Possession of prior information on the invasion target points is also limited. If counter-measure response points are selected incorrectly, the defense will fail and the general defense gain will be reduced. The defense goal of the defender is established differentially according to importance levels and potential vulnerabilities by node in the topology. However, the defender operates with a focus on successfully defending the occupation of all nodes and finally blocking the attacker. According to a separate scenario, a host invaded by an attacker can be restored by a cloned backup copy and formulated as a node occupied by the defender. Backup copies cloned by the host are snapshots after performing certain episodes. The process of isolating and storing them in an environment that can be invaded by attackers is executed.
  • Defender’s performance of defensive deception: The defender performs MTD and Decoy-based proactive deception in the topology based on signaling. However, since only a single host and one terminal can perform reactive response to the attacker that bypasses the deception and successfully invaded, an appropriate prevention point and a response point deployment strategy are needed to maximize limited resources.
  • Reward Standard: The attacker acquires rewards by episode based on the success of the CKC stage and the defender acquires rewards based on the attacker’s CKC protection success-exit-punishment. Through the foregoing, the reward is optimized into the maximum total reward value at the end of the final simulation.
  • Constantization and variabalization of rewards: The concept of rewards in the SOD2G framework is a constant. It is established based on CVE and CVSS. However, according to the scenario and topology structure selection, it is also converted into a variable that is different according to the attacker’s APT level and the defender’s deception-response level. The probability of transition between states is also set as a constant. However, the attack-defense level is fixed at a certain level. The constant can also change dynamically depending on the present situation of the state of attack-defense shares.
  • Encouraging rapid decision making: The micro-decision gradient is maximized according to successive spatiotemporal changes through a discount coefficient to prevent the fixation of the macro-equilibrium state. That is, the attacker and defender cannot indefinitely postpone the current decision for more than tens of episodes. Branch calculation space for obtaining the maximum reward at the present time is also limited in scope. This is carried out with a cut-off MILP optimization scheme.
  • Scenario ending condition: The attacker’s ultimate goal is to perform a CKC-based action of object (AoO) by reaching the currently established single or multiple invasion target point after the final exploit. The defender’s ultimate goal is defined as depleting the attacker’s APT attack attempts and related attack resources to force the attacker to completely exit from the current organization’s internal network, finally punishing the attacker. Upon completion of the simulation, rewards by episode and attack-defense probability values are returned.
  • Addition of the concept of security solution: unified threat management (UTM) is an integrated secondary security device in which IDS and IPS are mixed. It is performed for the purpose of attacker threat detection, blocking, and expelling. Firewall is a form of primary security equipment for user access control and the judgement of whether a user is legitimate, which is performed for the purpose of user validation, authentication, and authorization. In this way, security solutions are also upgraded, with a focus on horizontal diversity or vertical redundancy. In this case, potential vulnerabilities to define the APT attacker to penetrate from an external network to the organization’s internal network through relevant security solutions are assumed as CVE and CVSS. Unlike other hosts, the attacker’s acquisition of administrator privileges in security solution is limited not to be carried out.
Based on the foregoing, attack-defense scenarios are established as follows. Independent focus points related to social engineering deception and organizational legacy operation strategies are also defined in detail.
  • Scenario 1: ‘Organizational open network based on a single security solution in a non-critical domain’
    (1)
    Host configuration: Local mail host (Win7-1), remote mail host (Win7-2), mail server (Win server 16), DB server (Debian), and internal network DNS server (Win server 12).
    (2)
    Security equipment configuration: Integrated single UTM with some firewall functions.
    (3)
    Attacker’s goal: The best goal is to achieve DB hijacking-destruction after PFTPd remote code-based privilege escalation and acquire DB administrator privileges. The second-best goal is to carry out differential intrusion behaviors by host according to the level of vulnerability when it is judged that Debian-FTP cannot be compromised.
    (4)
    Defender’s goal: After securing the protection of hosts and terminals in all topologies, depleting all resources possessed by the APT attacker to achieve complete exit from the internal network.
    (5)
    Major attack sequence: Outside the topology → Attempt to break into the internal network → UTM infiltration and bypass → Intrusion and occupation of many other hosts are carried out concurrently when immediate access to Debian-FTP and search is not possible, → Intrusion of Debian-FTP and achievement of final occupation.
    (6)
    Major defense sequence: Perform proactive deception and monitoring of the entire topology → Perform attacker intrusion detection, reactive blocking, and exit → Achieve complete exit to the outside.
  • Scenario 2: ‘Independent multi-security solution-based non-critical organizational semi-closed network’
    (1)
    Host configuration, defender’s goal, and major defense sequence: same as [Scenario 1].
    (2)
    Security equipment configuration: horizontally diversified UTM and firewall.
    (3)
    Attacker’s goal: The best goal is to elevate privileges through heap buffer overflow remote code execution attack and obtain DNS server administrator privileges and achieve worm-centered network intra-control point occupation. The second-best goal is to conduct differential intrusion behaviors by host according to vulnerability when it is judged that malicious invasion through DNS service query in Win server 12 is impossible.
    (4)
    Major attack sequence: Outside the topology → Attempts to break into the internal network → Infiltration and bypass according to UTM or firewall function → Intrude and occupy many other hosts concurrently when immediate access to Win server 12 and search is impossible, → Win Server 12 invasion and achievement of final occupation.
  • Scenario 3: ‘Closed organizational network based on triple security solution in critical domain’
    (1)
    Host configuration: Additional definition of duplicated backup copies based on cloning of hosts in [Scenario 2]. This includes all live and unlive hosts.
    (2)
    Security equipment configuration: Three vertically redundant integrated UTMs and three firewalls.
    (3)
    Attacker’s goal: The best goal is to obtain DNS server administrator authority through a heap buffer overflow remote code execution vulnerability attack, occupy the worm-focused network intra-point of control, and achieve complete destruction of topology restoration points. The second-best goal is the same as the second-best goal in [Scenario 2].
    (4)
    Defender’s Goal: Same as [Scenario 2].
    (5)
    Major attack sequence: Outside the topology → Attempt to break into the internal network → Temporarily break through UTM or Firewall terminal configured based on triplet → Win server 12 DNS Major invasion and Backup image neutralization and destruction → Achieve final occupation.
    (6)
    Major defense sequence: Carry out proactive deception and monitoring of hosts and terminals in the entire topology → Carry out additional monitoring based on multiplexing and host image integrity guarantee at set time intervals → Carry out attacker intrusion detection, reactive blocking-exit, and normal restoration to a random snapshot → Achieve complete exit to the outside.
MITER CVE vulnerability and CVSS scores constantized and calculated by scenario are structured as detailed in Table 3 and converted into a DB. Each vulnerability was selected for the ease of performing zero-sum-based requests and feedback of attackers and defenders based on limited views and surface knowledge possessed at any point in the CVE list that yields high CVSS focusing on Windows and Linux OS. These vulnerability scores are also applied as standard indexes when quantitative rewards by episode are subtracted or added. In addition, it is formulated in detail as a potential attack point based on weak services and protocols based on Ipv4 and socket ports for important hosts such as DNS servers to induce intrusion behaviors.
Both attack graph based on the CVE-based attack contact point and defense graph of the defender are conceptualized as adaptive mutations according to the network and host-based organizational OSINT with the same deception sequences as shown in Figure 7. In this case, directly mutating the unique authorized network band group used for normal service supply as direct mutation targets of MTD is unsuitable for all continuous organization service provision, channel migration, and communication operation schemes between internal network servers. Accordingly, the virtual network communication channel is additionally expanded based on the network fingerprint range in OSINT-based Decoy promised in advance at random mutation time intervals between internal users or certain permitted outsiders. MTD mutation targets are then selected. In addition, a rule table is formulated so that the public network information given to the host and the false virtual network information can communicate with each other in a dynamic pair structure. The concept of judging whether or not the dynamic shifting of virtual network information should be performed according to changes in the external entropy such as the occupation of hosts following the attacker’s success in partial invasion is also applied.
The deception sequence corresponding to Figure 7a is the reactive sequence related to MTD and decoy deceptions by host for cases in which access to the legitimate server is requested based on the network information group such as IP, port, network-layer service, etc. previously collected by the attacker. If the network specification in the packet requested by the attacker is not permitted at the present time point, false internal surface or vulnerability information is transmitted to the attacker through deceptive signaling to induce a cognitive judgment in the episode to be biased to the defender dominantly. In addition, to intentionally deceive the attacker to easily possess an isolated sandbox and drift, the defender artificially creates a false network information group corresponding to the information group requested by the attacker in the direction to minimize the attacker’s suspicion, inserts it into the response packet, and transmits it. As a result, the deceived attacker gradually continues acting in the direction that maximize the defender’s gain in a relevant episode and therefore loses the possibility to reach its final goal. Figure 7b is similar to, but different from Figure 7a in that it is a reactive deception sequence through which the attacker does not attempt attacks with IP or port-based network information, but instead executes attacks in units of application-layer service such as protocols (e.g., SSH and FTP), SMB and NetBIOS-based internal message sharing, RDP-based remote monitoring, and so on.

4.2. Results

This section details results of the MTD and Decoy-based attack-defense simulation experiments performed by scenario within the proposed SOD2G. A comparative analysis between different studies conducted by major metrics is also performed. Individual scenarios and attack-defense sequences are characterized with a focus on sub-party-based non-critical OSINT specifications as listed in Table A1. Overall parameters related to the experiment are also established based on Table A2 and Table A3.

4.2.1. Comparative Results of Attacker-Defender Game Competition with Strategies

Figure 8, Figure 9 and Figure 10 show result sets normalized by POMDP status in Figure 3 based on MTD and Decoy sampling and related signaling techniques customized based on existing MTD deception and distribution policies [21,35,46] for the degree of success of defense according to changes in the defender’s final reward based on discount coefficients by scenario, as well as organizational mix newly added with organizational OSINT. Figure 11 shows a set of results normalized for the degree of success of attacks according to the discount coefficient-based attacker’s reward.
Figure 8a shows the degree of defense success efficiency of the defender in Scenario 1 based on the s0 state according to the discount factor that cuts off the range of judgment by player surfaced by episode to force them into limited microscopic views. Through the foregoing, to achieve the occupation of the final target invasion point or failure to reach as a result of the attacker finally being weeded out after reaching the maximum number of invasion attempts, the defender’s total reward decreases exponentially as discount factor increases. The decreased total reward is always the same regardless which deceptive technique is additionally applied as a major deceptive strategy. In this case, the reward shows a tendency to linearly decrease until the discount factor reaches 0.7. It is calculated to be −29, −40, and −67 when discount factors are 0.3, 0.5, and 0.7, respectively. It decreases rapidly and exponentially as the discount factor increases beyond 0.7. Thus, extreme approximate values are derived, such as −198 when the discount factor is 0.9. That is, in a competitive environment, the lower the minimum discount factor value for the defender to force the attacker to make quick judgments by episode, thereby securing predominance, the more advantageous the discount factor value is to the defender. This indicates that the discount factor should be optimized not to exceed 0.7, even if it is increased by the attacker. This tendency proves that when the degree of time-based deceptive signaling, which has been calculated to actively attenuate the spatiotemporal allowance for the attacker to determine the optimum branch in the presence of the defender while minimizing the formulation of effective attack surface information, is higher and the time interval for mutation is shorter, the attacker’s gain can be suppressed more and asymmetric inferiority can be forced further. This is also the case for Figure 8b–d, which are sets of partial comparison results for dynamic states other than s0. Finally, if the efficiency of the deceptive signaling scheme, which biases the attacker’s early cognitive judgment to the defender’s advantage, is improved as a single parameter, it is necessary to control the discount factor such that it does not exceed 0.7 for the position of the attacking player. Conversely, for the position of the deceptive defensive player, it is necessary to expand the market share in the competitive environment based on false signaling and feedback to increase the discount factor above 0.7.
Figure 8b–d with the same discount factor metric show defense success efficiency of the defender in Scenario 1 based on states s1, s2, and s3, respectively.
First, in Figure 8b, in the s1 state related to a single UTM deployment with improved network monitoring and access control functions, it was revealed that even if the discount factor that gave the attacker a high asymmetric predominance increases, defense success efficiency values of deception techniques except for Uniform Random did not decrease much. Discount factor values converged from 0.9 to between −20 and −30. In this case, Mixed derived an improvement in defense efficiency by 126% at the maximum and 21% at the minimum compared to Min-Max, which was the performance measure baseline. Organizational Mixed also produced improvements in defense efficiency by about 58% at the maximum and 16% at the minimum compared to Mixed. This tendency proves that since Uniform Random, which determines a single host as the target of concentration of the defender’s deceptive signaling and monitoring and reactive response with a uniform probability distribution, distributes invasion probability to upper entities that transmit unique and high ripple effects such as the UTM in Scenario 1 in an unconditionally equal manner, even for those that do not execute attacks, the attacker cannot judge that it can carry out an invasion in earnest only after infiltrating and breaking in through the defense solution. In addition, since Mixed and Organizational Mixed represent mixtures of Min-Max and Uniform Random and thereafter apply and strategize different probability distributions to different hosts according to degrees of changes in entropy such as the attacker’s invasion preference, observable-determinable topology state, and the attacker’s environment, they have further improved the deception defendability. It was also proved that Organizational Mixed derived an additional performance improvement compared to Mixed since it added more subdivided deceptive elements based on network, host, and service layers. It additionally secured substantiality as OSINT. This was similar in Scenario 2 of Figure 9 and Scenario 3 of Figure 10, except for the special tendency of Uniform Random. Finally, if it is intended to minimize the decreased efficiency of the deceptive defense for a single security solution between the internal network and the external network, the share should be ensured such that the defender’s inferior discount factor is also configured to be more than 0.75 but lower than 0.8 without selecting Uniform Random.
Next, in (c), although Uniform Random showed the most inefficient defense success degree even in the s2 state related to the achievement of a weak invasion through the early search in the topology and the first exploit, unlike before, it was revealed that the defense success efficiency rapidly decreased from when the discount factor reached 0.8 to converge between −55 and −75. In this case, Mixed derived an improvement in defense efficiency by 48% at the maximum and by 3% at the minimum compared to Min-Max. Organizational Mixed also resulted in improvements in defense efficiency by 22% at the maximum and 10% at the minimum compared to Mixed. This tendency proves that the efficiency of Uniform Random is reduced drastically due to the unique high ripple effect based on the internal communication channels and shared directories installed by the host on the Linux DB server, which is the final invasion target point of the attacker, along with the fact that the degree of specificity is high for both improved Mixed and Organizational Mixed. This is also the case in Scenario 2 of Figure 9 and Scenario 3 of Figure 10. Therefore, if it is intended to minimize the possibility of success of internal reconnaissance and early exploitation by the APT attacker who is performing initial surfacing to improve attack efficiency after the initial intrusion into the organization’s internal network, the attacker should be perturbed such that Uniform Random is not selected and that the attacker dominant discount factor is configured to be lower than 0.75.
Finally, (d), that is, the s3 state related to the achievement of the final invasion goal after the lateral movement and serial chain advancement following the initial exploit, has almost the same pattern as (c). Mixed showed an improvement of defense efficiency by 121% at the maximum and 3% at the minimum compared to Min-Max. Organizational Mixed also yielded improvements of defense efficiency by 62% at the maximum and 14% at the minimum compared to Mixed. Finally, if it is intended to highly prevent and actively block the achievement of the final invasion by the attacker who performs lateral movement and multiple rough privilege elevation attacks based on the contact point after the success of initial occupation in the organization’s internal network, it is judged that a large number of deceptive signals should be rapidly projected and made into noise without selecting Uniform Random, similarly to (c), so that the attacker dominant discount factor is configured to be equal to or higher than 0.75 but lower than 0.8.
Since Figure 9 and Figure 10, respectively, simulated competition based on Scenario 2 and Scenario 3, wherein concepts of multiple security solutions, hosts, and final invasion target points, to which a single ripple effect was distributed, were defined differently, a tendency for Min-Max (not Uniform Random) to show the lowest defense efficiency was commonly derived, unlike in Figure 8. The superiority relationship of the defense efficiency of other deception techniques was the same as that shown in Figure 8.
Figure 9a shows the degree of defense success efficiency of the defender in Scenario 2 based on the s1 state. In this case, after the discount factor reached 0.7, the defense success efficiency levels of all deception techniques decreased sharply and drastically to converge on −75 to −150 at 0.9. In addition, Mixed showed improvements in defense efficiency of 101% at the maximum and 4% at the minimum compared to Uniform Random. Organizational Mixed also yielded improvements in defense efficiency by 38% at the maximum and 7% at the minimum compared to Mixed. This tendency was attributable to the fact that as the single UTM solution in the previous scenario 1 was horizontally multiplexed as multiple UTM and firewall solutions, the ripple effect concentrated on lower-level hosts was dispersed such that a uniform distribution of invasion probability of Uniform Random was feasible. This was also proved by the fact that the final point of invasion was changed to a DNS server host that probabilistically distributed the transition ripple effects by host through mutual communication with security solutions other than the Linux DB server and an external publicly approved DNS service. Through the foregoing, it can be seen that to minimize the attenuation of defensive deception efficiency for vertically-multiplexed security solutions, the real share in the topology should be secured such that the attacker dominant discount factor is configured to be lower than 0.7 without selecting the Min-Max.
Next, (b) shows defense success efficiency in Scenario 2 based on the s2 state. Mixed improved defense efficiency by about 51% at the maximum and 14% at the minimum compared to Uniform Random, whereas Organizational Mixed improved defense efficiency by about 36% at the maximum and 23% at the minimum compared to Mixed. Finally, the method to actively reduce the possibility of attack success in the attacker’s first intrusion and initial exploitation attempt in the organizational topology of the DNS server host should be configured similarly to (a) such that the discount factor is maintained between 0.7 and 0.75.
(c) also shows defense success efficiency in the s3 state based on Scenario 2. It was judged that related countermeasures could be formulated, such as configuring the discount factor to be lower than 0.65 in an approach similar to (a) and (b).
Results of SOD2G experiments based on Scenario 3 related to the operation strategies of multiple UTM and firewall solutions triplicated based on vertical redundancy, horizontal diversity, and hosts considering restorability and resilience by cloning candidate images based on snapshots are compared and analyzed. Result sets are shown in Figure 10.
Figure 10a shows defense efficiency in the s1 state related to the deployment and operation of triplicated UTM and firewall solutions. In this case, Organizational Mixed improved defense efficiency by 39% at the maximum and 7% at the minimum compared to Mixed and converged as a sharp negative gradient when the discount factor was in the range of 0.75 to 0.8, whereby it was finally derived to be between −60 and −135. The tendency as such proved that the triplicated security operation environment was selected due to the fact that when an invasion had occurred, the defender’s risk in this environment was lower than in other environments while the attacker’s high robustness for solution search and breakthrough could be ensured.
Figure 10b shows the degree of defense efficiency in the s2 state related to the attacker’s initial reconnaissance and exploitation attempts on hosts inside the organization that have cloned snapshot images. Organizational Mixed achieved improvements in defense efficiency by 44% at the maximum and 21% at the minimum compared to Mixed. Defense efficiency sharply decreased when the discount factor was in the range from 0.75 to 0.8, finally converging to be between -90 and -170. This tendency proves that even if the restorability and resilience level by host are improved with redundancy-based additional snapshots, if invasion and occupations are performed within the length of the time interval for snapshot configuration, the environment will revert to an environment similar to the defender intelligence already possessed by the attacker such that the defender information validity remains preserved.
Figure 10c shows the degree of defense efficiency in the s3 state related to the attacker’s final invasion and achievement of the target point. Organizational Mixed derived improvement in defense efficiency by 59% at the maximum and 28% at the minimum compared to Mixed. Defense efficiency sharply dropped when the discount factor was in the range from 0.65 to 0.7, finally converging to be between −110 and −190. Through the foregoing, it is judged that internal deceptive signaling to secure a discount factor lower than 0.6 should precede in order to prevent an attacker from gaining predominance and adding rewards in a closed network in which security has finally been secured based on redundancy and diversity.
Figure 11 shows that the attack success efficiency improvement is derived based on the positive gradient as the attacker dominant discount factor increases based on the surfaced defender information through search and reconnaissance in Scenario 3. Combined with the fact that the sum of the attacker’s total reward and the defender’s total reward is always zero, it can therefore be concluded that attacks and defenses are always performed on a zero-sum basis.

4.2.2. Sensitivity Analysis by Deceptive Metrics

Next, sensitivity analyses were carried out in detail using major metrics for the concepts of LPC-MTD and HS-Decoy-based social engineering-type organizational deceptive defense. Figure 12 shows sets of analysis results regarding the possibility of success of deceptive signaling based on the increase in surface uncertainty and limited effectiveness following the defender’s active deception normalized by scenario. Figure 13 and Figure 14 show sets of results obtained by reclassifying metrics in Figure 12 into MTD and Decoy-based detailed metrics and normalizing them.
First, possibilities of success of deceptive signal projection in Figure 12a–c are probabilistic comparison metrics that strategize the active deception events that deceive the occupancy and surface by disrupting the attacker’s perception related to the invasion target point in the defender’s favor, or by inducing the attacker to a false host occupation environment with detailed deception elements such as MTD mutation-based disinformation and artificial exposure, Decoy-based cloning and mimicking, and OSINT information group, which are associated with the proactive defense and signaling component in Figure 2. These detailed elements are composed of detailed strategies in Organizational Mixed. Based on the foregoing, it can be seen that compared to the existing Mixed, Organizational Mixed achieved at least 23% higher deception efficiency by increasing the uncertainty of the surface information by host and reducing the effectiveness. It can achieve the stage at which it completely deceives the attacker earlier at the deception surfacing value of 0.7. This shows that since the existing Mixed simply focuses on selecting a transition path that maximizes the reward to be obtained in the next episode, and since the OSINT-based disinformation, artificial exposure concept, social engineering decoy, etc. to obtain the highest final reward even if some loss should be born have yet to be composed, and the relevant detailed deception has not been clearly defined in the presence of the defender, Organizational Mixed considering all of them exhibits higher deception efficiency. This also proves that when establishing deception policies by host, policies should not be established for naïve application to entire entities. They should be established for application to entities characterized as detailed entities layered with IPv4 and port, socket, network service, application services, and so on. Through these characteristics, it was finally revealed that Organizational Mixed had higher deception performance in SOD2G than the existing Mixed.
The possibility of MTD-based deception in Figure 13a–c of is a probabilistic comparison metric wherein the degree to which the attacker’s attack attempts are induced is atomically normalized so that the final invasion target host in the topology surfaced based on CVE vulnerability or decoy beacon in an episode where the LPC-MTD or OF-RHM-based existing MTD scheme is implemented toward defender dominance is isolated from the actual host with false information surfaced based on OSINTs by organization. This comparison metric is also associated with the proactive defense and signaling behavior component in Figure 2. Based on these correlation metrics, it can be seen that Organizational LPC-MTD has a five times higher possibility of deception through disinformation and artificial information disclosure on average compared to the existing OF-RHM as well as at least three times higher signaling-based intentional exposure rate on false defender surfaces involving OSINT on average. This demonstrates that, unlike the existing MTD, LPC-MTD reactively lowers the mutation strength according to the analyzed attacker’s attack routine to deceive the attacker by forcing a judgment such that it can successfully bypass the MTD along with the fact that the related reliability can be improved with disinformation and artificial information exposure based on the deceptive OSINT layered referring to the target host. That is, it can be also identified that the OF-RHM scheme is constructed to only have deception possibilities not exceeding 31% for Scenario 1, 68% for Scenario 2, and 97% for Scenario 3. However, Organizational LPC-MTD is operated in order to deceive the attacker while always maintaining 100% convergence. When the attacker carries out invasion, it is quickly reconstructed with a separate deceptive surface that is not much related to the previous surface and restored to converge on 100%. The foregoing finally shows that LPC-MTD has a higher deception possibility in SOD2G than existing OF-RHM.
Finally, the possibility of decoy-based deception in Figure 14a–c is represented by probabilistic comparison metrics obtained by normalizing the degree to which the attacker can be induced and thereafter maximally isolated under the intention of the defender according to true-false occupancy rates by player along with the concept of a limited view in an episode where HS-Decoy and the existing naive Decoy scheme are created, distributed, and managed. This comparison metric is associated with the proactive defense, signaling, and reactive defense components shown in Figure 2. Based on the foregoing, it can be identified that in the case of Organizational HS-Decoy, the possibility of deception through OSINT-based cloning and mimicking is 37% higher on average compared to the existing naive Decoy and the degree of improvement of decoy quality utilizing the actual characteristics of any legitimate host within the topology is at least 163% higher on average. A positive gradient is always maintained. This is based on the fact that unlike naive Decoy, in the case of HS-Decoy, the number of deceptive entities is large and the structure is stratified based on network, host, and service such that the degrees of redundancy and diversity are enhanced. The reliability is also enhanced based on OSINTs by organization. This can be confirmed through results in the condition wherein deception efficiency gaps according to the execution times by scenario linearly increase. This ultimately shows that HS-Decoy has a higher deception possibility in SOD2G than existing naive Decoy.

4.2.3. Optimized Values with Organizational Deceptive Game Models

Through the aforementioned Section 4.2.1 and Section 4.2.2, overall optimized values for maximizing the defensive deception efficiency of the defender by game, deception metric, and related parameter in the SOD2G framework as well as minimizing the asymmetric advantage of the attacker are simplified as detailed in Table 4.

5. Discussion

In this study, PBNE, BSSG, and POMDP-based SOD2Gs were proposed to improve the practical operability of defensive cyber deception and optimize defender’s deception strategies against attackers by organization. In addition, the attack-defense decision-making process of mixed LPC-MTD which selects the MTD mutation strength with the advantage of the defender and HS-Decoy, which applies a deception concept with the goal of deceiving the attacker based on OSINT, are diversified by scenario. Through the foregoing, deception attack-defense efficiency levels optimized by organization could be calculated based on game equilibrium. Independent defensive deception values necessary to minimize operational performance degradation and maximize security without additional procedures or separate application of dedicated protocols could also be formulated by scenario.
First, when the defense success efficiency against the discount factor for the attacker’s advantage was composed as a main comparison index and summarized, the organizational mixed showed an average performance improvement of 40% or more compared to the mixed by scenario. To minimize attacker dominance in the topology, the attacker discount factor should also be suppressed so as not to exceed 0.7 on average. Next, in the sensitivity analysis for optimizing game models in SOD2G, LPC-MTD, and HS-Decoy based on OSINT yielded improvements of at least 60% and 30% in deception efficiency, respectively, compared to the existing model. Optimal values by metric and by parameter were also produced by empirical calculations through comparative analysis.
However, the following improvement measures are needed to address limitations of this study.
  • Reliability and Substantial Issues: To prevent both potential legal disputes and research ethics problems, characteristics by organization intellectualized in the game should be pre-processed and extracted by element within the limited range in the third party-based public OSINT information group. However, this will soon be different from the unique policies and principles related to the actual organization, leading to clear practical differences from the organizational environment where open and closed networks are operated separately. In addition, even when OSINTs are collected and formulated, since individual organizations have different numbers of elements, vertical-horizontal relationships, ripple effects, and vulnerability, additional augmentation and normalization should be performed.
  • Scalability and operability issues: Since the proposed study conducted experiments focusing on minimizing the degradation of availability in the defender environment and maximizing the defender predominance based on disinformation and artificial information exposure, scalability issues might occur in scenarios or domains that are not considered. Therefore, rather than concretizing the attack-defense probability value at the scenario level, the probability value should be abstracted at the organizational domain classification level and the evaluation verification process should be newly constructed.
  • Issue of additional demand for strategy based on hyper-game: In an actual attack-defense environment, players can subjectively process and determine asymmetric dominance or subordinance information that is accessible at the present time point. This proves that under the premise that players have consistent views and that they cannot be represented solely by a naive game theory that models dynamic decision-making in an uncertain situation. This study tried to alleviate some conflicts of views between players by conceptualizing the disinformation-based partial signaling game tactic and dynamically changing the state-transition probability according to changes in the attack-defense state based on the deceptive signal. However, related potential side effects will still remain. Accordingly, considering recently studied hyper games [33], it is necessary to quantify decision-making when selecting the best strategy by considering all players’ subjective or wrong beliefs and their perceived uncertainty at any point of time. A strategy set selection calculation method that processes a large number of solution spaces required for modeling and obtains an optimal solution should also be configured separately.

6. Conclusions

In this study, a PBNE and BSSG-based zero-sum game foreground and POMDP state-transition background-based SOD2G characterized as organizational OSINTs were proposed. To this end, the decision-making process was non-sequentially stratified based on an incomplete and deceivable organizational environment and vulnerability information, such that it could construct proactive deception considering the organizational domain and simulate the related deception efficiency improvement. Defender deception strategies to achieve both high reactive response rates and low false-positive and false negative rates could also be calculated to converge on optimal values by scenario. In addition, to improve the practicality and operability related to the customized deception concept, it was possible to quantify the organizational game model based on the scenario. In future studies, to secure the empirical reliability of deception efficiency levels of the proposed SOD2G and the detailed model, and to formulate sophisticated cyber threat intelligences by organization, attack-defense scenarios should be constructed based on operational behaviors to diversify all related deceptive countermeasures.

Author Contributions

Conceptualization, S.S.; methodology, S.S.; software, S.S.; validation, S.S. and D.K.; formal analysis, S.S. and D.K.; investigation, S.S.; resources, S.S.; data curation, S.S.; writing—original draft preparation, S.S. and D.K.; writing—review and editing, S.S. and D.K.; visualization, S.S.; supervision, D.K.; project administration, S.S. and D.K.; funding acquisition, D.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by a grant (No. NRF-2020R1F1A1068774) of the National Research Foundation of Korea (NRF) funded by the Korea government (MEST).

Acknowledgments

This research was funded by a grant (No. NRF-2020R1F1A1068774) of the National Research Foundation of Korea (NRF) funded by the Korea government (MEST) and also supported by Kyonggi University’s Graduate Research Assistantship 2021.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Supplementary Data

Table A1. Overview of organizational deception knowledge with OSINT for MTD and decoy-based game competitions.
Table A1. Overview of organizational deception knowledge with OSINT for MTD and decoy-based game competitions.
Element of Deception KnowledgeDescription and Related Examples
Network topology and infrastructurePublic or private network infrastructure
ex) firewall-router-DMZ-NAC-host-service
Network deviceDevice for internal and external communication in organizational infrastructure ex) CISCO ISR router, CISCO catalyst switch
Security solutionSolution for detecting, blocking and deceiving attackers in topology and hosts
ex) Sophos XG firewall, Sophos SG UTM, Splunk SOAR (security orchestration, automation and response)
IPv4 addressPublic or private IPv4 with prifix and subnet rule
ex) 11.19.0.0/16, 137.0.0.0/13, 205.79.192.0/19
PortPublic socket for inbound and outbound communication with network service
ex) well-known, registered, and dynamic on 0~65535
OSOperating system fingerprint by each host
ex) Windows 7 SP1, Windows Server 2012 and 2016, Debian, Rocky Linux
ServiceProtocol-based services in network and system-layer
ex) FTP, SSH, DNS, NetBIOS, SMTP, RPC, HTTP/HTTPS
VulnerabilityCVE and CVSS 2.0-based vulnerabilities in topology and hosts
ex) CVE-2017-0144, CVE-2020-1350, CVE-2015-3306
Table A2. Examples of deceptive MTD and decoy-based major parameters with defender model in SOD2G.
Table A2. Examples of deceptive MTD and decoy-based major parameters with defender model in SOD2G.
MTD ParameterValueDecoy ParameterValue
Time slot length for periodic mutation (s)1–100,000Activation time (s)0–259,200
Mutation batch pool size64–2048Number of decoying hosts0–10
Bellman-based mutation sampling size8–2048Number of decoying services0–40
Number of surface views in topology0–100Number of decoying vulnerabilities0–20
Maximum number of branches in attack graph10Number of decoying beacons0–50
Maximum number of deceptive signaling50–200Number of decoying signals1–10
Mutation range of security solutions0–6Level of vulnerabilitiesL-M-H
Mutation range of IPv4 addresses28–236Maximum number of compromised decoys0–3
Mutation range of port numbers210–216Maximum number of defender sandboxes0–1
Mutation range of OS fingerprints0–10Degree of OSINT-based cloning scheme0.01–0.90
Mutation range of services0–20Degree of OSINT-based mimicking scheme0.01–0.80
Mutation range of vulnerabilities0–10Degree of enticingness with decoy0.30–1.00
Degree of OSINT-based disinformation0.01–1.00Degree of conspicuousness with decoy0.50–1.00
Degree of OSINT-based artificial disclosure0.01–1.00Degree of variability with decoy0.10–1.00
Degree of reliability of MTD-based signal0.01–1.00Degree of differentiability with decoy0.70–1.00
Table A3. Configuration of experimental game parameter for decision making each episode in SOD2G.
Table A3. Configuration of experimental game parameter for decision making each episode in SOD2G.
ParameterValue
Simulation time (s)3600–259,200
Number of simulation run1–10
Attack time (s)1800–129,600
Defense time (s)1800–129,600
Number of scenarios3
Number of CKC phases with attacker4–7
Maximum number of attacker compromise attempts1–10
Operating systemUbuntu 20.04 LTS
LanguagePython 3.9.2 (Anaconda)
Game solverGurobi optimizer 9.0, IBM CPLEX, Google Or-tools

References

  1. Zhu, M.; Anwar, A.H.; Wan, Z.; Cho, J.-H.; Kamhoua, C.; Singh, M.P. Game-Theoretic and Machine Learning-based Approaches for Defensive Deception: A Survey. arXiv 2021, arXiv:2101.10121. [Google Scholar]
  2. Fraunholz, D.; Anton, S.D.; Lipps, C.; Reti, D.; Krohmer, D.; Pohl, F.; Tammen, M.; Schotten, H.D. Demystifying deception technology: A survey. arXiv 2018, arXiv:1804.06196. [Google Scholar]
  3. Heckman, K.E.; Stech, F.J.; Schmoker, B.S.; Thomas, R.K. Denial and Deception in Cyber Defense. Computer 2015, 48, 36–44. [Google Scholar] [CrossRef]
  4. Heckman, K.E.; Stech, F.J.; Thomas, R.K.; Schmoker, B.; Tsow, A.W. Cyber Denial, Deception and Counter Deception. In Advances in Information Security; Springer: Cham, Switzerland, 2015; ISBN 978-3-319-25131-8. [Google Scholar]
  5. Al Amin, M.A.R.; Shetty, S.; Njilla, L.; Tosh, D.K.; Kamhoua, C. Hidden markov model and cyber deception for the prevention of adversarial lateral movement. IEEE Access 2021, 9, 49662–49682. [Google Scholar] [CrossRef]
  6. Cho, J.H.; Sharma, D.P.; Alavizadeh, H.; Yoon, S.; Ben-Asher, N.; Moore, T.J.; Kim, D.S.; Lim, H.; Nelson, F.F. Toward Proactive, Adaptive Defense: A Survey on Moving Target Defense. IEEE Commun. Surv. Tutor. 2020, 22, 709–745. [Google Scholar] [CrossRef] [Green Version]
  7. Pawlick, J.; Colbert, E.; Zhu, Q. A game-theoretic taxonomy and survey of defensive deception for cybersecurity and privacy. ACM Comput. Surv. 2019, 52, 1–28. [Google Scholar] [CrossRef] [Green Version]
  8. Park, K.; Woo, S.; Moon, D.; Choi, H. Secure cyber deception architecture and decoy injection to mitigate the insider threat. Symmetry 2018, 10, 14. [Google Scholar] [CrossRef] [Green Version]
  9. Cohen, F. The Use of Deception Techniques: Honeypots and Decoys Deception. Handb. Inf. Secur. 2006, 3, 646–655. [Google Scholar]
  10. Manadhata, P.K.; Wing, J.M. An attack surface metric. IEEE Trans. Softw. Eng. 2011, 37, 371–386. [Google Scholar] [CrossRef]
  11. Li, H.; Zheng, Z. Optimal Timing of Moving Target Defense: A Stackelberg Game Model. In Proceedings of the 2019 IEEE Military Communications Conference (MILCOM), Norfolk, VA, USA, 12–14 November 2019; pp. 1–6. [Google Scholar]
  12. Wang, S.; Shi, H.; Hu, Q.; Lin, B.; Cheng, X. Moving Target Defense for Internet of Things Based on the Zero-Determinant Theory. IEEE Internet Things J. 2020, 7, 661–668. [Google Scholar] [CrossRef]
  13. Ferguson-Walter, K.; Mauger, J.; Fugate, S.; Major, M. Game theory for adaptive defensive cyber deception. In Proceedings of the 6th Annual Symposium on Hot Topics in the Science of Security, Nashville, TN, USA, 1–3 April 2019; pp. 1–8. [Google Scholar]
  14. Hu, H.; Liu, J.; Tan, J.; Liu, J. SOCMTD: Selecting Optimal Countermeasure for Moving Target Defense Using Dynamic Game. KSII Trans. Internet Inf. Syst. 2020, 14, 4157–4175. [Google Scholar]
  15. Lei, C.; Ma, D.H.; Zhang, H.Q. Optimal strategy selection for moving target defense based on markov game. IEEE Access 2017, 5, 156–169. [Google Scholar] [CrossRef]
  16. Xiong, X.L.; Yang, L.; Zhao, G.S. Effectiveness Evaluation Model of Moving Target Defense Based on System Attack Surface. IEEE Access 2019, 7, 9998–10014. [Google Scholar] [CrossRef]
  17. Zheng, J.; Siami Namin, A. A Markov Decision Process to Determine Optimal Policies in Moving Target. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, Toronto, ON, Canada, 15–19 October 2018; pp. 2321–2323. [Google Scholar]
  18. Maleki, H.; Valizadeh, S.; Koch, W.; Bestavros, A.; Van Dijk, M. Markov modeling of moving target defense games. In Proceedings of the 2016 ACM Workshop on Moving Target Defense, Vienna, Austria, 24 October 2016; pp. 81–92. [Google Scholar]
  19. Huang, Y.; Huang, L.; Zhu, Q. Reinforcement Learning for Feedback-Enabled Cyber Resilience. arXiv 2021, arXiv:2107.00783. [Google Scholar]
  20. Sengupta, S.; Kambhampati, S. Multi-agent Reinforcement Learning in Bayesian Stackelberg Markov Games for Adaptive Moving Target Defense. arXiv 2020, arXiv:2007.10457. [Google Scholar]
  21. Sengupta, S.; Chakraborti, T.; Kambhampati, S. MTDeep: Boosting the Security of Deep Neural Nets Against Adversarial Attacks with Moving Target Defense. arXiv 2019, arXiv:1705.07213v3. [Google Scholar]
  22. Qian, Y.; Shao, Q.; Wang, J.; Lin, X.; Guo, Y.; Gu, Z.; Wang, B.; Wu, C. EI-MTD: Moving Target Defense for Edge Intelligence against Adversarial Attacks. arXiv 2020, arXiv:2009.10537. [Google Scholar]
  23. Lockheed Martin. GAINING THE ADVANTAGE, Applying Cyber Kill Chain Methodology to Network Defense. Available online: https://www.lockheedmartin.com/content/dam/lockheed-martin/rms/documents/cyber/Gaining_the_Advantage_Cyber_Kill_Chain.pdf (accessed on 15 September 2021).
  24. Dunlop, M.; Groat, S.; Urbanski, W.; Marchany, R.; Tront, J. MT6D: A moving target IPv6 defense. In Proceedings of the MILCOM 2011 Military Communications Conference, Baltimore, MD, USA, 7–10 November 2011; pp. 1321–1326. [Google Scholar]
  25. Sun, J.; Sun, K. DESIR: Decoy-enhanced seamless IP randomization. In Proceedings of the IEEE INFOCOM 2016—The 35th Annual IEEE International Conference on Computer Communications, San Francisco, CA, USA, 10–14 April 2016; pp. 1–6. [Google Scholar]
  26. Heartfield, R.; Loukas, G. A taxonomy of attacks and a survey of defence mechanisms for semantic social engineering attacks. ACM Comput. Surv. 2015, 48, 1–39. [Google Scholar] [CrossRef]
  27. Sood, A.K.; Enbody, R.J. Malvertising—Exploiting web advertising. Comput. Fraud Secur. 2011, 2011, 11–16. [Google Scholar] [CrossRef]
  28. Krombholz, K.; Hobel, H.; Huber, M.; Weippl, E. Advanced social engineering attacks. J. Inf. Secur. Appl. 2015, 22, 113–122. [Google Scholar] [CrossRef]
  29. Ismail, K.A.; Singh, M.M.; Mustaffa, N.; Keikhosrokiani, P.; Zulkefli, Z. Security Strategies for Hindering Watering Hole Cyber Crime Attack. Procedia Comput. Sci. 2017, 124, 656–663. [Google Scholar] [CrossRef]
  30. Glassman, M.; Kang, M.J. Intelligence in the internet age: The emergence and evolution of Open Source Intelligence (OSINT). Comput. Hum. Behav. 2012, 28, 673–682. [Google Scholar] [CrossRef]
  31. Williams, H.; Blum, I. Defining Second Generation Open Source Intelligence (OSINT) for the Defense Enterprise; RAND Corporation: Santa Monica, CA, USA, 2018; ISBN 9780833098832. [Google Scholar]
  32. Yudovich, H.; Lavi, N.; Bittan, S.; Kahana, T.; Sela, T. Open Source Intelligence Deceptions. U.S. Patent 10,333,976, 25 June 2019. Available online: https://patents.google.com/patent/US10333976B1/en (accessed on 15 September 2021).
  33. Wan, Z.; Cho, J.-H.; Zhu, M.; Anwar, A.H.; Kamhoua, C.; Singh, M.P. Foureye: Defensive Deception based on Hypergame Theory Against Advanced Persistent Threats. arXiv 2021, arXiv:2101.02863. [Google Scholar] [CrossRef]
  34. Huang, S.; Zhang, H.; Wang, J.; Huang, J. Markov Differential Game for Network Defense Decision-Making Method. IEEE Access 2018, 6, 39621–39634. [Google Scholar] [CrossRef]
  35. Sengupta, S.; Chowdhary, A.; Huang, D.; Kambhampati, S. Moving target defense for the placement of intrusion detection systems in the cloud. In Proceedings of the 9th International Conference on Decision and Game Theory for Security, GameSec 2018, Seattle, WA, USA, 29–31 October 2018; pp. 326–345. [Google Scholar]
  36. Woo, S.; Moon, D.; Youn, T.Y.; Lee, Y.; Kim, Y. CAN ID Shuffling Technique (CIST): Moving Target Defense Strategy for Protecting In-Vehicle CAN. IEEE Access 2019, 7, 15521–15536. [Google Scholar] [CrossRef]
  37. Yoon, S.; Cho, J.H.; Kim, D.S.; Moore, T.J.; Free-Nelson, F.; Lim, H. DESOLATER: Deep Reinforcement Learning-Based Resource Allocation and Moving Target Defense Deployment Framework. IEEE Access 2021, 9, 70700–70714. [Google Scholar] [CrossRef]
  38. Zhu, Q.; Basar, T. Game-theoretic approach to feedback-driven multi-stage moving target defense. In Proceedings of the International Conference on Decision and Game Theory for Security, FortWorth, TX, USA, 11–12 November 2013; pp. 246–263. [Google Scholar]
  39. Mireles, J.D.; Ficke, E.; Cho, J.H.; Hurley, P.; Xu, S. Metrics towards measuring cyber agility. IEEE Trans. Inf. Forensics Secur. 2019, 14, 3217–3232. [Google Scholar] [CrossRef] [Green Version]
  40. Ge, L.; Yu, W.; Shen, D.; Chen, G.; Pham, K.; Blasch, E.; Lu, C. Toward effectiveness and agility of network security situational awareness using moving target defense (MTD). In Sensors and Systems for Space Applications VII; International Society for Optics and Photonics: Bellingham, WA, USA, 2014; Volume 9085, p. 90850Q. [Google Scholar]
  41. Carter, K.M.; Okhravi, H.; Riordan, J. Quantitative Analysis of Active Cyber Defenses Based on Temporal Platform Diversity. arXiv 2014, arXiv:1401.8255. [Google Scholar]
  42. Carter, K.M.; Riordan, J.F.; Okhravi, H. A game theoretic approach to strategy determination for dynamic platform defenses. In Proceedings of the First ACM Workshop on Moving Target Defense, Scottsdale, AZ, USA, 3–7 November 2014; pp. 21–30. [Google Scholar]
  43. Hasan, M.M.; Rahman, M.A. Protection by Detection: A Signaling Game Approach to Mitigate Co-Resident Attacks in Cloud. In Proceedings of the 2017 IEEE 10th International Conference on Cloud Computing (CLOUD), Honolulu, HI, USA, 25–30 June 2017; pp. 552–559. [Google Scholar]
  44. Feng, X.; Zheng, Z.; Cansever, D.; Swami, A.; Mohapatra, P. A signaling game model for moving target defense. In Proceedings of the IEEE INFOCOM 2017—IEEE Conference on Computer Communications, Atlanta, GA, USA, 1–4 May 2017; pp. 1–9. [Google Scholar]
  45. Zhu, Q.; Clark, A.; Poovendran, R.; Basar, T. Deceptive routing games. In Proceedings of the 2012 IEEE 51st IEEE Conference on Decision and Control (CDC), Maui, HI, USA, 10–13 December 2012; pp. 2704–2711. [Google Scholar]
  46. Sengupta, S.; Vadlamudi, S.G.; Kambhampati, S.; Doupé, A.; Zhao, Z.; Taguinod, M.; Ahn, G.J. A game theoretic approach to strategy generation for moving target defense in web applications. AAMAS 2017, 1, 178–186. [Google Scholar]
  47. Li, H.; Shen, W.; Zheng, Z. Spatial-Temporal Moving Target Defense: A Markov Stackelberg Game Model. arXiv 2020, arXiv:2002.10390. [Google Scholar]
  48. Manadhata, P.K. Game Theoretic Approaches to Attack Surface Shifting. In Moving Target Defense II; Springer: New York, NY, USA, 2013; pp. 1–13. [Google Scholar]
  49. Bilinski, M.; Ferguson-Walter, K.; Fugate, S.; Gabrys, R.; Mauger, J.; Souza, B. You only lie twice: A multiround cyber deception game of questionable veracity. In Proceedings of the 10th International Conference on Decision and Game Theory for Security, GameSec 2019, Stockholm, Sweden, 30 October–1 November 2019; pp. 65–84. [Google Scholar]
  50. Durkota, K.; Lis, V.; Bo, B. Optimal Network Security Hardening Using Attack Graph Games. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 526–532. [Google Scholar]
  51. Anwar, A.H. A game-theoretic framework for dynamic cyber deception in Internet of Battlefield Things. In Proceedings of the 16th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, Houston, TX, USA, 12–14 November 2019; pp. 522–526. [Google Scholar]
  52. Pawlick, J.; Zhu, Q. Deception by Design: Evidence-Based Signaling Games for Network Defense. arXiv 2015, arXiv:1503.05458. [Google Scholar]
  53. Pawlick, J.; Colbert, E.; Zhu, Q. Modeling and Analysis of Leaky Deception using Signaling Games with Evidence. IEEE Trans. Inf. Forensics Secur. 2018, 14, 1871–1886. [Google Scholar] [CrossRef] [Green Version]
  54. Mohammadi, A.; Manshaei, M.H. A Game-Theoretic Analysis of Deception over Social Networks Using Fake Avatars. In Proceedings of the 7th International Conference on Decision and Game Theory for Security, GameSec 2016, New York, NY, USA, 2–4 November 2016; pp. 382–394. [Google Scholar]
  55. Casey, W.; Morales, J.A.; Mishra, B. Compliance Control: Managed Vulnerability Surface in Social-Technological Systems via Signaling Games. In Proceedings of the 7th ACM CCS International Workshop on Managing Insider Security Threats, Denver, CO, USA, 16 October 2015; pp. 53–62. [Google Scholar]
  56. Casey, W.; Andre, J.; Wright, E.; Zhu, Q.; Mishra, B. Compliance signaling games: Toward modeling the deterrence of insider threats. Comput. Math. Organ. Theory 2016, 22, 318–349. [Google Scholar] [CrossRef]
  57. Nan, S.; Brahma, S. Behavioral Cyber Deception: A Game and Prospect Theoretic Approach. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Waikoloa, HI, USA, 9–13 December 2019; pp. 5–11. [Google Scholar]
  58. Nan, S.; Brahma, S.; Kamhoua, C.A.; Njilla, L.L. On Development of a Game-Theoretic Model for Deception-Based Security. In Modeling and Design of Secure Internet of Things; IEEE: Piscataway, NJ, USA, 2020; pp. 123–140. [Google Scholar]
  59. Dimitriadis, C.K. Improving mobile core network security with honeynets. IEEE Secur. Priv. 2007, 5, 40–47. [Google Scholar] [CrossRef]
  60. Huang, L.; Zhu, Q. Dynamic Bayesian Games for Adversarial and Defensive Cyber Deception. In Autonomous Cyber Deception; Springer: Cham, Switzerland, 2019; pp. 75–97. [Google Scholar]
  61. Seo, S.; Kim, D. OSINT-based LPC-MTD and HS-decoy for organizational defensive deception. Appl. Sci. 2021, 11, 3402. [Google Scholar] [CrossRef]
Figure 1. Main overview of the proposed organizational deceptive game with social-engineering MTD and Decoy.
Figure 1. Main overview of the proposed organizational deceptive game with social-engineering MTD and Decoy.
Electronics 10 03012 g001
Figure 2. Detailed overview of zero-sum-based attack-defense competition with PBNE, BSSG, and signaling.
Figure 2. Detailed overview of zero-sum-based attack-defense competition with PBNE, BSSG, and signaling.
Electronics 10 03012 g002
Figure 3. Sub-overview of POMDP with attacker side in Scenario 1.
Figure 3. Sub-overview of POMDP with attacker side in Scenario 1.
Electronics 10 03012 g003
Figure 4. Overview and related topology with an organizational attack-defense strategy for Scenario 1.
Figure 4. Overview and related topology with an organizational attack-defense strategy for Scenario 1.
Electronics 10 03012 g004
Figure 5. Overview and related topology with an organizational attack-defense strategy for Scenario 2.
Figure 5. Overview and related topology with an organizational attack-defense strategy for Scenario 2.
Electronics 10 03012 g005
Figure 6. Overview and related topology with an organizational attack-defense strategy for Scenario 3.
Figure 6. Overview and related topology with an organizational attack-defense strategy for Scenario 3.
Electronics 10 03012 g006
Figure 7. Examples of defender’s episode-based deceptive sequences with MTD and Decoy concepts in SOD2G. (a) Mutation sequence based on network-layer elements. (b) Mutation sequence based on host-layer elements.
Figure 7. Examples of defender’s episode-based deceptive sequences with MTD and Decoy concepts in SOD2G. (a) Mutation sequence based on network-layer elements. (b) Mutation sequence based on host-layer elements.
Electronics 10 03012 g007
Figure 8. Comparison of the degree of defense success in each state with Scenario 1. (a) S0, (b) S1, (c) S2, and (d) S3.
Figure 8. Comparison of the degree of defense success in each state with Scenario 1. (a) S0, (b) S1, (c) S2, and (d) S3.
Electronics 10 03012 g008aElectronics 10 03012 g008b
Figure 9. Comparison of the degree of defense success in each state with Scenario 2. (a) S1, (b) S2, and (c) S3.
Figure 9. Comparison of the degree of defense success in each state with Scenario 2. (a) S1, (b) S2, and (c) S3.
Electronics 10 03012 g009
Figure 10. Comparison of the degree of defense success in each state with Scenario 3. (a) S1, (b) S2, and (c) S3.
Figure 10. Comparison of the degree of defense success in each state with Scenario 3. (a) S1, (b) S2, and (c) S3.
Electronics 10 03012 g010
Figure 11. Comparison of the degree of attack success in each state with Scenario 3. (a) S1, (b) S2, and (c) S3.
Figure 11. Comparison of the degree of attack success in each state with Scenario 3. (a) S1, (b) S2, and (c) S3.
Electronics 10 03012 g011
Figure 12. Comparative analysis of the overall success probability of signaling deception in SOD2G with each scenario. (a) Scenario 1, (b) Scenario 2, and (c) Scenario 3.
Figure 12. Comparative analysis of the overall success probability of signaling deception in SOD2G with each scenario. (a) Scenario 1, (b) Scenario 2, and (c) Scenario 3.
Electronics 10 03012 g012
Figure 13. Comparative analysis of deceptive efficiency of disinformation and disclosure-based loosely mutations. (a) Scenario 1, (b) Scenario 2, and (c) Scenario 3.
Figure 13. Comparative analysis of deceptive efficiency of disinformation and disclosure-based loosely mutations. (a) Scenario 1, (b) Scenario 2, and (c) Scenario 3.
Electronics 10 03012 g013
Figure 14. Comparative analysis of deceptive efficiency of cloning and mimicking-based hierarchical OSINT decoys. (a) Scenario 1, (b) Scenario 2, and (c) Scenario 3.
Figure 14. Comparative analysis of deceptive efficiency of cloning and mimicking-based hierarchical OSINT decoys. (a) Scenario 1, (b) Scenario 2, and (c) Scenario 3.
Electronics 10 03012 g014
Table 1. Taxonomy of existing defensive deception research studies and proposed deceptive game framework.
Table 1. Taxonomy of existing defensive deception research studies and proposed deceptive game framework.
ApproachSpecific TechniqueAdvantagesDisadvantages
Game-theoretic defensive deception with MTD [38,39,40,41,42,43,44,45,46,47,48]General game, Bayesian Stackelberg game, Stochastic gameFormalizing the interactions between attackers and defenders and providing methodologies for estimating deception effectiveness and security.
Determining the optimal MTD strategy based on learning and realizing the decision model.
Not working the best strategy devised by a defender for the irrational attacker.
Very high computational resources used to derive the optimal solution.
Causing uncertainty due to different interpretations of the same game depending on the subjective perception of each different player.
Difficulty in modeling cybersecurity and defensive cyber deception issues in real-world and practical environments.
Game-theoretic defensive deception with non MTD [49,50,51,52,53,54,55,56,57,58,59,60]Game with MDP and POMDP, Signaling game, Static gameRealizing improvements in intrusion detection and prevention schemes by collecting additional attack intelligence while protecting existing network and system components.
Possesses high flexibility, applicability, and rationality for the application of hierarchical deception concepts.
Difficulty in developing realistic honeypots that can effectively deceive attackers in terms of complexity and cost with the advent of increasingly intelligent and sophisticated attackers.
Difficulty in calculating optimal solutions for combining, managing, and deploying multiple elements of defensive cyber deception.
Not considering the negative effect on legitimate users due to insufficient access control and authorization.
Difficulty in modeling cybersecurity and defensive cyber deception issues in real-world and practical environments.
SOD2G
framework
- Deception
MTD with LPC-MTD, Decoy with HS-Decoy

- Social-engineering
Open-source intelligence (OSINT), Disinformation

- Game-based
Perfect Bayesian Nash equilibrium with zero-sum (PBNE), Bayesian stochastic Stackelberg game (BSSG), Partially signaling scheme for disinformation (PSG), Partially observable Markov decision process (PDMDP)
Description and Improvement
LPC-MTD is an MTD strategy selection concept that deliberately tolerates high-level organizational attacks while avoiding automated attacks that cause noise. Using this concept, flexibility and scalability issues and systematic deception management strategy issues can be alleviated, and MTD’s post-responsiveness can be secured.

HS-Decoy is an attacker cognitive bias induction concept composed of hierarchical OSINT and HUMINT-based false organization information. It uses the actual characteristics of each organization to increase the likelihood of an attacker’s earlier attempt compared to other decoys, while minimizing the attacker’s suspicion. Using this concept, it is possible to improve the post-responsiveness and scalability of decoys while alleviating the existing limitations of practical application.

The SOD2G framework is a two-player competitive game—which is composed of a PBNE and BSSG-based zero-sum game foreground and a POMDP state-transition background—which models the real time attack-defense process in the organizational environment with limited resources as a multi-stage temporospatial deception decision making process. With this framework, a limited view and surface can be constructed along with the disinformation deceptive signal push concept.
Table 2. Probability of transition and semi-constant reward value with each state and action in Scenario 1.
Table 2. Probability of transition and semi-constant reward value with each state and action in Scenario 1.
StateProbability of TransitionReward Value for Defender
S0 [ ( 1 ,   0 ,   0 ,   0 ) ] [−20]
S1 [ ( 0 ,   1 ,   0 ,   0 ) ( 0 ,   1 ,   0 ,   0 ) ( 0 ,   0.3 ,   0.7 ,   0 ) ( 0 ,   0.9 ,   0.1 ,   0 ) ] [ 0 3 10 10 ]
S2 [ ( 0 ,   0 ,   1 ,   0 ) ( 0 ,   0 ,   1 ,   0 ) ( 0 ,   0 ,   1 ,   0 ) ( 0 ,   0 ,   1 ,   0 ) ( 0 ,   0 ,   1 ,   0 ) ( 0 ,   0 ,   1 ,   0 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0.4 ,   0 ,   0.4 ,   0.2 ) ( 0.8 ,   0 ,   0.1 ,   0.1 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0.4 ,   0 ,   0.4 ,   0.2 ) ( 0.8 ,   0 ,   0.1 ,   0.1 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0 ,   0 ,   0.3 ,   0.7 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0.45 ,   0.15 ,   0.15 ,   0.25 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0.8 ,   0 ,   0.1 ,   0.1 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0.45 ,   0.15 ,   0.15 ,   0.25 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0.8 ,   0 ,   0.1 ,   0.1 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0.4 ,   0.2 ,   0.2 ,   0.2 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ( 0.8 ,   0 ,   0.1 ,   0.1 ) ( 0 ,   0 ,   0.4 ,   0.6 ) ] [ 0 3 3 5 5 5 9.3 9.3 3 5 5 5 10 3 10 5 5 5 7.5 3 3 7.5 5 5 10 3 3 5 10 5 10 3 3 5 5 10 ]
S3 [ ( 0 ,   0 ,   0 ,   1 ) ( 0 ,   0 ,   0 ,   1 ) ( 0.8 ,   0 ,   0 ,   0.2 ) ( 0.4 ,   0.4 ,   0 ,   0.2 ) ] [ 0 3 10 20 ]
Table 3. CVE-based vulnerability table for deception with organizational defender or CKC-based APT attack with attacker.
Table 3. CVE-based vulnerability table for deception with organizational defender or CKC-based APT attack with attacker.
CVE IDVulnerabilityRole in Org.Related NodeCVSS 2.0
CVE-2020-25223Remote code execution on web-admin in Sophos SGIDS/IPSUTM10.0
CVE-2020-12271Remote code execution on web-admin in Sophos SFOSAccess controlFirewall7.5
CVE-2017-0144EternalBlue. Remote code executionLocal hostWin7-19.3
CVE-2019-0708BlueKeep. Use-after-free, Remote code executionRemote hostWin7-210.0
CVE-2020-1350SigRed. Heap overflow, Remote code executionDNS serverWin server 201210.0
CVE-2020-0796SMBGhost. Remote code executionMail serverWin server 20167.5
CVE-2015-3306Remote command execution in ProFTPD 1.3.5PFTPD serverDebian-FTP server10.0
Table 4. Optimized defender values with game and deceptive metrics and parameters in SOD2G framework.
Table 4. Optimized defender values with game and deceptive metrics and parameters in SOD2G framework.
Metric and Related ParameterTargeted ScenarioStateValue
Attacker’s discount factorScenario 1S0<0.7
S1<0.75~0.8
S2<0.7~0.75
S3<0.75~0.8
Scenario 2S1<0.7
S2<0.7~0.75
S3<0.65
Scenario 3S1<0.75
S2<0.7
S3<0.6
Utility of deceptive surfaceScenario 1S3>0.6
Scenario 2S3>0.7
Scenario 3S3>0.75
Minimum mutation strength for organizational LPC-MTDScenario 1S3>0.2
Scenario 2S3>0.3
Scenario 3S3>0.15
Minimum decoying signal strength for organizational HS-DecoyScenario 1S3>0.75
Scenario 2S3>0.85
Scenario 3S3>1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Seo, S.; Kim, D. SOD2G: A Study on a Social-Engineering Organizational Defensive Deception Game Framework through Optimization of Spatiotemporal MTD and Decoy Conflict. Electronics 2021, 10, 3012. https://doi.org/10.3390/electronics10233012

AMA Style

Seo S, Kim D. SOD2G: A Study on a Social-Engineering Organizational Defensive Deception Game Framework through Optimization of Spatiotemporal MTD and Decoy Conflict. Electronics. 2021; 10(23):3012. https://doi.org/10.3390/electronics10233012

Chicago/Turabian Style

Seo, Sang, and Dohoon Kim. 2021. "SOD2G: A Study on a Social-Engineering Organizational Defensive Deception Game Framework through Optimization of Spatiotemporal MTD and Decoy Conflict" Electronics 10, no. 23: 3012. https://doi.org/10.3390/electronics10233012

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop