Next Article in Journal
Electrocardiogram Monitoring Wearable Devices and Artificial-Intelligence-Enabled Diagnostic Capabilities: A Review
Previous Article in Journal
The Classification of Movement in Infants for the Autonomous Monitoring of Neurological Development
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sine-Cosine-Adopted African Vultures Optimization with Ensemble Autoencoder-Based Intrusion Detection for Cybersecurity in CPS Environment

1
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Mathematics, Faculty of Sciences and Arts, King Khalid University, Muhayil Asir 63311, Saudi Arabia
3
Department of Software Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 103786, Riyadh 11543, Saudi Arabia
4
Department of Computer Science, Faculty of Computers and Information Technology, Future University in Egypt, New Cairo 11835, Egypt
5
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam Bin Abdulaziz University, AlKharj 11942, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4804; https://doi.org/10.3390/s23104804
Submission received: 31 January 2023 / Revised: 22 March 2023 / Accepted: 12 May 2023 / Published: 16 May 2023
(This article belongs to the Section Sensor Networks)

Abstract

:
A Cyber-Physical System (CPS) is a network of cyber and physical elements that interact with each other. In recent years, there has been a drastic increase in the utilization of CPSs, which makes their security a challenging problem to address. Intrusion Detection Systems (IDSs) have been used for the detection of intrusions in networks. Recent advancements in the fields of Deep Learning (DL) and Artificial Intelligence (AI) have allowed the development of robust IDS models for the CPS environment. On the other hand, metaheuristic algorithms are used as feature selection models to mitigate the curse of dimensionality. In this background, the current study presents a Sine-Cosine-Adopted African Vultures Optimization with Ensemble Autoencoder-based Intrusion Detection (SCAVO-EAEID) technique to provide cybersecurity in CPS environments. The proposed SCAVO-EAEID algorithm focuses mainly on the identification of intrusions in the CPS platform via Feature Selection (FS) and DL modeling. At the primary level, the SCAVO-EAEID technique employs Z-score normalization as a preprocessing step. In addition, the SCAVO-based Feature Selection (SCAVO-FS) method is derived to elect the optimal feature subsets. An ensemble Deep-Learning-based Long Short-Term Memory–Auto Encoder (LSTM-AE) model is employed for the IDS. Finally, the Root Means Square Propagation (RMSProp) optimizer is used for hyperparameter tuning of the LSTM-AE technique. To demonstrate the remarkable performance of the proposed SCAVO-EAEID technique, the authors used benchmark datasets. The experimental outcomes confirmed the significant performance of the proposed SCAVO-EAEID technique over other approaches with a maximum accuracy of 99.20%.

1. Introduction

A Cyber-Physical System (CPS) is a type of computing system combined with physical gadgets and can be broadly utilized in various areas, namely, energy, manufacturing, safety management, and traffic control [1]. The most common enabler for the intelligence sector is the integration of the Cloud Computing (CC) technique and CPS, which remains a common trend with numerous real-time cases, for example, Small and Medium Enterprises (SMEs), supporting industrial cluster collaboration with business cooperation and cloud manufacturing service platforms [2]. With the help of CC, more optimized techniques are constituted to enrich the robustness and reliability of the system and collaborate with other systems in order to enlarge the efficiency of the functions at limited data usage for CPSs. Although Information and Communication Technology (ICT) is an advanced part of CPSs, cybersecurity is considered to be a challenging issue in many domains. Intrusion is one of the primary complications encountered in CPSs [3]. In the past, special attention has been paid to the development of secure CPSs. Further, efforts have been taken to maximize the integrity of CPSs with Intrusion Detection (ID), which has become a predominant application. Generally, an Intrusion Detection System (IDS) is used to prevent attacks in an efficient manner [4]. An IDS is referred to as a tool that detects or classifies cyberattacks in a network or host by implementing some identification approaches. There are two classes present in IDSs, such as Anomaly-based IDSs and Signature-based IDSs (AIDSs and SIDSs). In SIDSs, attacks are identified with regard to the predefined sign or pattern of attacks. In AIDS networks, traffic patterns are monitored and compared against regular or normal paradigms in the network so as to identify intrusions [5].
Several AI-based IDS approaches have been devised so far to ensure the security of CPSs. In spite of the fact that the presented methods show good performance, they are grounded on the assumption that the dataset reflects the real-time situations of cyberattacks [6]. However, on a real-time basis, users access datasets with limited examples of cyberattacks. As cyberattacks evolve in terms of complexity and volume, Machine Learning (ML) techniques have been applied to manage various malicious performance and cybersecurity attacks. The CPS unites the calculation with that of the physical process [7]. The embedded network computer controls and monitors the physical processes, normally with feedback loops in which the physical processes affect the computation simultaneously. In general, ML techniques are prone to data pollution attacks. Therefore, it is important to enhance network security and achieve a strong ML-based network method in the development of CPSs [8]. ML approaches are broadly leveraged in the detection of cyber intrusions due to their timely and automatic manner of action. The opportunity to make an adaptable and scalable detection system is offered by DL methods. The DL method is utilized with unsupervised and supervised techniques [9,10]. Unsupervised techniques are used to make labels for non-labeled samples.
The current study presents a Sine-Cosine-Adopted African Vultures Optimization with Ensemble Autoencoder-based Intrusion Detection (SCAVO-EAEID) technique for cybersecurity in the CPS environment. At the primary level, the proposed SCAVO-EAEID technique employs the Z-score normalization process as a preprocessing step. Then, the SCAVO-based Feature Selection (SCAVO-FS) method is applied to elect the optimal feature subsets. This step shows the novelty of the work. For intrusion detection, the ensemble Deep-Learning-based Long Short-Term Memory–Auto Encoder (LSTM-AE) model is employed. Finally, the Root Mean Square Propagation (RMSProp) optimizer is used for the hyperparameter tuning of the LSTM-AE model. To demonstrate the remarkable performance of the proposed SCAVO-EAEID technique, benchmark datasets were used. In short, the key contributions of the current study are summarized herewith.
  • An automated SCAVO-EAEID technique comprising Z-score normalization, the SCAVO-FS technique, LSTM-AE-based intrusion detection, and the RMSProp optimizer is developed for intrusion detection in the CPS environment. To the best of the researchers’ knowledge, no researchers have proposed the SCAVO-EAEID technique in the literature.
  • A new SCAVO-FS technique has been designed by integrating the sine-cosine scaling factor and the AVO algorithm for the repositioning of the vultures at the end of the iterations.
  • Both the RMSProp optimizer and the LSTM-AE model are employed in this study for the intrusion detection process.
  • The performance of the proposed SCAVO-EAEID technique was validated using two benchmark datasets such as the NSL-KDD 2015 and CICIDS2017 datasets.
The rest of the paper is organized as follows. Section 2 discusses the related works, and Section 3 offers the proposed model. Then, Section 4 provides the analytical results, and Section 5 concludes the paper.

2. Related Works

Huang et al. [11] introduced a new federated Execution and Evaluation dual network model (EEFED), which allows different federal participants to identify the local detection model. This phenomenon undermines the primary objective of Federated Learning (FL). Mansour [12] proposed a novel Poor and Rich Optimization with the DL method for BC-Assisted IDS in CPS Environments (PRO-DLBIDCPS). At first, the model implemented the Adaptive Harmony Search Algorithm (AHSA)-based FS method for an appropriate selection of the feature subset. The PRO technique with the Attention-based Bi-Directional Gated RNN (ABi-GRNN) system was used in this study for both the detection and classification of the intrusions. Henry et al. [13] introduced a method integrating CNN and GRU in which both combinations were used for the optimization of the network parameters. In this work, the author utilized the CICIDS-2017 benchmark datasets.
Ortega-Fernandez et al. [14] introduced the Network IDS (NIDS) model based on the DAE, trained with network flow data. This model had a benefit, i.e., no need to have previous knowledge about the underlying architecture or the network topology. The experimental result showed that the presented method was capable of detecting the anomalies, caused by distributed DoS attacks. The proposed method provided a low false alarm rate and high detection accuracy. It also outperformed the other methods and acted as a baseline and a state-of-the-art model for the unsupervised learning model. Likewise, the DAE model is capable of detecting abnormal behaviors in legitimate devices after an attack. Wang et al. [15] introduced a knowledge distillation method-based Triplet CNN technique to heavily enhance the speed and improve the AD performance of the model for industrial CPS in addition to the reduction of model complexity. In particular, during the training, a robust model loss function was devised to enhance the network’s stability. A novel NN training model named K-fold cross-training was developed in this study to enhance anomaly detection performance.
Mittal et al. [16] introduced a new technique for the IDS. The presented technique applied a new variant of the gravitational search technique to attain the optimum cluster. In the presented method, Kbest was adapted as an exponentially-declining function with logistic-mapping-based chaotic behaviors. Presekal et al. [17] developed a new technique for online cyberattack awareness. This technique improves the power grid resilience and assists the power system operators during localization and identification processes of the active attack locations in the Operational Technology (OT) network on a real-time basis. The presented technique used a hybrid DL mechanism, i.e., deep convolutional network with Graph Convolutional LSTM (GC-LSTM), for time-series classification-related AD.
Though several ML and DL models have been proposed earlier for intrusion detection and classification, a need still exists to enhance the classification performance. Owing to the continuous evolution of the model, the number of parameters in DL models also increases quickly, bringing model overfitting issues. Since the trial-and-error method is a tedious and erroneous process for hyperparameter tuning, the metaheuristic algorithms are applied. Therefore, in this work, the authors employ the BO algorithm for the parameter selection of the LSTM-AE model.

3. Proposed Model

In this study, a new SCAVO-EAEID method has been developed for the classification of intrusions in the CPS environment. The SCAVO-EAEID algorithm performs the detection of intrusions in the CPS environment using major sub-processes namely, the LSTM-AE-based classification, Z-score normalization, SCAVO-FS technique, and the RMSProp optimizer-based hyperparameter tuning. The workflow of the proposed model is demonstrated in Figure 1.

3.1. Data Used

In the current research work, the proposed model was experimentally validated upon two benchmark datasets such as the NSLKDD2015 (https://www.unb.ca/cic/datasets/nsl.html, accessed on 12 February 2023) and CICIDS2017 datasets (https://www.unb.ca/cic/datasets/ids-2017.html, accessed on 12 February 2023). The NSL-KDD 2015 dataset has a total of 125,973 samples with 41 features. The NSL-KDD 2015 dataset does not include any redundant records in the training set. So, the classifiers remain unbiased towards more frequent records. Likewise, CICIDS2017 has a total of 2,830,743 samples with 80 features. The CICIDS2017 dataset contains both benign and the most up-to-date common attacks, which resemble the true real-world data (PCAPs). It also includes the network traffic analysis results with the help of CICFlowMeter and labeled flows based on a few parameters such as the time stamp, source, destination IPs, source and destination ports, protocols, and attack.

3.2. Data Preprocessing

At first, the SCAVO-EAEID technique applies the Z-score normalization as a preprocessing step. This procedure is leveraged to attain the normalized values or ranges of the input dataset, from the unstructured dataset, by applying the concepts such as mean and standard deviation [18]. The normalized values or ranges can be obtained by dividing the presented data of every gate using standard deviation values and then subtracting the mean of every gate from that value. Equations (1) and (2) are utilized to map the value of the new input between the target ranges, i.e., [ x , y ] .
A Z a ( t ) = A t A j t A j ( t )
A Z a d ( t ) = c + ( d c ) ( A Z S ( t ) A Z S m i n ) A Z S m a x A Z S m i n
Here, A Z a d ( t ) denotes the scaling value of the z -score-normalized field A Z a ( t ) . A i ( t ) and A j ( t ) are evaluated from the training datasets, which represent the standard deviation and mean at every gate time t , whereas A Z S m i n and A Z S m a x denote the absolute initial and final gate values that are noticed for the gate period during the testing dataset.

3.3. Processes Involved in the SCAVO-FS Technique

In this work, the SCAVO-FS system is derived to elect the optimal feature subsets. The AVO algorithm is stimulated by the navigational and foraging behaviors of the African vultures [19]. The biological nature of these vultures with regard to competing and searching for food is outlined in four different phases as follows. Consider N number of vultures in the atmosphere, which describes the amount of population, i.e., n = { 1,2 , , N } . In the following equation, the fitness function of every location is evaluated.
p n = F n n = 1 N F n
Here, p n represents the probability of choosing either the first or second group, F n denotes the fitness function of the n th location [19]. Next, the formation of the 1st and 2nd groups during all the iterations is attained as follows.
R ( i t ) = f i r s t   g r o u p , p n = L 1 s e c o n d   g r o u p , p n = L 2
Here, the ranges of L 1 and L 2 are 0 L 1 , L 2 1 and L 1 + L 2 = 1 , respectively. The satiated vulture with sufficient energy can move a long distance to find food, whereas a hungry one cannot fly longer as denoted below.
A = ( 2 × r a n d 1 + 1 ) × x × ( 1 i t I T m a x ) + y
y = h × ( s i n z ( Π 2 × i t I T m a x ) + c o s ( Π 2 × i t I T m a x ) 1 )
In Equations (5) and (6), A denotes the vulture with high energy, i t and I T m a x denote the present and the maximal iterations, correspondingly,  h and r a n d 1 indicate the random number in the range of 1 , 1 , 2 , 2 , and [ 0 , 1 ] , correspondingly; and z describes the probability of entering the exploration phase.
The procedure of seeking food by African vultures defines the exploration stage in which the parameter p 1 , 0 p 1 1 defines the selection of strategy.
P ( i t + 1 ) = ( 6 ) , p 1 r a n d 2 ( 8 ) , p 1 < r a n d 2
P ( i t + 1 ) = R ( i t ) D ( i t ) × A
D ( i t ) = | q × R ( i t ) P ( i t ) |
In this expression, P ( i t + l ) represents the location vector of the vulture in the second iteration. q = 2 × r a n d 3 , where r a n d 3 denotes the randomly generated value in the range of [ 0 , 1 ] .
P i t + 1 = R i t A + r a n d 4 × u b 1 b × r a n d 5 + 1 b
u b and l b denote the upper and lower boundaries correspondingly, while R a n d 4 and R a n d 5 indicate the randomly generated integers that lie in the range of 0 to 1.
The exploitation phase includes two stages with dissimilar strategies. The selection of any method relies on both p 2 and p 3 parameters, while its values lie in the range of 0 to 1. If | F | ranges between 0.5 and 1 , then the exploitation phase enters the initial phase, which defines the siege fight and rotating flight strategies.
P ( i t + 1 ) = ( 10 ) , p 2 r a n d 6 ( 11 ) , p 2 < r a n d 6
Here, r a n d 6 denotes a randomly generated value in the range of [0, 1]. The solution to Equation (11) is given below.
P ( i t + l ) = D ( i t ) × ( A + r a n d 7 ) d ( i t )
D ( i t ) = R ( i t ) P ( i t )
Next, the rotational flight of the vulture can be modeled as given below.
P ( i t + l ) = R ( i t ) ( M 1 + M 2 )
M 1 = R ( i t ) × ( r a n d 8 × P ( i t ) 2 π ) × c o s ( P ( i t ) )
M 2 = R ( i t ) × ( r a n d 9 × P ( i t ) 2 π ) × s i n ( P ( i t ) )
Here, r a n d 8 and r a n d 9 denote the two randomly-generated integers in the range of [0, 1]. If F > 0.5 , then the exploitation phase enters the second phase, which describes the aggressive siege and accumulation fight strategies for finding the food. Based on the following condition, any strategy can be selected.
P ( i t + 1 ) = ( 16 ) , p 3 r a n d 7 ( 19 ) , p 3 < r a n d 7
where
P ( i t + 1 ) = B 1 + B 2 2
B 1 = B e s t v u l t u r e 1 ( i t ) B e s t v u l t u r e 1 ( i t ) × P ( i t ) B e s t v u l t u r e 1 ( i t ) × P ( i t ) 2 × A
B 2 = B e s t v u l t u r e 2 ( i t ) B e s t v u l t u r e 2 ( i t ) × P ( i t ) B e s t v u l t u r e 2 ( i t ) × P ( i t ) 2 × A
B e s t v u l t u r e 1 (it) and B e s t v u l t u r e 2 ( i t ) represent the better vultures of the first and second groups, correspondingly as shown below.
P ( i t + 1 ) = R ( i t ) | d ( i t ) | × A × L e v y ( d )
Here, d 1 signifies the dimension of the problem. Levy( d 1 ) is evaluated as given below.
Levy ( x ) = 0.01 × u × σ | v | 1 / β , σ = ( Y ( 1 + β ) × s i n ( π β 2 ) Y ( 1 + 2 β ) × β × 2 ( β 1 2 ) ) 1 / β
The best possible solution is not known at the initial stage of the AVO algorithm. Therefore, it is better to use a large step in the beginning, which might generate the calculation process farther from the optimum location [20]. Consequently, the scaling variable is used for changing the situation in the initial phase. In the SCAVO algorithm, the vulture is repositioned at the end of every iteration based on sine- and cosine-adapted scaling factors, as given below [20].
P ( i t + 1 ) N e w = P ( i t + 1 ) × S C a S F
Now, the SCaSF denotes the scaling factor as follows
S C a S F = s i n W 1 W 2 i t M a x i t i f   R N D I < 0.5 cos W 1 W 2 i t M a x i t i f   R N D I 0.5
In Equation (24), RNDI indicates a randomly-generated value; W 1 and W 2 denote the weighting factors, and i t   a n d   M a x _ i t indicate the current and maximal iteration, respectively. The scaling factor is used to control the vulture’s development at the initial stage. Different upsides of W are tried and appointed for a proper choice of W 1 and W 2 . It is to be noted that the best possible results are attained when W 1 and W 2 are selected as 10 and 9 correspondingly. The scaling factor has changed the vulture’s position at the underlying period of the pursuit interaction, thus increasing the hunting capability of the AVO technique further. The calculation should provide the option to locate the best location between two locations for the exploitation of the search range. These cycles ensure the best exploitation and exploration capabilities of the calculation.
The fitness function considers the number of features selected and the accuracy of the classifier. It reduces the size of features selected and increases the classification accuracy. Consequently, the subsequent fitness function is used to evaluate the individual solution.
F i t n e s s = α     E r r o r R a t e + 1 α     # S F # A l l _ F
Here, α , which is usually set to 0.9, is used to control the importance of subset length and classification quality. ErrorRate is the classification error rate based on the number of features selected. ErrorRate can be evaluated as the percentage of incorrect classifications to the whole number of classifications made, and its values lie in the range of 0 to 1. ErrorRate is the complement of classification accuracy. # S F denotes the total features selected, and # A l l _ F represents the overall number of features in the original data.

3.4. Classification Model

For accurate classification of the intrusions, the LSTM-AE model is leveraged for both the identification and the classification of the intrusions. The deep RNN (particularly LSTM) model is the underlying structure of the DL model that is applied to time-series and sequential data to learn the features and patterns. But, the LSTM [21], out of the RNN method, contains memory cells for pattern recognition that is dependent on short -and long-term input datasets. These are beneficial in the detection and forecasting of the outliers in time-series datasets [22]. The LSTM cell comprises three memory gates such as the forget, input (update), and output gates.
F t = δ ( M f ( h t 1 , x t ) + B f )
In Equation (26), B and M denote the bias and weight of the LSTM, respectively. Furthermore, the dataset fed as input to the LSTM cells gets upgraded by the input gate ( I t ) .
l t = δ ( M j ( h t 1 ) x t ) + B i )
C ¯ t = t a n h ( M c ( h t 1 , x t ) + B c )
At last, the output gate plays its role by transmitting the dataset created to the output cell ( H t ) or the succeeding state vector ( C t ) .
C t = F t C t 1 + l t
H t = δ ( M o ( h t 1 ) x t ) + B o ) t a n h ( C t )
In this expression, x t stands for input dataset, h t 1 denotes the prior short-term state, and B and M represent the bias and weight matrices, correspondingly. Meanwhile, the LSTM model has a state vector C t that includes t a n h and s i g m o i d functions. The model evaluates the gradient error at every time point, whereas the derivative items of t a n h and s i g m o i d functions become additive. It prevents the model from suffering gradient disappearing problems. Unlike other gradient descent models, the LSTM exploits a mini-batch method at the time of training the data samples. Usually, the LSTM model comprises a single layer of cells, whereas the ensemble and a combination of numerous LSTM models increase the depth of the layer. This in turn increases the model’s performance and accuracy in training. It also helps in identifying the short- and long-term continuous patterns and big datasets.
On top of that, the AE method is utilized with LSTM for training the long-term patterns and the most important features. AE is an unsupervised type of ANN that intends to learn the essential hidden representation of the datasets by decoding and encoding processes. The output and the input datasets are compared to evaluate the differences. In the case of large differences, it shows that the reconstruction loss is higher. Based on this output, it can be assumed that the model can handle the reconstructed dataset. Accordingly, the data are recognized as irregular.
The LSTM-AE is an application of AE in which the LSTM cells are employed in the encoder–decoder layer. This setup brings the advantages of both methods for time-series or sequential datasets. In this work, the LSTM-AE is employed because it provides certain advantages over the normal (regular) AEs, for instance, LSTM-AE can handle sequence data as input (time-series dataset), whereas the normal AE cannot accept a sequential sample as the input dataset. Furthermore, the LSTM-AE models take a wider range of input lengths (short- or long-term), while on the other hand, the regular AE takes only a fixed size of the input dataset. At last, the data dimension increases, and the computation becomes complicated, since the long- and short-term dependence on time in previous data affects the current IIoT data. To resolve these problems, LSTM-AE is applied for the extraction of the fundamental feature with benefits over normal AEs. The structure of the AE is shown in Figure 2.
The AE model comprises the output layer, input layer, and Hidden Layer (HL), whereas the interconnected layer is made up of an LSTM cell to create the output layer dataset. The HL takes the sample from various sampling times to estimate and calculate the impact on the succeeding sample datasets at another sampling period. The next time point values are attained as the output values by modeling and integrating the impact as the forecasted values. The respective sample of the HL is characterized by an equation in which the sample of x i remains the input instance of the data sample from X ,   N w signifies the weight matrices, and P w characterizes the bias vector between the input layer and the HLs. The function δ updates the next input layer to make C i as the output value in the AE architecture.
C j = δ ( N w . x j + P w )

3.5. Hyperparameter Tuning Model

At last, the RMSProp optimizer is exploited for the hyperparameter tuning of the LSTM-AE model. In the vertical direction, the RMSProp optimizer [22] restricts the oscillation. Thus, in the horizontal direction, the learning rate gets improved and the algorithm takes a large step in converging at a faster rate. The RMSProp calculation is given as follows. The value of the momentum is represented as beta and is set to 0.9 [22].
v d w = β v d w + ( 1 β ) d w 2
v d b = β v d w + ( 1 β ) d b 2
W = W α d w v d w + ε
b = b α d b v d b + ε
In backward propagation process, d W and db are used to update the W and b parameters with the help of the following expression [22]:
W = W l e a r n i n g   r a t e     d W
b = b l e a r n i n g   r a t e     d b
Let us assume the exponentially-weighted average square values of d W and d b instead of independently using the d W and db values for all the epochs [22].
S d W = β     S d W + ( 1 β )     d W 2
S d b = β     S d b + ( 1 β )     d b 2
Here, β represents the hyperparameter in the range of 0 to 1. The newly weighed average can be formed with the help of current value square, weights and the average of the previous values. The parameters will be updated after the evaluation of the exponentially-weighted averages [22].
W = W l e a r n i n g   r a t e     d W / s q r t ( S )
b = b l e a r n i n g   r a t e     d b / s q r t ( S )
S d W is relatively lower in such a way that it is divided by d W . Here, S d b is relatively higher so that when it is divided by d b with a comparatively large number, it slows down the update on the vertical dimension.

4. Results Analysis

The performance of the SCAVO-EAEID method was experimentally validated on two datasets such as NSL-KDD 2015 and CICIDS 2017. The proposed model was simulated in the Python 3.6.5 tool on a PC configured with these specifications; i5-8600k, GeForce 1050Ti 4GB, 16 GB RAM, 250 GB SSD, and 1 TB HDD. The parameter settings are given as follows: learning rate, 0.01; dropout, 0.5; batch size, 5; epoch count, 50; and activation, ReLU.
Table 1 reports the best cost outcomes of the proposed SCAVO-FS method and other FS algorithms on two datasets. The experimental values indicate that the proposed SCAVO-FS technique achieved the optimal best cost values such as 0.05101 and 0.41204 under the NSL-KDD-2015 and CICIDS-2017 datasets, correspondingly.
In Table 2, the FS results are represented in terms of several selected features. The results indicate the promising performance of the presented SCAVO-FS technique. Moreover, it is recognized that the SCAVO-FS technique effectually selected 41 and 80 features under NSL-KDD-2015 and CICIDS-2017 datasets, correspondingly.
In Table 3 and Figure 3, the experimental outcomes accomplished by the proposed SCAVO-EAEID method upon the NSL-KDD dataset are portrayed. The outcomes indicate that the SCAVO-EAEID technique achieved increased values under all the training set/testing set (TRS/TSS) instances. For instance, with 40:60 of TRS/TSS, the SCAVO-EAEID technique attained an a c c u y of 98.70%, p r e c n of 99.16%, r e c a l of 96.79%, and F s c o r e of 97.69%. Meanwhile, with 50:50 of TRS/TSS, the SCAVO-EAEID technique accomplished an a c c u y of 98.74%, p r e c n of 99.24%, r e c a l of 98.14%, and F s c o r e of 99.53%. Finally, with 80:20 of TRS/TSS, the SCAVO-EAEID technique acquired an a c c u y of 99.20%, p r e c n of 99.58%, r e c a l of 99.42%, and F s c o r e of 99.84%.
The TACC and VACC values, achieved by the proposed SCAVO-EAEID system upon the NSL-KDD 2015 dataset are shown in Figure 4. The outcomes exhibit that the SCAVO-EAEID method yielded better performance with maximum TACC and VACC values. It is noticeable that the SCAVO-EAEID methodology gained the maximum TACC outcomes.
The TLS and VLS values, accomplished by the proposed SCAVO-EAEID system upon the NSL-KDD 2015 dataset, are shown in Figure 5. The results display that the SCAVO-EAEID approach showcased a superior performance with low TLS and VLS values. It is noticeable that the SCAVO-EAEID method achieved the least VLS outcomes.
In Table 4 and Figure 6, the experimental outcomes of the SCAVO-EAEID method and other techniques upon the CICIDS-2017 dataset are portrayed. The outcomes indicate that the SCAVO-EAEID method achieved improved values under all the TRS/TSS instances. For instance, with 40:60 of TRS/TSS, the SCAVO-EAEID technique attained an a c c u y of 98.70%, p r e c n of 99.16%, r e c a l of 96.79%, and F s c o r e of 97.69%. Meanwhile, with 50:50 of TRS/TSS, the SCAVO-EAEID technique accomplished an a c c u y of 98.74%, p r e c n of 99.24%, r e c a l of 98.14%, and F s c o r e of 99.53%. Finally, with 80:20 of TRS/TSS, the SCAVO-EAEID method achieved an a c c u y of 99.10%, p r e c n of 97.92%, r e c a l of 98.54%, and F s c o r e of 98.64%.
The TACC and VACC values, achieved by the proposed SCAVO-EAEID method upon the CICIDS-2017 dataset are shown in Figure 7. The outcomes demonstrate that the SCAVO-EAEID method achieved a superior performance with maximum TACC and VACC values. Notably, the SCAVO-EAEID methodology attained the highest TACC outcomes.
The TLS and VLS values, acquired by the proposed SCAVO-EAEID technique upon the CICIDS-2017 dataset, are portrayed in Figure 8. The results confirm that the SCAVO-EAEID method achieved a superior performance with low TLS and VLS values. Notably, the proposed SCAVO-EAEID method displayed the minimal VLS outcomes.
A comparative a c c u y examination was conducted between the proposed SCAVO-EAEID technique and other recent methods [12,23] and the results are shown in Table 5 and Figure 9. The outcomes infer that the SCAVO-EAEID technique accomplished the maximum a c c u y of 99.20%. Contrastingly, the rest of the models such as PRO-DLBIDCPS, BBFO-GRU, DT Model, MLIDS Model, CSPSO Model, CO Model, DNN-SVM Model, GA-Fuzzy, FCM Model, and GBT Model attained minimum a c c u y values such as 99.00%, 98.79%, 96.85%, 94.02%, 74.98%, 98.47%, 93.31%, 97.51%, 97.4%, and 84.64%, respectively.
To exhibit the enhanced performance of the SCAVO-EAEID technique, a brief time complexity analysis was conducted and the results are shown in Table 6. The outcomes infer that the DNN-SVM, GA-Fuzzy, FCM Model, GBT Model, BBFO-GRU, MLIDS, and CSPSO techniques demanded higher TRT and TST values. However, the SCAVO-EAEID technique accomplished a superior performance with minimal TRT and TST values such as 0.542 min and 0.246 min, respectively. These results highlight the supremacy of the proposed SCAVO-EAEID technique.

5. Conclusions

In this study, a new SCAVO-EAEID system has been introduced for intrusion classification in the CPS environment. The presented SCAVO-EAEID method emphasizes the detection of intrusions in the CPS environment using major sub-processes, namely, LSTM-AE-based classification, Z-score normalization, the SCAVO-FS technique, and RMSProp optimizer-based hyperparameter tuning. At the primary level, the SCAVO-EAEID technique applies Z-score normalization as a preprocessing step. Moreover, the SCAVO-FS technique is derived to elect the optimal feature subsets. Furthermore, the LSTM-AE model is applied for the detection and classification of intrusions. At last, the RMSProp optimizer is used for hyperparameter tuning of the LSTM-AE model. To demonstrate the remarkable performance of the proposed SCAVO-EAEID technique, two benchmark datasets were used. The experimental outcomes reiterated the significant performance of the proposed SCAVO-EAEID technique over other approaches. In the future, the performance of the presented method can be enhanced using metaheuristics-based feature selection techniques.

Author Contributions

Data curation, M.I.A.; Formal analysis, F.A.-M. and S.D.; Investigation, M.M.; Methodology, L.A. and F.A.-M.; Project administration, A.M.H.; Resources, S.A. and S.D.; Software, M.I.A.; Supervision, L.A.; Visualization, H.M.; Writing—original draft, L.A., F.A.-M., M.M., H.M. and A.M.H.; Writing—review & editing, M.M., A.M.H., M.I.A. and S.D. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through the Large Groups Project under grant number (RGP2/35/44). Princess Nourah bint Abdulrahman University Researchers Supporting Project Number (PNURSP2023R349), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. Research Supporting Project number (RSPD2023R787), King Saud University, Riyadh, Saudi Arabia. This study is supported via funding from Prince Sattam bin Abdulaziz University project number (PSAU/2023/R/1444).

Institutional Review Board Statement

Not aplicable.

Informed Consent Statement

Not aplicable.

Data Availability Statement

Data sharing not applicable to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through the contributions of all authors. All authors have given approval to the final version of the manuscript.

References

  1. Bashendy, M.; Tantawy, A.; Erradi, A. Intrusion response systems for cyber-physical systems: A comprehensive survey. Comput. Secur. 2022, 124, 102984. [Google Scholar] [CrossRef]
  2. Pham, L.N.H. Exploring Cyber-Physical Energy and Power System: Concepts, Applications, Challenges, and Simulation Approaches. Energies 2023, 16, 42. [Google Scholar] [CrossRef]
  3. Alzahrani, A.O.; Alenazi, M.J. Designing a Network Intrusion Detection System Based on Machine Learning for Software Defined Networks. Future Internet 2021, 13, 111. [Google Scholar] [CrossRef]
  4. Disha, R.A.; Waheed, S. Performance analysis of machine learning models for intrusion detection system using Gini Impurity-based Weighted Random Forest (GIWRF) feature selection technique. Cybersecurity 2022, 5, 1–22. [Google Scholar] [CrossRef]
  5. Megantara, A.A.; Ahmad, T. A hybrid machine learning method for increasing the performance of network intrusion detection systems. J. Big Data 2021, 8, 142. [Google Scholar] [CrossRef]
  6. Almomani, O. A Feature Selection Model for Network Intrusion Detection System Based on PSO, GWO, FFA and GA Algorithms. Symmetry 2020, 12, 1046. [Google Scholar] [CrossRef]
  7. Almomani, O. A Hybrid Model Using Bio-Inspired Metaheuristic Algorithms for Network Intrusion Detection System. Comput. Mater. Contin. 2021, 68, 409–429. [Google Scholar] [CrossRef]
  8. Mohammad, A.H.; Alwada’n, T.; Almomani, O.; Smadi, S.; ElOmari, N. Bio-inspired Hybrid Feature Selection Model for Intrusion Detection. Comput. Mater. Contin. 2022, 73, 133–150. [Google Scholar] [CrossRef]
  9. Almaiah, M.A.; Hajjej, F.; Ali, A.; Pasha, M.F.; Almomani, O. A Novel Hybrid Trustworthy Decentralized Authentication and Data Preservation Model for Digital Healthcare IoT Based CPS. Sensors 2022, 22, 1448. [Google Scholar] [CrossRef]
  10. Alohali, M.A.; Al-Wesabi, F.N.; Hilal, A.M.; Goel, S.; Gupta, D.; Khanna, A. Artificial intelligence enabled intrusion detection systems for cognitive cyber-physical systems in industry 4.0 environment. Cogn. Neurodyn. 2022, 16, 1045–1057. [Google Scholar] [CrossRef]
  11. Huang, X.; Liu, J.; Lai, Y.; Mao, B.; Lyu, H. EEFED: Personalized Federated Learning of Execution & Evaluation Dual Network for CPS Intrusion Detection. IEEE Trans. Inf. Forensics Secur. 2022, 18, 41–56. [Google Scholar]
  12. Mansour, R.F. Artificial intelligence based optimization with deep learning model for blockchain enabled intrusion detection in CPS environment. Sci. Rep. 2022, 12, 12937. [Google Scholar] [CrossRef] [PubMed]
  13. Henry, A.; Gautam, S.; Khanna, S.; Rabie, K.; Shongwe, T.; Bhattacharya, P.; Sharma, B.; Chowdhury, S. Composition of Hybrid Deep Learning Model and Feature Optimization for Intrusion Detection System. Sensors 2023, 23, 890. [Google Scholar] [CrossRef] [PubMed]
  14. Ortega-Fernandez, I.; Sestelo, M.; Burguillo, J.C.; Piñón-Blanco, C. Network intrusion detection system for DDoS attacks in ICS using deep autoencoders. Wirel. Netw. 2023, 1–17. [Google Scholar] [CrossRef]
  15. Wang, Z.; Li, Z.; He, D.; Chan, S. A lightweight approach for network intrusion detection in industrial cyber-physical systems based on knowledge distillation and deep metric learning. Expert Syst. Appl. 2022, 206, 117671. [Google Scholar] [CrossRef]
  16. Mittal, H.; Tripathi, A.K.; Pandey, A.C.; Alshehri, M.D.; Saraswat, M.; Pal, R. A new intrusion detection method for cyber–physical system in emerging industrial IoT. Comput. Commun. 2022, 190, 24–35. [Google Scholar] [CrossRef]
  17. Presekal, A.; Stefanov, A.; Rajkumar, V.S.; Palensky, P. Attack Graph Model for Cyber-Physical Power Systems using Hybrid Deep Learning. IEEE Trans. Smart Grid 2023, in press. [CrossRef]
  18. Choubey, S.; Barde, S.; Badholia, A. Enhancing the prediction efficiency of virus borne diseases using enhanced backpropagation with an artificial neural network. Meas. Sens. 2022, 24, 100505. [Google Scholar] [CrossRef]
  19. Abdollahzadeh, B.; Gharehchopogh, F.S.; Mirjalili, S. African vultures optimization algorithm: A new nature-inspired metaheuristic algorithm for global optimization problems. Comput. Ind. Eng. 2021, 158, 107408. [Google Scholar] [CrossRef]
  20. Nayak, S.R.; Khadanga, R.K.; Panda, S.; Sahu, P.R.; Padhy, S.; Ustun, T.S. Participation of Renewable Energy Sources in the Frequency Regulation Issues of a Five-Area Hybrid Power System Utilizing a Sine Cosine-Adopted African Vulture Optimization Algorithm. Energies 2023, 16, 926. [Google Scholar] [CrossRef]
  21. Yazdinejad, A.; Kazemi, M.; Parizi, R.M.; Dehghantanha, A.; Karimipour, H. An ensemble deep learning model for cyber threat hunting in industrial internet of things. Digit. Commun. Netw. 2022, 9, 101–110. [Google Scholar] [CrossRef]
  22. Babu, D.V.; Karthikeyan, C.; Kumar, A. Performance Analysis of Cost and Accuracy for Whale Swarm and RMSprop Optimizer. IOP Conf. Ser. Mater. Sci. Eng. 2020, 993, 012080. [Google Scholar] [CrossRef]
  23. Althobaiti, M.M.; Kumar, K.P.M.; Gupta, D.; Kumar, S.; Mansour, R.F. An intelligent cognitive computing based intrusion detection for industrial cyber-physical systems. Measurement 2021, 186, 110145. [Google Scholar] [CrossRef]
Figure 1. Working Principle of the SCAVO-EAEID technique.
Figure 1. Working Principle of the SCAVO-EAEID technique.
Sensors 23 04804 g001
Figure 2. Structure of AE.
Figure 2. Structure of AE.
Sensors 23 04804 g002
Figure 3. Overall classification outcomes of the proposed SCAVO-EAEID technique and other techniques on the NSL-KDD dataset.
Figure 3. Overall classification outcomes of the proposed SCAVO-EAEID technique and other techniques on the NSL-KDD dataset.
Sensors 23 04804 g003
Figure 4. TACC and VACC analytical outcomes of the SCAVO-EAEID technique on the NSL-KDD dataset.
Figure 4. TACC and VACC analytical outcomes of the SCAVO-EAEID technique on the NSL-KDD dataset.
Sensors 23 04804 g004
Figure 5. TLS and VLS analytical outcomes of the SCAVO-EAEID technique on the NSL-KDD dataset.
Figure 5. TLS and VLS analytical outcomes of the SCAVO-EAEID technique on the NSL-KDD dataset.
Sensors 23 04804 g005
Figure 6. Overall classification outcomes of the SCAVO-EAEID and other techniques on the CICIDS-2017 dataset.
Figure 6. Overall classification outcomes of the SCAVO-EAEID and other techniques on the CICIDS-2017 dataset.
Sensors 23 04804 g006
Figure 7. TACC and VACC analytical outcomes of the SCAVO-EAEID technique upon the CICIDS-2017 dataset.
Figure 7. TACC and VACC analytical outcomes of the SCAVO-EAEID technique upon the CICIDS-2017 dataset.
Sensors 23 04804 g007
Figure 8. TLS and VLS analytical outcomes of the SCAVO-EAEID method upon the CICIDS-2017 dataset.
Figure 8. TLS and VLS analytical outcomes of the SCAVO-EAEID method upon the CICIDS-2017 dataset.
Sensors 23 04804 g008
Figure 9. Overall a c c u y outcomes of the SCAVO-EAEID and other recent techniques.
Figure 9. Overall a c c u y outcomes of the SCAVO-EAEID and other recent techniques.
Sensors 23 04804 g009
Table 1. Best cost results of the SCAVO-FS technique and other techniques.
Table 1. Best cost results of the SCAVO-FS technique and other techniques.
Best Cost
MethodsNSL-KDD-2015CICIDS-2017
SCAVO-FS0.051010.41204
AHSA-FS0.054330.04311
BBFO-FS0.073820.06445
BFO-FS0.093710.08753
SSO-FS0.103840.09422
WOA-FS0.119400.11790
Table 2. FS Results of the proposed SCAVO-FS technique and other techniques.
Table 2. FS Results of the proposed SCAVO-FS technique and other techniques.
Number of Selected Features
MethodsNSL-KDD-2015CICIDS-2017
Total Features4180
SCAVO-FS1417
AHSA-FS1519
BBFO-FS1824
BFO-FS1930
SSO-FS2028
WOA-FS2033
Table 3. Classification outcomes of the proposed SCAVO-EAEID technique and other techniques on the NSL-KDD dataset.
Table 3. Classification outcomes of the proposed SCAVO-EAEID technique and other techniques on the NSL-KDD dataset.
Training/Testing Phase (%)AccuracyPrecisionRecallF-Score
40:60
SCAVO-EAEID98.7099.1698.1399.23
PRO-DLBIDCPS98.2998.8097.7498.91
BBFO-GRU Model97.9298.4497.4298.41
Optimal GRU Algorithm97.4498.2197.0298.05
GRU Algorithm97.1697.8596.7997.69
50:50
SCAVO-EAEID98.7499.2498.1499.53
PRO-DLBIDCPS98.4899.0397.9299.30
BBFO-GRU Model98.1298.7397.6598.96
Optimal GRU Algorithm97.9298.3297.2798.53
GRU Algorithm97.6397.8796.8098.27
60:40
SCAVO-EAEID98.9199.5098.1799.71
PRO-DLBIDCPS98.4199.1597.9099.30
BBFO-GRU Model97.9698.7197.5498.87
Optimal GRU Algorithm97.6298.3497.2198.60
GRU Algorithm97.2597.9996.8698.40
70:30
SCAVO-EAEID98.9599.5099.1299.81
PRO-DLBIDCPS98.699.1598.8199.58
BBFO-GRU Model98.3398.9398.4599.19
Optimal GRU Algorithm98.0298.4497.9998.69
GRU Algorithm97.6998.1697.6298.29
80:20
SCAVO-EAEID99.2099.5899.4299.84
PRO-DLBIDCPS99.0099.1299.0399.41
BBFO-GRU Model98.7998.8998.5598.95
Optimal GRU Algorithm98.4998.4798.2498.52
GRU Algorithm98.2498.1697.9198.26
Table 4. Classification outcomes of the SCAVO-EAEID and other techniques on the CICIDS-2017 dataset.
Table 4. Classification outcomes of the SCAVO-EAEID and other techniques on the CICIDS-2017 dataset.
Training/Testing Phase (%)AccuracyPrecisionRecallF-Score
40:60
SCAVO-EAEID99.0498.9798.1799.26
PRO-DLBIDCPS98.7398.6397.9198.76
BBFO-GRU Model98.2698.3897.6598.42
Optimal GRU Algorithm97.7997.9097.2798.18
GRU Algorithm97.4997.5197.0797.72
50:50
SCAVO-EAEID99.1199.5298.4799.46
PRO-DLBIDCPS98.6299.1398.2399.07
BBFO-GRU Model98.3698.8497.7498.69
Optimal GRU Algorithm98.0898.5697.2998.37
GRU Algorithm97.6298.2896.8898.11
60:40
SCAVO-EAEID98.7699.3298.1499.51
PRO-DLBIDCPS98.4398.8997.6899.01
BBFO-GRU Model98.0198.4997.4098.63
Optimal GRU Algorithm97.6397.9996.9298.31
GRU Algorithm97.2997.5996.6197.92
70:30
SCAVO-EAEID99.1899.5499.4299.62
PRO-DLBIDCPS98.8399.2799.1499.21
BBFO-GRU Model98.5198.9398.7098.72
Optimal GRU Algorithm98.0798.7198.3398.42
GRU Algorithm97.8198.3698.0798.03
80:20
SCAVO-EAEID99.1099.6799.8299.73
PRO-DLBIDCPS98.6099.2399.5599.51
BBFO-GRU Model98.2598.8499.2499.29
Optimal GRU Algorithm97.7998.4098.9899.01
GRU Algorithm97.5297.9298.5498.64
Table 5. Comparative a c c u y analysis outcomes achieved by the proposed SCAVO-EAEID technique and other techniques.
Table 5. Comparative a c c u y analysis outcomes achieved by the proposed SCAVO-EAEID technique and other techniques.
MethodsAccuracy (%)
SCAVO-EAEID99.20
PRO-DLBIDCPS Model [12]99.00
BBFO-GRU Model [23]98.79
DT Model [12]96.85
MLIDS Model [12]94.02
CSPSO Model [12]74.98
CO Model [12]98.47
DNN-SVM Model [12]93.31
GA-Fuzzy [12]97.51
FCM Model [12]97.4
GBT Model [12]84.64
Table 6. Time complexity analysis outcomes of the SCAVO-EAEID and other techniques.
Table 6. Time complexity analysis outcomes of the SCAVO-EAEID and other techniques.
MethodsTraining Time (min)Testing Time (min)
SCAVO-EAEID0.5420.246
PRO-DLBIDCPS Model [12]0.7520.381
BBFO-GRU Model [23]1.1060.363
DT Model [12]0.8880.677
MLIDS Model [12]1.2120.331
CSPSO Model [12]1.2420.425
CO Model [12]0.8020.572
DNN-SVM Model [12]1.3840.996
GA-Fuzzy [12]1.3510.444
FCM Model [12]1.7490.873
GBT Model [12]1.4630.875
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Almuqren, L.; Al-Mutiri, F.; Maashi, M.; Mohsen, H.; Hilal, A.M.; Alsaid, M.I.; Drar, S.; Abdelbagi, S. Sine-Cosine-Adopted African Vultures Optimization with Ensemble Autoencoder-Based Intrusion Detection for Cybersecurity in CPS Environment. Sensors 2023, 23, 4804. https://doi.org/10.3390/s23104804

AMA Style

Almuqren L, Al-Mutiri F, Maashi M, Mohsen H, Hilal AM, Alsaid MI, Drar S, Abdelbagi S. Sine-Cosine-Adopted African Vultures Optimization with Ensemble Autoencoder-Based Intrusion Detection for Cybersecurity in CPS Environment. Sensors. 2023; 23(10):4804. https://doi.org/10.3390/s23104804

Chicago/Turabian Style

Almuqren, Latifah, Fuad Al-Mutiri, Mashael Maashi, Heba Mohsen, Anwer Mustafa Hilal, Mohamed Ibrahim Alsaid, Suhanda Drar, and Sitelbanat Abdelbagi. 2023. "Sine-Cosine-Adopted African Vultures Optimization with Ensemble Autoencoder-Based Intrusion Detection for Cybersecurity in CPS Environment" Sensors 23, no. 10: 4804. https://doi.org/10.3390/s23104804

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop