Next Article in Journal
A Hybrid GA/ML-Based End-to-End Automated Methodology for Design Acceleration of Wireless Communications CMOS LNAs
Previous Article in Journal
Ti/HfO2-Based RRAM with Superior Thermal Stability Based on Self-Limited TiOx
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Intrusion Detection System in the Cloud Using Fusion Feature Selection Approaches and an Ensemble Classifier

1
Department of Computer Science and Engineering, Faculty of Engineering, C. V. Raman Global University, Bhubaneswar 752054, India
2
Department of Information Systems, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
3
Department of Computer Engineering and Applications, GLA University, Mathura 281406, India
4
Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Guntur 522302, India
5
Department of Software Engineering, LUT University, 53850 Lappeenranta, Finland
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(11), 2427; https://doi.org/10.3390/electronics12112427
Submission received: 27 April 2023 / Revised: 17 May 2023 / Accepted: 23 May 2023 / Published: 27 May 2023

Abstract

:
The application of cloud computing has increased tremendously in both public and private organizations. However, attacks on cloud computing pose a serious threat to confidentiality and data integrity. Therefore, there is a need for a proper mechanism for detecting cloud intrusions. In this paper, we have proposed a cloud intrusion detection system (IDS) that is focused on boosting the classification accuracy by improving feature selection and weighing the ensemble model with the crow search algorithm (CSA). The feature selection is handled by combining both filter and automated models to obtain improved feature sets. The ensemble classifier is made up of machine and deep learning models such as long short-term memory (LSTM), support vector machine (SVM), XGBoost, and a fast learning network (FLN). The proposed ensemble model’s weights are generated with the CSA to obtain better prediction results. Experiments are executed on the NSL-KDD, Kyoto, and CSE-CIC-IDS-2018 datasets. The simulation shows that the suggested system attained more satisfactory results in terms of accuracy, recall, precision, and F-measure than conventional approaches. The detection rate and false alarm rate (FAR) of different attack types was more efficient for each dataset. The classifiers’ performances were also compared individually to the ensemble model in terms of the false positive rate (FPR) and false negative rate (FNR) to demonstrate the ensemble model’s robustness.

Graphical Abstract

1. Introduction

Cloud computing is a technology that provides internet services on demand with paid resources based on usage [1,2]. As cloud computing does not require large investments in infrastructure and development, government and private organizations have started to utilize the advantages of the cloud by deploying it in their operations [3,4]. The security of cloud computing is of prime importance, since individuals’ and entities’ information is stored in cloud data centers, which perhaps face security threats if a network is infiltrated by an attacker [5]. Since the network acts as the backbone of the cloud and facilitates the provision of cloud services to clients, any threats or weaknesses in the network directly affect the overall security and success of the cloud. Thus, safeguarding the network against any potential threats is a significant matter. The cloud utilizes an array of cybersecurity methods, including firewalls, Intrusion Prevention Systems (IPSs), and Intrusion Detection Systems (IDSs), to tackle various security problems. Some of the attacks faced by cloud computing are denial-of-service (DoS) attacks, routing information attacks, distributed denial-of-service (DDoS) attacks, SQL injection attacks, cross-site scripting XSS attacks, etc. [6,7]. The foremost security challenge in cloud computing involves the detection and thwarting of network attacks. Lately, network intrusions have escalated due to insufficient countermeasures, and an IDS can address these security concerns. However, before deploying any IDS model in a cloud environment, it is crucial to ensure that the proposed system is well developed and performs effectively. Consequently, this paper aims to create such an efficient IDS.
An IDS is used to detect an intrusion in a network, monitor all packets, and decide whether any of the incoming and outgoing packets have been affected by the intrusion [8]. In conventional intrusion detection models, statistical and knowledge-based approaches were employed. These approaches struggled with the detection of unknown attacks and also faced difficulties processing large amounts of network traffic data [9]. Machine learning offers significant potential as a technique capable of developing a range of robust strategies to reinforce the security of systems such as cloud computing, IoT, and any other network-based system. A notable application of machine learning is in the design of an IDS. The role of this system is to analyze network traffic, distinguish between normal and abnormal behaviors, and appropriately categorize them [10]. ML approaches are considered to be better than conventional models since ML models possess the capability to learn complex traffic patterns and can precisely identify the attack [11]. Basic ML models face limitations when intrusion is more complex and diverse [8]. Studies conducted on conventional approaches show that ensemble models provide better results than single ML-based classifiers. Ensemble-based models are generated from numerous ML and DL models, so they offer a higher accuracy and a reduced false alarm rate (FAR) [12]. The ensemble-based models possess higher data processing capacities and better classification rates. Ensemble-based cloud intrusion detection has improved accuracy with a feature set formed by combining both the filter-based and automated feature selection models. Stacked autoencoder (SAE)-based automated feature selection helps to minimize the dimensions of the feature set by eliminating redundant and irrelevant features [13]. Conventional approaches have not focused on generating a comprehensive feature set. Since the selection of a better feature set ensures better classification results, both the filter and automated features are combined to make a set of features in this research. Our ensemble model employs classifiers such as SVM, LSTM, XGBoost, and FLN. The weights of the classifiers are optimized with CSA to enhance the classification results. In short, our main contributions are as follows:
  • An ensemble model is developed with the ML and DL models with SVM, LSTM, XGBoost, and FLN.
  • A comprehensive feature set is generated with the help of a filter and automated feature selection.
  • The weights of the selected ML and DL models are selected with the CSA.
  • We overcome the issue of unbalanced datasets by utilizing the SMOTE algorithm (Synthetic Minority Oversampling Technique), with the purpose of enhancing the detection accuracy for minority-type attacks.
  • The experiment is validated using old and modern benchmark datasets (NSL-KDD, Kyoto, and CSE-CIC-IDS-2018), that reflect a real-world environment involving the latest attacks. The outcomes are calculated in terms of accuracy, recall, false positive rate, false negative rate, precision, and F-measure.
We organize the rest of our work as follows: Section 2 details the related work, Section 3 explains the proposed model, Section 4 illustrates the experimental results, and Section 5 ends with the conclusions and future work.

2. Related Work

Ensemble techniques are built on the notion of combining several classifiers to achieve a competitive edge. Each classifier comes with its own set of benefits and drawbacks. Some modules may be good at detecting certain types of attacks but not so good at detecting others. The ensemble approach picks the best result using a voting mechanism. Various studies have illustrated that ensemble approaches generally give better, more accurate outcomes than a single model [14]. Kushwah and Ranga [15] presented a voting extreme learning machine (V-ELM) for detecting DDOS attacks in cloud computing, which uses a large number of ELMs instead of a single ELM. Majority voting combines the results of all ELMs to obtain the final results. With multiple ELMs, the accuracy of detection increases and false alarms decrease. Srilatha and Shyam [8] proposed a cloud intrusion detection technique by combining kernel fuzzy c-means clustering (KFCM) and an optimal type 2 fuzzy neural network (O2TFNN). Lion optimization was employed to select the optimal parameters of the T2FNN. The performance of the method was simulated with the NSL-KDD dataset based on precision, recall, and F-measure. Xu et al. [9] suggested an intrusion detection model with a deep neural network (DNN). The improved LSTM model used in the proposed approach consists of gated recurrent units (GRU) along with a multilayer perceptron. The proposed LSTM-based approach achieved a higher detection rate for KDD 99 and NSL-KDD datasets. Mighan and Kahani [11] offered a deep-learning-based intrusion detection technique. A stacked autoencoder extracts the features, and classification approaches such as SVM, random forest (RF), decision forests, and naive Bayes were employed to obtain the classification results. The UNB ISCX 2012 dataset was evaluated with the proposed approach regarding accuracy, precision, sensitivity, and F-measure, in which it attained a better accuracy and reduced time consumption. Mayuranathan et al. [12] proposed an intrusion detection system using a restricted Boltzmann machine (RBM) model. The optimal features are selected with the random harmony search model, and the detection rate is enhanced with the seven extra layers in the RBM. In addition, the hyper-parameters are optimized for better results. Gaussian distribution replaces probability distribution in the proposed RBM model. Krishnaveni et al. [16] introduced an ensemble model for network intrusion with an efficient feature selection approach, in which the feature sets are reduced using univariate feature selection and classifiers such as SVM, naive Bayes, logistic regression, and decision trees are used as an ensemble model, with the class determined by a majority vote. The performance of this methodology, UEFFS, was evaluated with the NSL-KDD, Kyoto 2006, and real-time honeypot datasets, recording a high accuracy with a lower false alarm rate.
Thaseen et al. [17] demonstrated an ensemble of LSTM with a genetic algorithm (GA), in which the GA selects the features effectively. The ensemble of LSTM employs a voting approach based on the average probability combination rule. The intrusion detection technique evaluates the proposed model with criterion datasets, namely NSL-KDD and UNSW-NB15, in which a higher accuracy and detection rate and a minimum false alarm rate were achieved. The authors of [18] proposed an ensemble model for intrusion detection where the ensemble model is robust and able to produce results with less computation overhead. Parul Singh and Virendra Ranga [19] illustrated an ensemble learning approach in cloud computing, which used four machine learning techniques, namely boosted tree, bagged tree, subspace discriminant, and RUS boosted. The performance of the proposed approach was simulated with the CICIDS 2017 dataset and exhibited better results based on the detection rate and execution time. Mehanovic et al. [20] displayed an efficient feature selection for intrusion detection. The features were reduced with ML approaches such as SVM, artificial neural networks (ANNs), random tree (RT), logistic regression, and naive Bayes. The parallel genetic algorithm was implemented with an open-source MapReduce library that helps select better features. The authors of [21] presented a network intrusion detection mechanism where a CNN model based on contextual feature extraction improves the accuracy of the IDS. The approach reduced the feature set before feeding it into a CNN to reduce the computational time. Tummalapalli and Chakravarthy [22] exhibited Bayesian fuzzy clustering and a two-level gravitational group search-based support vector neural network (GG-SVNN). In this approach, intrusion information is gathered with a level 1 classifier, and level 2 classifier makes the final decision on the presence or absence of intruders. The proposed approach achieved a better accuracy and a lower false alarm rate.
Amali Angel Punitha and Indumathi [23] proposed ensemble-based attack detection. The approach consists of centralized cloud information and accountability integrity with an imperialist competitive key generation algorithm (CCIAI-ICKGA) proposed for key generation. The attack detection rate is high and the computation time is less than the current methods. Su et al. [24] employed a DL model to detect intrusion in the cloud network, which uses the BAT model as a combination of bidirectional LSTM and an attention mechanism. The local features of the traffic data are captured by the multiple convolutional layers. Softmax classifies the network traffic. The experimental results proved that the approach achieved a higher accuracy. Bhati et al. [25] implemented an ensemble-based intrusion detection approach with XGBoost. The ensemble-based XGBoost focused on improving the selection of features and the accuracy. The bias–variance tradeoff is smoothened with the tree-boosting machine learning algorithm. The approach was implemented on the KDDcup99 dataset and achieved an accuracy of 99.95%. Suman et al. [26] presented a multi-objective optimization (MOO) to select features. The optimal feature sets were obtained with the nondominated sorting genetic algorithm (NSGA-II). The optimal feature sets were applied to ML classifiers such as decision tree, SVM, RF, k-nearest neighbor (KNN), Adaboost, etc., to determine the effect of optimized features on various ML classifiers. The approach was simulated with various datasets such as Kyoto 2006+, NSL-KDD, and KDD-99. Rajagopal et al. [27] recommended a stacking-ensemble-based paradigm for intrusion detection systems (IDS). They stressed that this limitation is a significant component in a stacked ensemble’s performance when compared to the single best learner. Lopez et al. [28] introduced an NIDS utilizing a conditional variational autoencoder (CVAE). Their proposal marked the first application of a CVAE, which offered the ability to retrieve missing categorical features from incomplete training datasets. The proposed solution features a unique architecture that embeds intrusion labels into the decoder layers. When tested on the NSL-KDD dataset, the model outperformed other well-known classifiers, achieving an accuracy rate of up to 99%.
Zhou et al. [29] presented an IDS approach that leverages feature selection and ensemble learning methods. Initially, they proposed a correlation-based feature selection combined with a bat algorithm (CFS-BA), aimed at identifying the ideal subset. Subsequently, they designed an ensemble classifier using C4.5, random forest (RF), and forest by penalizing attributes (Forest PA), along with the AOP rule to construct the classification model. Ultimately, they applied a voting strategy to merge the probability distributions of foundational learners for detecting attacks. The validation was conducted using the NSL-KDD, AWID, and CIC-IDS2017 datasets. Although the overall performance was superior, the accuracy of the proposal for minority classes (R2L and U2R) in the NSL-KDD dataset was not satisfactory due to inadequate addressing of the imbalanced data issue. Balyan et al. [30] presented an effective hybrid network-based IDS (HNIDS) model, created using a combination of the enhanced genetic algorithm, particle swarm optimization (EGA-PSO), and the improved random forest (IRF) method. Initially, the HNIDS applied the hybrid EGA-PSO method to bolster minor data instances, thereby creating a balanced dataset. The PSO method assists in vector enhancement, while the GA is improved with a multi-objective function for optimal feature selection. Subsequently, the IRF method is employed as a classifier. The model’s efficiency was evaluated using the NSL-KDD dataset in both binary and multi-class classification scenarios. Bakro et al. [31] suggested an IDS that utilizes filter methods (information gain, gain ratio, and chi-square) to select relevant features and employs the SVM model for classification. However, they did not tackle the issue of unbalanced data. Their model underwent testing using the NSL-KDD, Kyoto, and CSE-CIC-IDS-2018 datasets.
From the preceding relevant works and methodologies outlined in Tables 7–9, we noted some constraints, such as the inadequate handling of unbalanced data issues and the use of a single classifier instead of an ensemble approach, which may not yield satisfactory performance across all categories of the used dataset. Moreover, we have discerned that designing an efficient intrusion detection system necessitates well-prepared datasets, and issues related to imbalanced data must be adequately addressed. Additionally, the reduction of high-dimensional datasets through feature selection methods also plays a pivotal role in the success of the IDS model; therefore, we found that the majority of these studies employed various techniques to select optimal features, including wrapper, filter, metaheuristic, and unsupervised models, among others. As a result, in our research, we constructed our proposal by leveraging the strengths and addressing the potential weaknesses of other studies to achieve superior results.

3. Methodology

The proposed cloud intrusion detection system consists of pre-processing to reduce the class imbalance problem. The final feature set is formed with the combination of filter and automatic feature selection approaches.
Figure 1 presents a model of the proposed technique. The ML and DL models are trained with the feature set. Then, the weights of the ML and DL models are optimized with the CSA approach, and the final classification results are obtained from the ensemble model using the weighted voting concept.

3.1. Pre-Processing

This section details the pre-processing of the data in order to prepare them for feature selection. The creation of a better training dataset is also required [32]. The dataset contains a variety of symbolic, binary, numerical data, etc. Some of them may have imperfect and inconsistent values that lead to unsatisfactory outcomes. The proposed system only takes clean and numeric inputs; thus, the dataset must be prepared first. Therefore, we implemented the following steps:
  • Removing the noisy, incomplete outliers and irrelevant, duplicate values and deleting features: In NSL-KDD we deleted features such as “difficulty_level”, and in the Kyoto dataset, we omitted features such as “Source IP Address”, “Destination IP Address”, etc. Furthermore, “Timestamp”, “Destination Address”, “Source Address”, “Source Port”, and other features were removed in CSE-CIC-IDS 2018 [33], that means we only retain one useful value.
  • Feature encoding, also known as feature mapping, which means transforming categorical variables into numerical form. There are two methods used for this, one-hot encoding and ordinal encoding, in which one-hot encoding yields a higher classifier performance than ordinal coding [34]. As a result, one-hot encoding was adopted in this study. Protocol type, service, and flag are three symbolical characteristics in the NSL-KDD dataset. For example, the feature of protocol type has three category values (TCP, UDP, and ICMP). They form binary vectors (1, 0, 0), (0, 1, 0), and (0, 0, 1) after one-hot encoding [35], and we will perform the same operation on the Kyoto dataset. In the CSE-CIC-IDS 2018 dataset, we implement one-hot encoding on a protocol, that is a nominal feature that has three groups (protocol 0.0: hop-by-hop IPv6 “HOPOPT”, protocol 6.0: TCP, and protocol 17.0: UDP) [36].
  • Feature scaling, also known as data normalization, which is a technique to convert the whole scope of values from a set of features into a predetermined range. Common methods of feature scaling are normalization and standardization. Standardization is called also Z-score normalization, where a single standard deviation and values centered around the mean indicate that the attribute’s mean tends to zero, and the resulting distribution has a standard deviation of one unit. Normalization or Min-Max scaling is the process of shifting and rescaling values so that they fall in the range [0,1], which offers satisfying outcomes in the AE procedure [36]. In our research, we used Min-Max scaling, shown in the following equation:
    X j * = ( X j X min ) / ( X max X min )
    where X j is the initial value, X j * are the data after the process, X min is the impacting data sequence’s lowest value, and X max is said sequence’s highest value [37].
  • Depending on the attack type, the label/class field must be gathered. The class column is split into two groups for detection: normal and abnormality. Regarding attack classification, the class column is separated into types as follows: 4 types (DOS, Probe, R2L, and U2R) in NSL-KDD [38], 2 types (known and unknown attacks) in the Kyoto dataset, and 14 types (DDOS attacks-LOIC-HTTP, DDOS attack-HOIC, DDOS attack-LOIC-UDP, FTP-Brute Force, SSH-Brute force, Brute Force-XSS, Brute Force-Web, DOS attacks-SlowHTTPTest, DOS attacks-Hulk, DOS attacks-GoldenEye, DOS attacks-Slowloris, Infilteration, Bot, and SQL Injection) in the CSE-CIC-IDS 2018 dataset.
  • Feature correlation. This is a useful technique for feature engineering, and is a statistical method that defines the relevance between one or more variables in order to detect the related important features and only keep these features [39].
After that, the pre-processed data are split into two sets: a training dataset and a test dataset. In order to address the class imbalance challenge which affects the performance of the classifiers for a minority of attacks [40], we employ the Synthetic Minority Over-Sampling Technique (SMOTE) by expanding the training dataset’s mitotic instances [41]. In the following step, we extract the features from the training dataset that will be used to train the proposed model. The test dataset, which generates various classifications, is then used to test the trained proposal.

3.2. Feature Selection

In the proposed method, after pre-processing, feature selection is handled. Feature selection (FS) is considered a pre-processing approach that aids in the selection of suitable features. Feature selection methods minimize data dimensionality by removing redundant and irrelevant attributes [13]. When irrelevant features are added to the classification process, not only is the accuracy of the process reduced, but it also takes up more space and time to execute. As a result, feature selection is critical since it allows for a deeper understanding of data by holding only the relevant appropriate features, therefore improving the classifier’s accuracy, speed, and predictive capacity [42]. One feature selection approach is the filter technique, which shortlists the essential attributes first, irrespective of any classification method. Variable ranking approaches are used in filter methods to score the variables. The most important variables are selected from a list of variables, leading to the elimination of less relevant attributes [43]. Therefore, the final feature vector is obtained from the filter and automated feature selection.

3.2.1. Filter Methods

Filter-based feature selection methods such as information gain (IG), gain ratio, chi-squared, and symmetric uncertainty are used to select the features, and are given as follows:

Information Gain

Information gain (IG) is a well-known filter-based feature selection approach. IG can reduce the noise caused due to irrelevant features. The execution time and computational complexity of IG are low. Features with a high IG are relevant. IG calculates the entropy change after utilizing the attribute, and thus it illustrates the importance of a specific feature. The impurity of a feature is measured in entropy. The higher the entropy, the more information the feature has; therefore, the entropy determines the best feature, which means that it will help predict the class label [26]. The entropy of a sample S is calculated according to the following:
E n t r o p y ( S ) = i n P i log 2 P i
where n and P refer to the number of classification values and the number of samples in class i, respectively. The information gain is given as:
I G ( S , A ) = E n t r o p y ( S ) V a l u e s ( A ) S n S E n t r o p y S n
where A is the attribute, V a l u e s ( A ) is the attribute’s potential value, S n is the number of samples for the value n, and S is the total number of samples. Entropy S n is the entropy for a sample that has a value of n. According to [42], the IG of feature A is given as [44]:
I G ( A ) = I n f o ( M ) I n f o A ( M )
where the entropy of the absolute dataset is I n f o ( M ) and the entropy of attribute A is I n f o A ( M ) .

Gain Ratio

Gain ratio (GR)-based feature selection improves the IG method. The GR is high when the data are evenly spread and it takes a small value when data come from only one branch of the attribute [13]. The GR is determined with the help of split information [16].
S p l i t i n f o A ( M ) = M n M log 2 M n M
where M is the number of probable values x can select and M n is the number of actual values of feature x. The gain ratio is obtained as:
G a i n r a t i o ( A ) = I G ( A ) / S p l i t I n f o ( A )

Chi-Squared

The chi-square method is one of the feature selection approaches that is used to find the relationship between two variables. The independence between two events can be found using this approach, and it may help to determine a feature’s independence from its class [16]. The chi-squared is given as:
γ 2 = i z O i z E i z 2 / E i z
where i and z represent two variables, O and E are the observed and expected value, respectively, and γ 2 indicates the Chi-squared value.

Symmetric Uncertainty

This form of feature selection is employed to evaluate the ranking of the outcome. The symmetric uncertainty is given as [16]:
S U ( i , j ) = 2 I G i i j j I G i i j j H ( i ) H ( i ) + H ( j )
where H ( i ) and H ( j ) represent the entropy of features i and j, respectively.

3.2.2. Automated Feature Selection with Stacked Autoencoder

An autoencoder is a common unsupervised DL approach that belongs to the family of artificial neural networks with three layers: input, hidden, and output layers [11]. The objective of the autoencoder is to obtain a similar reconstructed output and input, thereby attaining feature selection and a dimensionality reduction of the input. Autoencoders are able to learn efficient and compressed representations for the collection of data. In an AE, the data are reconstructed after computations based on the concept of learning the best features and matching the output to the input as closely as feasible [45]. Stacked AEs are one of the AE types designed for automated feature selection [46]. In an SAE, the output of the first AE is the input of the second one. This is repeated based on the depth of the architecture. The original information is encoded by the encoder to acquire high-level features in the middle layer, and the decoder reconstructs the input information. The hidden layer for the original training data x is given as follows [47]:
h i = f W 1 T x i + b 1
The activation function is represented by f = tanh . The output obtained with the decoding function is calculated as:
z i = W 2 T h i + b 2 x i
where W 1 T and W 2 T are the weight matrices and b 1 and b 2 are the different bias vectors. In order to train the AE, the expression is minimized as follows:
L X , Z = 1 1 2 i = 1 n x i z i 2 i = 1 n x i z i 2
where x i and z i denote the ith element of x and z, respectively, and n signifies the input set denoted by x i where i = 1 n .
Prior information is ignored in the AE, while an SAE learns a more satisfactory representation of the input data than a single AE. The hidden neurons in the proposed SAE for the two encoder layers are 50 and 30, respectively, as shown in Figure 2. An SAE functions essentially as a multi-layer AE. In this setup, the output from one layer provides the input for the subsequent layer. Typically, an SAE undergoes layer-by-layer training, a procedure commonly known as pretraining. Following this stage, the pre-trained network weights are amalgamated to establish the ultimate network weights. When using an SAE for feature selection, we begin by training the SAE on the input data, after which we utilize the encoded representations produced by the SAE as the new feature set. This new feature set possesses a smaller dimensionality and includes the most crucial attributes of the input data. Consequently, this can enhance the efficiency and performance of the classification model. The filter and automated features of SAE are combined to form the feature vector.

3.3. Classification Using the Ensemble Learning Approach

After the feature selection process, we propose an ensemble learning approach that includes several base classifiers, namely LSTM, SVM, XGBoost, and FLN. Each of these classifiers is trained using optimally selected features to solve the same problem, and their outputs are then combined to enhance the overall results. Two commonly used methods for combining predictions from different models are majority voting and weighted voting. Majority voting is ideal when all models in the ensemble have equal reliability. Conversely, if certain models consistently outperform others, weighted voting becomes the superior choice. In our ensemble learning approach, we have chosen to employ weighted voting due to the different performance levels of our base classifiers. Before detailing our chosen weighted voting method in Section 3.4, we provide a brief explanation of each model utilized in our proposed ensemble below.

3.3.1. LSTM

Long short-term memory is a form of recurrent neural network (RNN) that belongs to the family of supervised DL [48]. Compared to other RNNs, LSTM offers more advantages [33]. LSTM was developed to tackle the vanishing gradient problem that is encountered in RNNs through a concept known as gates. These gates perform an important role in prediction because they learn to keep relevant data while discarding irrelevant inputs [17]. LSTM consists of input, hidden, and output layers. One cell state layer and three gate layers make up the hidden layer: an update cell state, a forget gate layer, an input gate layer, and an output gate layer [49]. The mathematical model of the cell state layer and the three gate layers is given as [50]:
I n p u t g a t e : i t = σ W i X t + U i S t 1 F o r g e t g a t e : f t = σ W f X t + U f S t 1 O u t p u t g a t e : o t = σ W o X t + U o S t 1 T h e m e m o r y c e l l s n e w s t a t e : S ˜ t = tanh W c X t + U c S t 1 T h e m e m o r y c e l l s f i n a l s t a t e : S t = f t S t 1 + i t S t ˜ T h e m e m o r y u n i t s f i n a l o u t p u t : O t = S o f t max o t tanh S t
where W i , W f , W o , W c , U i , U f , U o , and U c are weight matrices, X t is the input vector, “∘” represents the Hadamard product, and “ σ ” denotes Sigmoid. In the final output layer, we utilize Softmax for the classification [9].

3.3.2. SVM

Support vector machine is a supervised ML technique utilized for classification and regression. The search for a max-margin separation hyperplane in the n-dimension space of features is the used approach in SVMs; therefore, we can obtain satisfying outcomes regardless of the size of the dataset [51]. Using support vectors, the hyperplane acts as a decision limit [20]. Both linear and nonlinear issues are addressed using this method. Kernel functions are applied in nonlinear situations; kernel techniques are commonly utilized in SVMs and other machine learning models [13]. SVMs were the most effective among ML algorithms in past years, until the appearance of DL [33]. SVMs are also employed in multi-class problems. The hyperplane is given as [52]:
H x = w T x + b
where x, w, and b denote the input, weight, and bias values, respectively. The objective of SVM is calculated as follows [53]:
M i n i m i z e : 1 2 w + C i μ i
S u b j e c t t o : z i w T x i + w o 1
where μ i is the training error, z i is the class label, x i is the feature vector, i is the sample number, and C is the cost parameter.

3.3.3. XGBoost

Extreme gradient boosting utilizes a tree ensemble model, which means XGBoost is a version of a multiple decision tree, because one tree may not be sufficient to provide satisfactory outcomes [54]. Decision trees belong to the class of supervised ML methods that are employed in regression and classification. To overcome the over-fitting issue, the DT algorithm automatically chooses the optimal attributes and removes irrelevant data to create a decision tree [45]. XGBoost can reduce the processing time and also possesses a higher accuracy. The accuracy is enhanced because XGBoost uses a second-order Taylor expansion [33]. The calculation is parallelized and the speed is increased as the feature is sorted and blocked by XGBoost. The output is predicted by XGBoost for a dataset that has k features. The N additive function to predict the output is given as follows [25]:
y i ^ = ϕ x i = n = 1 N f n x i , f n F
where f represents the tree structure and D is the ensemble model of the tree. The objective is minimized according to the following:
L t = i = 1 n g i f t x i + 1 2 h i f t 2 x i + Ω f t

3.3.4. FLN

Fast-learning networks are considered a double parallel forward neural network (FNN) in which a direct connection is added along with the input and output layers. There are three layers to the FLN: input, hidden, and output layers. The nonlinear relationship from the hidden to the output layer is merged in an FLN, as is the linear relationship from the input to the output layer. FLNs are considered a development of extreme learning machine (ELM) algorithms, where ELM algorithms were proposed to avoid the drawbacks of ANNs which is one of the supervised ML models. FLNs demonstrate excellent performance while consuming less time [45]. The FLN model is given as follows [37]:
T i = a X i + g W j X i + b j β j , i = 1 , 2 , q
In an FLN model, the weights are randomly created along with the hidden layer thresholds. The hidden layer output matrix H is calculated. The output weight matrix is finally formed, where we use SoftMax in the output layer with respect to the classification.

3.4. Crow Search Algorithm (CSA) and Weighted Voting

The objective of ensemble-based classification is to create a model with a high prediction capacity system. Each of the individual classifiers varies regarding the number of layers, pooling, filter size, etc. The predictions of individual DL and ML models are combined with the voting technique. In an ensemble model, majority voting and weighted voting are generally used. In majority voting, each classifier in the ensemble casts one vote. The class label that receives more than half of these votes, drawn from the set of predicted class labels from different classifiers, is then selected. In the weighted voting approach, base classifiers are assigned differing weights, reflecting their performance. Each classifier’s vote is then multiplied by this weight. The class that accumulates the highest total weighted vote becomes the ensemble’s final prediction. Therefore, the weighted voting assignment of weights on the classifiers has a considerable effect on the prediction of the ensemble model. The optimized weights are not randomly generated; the weights are obtained with an objective function so that weights will be tuned to improve the performance of the classifiers. Weight optimization is carried out with several optimization techniques such as genetic algorithms, particle swarm optimization, and differential evolution algorithms. Here, weighted voting with the CSA algorithm is proposed.
The prediction of individual models can be handled inside the ensemble. The ensemble takes into account all the classifiers’ prediction probabilities and decides the final prediction result based on the majority of the prediction probability results. The weight-based voting approach is chosen to assign weights to the classifiers to select the final prediction. The weights are decided with a meta-heuristic algorithm: the crow search algorithm (CSA). The CSA works based on the foraging pattern of crows. The crows represent the solution in this optimization approach. The initial position is randomly assigned, and the fitness value is calculated for the memorized position [55]. In the CSA, based on two conditions, the positions of the crows are determined. In the first condition, a crow is unaware that it is being followed by another crow to the location of the hidden food. In the second condition, the crow possesses the knowledge that another crow is following it, and to avoid theft, the crow will move to some other position in the search space [56]. The position of the crows in the search space is given as follows:
P i = W b 1 , W b 2 , . . . . . W b n
where P i is the position of the ith crow and W b n is the weight of the nth base classifier. The objective function of the weighted voting with the CSA is:
f P i = 1 k = 1 t P n a n 2 s
where p n is the predicted value of the class after the application of CSA-based weighted voting, a n is the actual value of the class, and s is the number of test samples. Based on the objective function, the CSA updates the crow’s position. The optimal weights are thus obtained from the CSA-based weighted voting approach for every classifier in the ensemble model. A final prediction is carried out with the optimized weights through weighted majority voting of the results obtained from the individual classifiers.

4. Experimental Results

In this part, we detail the datasets that we utilized, followed by the performance metrics. The outcomes of our proposal are then discussed, and finally, the constraints are shown. Our proposal was executed using libraries such as Scikit-learn and Tensorflow (ML-DL framework), and we used Imblearn to address the unbalanced dataset issue. The simulations were carried out in Python 3.9.0, with the Google Co-laboratory Pro platform with 25 GB RAM on an Intel Core i5 processor (ASUS, Bhubaneswar, India).

4.1. Dataset Description

4.1.1. NSL-KDD

This dataset is an upgraded version of KDD’99 that was developed by Tavallaee et al. [57] to handle some of the innate drawbacks of the KDD’99 dataset, where duplicate and redundant records have been removed in order to reduce the issue of data bias to make benchmarking learning algorithms more acceptable and realistic [58]. Thus, it is most commonly utilized in much research on intrusion detection [45]. We utilized KDDTrain+ and KDDTest+ as the training and test sets in the NSL-KDD dataset [59], which makes it useful to perform experiments. Every record is labeled as normal or abnormal based on the 42-feature set, and the abnormal type is divided into four kinds, namely DOS, probe attacks, remote to local (R2L) attacks, and user to root (U2R) attacks [57]. Although this dataset is perhaps not a perfect expression of real computer networks, we feel it is still a useful baseline for researchers to evaluate alternative methodologies [60]. Table 1 represents the NSL-KDD.

4.1.2. Kyoto

The Kyoto dataset was collected between 2006 and 2015 from a honeypot system [61], email servers, web crawlers, darknet sensors, and other servers installed at Kyoto University. In our research, we used the most recent dataset that contains traffic statistics through to 2015. The dataset has 24 features, 14 of which were gathered from KDD-99, and the other 10 are modern features that allow us to more properly investigate what occurred in our networks. Each record is classified according to a label, where 1 refers to normal traffic, −1 points out known attacks, and −2 indicates an unknown attack [26]. In the Kyoto dataset, there is no clear demarcation between the training and test sets and the dataset is extremely huge. As a result, we chose a random subset to test our assumptions, which is shown in Table 2.

4.1.3. CSE-CIC-IDS-2018

The Canadian Institute of Cybersecurity (CIC) with the Communications Security Establishment (CSE) created this dataset, which is a modernised version of CIC-IDS 2017 [62]. It is the most up-to-date launched dataset for intrusion detection research, which reflects the present real network environment, including seven traffic categories [36]: Benign, Botnet, Web Attacks, DDoS, DoS, Brute Force, Infilteration, and 14 types of labeled attacks with 84 features [63]. Additionally, there is no formal distinction between training and test sets in CSE-CIC-IDS 2018, and it is an excessively large dataset. Therefore, we randomly selected a subset to test our proposal, which is indicated in Table 3.

4.2. Performance Metrics

In this passage, we explain the evaluation measures for assessing the performance of our proposed model, which are calculated based on the various variables in the confusion matrix, a two-dimensional matrix that describes the actual/predicted classes, as is presented in Table 4 [45]. Correct identification of the data instances by classifiers as an attack is denoted a True Positive (TP) case, while if correctly identifying it as normal it is denoted a True Negative (TN). However, incorrect identification of the data instances by classifiers as an attack is denoted a False Positive (FP), while if correctly identifying as normal it is denoted a False Negative (FN) [31].
  • Accuracy (ACC): It is the proportion of instances that have been correctly detected to the overall number of instances. It is determined as follows:
    A C C = T P + T N T P + T N + F P + F N
  • Recall (R): It is the proportion of successfully recognized normal instances to the total number. Furthermore, it is called the detection rate (DR), sensitivity (S), or true positive rate (TPR). It is calculated according to:
    R = T P T P + F N
  • False Alarm Rate (FAR): It calculates the ratio of records that are mistakenly detected as attacks to all records. Furthermore, it is names the false positive rate (FPR). It is estimated as follows:
    F A R = F P F P + T N
  • False Negative Rate (FNR): It is also known as the missed alarm rate (MAR), and is the proportion of instances that are wrongly detected as normal to all instances. It is given as follows:
    F N R = F N F N + T P
  • Precision (P): It is the ratio of accurately predicted attacks to all attack samples, which provides how accurate predicted positive values are. It is evaluated as follows:
    P = T P T P + F P
  • F-measure (F): It is utilized to assess the accuracy of a detection system taking into account precision and recall, also known as the F1 score. It is computed according to:
    F m e a s u r e = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l

4.3. Results Discussion

By the simulation, the performance metrics of multi-class classification are calculated for our proposed model in terms of accuracy and the macro-average of recall, precision, and F1-score.
Table 5 presents the performance of the various feature selection approaches such as filter methods, stacked autoencoder, and fused features. It is observed that the fused feature method has succeeded in exhibiting better results than the others.
Table 6 presents the performance of the proposed method with and without the CSA. The proposal without the CSA used random weights, while with the CSA approach optimized weights have been generated. Therefore, CSA-based weighted voting showed a better performance than the method without the CSA.
Table 7, Table 8 and Table 9 show a performance comparison between previous works and our proposed ensemble-based cloud intrusion detection. The studies presented in Table 7, Table 8 and Table 9 all used one of the following:
  • Feature selection or feature extraction methods in addition to one of the machine learning models for classification.
  • Deep learning models to determine and classify features.
  • A method to handle the problem of imbalanced datasets along with ML or DL models.
  • Ensemble learning model based on majority voting.
  • One of the optimization algorithms in addition to ML or DL models.
Our proposal differs from others because it begins with preparing the dataset and addressing the issue of imbalance. Next, we formed a comprehensive feature set using feature selection approaches, which was then fed into ML and DL algorithms. Additionally, the results were optimized via the crow search algorithm according to the weight-based voting approach to obtain a high performance in our ensemble learning model.
Table 7. Comparison of the proposed model with state-of-the-art methodologies on the NSL-KDD dataset.
Table 7. Comparison of the proposed model with state-of-the-art methodologies on the NSL-KDD dataset.
No.Reference No.Year PublishedMethodologyACCRPF
1 [64]2019FFDNN86.62---
2 [65]2019ICVAE-DNN85.9777.4397.3986.27
3 [66]20195-layers DNN78.5078.5081.0076.50
4 [67]2019WGAN-GP80.80---
5 [68]2019AE-RL80.1680.1679.7479.40
6 [69]2019CNN-1D78.97---
7 [70]2019AFSA-GA-PSO-DBN82.36---
8 [71]2019FNN-LSO94.0489.8397.4393.05
9 [72]2019MDPCA-DBN82.0870.5197.2781.75
10 [73]2019RNN-ABC95.6295.84--
11 [74]2019GA+DBN-97.6797.36-
12 [24]2020BAT-MC84.25---
13 [29]2020CFS-BA-Voting(C4.5,RF,ForestPA)99.8199.8099.899.8
14 [33]2020DSSTE+AlexNet82.8482.7883.9481.66
15 [40]2020AESMOTE82.0982.0984.1182.43
16 [75]2020AE87.0082.0487.8581.21
17 [76]2020FCM-SMO86.0088.4084.7086.50
18 [77]2020C5+OC-SVM83.24---
19 [78]2020Multi-CNN fusion81.33---
20 [79]2020CNN-BiLSTM83.5884.4985.8285.14
21 [80]2020DRNN92.1894.2790.2392.29
22 [81]2020GA-KELM-94.01--
23 [8]2021T2FNN-97.3098.5096.00
24 [16]2021UEFFS96.0697.90--
25 [18]2021MFFSEM84.3396.4374.6184.13
26 [20]2021MapReduce+GA+Random Tree90.45---
27 [21]2021CAFE-CNN83.43---
28 [36]2021PTDAE+DNN83.3383.3386.0282.04
29 [82]2021OCNN-HMLSTM90.6795.1986.7191.46
30 [83]2021I-SiamIDS80.00---
31 [30]2022IG+GR+CS-SVM88.1590.4582.8783.48
32 [84]2022ABC-BWO-CONV-LSTM98.6710097.4898.73
33 [85]2022CP-GWO-O-LSTM96.3898.6397.5998.11
34 [31]2023EGA-PSO-IRF98.0988.5396.2491.87
Our Proposed Ensemble99.0199.0899.9599.51
Table 8. Comparison of the proposed ensemble with conventional methodologies on the Kyoto dataset.
Table 8. Comparison of the proposed ensemble with conventional methodologies on the Kyoto dataset.
NoReference NoYear PublishedMethodologyACCRPF
1 [86]2014CSV-ISVM-90.14--
2 [87]2015OS-ELM96.3797.9595.8096.86
3 [88]2018VAE-Label-75.3097.5085.00
4 [89]2018BA-ELM97.9698.75--
5 [66]20195-layers DNN88.5096.4091.3093.80
6 [90]2019HIDS (NBFS-OSVM-PKNN)-94.7556.89-
7 [31]2023IG+GR+CS-SVM96.4296.2390.5392.96
Our Proposed Ensemble98.9998.9398.1698.53
Table 9. Comparison of the proposed ensemble with recent methods on the CSE-CIC-IDS 2018 dataset.
Table 9. Comparison of the proposed ensemble with recent methods on the CSE-CIC-IDS 2018 dataset.
NoReference NoYear PublishedMethodologyACCRPF
1 [91]2019SMOTE-LSTM+AM96.2096.0096.00-
2 [92]2019CNN95.14---
3 [33]2020DSSTE+miniVGGNet96.9996.9797.4697.04
4 [58]2020RBM96.5594.00--
5 [93]2020DNN+PSO95.0098.20--
6 [94]2020DNN90.2559.0065.00-
7 [36]2021PTDAE+DNN95.7995.7995.3895.11
8 [95]2021HCRNN97.7597.1296.3397.60
9 [84]2022ABC-BWO-CONV-LSTM98.2598.6797.4898.18
10 [31]2023IG+GR+CS-SVM99.8992.9393.0292.97
Our Proposed Ensemble99.9999.8799.9699.91
From Table 7, Table 8 and Table 9, it can be inferred that the proposed ensemble model outperforms other strategies. Compared to others, the proposed work assigns more importance to feature selection. Feature selection plays a crucial role in improving performance metrics. Our approach was tested on multiple benchmark datasets during validation, unlike some other works, and this provides the benefit of including a variety of attacks. The proposed ensemble-based intrusion detection system focuses on generating a feature set and optimizing weights. The feature selection technique incorporates both filter and automated approaches to create an extensive feature set, and the usage of the ensemble model over a single classifier results in a better accuracy.
From Figure 3a, it can be inferred that the proposed ensemble model has the lowest FPR value, making it the best-performing model among those tested in terms of minimizing false positive detections. Conversely, the XGBoost model has the highest FPR value, indicating a higher likelihood of misclassifying benign network activities as attacks. The other models, SVM, LSTM, and FLN, follow in that order. Figure 3b demonstrates that the proposed ensemble model has the lowest FNR value, establishing it as the top-performing model among those evaluated for reducing false negative detections. In contrast, the XGBoost model possesses the highest FNR value, suggesting an increased probability of incorrectly categorizing attacks as normal activities. The remaining models, SVM, LSTM, and FLN, are ranked in that order according to their FNR values.
Overall, Figure 3 presents a comparison of FPRs and FNRs between each of the individual models (SVM, XGBoost, LSTM, and FLN) used in the ensemble model and our proposal tested on the NSL-KDD dataset. A performance comparison was conducted to show the strength of the ensemble model over individual models in the ensemble. The performance comparison shows that the FPR and the FNR are reduced when compared with the individual models.
Figure 4a reveals that the proposed ensemble model exhibits the lowest FPR value, while the XGBoost model presents the highest FPR value. As for the remaining models, the FPR values decrease in the order SVM, FLN, and LSTM. Notably, the FPR values for SVM and FLN are identical. Similarly, Figure 4b explains the FNR values in the same manner.
The overall FPR and FNR comparisons for the Kyoto dataset are shown in Figure 4. The ensemble model’s FPR and FNR are lower than the individual ML models, as can be observed.
In Figure 5, the performance of the FPR and FNR of the CSE-CIC-IDS 2018 shows a decline compared to the ML and DL models used in the ensemble. Figure 5a illustrates that the proposed ensemble model showcases the minimum FPR value, whereas the XGBoost model has the maximum FPR value. Regarding the other models, SVM, LSTM, and FLN, they are arranged according to their individual FPR values. It is worth noting that the FPR values for LSTM and FLN are the same. In a similar fashion, Figure 5b describes the FNR values.
Figure 3, Figure 4 and Figure 5 show that the proposed ensemble model was able to achieve a lower FPR and FNR, indicating improved classification accuracy. The occurrence of falsely reporting data as an intrusion is reduced in all types of attacks. In summary, a lower FPR signifies that the model has a reduced chance of incorrectly identifying benign network activities as malicious. Similarly, a lower FNR indicates that the model has a decreased likelihood of misclassifying attacks as benign network activities.
Figure 6, Figure 7 and Figure 8 illustrate the confusion matrices of our proposed ensemble. Each dataset (NSL-KDD, Kyoto, and CSE-CIC-IDS 2018) has its own confusion matrix representing the performance of our proposed ensemble. In the case of these datasets (NSL-KDD, Kyoto, and CSE-CIC-IDS 2018), the proposed ensemble is trained to predict the type of network attack or normal traffic from a total of 5, 3, and 15 categories (including normal traffic), respectively. These datasets are used for evaluating intrusion detection systems. The confusion matrix is a square matrix, where each row represents the actual class (true labels) and each column represents the predicted class (predicted labels). The diagonal elements represent correct predictions, while the off-diagonal elements represent misclassifications.
Figure 6 shows the confusion matrix of the NSL-KDD dataset. In this matrix, the first row and the first column represent the number of DoS attack instances. The proposed model has correctly predicted 7459 DoS attack instances with only one misclassification, where it predicted the normal category instead. The second row and the second column represent the number of probe attack instances. The proposed model accurately predicted 2421 instances of this attack type without any misclassifications. The same analysis can be applied to the remaining three types.
Figure 7 displays the Kyoto dataset’s confusion matrix. Within this matrix, the first row and column correspond to the count of known attack instances. The suggested system accurately identified 75,759 instances of known attacks. Nevertheless, 52 and 134 misclassifications occurred where the model predicted normal traffic and unknown attack categories, respectively. The second row and column denote the number of normal traffic instances. The proposed system successfully predicted 10,987 normal traffic instances. However, there are 263 instances where the model incorrectly identified a known attack. No cases were misclassified as the unknown attack category. This analysis can be extended to the unknown attack type as well.
Figure 8 showcases the confusion matrix for the CSE-CIC-IDS 2018 dataset. Within this matrix, the first row and column pertain to the count of benign traffic instances. The recommended system precisely identified 36,003 instances of benign traffic, with no misclassifications. The second row and column signify the number of bot attack instances. The proposed system effectively predicted 3759 bot attack instances, also without any misclassifications. This analysis can be similarly applied to the remaining thirteen types of attacks.
In summary, the confusion matrices demonstrate that the proposed ensemble system boasts a high degree of accuracy in detecting various network attack types and normal traffic, exhibiting outstanding performance on all three datasets. The most correct predictions lie on the diagonal, with only a minimal number of misclassifications. For a thorough assessment of the proposed system’s performance, it is crucial to take into account additional metrics such as recall, FAR, FNR, precision, and F1-score, alongside the confusion matrix. These metrics are detailed in Table 10, Table 11 and Table 12.
Based on Figure 6, the performance metrics of the classes in the NSL-KDD dataset are calculated, as shown in this Table 10.
Figure 9 shows that our approach outperformed other traditional works in terms of detection rate, false alarm rate, and accuracy for the various types of classes in the NSL-KDD dataset. Most of the attack types were detected with a higher detection rate and a reduced FAR.
The performance metrics of the classes in the Kyoto dataset were determined using Figure 7, as presented in Table 11, which shows that normal, known, and unknown attacks were detected with a high performance. To our knowledge, we did not find any previous work containing a similar analysis of performance metrics for the Kyoto dataset to compare the results of our work with.
Figure 8 was used to generate the performance metrics for the classes in the CSE-CIC-IDS-2018 dataset, as shown in Table 12.
The following Table 13 displays that the results of our methodology have exceeded the conventional works regarding the detection rate, false alarm rate, and accuracy for the various types of classes in the CSE-CIC-IDS-2018 dataset.
However, we did not focus on reducing the consumption of computational and temporal resources in our system; our priority was on designing a comprehensive model that overcomes some of the most important drawbacks that were presented in previous studies and that can be implemented on the most three popular datasets for detecting the largest number of attacks with an outstanding performance according to the comparison in Table 7, Table 8 and Table 9 in terms of accuracy, recall, precision, and F-measure.

4.4. Limitations and Constraints of the Study

The limitations and the constraints of the study are explained below:
  • Lack of a systematically collected dataset: Creating a dataset is an expensive procedure that requires a lot of money and high-level expertise. As a result, one of the key difficulties for IDS is the systematic creation of an up-to-date dataset with sufficient examples of practically all the intruder types. To aid the research community, the dataset should be regularly updated to incorporate the most recent attack records. In the current work, datasets with older (NSL-KDD and Kyoto) and newer (CSE-CIC-IDS-2018) attacks were used to test and validate the proposal. However, we emphasize the requirement for an up-to-date comprehensive dataset that represents new attacks on real environment networks that exist today. By including the definition of the largest number of intrusions in a dataset, the system will be able to discover additional patterns and offer a defense against the greatest number of zero-day attacks.
  • An unbalanced dataset has an impact on performance: It is noted from the current study that we used the SMOTE technique to increase the number of minority attack instances to balance the dataset, which led to an increase in the data size and thus the computational resources needed. Therefore, we can use random undersampling to decrease the number of instances in the majority class, then use any existing technologies such as SMOTE to oversample the minority class to balance the class distribution. However, we may try alternative approaches, such as the Difficult Set Sampling Technique (DSSTE), Adaptive Synthetic Sampling (ADASYN), RandomOverSampler (ROS), etc, to decrease the dataset’s unbalance ratio for a better performance.
  • Unknown efficiency in a real-world environment: The proposal was evaluated in a lab setting based on publicly available datasets. However, it was not put to the test in a real-world setting. Therefore, it is not yet clear how it will function in real situations. Thus, to ensure that the suggested strategy works effectively for contemporary networks, it should be tested in a real-time setting after being evaluated in a lab setting.
  • Resource consumption: DL models utilize power to learn features deeply, which provides outstanding results in detecting attacks; however, it demands a lot of time, storage, and computational resources. Bio-inspired algorithms can be used in the future to improve the effectiveness and intelligence of feature selection, reduce computational resources, and achieve better results.

5. Conclusions and Future Work

The security of cloud computing is of utmost importance with the rise in the application of cloud computing in the majority of public and private organizations. The proposed intrusion detection system focused on designing a model that can improve intrusion detection with a higher accuracy using some advancements in the feature selection part by combining filter and automated feature selection and optimizing weights with the CSA in the ensemble model. This resulted in better results than other state-of-the-art models. The proposed approach is capable of detecting various types of attacks with a higher detection rate and a reduced false alarm rate.
In our future work, we will strive to compute the consumption of computational resources and perform a comparison with other approaches. Furthermore, bio-inspired algorithms can be implemented to optimize the value. In verification, we can use the most recent datasets that have a wide spectrum of attacks that reflect real networks, which will be prepared well to create a balanced dataset for a better performance.

Author Contributions

M.B. and R.R.K. contributed the main idea of qualitative analysis for the research and wrote the original manuscript. A.A.A., S.K.B., Z.A., N.P., S.K. and A.A. contributed significantly to improving the technical and grammatical contents of the manuscript. A.A.A., S.K.B., Z.A., N.P., S.K. and A.A. reviewed the manuscript and provided valuable suggestions to further refine the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

Researchers Supporting Project number (RSP2023R476), King Saud University, Riyadh, Saudi Arabia.

Data Availability Statement

The datasets are public and already cited. The data used to support the findings of this study are available from the corresponding author upon request.

Acknowledgments

The authors would like to acknowledge the Researchers Supporting Project number (RSP2023R476), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kumar, R.R.; Tomar, A.; Shameem, M.; Alam, M.D. Optcloud: An optimal cloud service selection framework using QoS correlation lens. Comput. Intell. Neurosci. 2022, 2022, 2019485. [Google Scholar] [CrossRef] [PubMed]
  2. Kumar, R.R.; Shameem, M.; Khanam, R.; Kumar, C. A hybrid evaluation framework for QoS based service selection and ranking in cloud environment. In Proceedings of the 2018 15th IEEE India Council International Conference (INDICON), Coimbatore, India, 16–18 December 2018; pp. 1–6. [Google Scholar] [CrossRef]
  3. Kumar, R.R.; Shameem, M.; Kumar, C. A computational framework for ranking prediction of cloud services under fuzzy environment. Enterp. Inf. Syst. 2022, 16, 167–187. [Google Scholar] [CrossRef]
  4. Akbar, M.A.; Shameem, M.; Mahmood, S.; Alsanad, A.; Gumaei, A. Prioritization based taxonomy of cloud-based outsource software development challenges: Fuzzy AHP analysis. Appl. Soft Comput. 2020, 95, 106557. [Google Scholar] [CrossRef]
  5. Bakro, M.; Bisoy, S.K.; Patel, A.K.; Naal, M.A. Performance Analysis of Cloud Computing Encryption Algorithms. In Advances in Intelligent Computing and Communication; Springer: Singapore, 2021; pp. 357–367. [Google Scholar] [CrossRef]
  6. Bhushan, K.; Gupta, B.B. Security challenges in cloud computing: State-of-art. Int. J. Big Data Intell. 2017, 4, 81–107. [Google Scholar] [CrossRef]
  7. Bakro, M.; Bisoy, S.K.; Patel, A.K.; Naal, M.A. Hybrid Blockchain-Enabled Security in Cloud Storage Infrastructure Using ECC and AES Algorithms. In Blockchain Based Internet of Things; Springer: Singapore, 2022; pp. 139–170. [Google Scholar] [CrossRef]
  8. Srilatha, D.; Shyam, G.K. Cloud-based intrusion detection using kernel fuzzy clustering and optimal type-2 fuzzy neural network. Clust. Comput. 2021, 24, 2657–2672. [Google Scholar] [CrossRef]
  9. Xu, C.; Shen, J.; Du, X.; Zhang, F. An intrusion detection system using a deep neural network with gated recurrent units. IEEE Access 2018, 6, 48697–48707. [Google Scholar] [CrossRef]
  10. Abbas, G.; Mehmood, A.; Carsten, M.; Epiphaniou, G.; Lloret, J. Safety, Security and Privacy in Machine Learning Based Internet of Things. J. Sens. Actuator Netw. 2022, 11, 38. [Google Scholar] [CrossRef]
  11. Mighan, S.N.; Kahani, M. A novel scalable intrusion detection system based on deep learning. Int. J. Inf. Secur. 2021, 20, 387–403. [Google Scholar] [CrossRef]
  12. Mayuranathan, M.; Murugan, M.; Dhanakoti, V. Best features based intrusion detection system by RBM model for detecting DDoS in cloud environment. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 3609–3619. [Google Scholar] [CrossRef]
  13. Arora, N.; Kaur, P.D. A Bolasso based consistent feature selection enabled random forest classification algorithm: An application to credit risk assessment. Appl. Soft Comput. 2020, 86, 105936. [Google Scholar] [CrossRef]
  14. Mirza, A.H. Computer network intrusion detection using various classifiers and ensemble learning. In Proceedings of the 2018 26th Signal Processing and Communications Applications Conference (SIU), Izmir, Turkey, 2–5 May 2018; pp. 1–4. [Google Scholar] [CrossRef]
  15. Kushwah, G.S.; Ranga, V. Voting extreme learning machine based distributed denial of service attack detection in cloud computing. J. Inf. Secur. Appl. 2020, 53, 102532. [Google Scholar] [CrossRef]
  16. Krishnaveni, S.; Sivamohan, S.; Sridhar, S.; Prabakaran, S. Efficient feature selection and classification through ensemble method for network intrusion detection on cloud computing. Clust. Comput. 2021, 24, 1761–1779. [Google Scholar] [CrossRef]
  17. Thaseen, I.S.; Chitturi, A.K.; Al-Turjman, F.; Shankar, A.; Ghalib, M.R.; Abhishek, K. An intelligent ensemble of long-short-term memory with genetic algorithm for network anomaly identification. Trans. Emerg. Telecommun. Technol. 2020, 33, e4149. [Google Scholar] [CrossRef]
  18. Zhang, H.; Li, J.L.; Liu, X.M.; Dong, C. Multi-dimensional feature fusion and stacking ensemble mechanism for network intrusion detection. Future Gener. Comput. Syst. 2021, 122, 130–143. [Google Scholar] [CrossRef]
  19. Singh, P.; Ranga, V. Attack and intrusion detection in cloud computing using an ensemble learning approach. Int. J. Inf. Technol. 2021, 13, 565–571. [Google Scholar] [CrossRef]
  20. Mehanović, D.; Kečo, D.; Kevrić, J.; Jukić, S.; Miljković, A.; Mašetić, Z. Feature selection using cloud-based parallel genetic algorithm for intrusion detection data classification. Neural Comput. Appl. 2021, 33, 11861–11873. [Google Scholar] [CrossRef]
  21. Shams, E.A.; Rizaner, A.; Ulusoy, A.H. A novel context-aware feature extraction method for convolutional neural network-based intrusion detection systems. Neural Comput. Appl. 2021, 33, 13647–13665. [Google Scholar] [CrossRef]
  22. Tummalapalli, S.R.K.; Chakravarthy, A. Intrusion detection system for cloud forensics using bayesian fuzzy clustering and optimization based SVNN. Evol. Intell. 2021, 14, 699–709. [Google Scholar] [CrossRef]
  23. Punitha, A.; Indumathi, G. A novel centralized cloud information accountability integrity with ensemble neural network based attack detection approach for cloud data. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 4889–4900. [Google Scholar] [CrossRef]
  24. Su, T.; Sun, H.; Zhu, J.; Wang, S.; Li, Y. BAT: Deep learning methods on network intrusion detection using NSL-KDD dataset. IEEE Access 2020, 8, 29575–29585. [Google Scholar] [CrossRef]
  25. Bhati, B.S.; Chugh, G.; Al-Turjman, F.; Bhati, N.S. An improved ensemble based intrusion detection technique using XGBoost. Trans. Emerg. Telecommun. Technol. 2021, 32, e4076. [Google Scholar] [CrossRef]
  26. Suman, C.; Tripathy, S.; Saha, S. Building an effective intrusion detection system using unsupervised feature selection in multi-objective optimization framework. arXiv 2019, arXiv:1905.06562. [Google Scholar] [CrossRef]
  27. Rajagopal, S.; Kundapur, P.P.; Hareesha, K.S. A stacking ensemble for network intrusion detection using heterogeneous datasets. Secur. Commun. Netw. 2020, 2020, 4586875. [Google Scholar] [CrossRef]
  28. Lopez-Martin, M.; Carro, B.; Sanchez-Esguevillas, A.; Lloret, J. Conditional variational autoencoder for prediction and feature recovery applied to intrusion detection in IoT. Sensors 2017, 17, 1967. [Google Scholar] [CrossRef]
  29. Zhou, Y.; Cheng, G.; Jiang, S.; Dai, M. Building an efficient intrusion detection system based on feature selection and ensemble classifier. Comput. Netw. 2020, 174, 107247. [Google Scholar] [CrossRef]
  30. Balyan, A.K.; Ahuja, S.; Lilhore, U.K.; Sharma, S.K.; Manoharan, P.; Algarni, A.D.; Elmannai, H.; Raahemifar, K. A hybrid intrusion detection model using ega-pso and improved random forest method. Sensors 2022, 22, 5986. [Google Scholar] [CrossRef]
  31. Bakro, M.; Kumar, R.R.; Bisoy, S.K.; Addas, M.O.; Khamis, D. Developing a Cloud Intrusion Detection System with Filter-Based Features Selection Techniques and SVM Classifier. In Proceedings of the Computing, Communication and Learning: First International Conference (CoCoLe 2022), Warangal, India, 27–29 October 2022; Springer: Cham, Switzerland, 2023; pp. 15–26. [Google Scholar] [CrossRef]
  32. Wang, D.; Zhang, Z.; Bai, R.; Mao, Y. A hybrid system with filter approach and multiple population genetic algorithm for feature selection in credit scoring. J. Comput. Appl. Math. 2018, 329, 307–321. [Google Scholar] [CrossRef]
  33. Liu, L.; Wang, P.; Lin, J.; Liu, L. Intrusion detection of imbalanced network traffic based on machine learning and deep learning. IEEE Access 2020, 9, 7550–7563. [Google Scholar] [CrossRef]
  34. Potdar, K.; Pardawala, T.S.; Pai, C.D. A comparative study of categorical variable encoding techniques for neural network classifiers. Int. J. Comput. Appl. 2017, 175, 7–9. [Google Scholar] [CrossRef]
  35. Potluri, S.; Diedrich, C. Accelerated deep neural networks for enhanced intrusion detection system. In Proceedings of the 2016 IEEE 21st International Conference on Emerging Technologies and Factory Automation (ETFA), Berlin, Germany, 6–9 September 2016; pp. 1–8. [Google Scholar] [CrossRef]
  36. Kunang, Y.N.; Nurmaini, S.; Stiawan, D.; Suprapto, B.Y. Attack classification of an intrusion detection system using deep learning and hyperparameter optimization. J. Inf. Secur. Appl. 2021, 58, 102804. [Google Scholar] [CrossRef]
  37. Ren, F.; Long, D. Carbon emission forecasting and scenario analysis in Guangdong Province based on optimized Fast Learning Network. J. Clean. Prod. 2021, 317, 128408. [Google Scholar] [CrossRef]
  38. Wang, Z. Deep learning-based intrusion detection with adversaries. IEEE Access 2018, 6, 38367–38384. [Google Scholar] [CrossRef]
  39. Xiao, P.; Qu, W.; Qi, H.; Li, Z. Detecting DDoS attacks against data center with correlation analysis. Computer Communications 2015, 67, 66–74. [Google Scholar] [CrossRef]
  40. Ma, X.; Shi, W. Aesmote: Adversarial reinforcement learning with smote for anomaly detection. IEEE Trans. Netw. Sci. Eng. 2020, 8, 943–956. [Google Scholar] [CrossRef]
  41. Sigirci, I.O.; Albayrak, A.; Bilgin, G. Detection of mitotic cells in breast cancer histopathological images using deep versus handcrafted features. Multimed. Tools Appl. 2021, 81, 13179–13202. [Google Scholar] [CrossRef]
  42. Wang, W.; Liu, J.; Pitsilis, G.; Zhang, X. Abstracting massive data for lightweight intrusion detection in computer networks. Inf. Sci. 2018, 433, 417–430. [Google Scholar] [CrossRef]
  43. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  44. Omuya, E.O.; Okeyo, G.O.; Kimwele, M.W. Feature selection for classification using principal component analysis and information gain. Expert Syst. Appl. 2021, 174, 114765. [Google Scholar] [CrossRef]
  45. Ahmad, Z.; Shahid Khan, A.; Wai Shiang, C.; Abdullah, J.; Ahmad, F. Network intrusion detection system: A systematic study of machine learning and deep learning approaches. Trans. Emerg. Telecommun. Technol. 2021, 32, e4150. [Google Scholar] [CrossRef]
  46. Atashgahi, Z.; Sokar, G.; van der Lee, T.; Mocanu, E.; Mocanu, D.C.; Veldhuis, R.; Pechenizkiy, M. Quick and robust feature selection: The strength of energy-efficient sparse training for autoencoders. Mach. Learn. 2021, 111, 377–414. [Google Scholar] [CrossRef]
  47. Yu, M.; Quan, T.; Peng, Q.; Yu, X.; Liu, L. A model-based collaborate filtering algorithm based on stacked AutoEncoder. Neural Comput. Appl. 2022, 34, 2503–2511. [Google Scholar] [CrossRef]
  48. Sai, A.V.; Hitesh, M.S.V.; Jadala, V.C.; Pasupuleti, S.K.; Raju, S.H.; Shameem, M. Flower Identification and Classification applying CNN through Deep Learning Methodologies. In Proceedings of the 2022 International Mobile and Embedded Technology Conference (MECON), Noida, India, 10–11 March 2022; pp. 173–180. [Google Scholar] [CrossRef]
  49. Ma, Z.; Guo, S.; Xu, G.; Aziz, S. Meta learning-based hybrid ensemble approach for short-term wind speed forecasting. IEEE Access 2020, 8, 172859–172868. [Google Scholar] [CrossRef]
  50. Zhao, Z.; Chen, W.; Wu, X.; Chen, P.C.; Liu, J. LSTM network: A deep learning approach for short-term traffic forecast. IET Intell. Transp. Syst. 2017, 11, 68–75. [Google Scholar] [CrossRef]
  51. Liu, H.; Lang, B. Machine learning and deep learning methods for intrusion detection systems: A survey. Appl. Sci. 2019, 9, 4396. [Google Scholar] [CrossRef]
  52. Soumaya, Z.; Taoufiq, B.D.; Benayad, N.; Yunus, K.; Abdelkrim, A. The detection of Parkinson disease using the genetic algorithm and SVM classifier. Appl. Acoust. 2021, 171, 107528. [Google Scholar] [CrossRef]
  53. Deepak, S.; Ameer, P. Automated categorization of brain tumor from mri using cnn features and svm. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 8357–8369. [Google Scholar] [CrossRef]
  54. Nobre, J.; Neves, R.F. Combining principal component analysis, discrete wavelet transform and XGBoost to trade in the financial markets. Expert Syst. Appl. 2019, 125, 181–194. [Google Scholar] [CrossRef]
  55. Adamu, A.; Abdullahi, M.; Junaidu, S.B.; Hassan, I.H. An hybrid particle swarm optimization with crow search algorithm for feature selection. Mach. Learn. Appl. 2021, 6, 100108. [Google Scholar] [CrossRef]
  56. Askarzadeh, A. A novel metaheuristic method for solving constrained engineering optimization problems: Crow search algorithm. Comput. Struct. 2016, 169, 1–12. [Google Scholar] [CrossRef]
  57. Tavallaee, M.; Bagheri, E.; Lu, W.; Ghorbani, A.A. A detailed analysis of the KDD CUP 99 data set. In Proceedings of the 2009 IEEE Symposium on Computational Intelligence for Security and Defense Applications, Ottawa, ON, Canada, 8–10 July 2009; pp. 1–6. [Google Scholar] [CrossRef]
  58. Ferrag, M.A.; Maglaras, L.; Moschoyiannis, S.; Janicke, H. Deep learning for cyber security intrusion detection: Approaches, datasets, and comparative study. J. Inf. Secur. Appl. 2020, 50, 102419. [Google Scholar] [CrossRef]
  59. Canadian Institute for Cybersecurity, University of New Brunswick. NSL-KDD | Datasets | Research | Canadian Institute for Cybersecurity | UNB. 2009. Available online: https://www.unb.ca/cic/datasets/nsl.html (accessed on 3 March 2022).
  60. Dhanabal, L.; Shantharajah, S. A study on NSL-KDD dataset for intrusion detection system based on classification algorithms. Int. J. Adv. Res. Comput. Commun. Eng. 2015, 4, 446–452. [Google Scholar]
  61. Kyoto University. Traffic Data from Kyoto University Honeypots. 2006. Available online: http://www.takakura.com/Kyoto_data/new_data201704/ (accessed on 3 March 2022).
  62. Canadian Institute for Cybersecurity. IDS 2018 Datasets Canadian Institute for Cybersecurity. 2018. Available online: https://www.unb.ca/cic/datasets/ids-2018.html (accessed on 3 March 2022).
  63. Kilincer, I.F.; Ertam, F.; Sengur, A. Machine learning methods for cyber security intrusion detection: Datasets and comparative study. Comput. Netw. 2021, 188, 107840. [Google Scholar] [CrossRef]
  64. Kasongo, S.M.; Sun, Y. A deep learning method with filter based feature engineering for wireless intrusion detection system. IEEE Access 2019, 7, 38597–38607. [Google Scholar] [CrossRef]
  65. Yang, Y.; Zheng, K.; Wu, C.; Yang, Y. Improving the classification effectiveness of intrusion detection by using improved conditional variational autoencoder and deep neural network. Sensors 2019, 19, 2528. [Google Scholar] [CrossRef] [PubMed]
  66. Vinayakumar, R.; Alazab, M.; Soman, K.; Poornachandran, P.; Al-Nemrat, A.; Venkatraman, S. Deep learning approach for intelligent intrusion detection system. IEEE Access 2019, 7, 41525–41550. [Google Scholar] [CrossRef]
  67. Wang, J.T.; Wang, C.H. High performance WGAN-GP based multiple-category network anomaly classification system. In Proceedings of the 2019 International Conference on Cyber Security for Emerging Technologies (CSET), Doha, Qatar, 27–29 October 2019; pp. 1–7. [Google Scholar] [CrossRef]
  68. Caminero, G.; Lopez-Martin, M.; Carro, B. Adversarial environment reinforcement learning algorithm for intrusion detection. Comput. Netw. 2019, 159, 96–109. [Google Scholar] [CrossRef]
  69. Verma, A.K.; Kaushik, P.; Shrivastava, G. A network intrusion detection approach using variant of convolution neural network. In Proceedings of the 2019 International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 17–19 July 2019; pp. 409–416. [Google Scholar] [CrossRef]
  70. Wei, P.; Li, Y.; Zhang, Z.; Hu, T.; Li, Z.; Liu, D. An optimization method for intrusion detection classification model based on deep belief network. IEEE Access 2019, 7, 87593–87605. [Google Scholar] [CrossRef]
  71. Benmessahel, I.; Xie, K.; Chellal, M.; Semong, T. A new evolutionary neural networks based on intrusion detection systems using locust swarm optimization. Evol. Intell. 2019, 12, 131–146. [Google Scholar] [CrossRef]
  72. Yang, Y.; Zheng, K.; Wu, C.; Niu, X.; Yang, Y. Building an effective intrusion detection system using the modified density peak clustering algorithm and deep belief networks. Appl. Sci. 2019, 9, 238. [Google Scholar] [CrossRef]
  73. Qureshi, A.U.; Larijani, H.; Mtetwa, N.; Javed, A.; Ahmad, J. RNN-ABC: A new swarm optimization based technique for anomaly detection. Computers 2019, 8, 59. [Google Scholar] [CrossRef]
  74. Zhang, Y.; Li, P.; Wang, X. Intrusion detection for IoT based on improved genetic algorithm and deep belief network. IEEE Access 2019, 7, 31711–31722. [Google Scholar] [CrossRef]
  75. Ieracitano, C.; Adeel, A.; Morabito, F.C.; Hussain, A. A novel statistical analysis and autoencoder driven intelligent intrusion detection approach. Neurocomputing 2020, 387, 51–62. [Google Scholar] [CrossRef]
  76. Samriya, J.K.; Kumar, N. A novel intrusion detection system using hybrid clustering-optimization approach in cloud computing. Mater. Today Proc. 2020. [Google Scholar] [CrossRef]
  77. Khraisat, A.; Gondal, I.; Vamplew, P.; Kamruzzaman, J.; Alazab, A. Hybrid intrusion detection system based on the stacking ensemble of c5 decision tree classifier and one class support vector machine. Electronics 2020, 9, 173. [Google Scholar] [CrossRef]
  78. Li, Y.; Xu, Y.; Liu, Z.; Hou, H.; Zheng, Y.; Xin, Y.; Zhao, Y.; Cui, L. Robust detection for network intrusion of industrial IoT based on multi-CNN fusion. Measurement 2020, 154, 107450. [Google Scholar] [CrossRef]
  79. Jiang, K.; Wang, W.; Wang, A.; Wu, H. Network intrusion detection combined hybrid sampling with deep hierarchical network. IEEE Access 2020, 8, 32464–32476. [Google Scholar] [CrossRef]
  80. Almiani, M.; AbuGhazleh, A.; Al-Rahayfeh, A.; Atiewi, S.; Razaque, A. Deep recurrent neural network for IoT intrusion detection system. Simul. Model. Pract. Theory 2020, 101, 102031. [Google Scholar] [CrossRef]
  81. Ghasemi, J.; Esmaily, J.; Moradinezhad, R. Intrusion detection system using an optimized kernel extreme learning machine and efficient features. Sādhanā 2020, 45, 1–9. [Google Scholar] [CrossRef]
  82. Kanna, P.R.; Santhi, P. Unified deep learning approach for efficient intrusion detection system using integrated spatial–temporal features. Knowl.-Based Syst. 2021, 226, 107132. [Google Scholar] [CrossRef]
  83. Bedi, P.; Gupta, N.; Jindal, V. I-SiamIDS: An improved Siam-IDS for handling class imbalance in network-based intrusion detection systems. Applied Intelligence 2021, 51, 1133–1151. [Google Scholar] [CrossRef]
  84. Kanna, P.R.; Santhi, P. Hybrid intrusion detection using mapreduce based black widow optimized convolutional long short-term memory neural networks. Expert Syst. Appl. 2022, 194, 116545. [Google Scholar] [CrossRef]
  85. Dora, V.; Lakshmi, V.N. Optimal feature selection with CNN-feature learning for DDoS attack detection using meta-heuristic-based LSTM. Int. J. Intell. Robot. Appl. 2022, 6, 323–349. [Google Scholar] [CrossRef]
  86. Chitrakar, R.; Huang, C. Selection of candidate support vectors in incremental SVM for network intrusion detection. Comput. Secur. 2014, 45, 231–241. [Google Scholar] [CrossRef]
  87. Singh, R.; Kumar, H.; Singla, R. An intrusion detection system using network traffic profiling and online sequential extreme learning machine. Expert Syst. Appl. 2015, 42, 8609–8624. [Google Scholar] [CrossRef]
  88. Malaiya, R.K.; Kwon, D.; Kim, J.; Suh, S.C.; Kim, H.; Kim, I. An empirical evaluation of deep learning for network anomaly detection. In Proceedings of the 2018 International Conference on Computing, Networking and Communications (ICNC), Maui, HI, USA, 5–8 March 2018; pp. 893–898. [Google Scholar] [CrossRef]
  89. Shen, Y.; Zheng, K.; Wu, C.; Zhang, M.; Niu, X.; Yang, Y. An ensemble method based on selection using bat algorithm for intrusion detection. Comput. J. 2018, 61, 526–538. [Google Scholar] [CrossRef]
  90. Saleh, A.I.; Talaat, F.M.; Labib, L.M. A hybrid intrusion detection system (HIDS) based on prioritized k-nearest neighbors and optimized SVM classifiers. Artif. Intell. Rev. 2019, 51, 403–443. [Google Scholar] [CrossRef]
  91. Lin, P.; Ye, K.; Xu, C.Z. Dynamic network anomaly detection system by using deep learning techniques. In Proceedings of the International Conference on Cloud Computing; Springer: Cham, Switzerland, 2019; pp. 161–176. [Google Scholar] [CrossRef]
  92. Kim, J.; Shin, Y.; Choi, E. An intrusion detection model based on a convolutional neural network. J. Multimed. Inf. Syst. 2019, 6, 165–172. [Google Scholar] [CrossRef]
  93. Farhan, R.I.; Maolood, A.T.; Hassan, N.F. Optimized Deep Learning with Binary PSO for Intrusion Detection on CSE-CIC-IDS2018 Dataset. J. Al-Qadisiyah Comput. Sci. Math. 2020, 12, 16–27. [Google Scholar]
  94. Farhan, R.I.; Maolood, A.T.; Hassan, N. Performance analysis of flow-based attacks detection on CSE-CIC-IDS2018 dataset using deep learning. Indones. J. Electr. Eng. Comput. Sci. 2020, 20, 1413–1418. [Google Scholar] [CrossRef]
  95. Khan, M.A. HCRNNIDS: Hybrid convolutional recurrent neural network-based network intrusion detection system. Processes 2021, 9, 834. [Google Scholar] [CrossRef]
Figure 1. Proposed Block Diagram.
Figure 1. Proposed Block Diagram.
Electronics 12 02427 g001
Figure 2. Stacked autoencoder.
Figure 2. Stacked autoencoder.
Electronics 12 02427 g002
Figure 3. FPR and FNR comparison between the individual models and ensemble model for the NSL-KDD dataset. (a) FPR and (b) FNR.
Figure 3. FPR and FNR comparison between the individual models and ensemble model for the NSL-KDD dataset. (a) FPR and (b) FNR.
Electronics 12 02427 g003
Figure 4. FPR and FNR comparison between the individual models and ensemble model for the Kyoto dataset. (a) FPR and (b) FNR.
Figure 4. FPR and FNR comparison between the individual models and ensemble model for the Kyoto dataset. (a) FPR and (b) FNR.
Electronics 12 02427 g004
Figure 5. FPR and FNR comparison between individual models and ensemble model for CSE-CIC-IDS 2018. (a) FPR and (b) FNR.
Figure 5. FPR and FNR comparison between individual models and ensemble model for CSE-CIC-IDS 2018. (a) FPR and (b) FNR.
Electronics 12 02427 g005
Figure 6. Confusion matrix of the NSL-KDD dataset.
Figure 6. Confusion matrix of the NSL-KDD dataset.
Electronics 12 02427 g006
Figure 7. Confusion matrix of Kyoto dataset.
Figure 7. Confusion matrix of Kyoto dataset.
Electronics 12 02427 g007
Figure 8. Confusion matrix of the CSE-CIC-IDS 2018 dataset.
Figure 8. Confusion matrix of the CSE-CIC-IDS 2018 dataset.
Electronics 12 02427 g008
Figure 9. Comparison of the performance of various categories between conventional models and the proposed ensemble model for the NSL-KDD dataset [24,29,36,66,67,68,69,70,74,78,81,84].
Figure 9. Comparison of the performance of various categories between conventional models and the proposed ensemble model for the NSL-KDD dataset [24,29,36,66,67,68,69,70,74,78,81,84].
Electronics 12 02427 g009aElectronics 12 02427 g009b
Table 1. Composition of the NSL-KDD dataset.
Table 1. Composition of the NSL-KDD dataset.
KDDTrain+KDDTest+
Normal67,3439711
DOS45,9277460
Probe11,6562421
R2L9952885
U2R5267
Total125,97322,544
Table 2. Composition of the Kyoto dataset.
Table 2. Composition of the Kyoto dataset.
TrainTest
Normal45,26011,250
Known attack303,41275,945
Unknown attack11,0952747
Total359,76789,942
Table 3. Composition of the CSE-CIC-IDS 2018 dataset.
Table 3. Composition of the CSE-CIC-IDS 2018 dataset.
CategoryAttack TypeTrainingTest
Benign-144,19836,003
BotnetBot14,6023759
Brute Force-Web611160
Web AttackBrute Force-XSS23051
SQL Injection8724
DDOS attack-HOIC34,3768536
DDOS AttackDDOS attack-LOIC-UDP1730450
DDOS attacks-LOIC-HTTP28,9067215
DOS AttackDoS attacks-GoldenEye81852059
DOS attacks-Hulk23,1135796
DOS attacks-SlowHTTPTest70681753
DOS attacks-Slowloris4003992
Brute forceFTP-BruteForce99982494
SSH-Bruteforce99732439
InfilterationInfilteration81072066
Total-295,18773,797
Table 4. Confusion matrix.
Table 4. Confusion matrix.
Predicted (Attack)Predicted (Normal)
Actual (Attack)True Positive (TP)False Negative (FN)
Actual (Normal)False Positive (FP)True Negative (TN)
Table 5. Comparison of performance with filter-based feature selection, automated feature selection with SAE, and fused features.
Table 5. Comparison of performance with filter-based feature selection, automated feature selection with SAE, and fused features.
MethodDatasetNo. of FeaturesACCRPF
Filter methodsNSL-KDD5097.9990.4296.4493.13
Kyoto1096.5295.8194.4294.88
CSE-CIC-IDS-2018698.9092.7599.5095.17
Automated with
SAE methods
NSL-KDD9398.4194.3398.3396.17
Kyoto3397.6796.9696.1496.43
CSE-CIC-IDS-2018699.0297.8699.9198.79
Fused featuresNSL-KDD4599.0199.0899.9599.51
Kyoto1598.9998.9398.1698.53
CSE-CIC-IDS-2018699.9999.8799.9699.91
Table 6. Comparison of performance when the CSA is not used and when the CSA is used.
Table 6. Comparison of performance when the CSA is not used and when the CSA is used.
MethodDatasetACCRPF
NSL-KDD97.0190.3195.4592.61
Without CSAKyoto96.1995.5291.9293.25
CSE-CIC-IDS-201898.1896.1193.2192.69
NSL-KDD99.0199.0899.9599.51
With CSAKyoto98.9998.9398.1698.53
CSE-CIC-IDS-201899.9999.8799.9699.91
Table 10. Performance metrics of various classes by the proposed ensemble model for the NSL-KDD dataset.
Table 10. Performance metrics of various classes by the proposed ensemble model for the NSL-KDD dataset.
Attack TypeACCRFARFNRPF
Normal99.9899.980.020.0299.9799.97
DOS99.9999.990.010.0199.9999.99
Probe99.99100.000.010.0099.9299.96
R2L99.9799.900.020.1099.9099.90
U2R99.9995.520.004.48100.0097.71
Table 11. Performance metrics of types of categories by the proposed ensemble for the Kyoto dataset.
Table 11. Performance metrics of types of categories by the proposed ensemble for the Kyoto dataset.
Attack TypeACCRFARFNRPF
Normal99.6597.660.072.3499.5398.59
Known Attack99.4899.762.000.2499.6399.69
Unknown Attack99.8399.380.150.6295.3297.31
Table 12. Performance metrics of types of classes by the proposed ensemble for the CSE-CIC-IDS-2018 dataset.
Table 12. Performance metrics of types of classes by the proposed ensemble for the CSE-CIC-IDS-2018 dataset.
Attack TypeACCRFARFNRPF
Benign100.00100.000.000.00100.00100.00
Bot100.00100.000.000.00100.00100.00
Brute Force-Web100.00100.000.000.0099.3899.69
Brute Force-XSS100.0098.040.001.96100.0099.01
SQL Injection100.00100.000.000.00100.00100.00
DDOS attack-HOIC100.00100.000.000.00100.00100.00
DDOS attack-LOIC-UDP100.00100.000.000.00100.00100.00
DDoS attacks-LOIC-HTTP100.00100.000.000.00100.00100.00
DoS attacks-GoldenEye100.00100.000.000.00100.00100.00
DoS attacks-Hulk100.00100.000.000.00100.00100.00
DoS attacks-SlowHTTPTest100.00100.000.000.00100.00100.00
DoS attacks-Slowloris100.00100.000.000.00100.00100.00
FTP-BruteForce100.00100.000.000.00100.00100.00
SSH-Bruteforce100.00100.000.000.00100.00100.00
Infilteration100.00100.000.000.00100.00100.00
Table 13. Comparison of the performance of various classes between the previous works and the proposed ensemble method for the CSE-CIC-IDS-2018 dataset.
Table 13. Comparison of the performance of various classes between the previous works and the proposed ensemble method for the CSE-CIC-IDS-2018 dataset.
Attack TypeOur Proposed[36] 2021[58] 2020
ACCRFARACCRFARR
Benign100.00100.000.0097.4899.464.48-
Bot100.00100.000.00100.0099.970.0096.19
Brute Force-Web100.00100.000.0099.9860.000.0082.22
Brute Force-XSS100.0098.040.0099.9974.420.0083.16
SQL Injection100.00100.000.00100.0043.750.00100.00
DDOS attack-HOIC100.00100.000.00100.00100.000.0097.54
DDOS attack-LOIC-UDP100.00100.000.0099.99100.000.0196.15
DDoS attacks-LOIC-HTTP100.00100.000.0099.9899.820.0096.18
DoS attacks-GoldenEye100.00100.000.00100.0099.930.0092.01
DoS attacks-Hulk100.00100.000.00100.0099.990.0091.32
DoS attacks-SlowHTTPTest100.00100.000.0098.3351.990.4393.31
DoS attacks-Slowloris100.00100.000.00100.00100.000.0097.04
FTP-BruteForce100.00100.000.0098.3388.161.29100.00
SSH-Bruteforce100.00100.000.00100.0099.970.00100.00
Infilteration100.00100.000.0097.5023.840.2796.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bakro, M.; Kumar, R.R.; Alabrah, A.A.; Ashraf, Z.; Bisoy, S.K.; Parveen, N.; Khawatmi, S.; Abdelsalam, A. Efficient Intrusion Detection System in the Cloud Using Fusion Feature Selection Approaches and an Ensemble Classifier. Electronics 2023, 12, 2427. https://doi.org/10.3390/electronics12112427

AMA Style

Bakro M, Kumar RR, Alabrah AA, Ashraf Z, Bisoy SK, Parveen N, Khawatmi S, Abdelsalam A. Efficient Intrusion Detection System in the Cloud Using Fusion Feature Selection Approaches and an Ensemble Classifier. Electronics. 2023; 12(11):2427. https://doi.org/10.3390/electronics12112427

Chicago/Turabian Style

Bakro, Mhamad, Rakesh Ranjan Kumar, Amerah A. Alabrah, Zubair Ashraf, Sukant K. Bisoy, Nikhat Parveen, Souheil Khawatmi, and Ahmed Abdelsalam. 2023. "Efficient Intrusion Detection System in the Cloud Using Fusion Feature Selection Approaches and an Ensemble Classifier" Electronics 12, no. 11: 2427. https://doi.org/10.3390/electronics12112427

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop