Next Article in Journal
Potential of Laurel (Laurus nobilis L.) Leaf Polyphenols for Modulation of Body Composition
Next Article in Special Issue
Early Forest Fire Detection Using a Protocol for Energy-Efficient Clustering with Weighted-Based Optimization in Wireless Sensor Networks
Previous Article in Journal
A 1 × 2 Two-Dimensional Slanted Grating Based on Double-Layer Cylindrical Structure
Previous Article in Special Issue
Road Scanner: A Road State Scanning Approach Based on Machine Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Cloud Intrusion Detection Systems Based on DNN Using Backpropagation and PSO on the CSE-CIC-IDS2018 Dataset

Department of Information Technology, College of Computer, Qassim University, Buraydah 52571, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(4), 2276; https://doi.org/10.3390/app13042276
Submission received: 19 December 2022 / Revised: 2 February 2023 / Accepted: 8 February 2023 / Published: 10 February 2023

Abstract

:
Cloud computing (CC) is becoming an essential technology worldwide. This approach represents a revolution in data storage and collaborative services. Nevertheless, security issues have grown with the move to CC, including intrusion detection systems (IDSs). Intruders have developed advanced tools that trick the traditional IDS. This study attempts to contribute toward solving this problem and reducing its harmful effects by boosting IDS performance and efficiency in a cloud environment. We build two models based on deep neural networks (DNNs) for this study: the first model is built on a multi-layer perceptron (MLP) with backpropagation (BP), and the other is trained by MLP with particle swarm optimization (PSO). We use these models to deal with binary and multi-class classification on the updated cybersecurity CSE-CIC-IDS2018 dataset. This study aims to improve the accuracy of detecting intrusion attacks for IDSs in a cloud environment and to enhance other performance metrics. In this study, we document all aspects of our experiments in depth. The results show that the best accuracy obtained for binary classification was 98.97% and that for multi-class classification was 98.41%. Furthermore, the results are compared with those from the related literature.

1. Introduction

Over the last decade, we have seen considerable advances in CC. Today, CC services have rich features, such as resource pooling and large networks. Cloud providers offer a platform, services, applications, and infrastructure to the cloud’s users. The concept of a decentralized cloud is changing the face of the internet and decreasing the need for hardware and software. On the user side, the only requirement is being able to navigate the user interface of the CC system.
Today, cloud technology is rapidly growing with a considerably increasing number of users, cloud service providers, and threats that target users and providers. As a result, the uploading of sensitive data to cloud storage by users poses many security issues regarding accessibility, availability, confidentiality, and integrity. Moreover, without perfect protection, cloud services’ uninterrupted provision leads to a risk of intrusion [1].
The cloud environment is vulnerable to intrusion attacks because of its distributed nature. An IDS can be used to ensure the security of the CC by automatically checking the logs, network flow, and configurations. Nevertheless, traditional IDSs that are host- or network-based are not practical for CC since these IDSs are unable to detect hidden attack paths. This makes traditional IDSs inconvenient for CC because they are not designed with the specific context of CC to provide the necessary protection in such an environment. As a result, using traditional IDS techniques in the cloud environment leads to problems that do not exist in the conventional CC [2].
The two main methods of IDS based on detection are misuse (signature) and anomaly. In anomaly, a profile is created by recording normal behaviors, and then anomalies are determined from the regular pattern that can be marked as intrusions. Signature (misuse) detection is achieved with a pattern relating to known attacks or signatures of vulnerabilities. When the IDS detects malicious code matching these signatures, it raises the alarm and marks this code as an intrusion [2].
The vulnerabilities of IDS in the cloud environment can be solved using multiple evolutionary computing and programming techniques. Therefore, current strategies are geared toward the use of a combination of several methods together in one system to build robust and efficient IDSs [3].
Currently, researchers in this field are experimenting with novel methods using metaheuristic algorithms, which attempt to combine the benefits of these algorithms and artificial neural networks (ANNs) to mitigate any significant downside of IDSs [4].
Over the past few years, the applications of DNNs and metaheuristic optimization algorithms have become quite effective. Furthermore, there are different methods for using optimization algorithms in the IDSs. One of these methods uses the PSO algorithm, which simulates self-learning and social organization in groups of animals, such as birds and fish while finding food [5].
BP is a widely used algorithm for training MLP feedforward neural networks in order to adjust the model’s weights and biases. Due to its capacity for learning and flexibility, it has been effectively used in many applications [6]. Furthermore, the PSO algorithm has a powerful capacity for exploitation as well as a quick speed of convergence [7]. Furthermore, using the PSO with neural networks gives better outcomes for IDSs [1].
This study attempts to contribute the previous efforts of researchers in this field using different strategies and an updated dataset to improve the performance of IDSs. The contributions of this study are summarized as follows:
  • We performed a comprehensive empirical study on IDSs using MLP-BP and MLP-PSO techniques to enhance performance metric scores and determine which model is more powerful in the cloud environment.
  • We tried to provide a general analysis of the CSE-CIC-IDS-2018 dataset, and we used it to validate the proposed models for both binary and multi-class classification.
  • We tried to address the shortcomings in the literature and avoid them in our experiments, as well as comparing our results against them.
  • We used several evaluation metrics to provide more detailed analysis and a full view of the proposed models’ performance.
  • In the final analysis, our results show an improvement in performance metric scores, as shown in Section 5.
The rest of this paper is organized as follows: Section 2 discusses related works. Section 3 presents the necessary background information. In Section 4, we show the proposed methods and experimental design. Section 5 presents the results. Section 6 discusses our experimental results and compares them with those in the related literature. Finally, Section 7 concludes this study, followed by future research directions.

2. Related Works

In this section, we will discuss the literary works related to this study, especially those based on DNNs for IDSs, which use the CSE-CIC-IDS2018 dataset. Several researchers have used this approach, with one goal: making the IDSs in the CC more efficient and robust by increasing the model scores (accuracy, precision, etc.).
Table 1 and Table 2 show a list of these studies’ aims, methodologies, environments, and best performance scores. Although all these studies used the CSE-CIC-IDS2018 dataset, each has its own implementation strategy for the proposed IDSs; the features selection process may differ from study to study, in addition to the number of features and samples, and approaches to data preprocessing and splitting. Furthermore, on the level of the DNN architecture, the tuning of the hyperparameters, the number of layers and neurons, and the activation and optimization functions may vary.

2.1. A Semi-Self-Taught Network Intrusion Detection System [8]

The authors used the denoising autoencoder (DAE) and the fuzzy c-means algorithm to build their IDS. This strategy was chosen to address missing values and duplicate data because the different class labels for duplicate or misclassified instances cause class noise. The ratio for training–validation–testing was 70%-15%-15%. The performance scores were 97.90% for accuracy and 98.00% for precision and recall. The limitations of this study include the lack of detail about the experiment and the use of only one classifier.

2.2. Intrusion Detection of Imbalanced Network Traffic Based on Machine Learning and Deep Learning [9]

The authors used a classical classification algorithm comprising machine learning (ML) and deep learning (DL), including random forest (RF), support vector machine (SVM), XGBoost, long short-term memory (LSTM), AlexNet, and Mini-VGGNet. The network was trained on the NSL-KDD and CSE-CIC-IDS2018 datasets. Data preprocessing involved the removal of the timestamp, destination address, source address, source port, etc., from the CSE-CIC-IDS2018. The training–testing ratio was 80%-20%. The highest accuracy was 96.99%, with a precision of 97.46% and 96.97% for recall. The shortcoming of this research is the absence of specific essential details about the experiments.

2.3. Deep Learning Methods in Network Intrusion Detection: A Survey and an Objective Comparison [10]

The authors used four models based on DNN, AE, deep belief network (DBN), and LSTM. These models were trained on KDD99, NSL-KDD, CIC-IDS2017, and CSE-CIC-IDS2018. The outcomes demonstrate the DNN model’s efficacy on all four datasets. Concerning CSE-CIC-IDS2018, the highest accuracy for the DNN was 98.38%, corresponding to a precision of 97.79% and a recall of 97.74%. Some of the PCs involved in the experiments utilized GPU acceleration. The absence of performance scores for multi-class classification was a limitation of this study.

2.4. Attack Classification of an Intrusion Detection System Using Deep Learning and Hyperparameter Optimization [11]

The authors developed a hybrid deep-structured model using DAE for training and DNN for attack classification. They used an autoencoder as an input and a DNN as an output classifier. The raining–testing ratio was 10%-2.5%. Despite the contribution of this helpful study, the accuracy (95.79%) was the lowest among other related works, and the training–testing ratio was very small.

2.5. Network Anomaly Detection with Temporal Convolutional Network and U-Net Model [12]

The authors proposed three models to make the classification of network threats more accurate and to compare the datasets used in their study. The first model used a U-shaped network (U-Net), one of the convolutional neural networks (CNNs). The second model was based on a temporal convolutional network (TCN). The last model was based on a combination of TCN and LSTM. Furthermore, they trained the proposed models on the KDD99 and CSE-CIC-IDS2018 datasets. The ratio for training–validation–testing was 72%-18%-10%. The highest accuracy score was 97.77% based on the TCN-LSTM model, with a precision of 97.94% and 97.53% for recall. The lack of performance scores for binary classification was a limitation of this research.

2.6. Cyber Threat Intelligence Using PCA-DNN Model to Detect Abnormal Network Behavior [13]

The authors used two models based on DNN and DNN with principal component analysis (PCA). Furthermore, they used one-hot encoding and min–max scaling in the data preprocessing phase. They reduced the training time by half while maintaining the accuracy rate using DNN-PCA with only 12 features from the dataset. The training–testing ratio was 80%-20%. From the confusion matrix in the results, we can quickly achieve the highest accuracy of 97.93%, with a precision of 99.97% and a recall of 97.42%. A disadvantage was the absence of performance scores covering the multi-class classification.

3. Background

This section briefly presents a background of the IDS, DNN, PSO, and CSE-CIC-IDS2018 dataset, as shown in the following subsections.

3.1. Intrusion Detection System

An IDS is a piece of software, piece of hardware, or system that can detect harmful activity or protect the system from insider and outsider attacks on the installed device or network. Today, IDSs are an essential tool that network administrators use to protect their systems from cyber and intrusion attacks. The following Figure 1 shows the classification of cloud-based IDSs.

3.1.1. Cloud-Based IDS Types

IDSs in the cloud can be classified into four main types, as follows:
  • Host IDS (HIDS): The HIDS operates by wiping and evaluating the records gathered from an individual computer system. The HIDS monitors system activities, such as system logs and CPU usage, to detect if the system has been attacked, and sends an alarm to the administrators [14].
  • Network IDS (NIDS): The NIDS observes and analyzes network flow and examines packet flow to detect suspicious activities, such as port scanning. NIDS efficiently detects network attacks, such as denial of service (DoS) [15].
  • Distributed IDS (DIDS): The DIDS contains various IDSs, for instance, HIDS and NIDS, that may communicate between themselves or with a centralized server for network monitoring. It may contain two main types of IDS methods. DIDS has the advantage of detecting known and unknown attacks [15].
  • Hypervisor IDS: The hypervisor IDS was created to run a virtual machine (VM), which observes communication between the VM and hypervisors. This type has the advantage of providing state information and event monitoring [16].

3.1.2. Cloud-Based IDS Methods

IDS in the cloud can be classified into two main methods, with a third possibility available, as follows:
  • Signature-based detection: This approach is indicated as misuse- or rule-based. It is based on documented signatures of previous attacks and aims to identify applicable intrusions by analyzing data in opposition to these predefined guidelines. Data compliance with the predefined guidelines could be identified as attacks and invoke an alarm. This system is not capable of detecting unknown attacks [17].
  • Anomaly-based detection: This system requires modern network users to be compared with a baseline of historically learned ordinary community behavior. The baseline is built using observations of the community and hosts for a prolonged period of profile creation. Any enormous perversion from this baseline is identified as an anomaly and invokes an alert. This system is capable of detecting unknown attacks [17].
  • Specification-based detection: This system is a hybrid of signature and anomaly, using the signature approach to detect known attacks and the anomaly approach to detect unknown attacks [18].

3.2. Deep Neural Network

ANNs are computer programs that are biologically inspired and are intended to imitate how the human brain processes information and obtains knowledge by identifying the patterns and correlations included within the data through experience [19]. ANNs comprise several individual units called neurons, which are coupled with weights to create the neural structure of the network, arranged in layers [20].
Nowadays, the term DNN is used to refer to a neural network with multiple hidden layers [21], and DNNs are categorized as a subset of ML [22,23]. Figure 2 shows the relationship between artificial intelligence (AI), ML, and DL.
In a DNN, when the weights of hidden layers are fully connected, this is called MLP [24]. In MLP, each neuron of the network has an activation function (such as softsign, tanh, relu, etc.). MLPs are categorized under feedforward algorithms in which the data flow from the input, through the hidden layers, to the output in one direction [25].
If the algorithm just calculates the weighted sum in each neuron and moves to the next layer, it cannot learn the weights used to minimize the cost function. Unless the algorithm iterates, there is no effective learning. To do this, we need to use the BP as a learning algorithm, which allows the MLP to adjust the network’s weights iteratively by using gradient descent to make the cost function as low as possible [26,27].
To summarize, DNN refers to an ANN with multiple hidden layers [28], and feedforward refers to the ANN’s data flow in one direction from the input to the output [29]. MLP refers to an ANN with fully connected neurons and the use of a kind of activation function [30]. Using gradient descent, BP allows the MLP to adjust the weights to iteratively minimize the cost function [31]. Figure 3 describes the architecture of MLP.

3.3. Particle Swarm Optimization

In 1995, Kennedy and Eberhart proposed the PSO algorithm [32], the goal of which was to solve optimization problems. It is a metaheuristic algorithm categorized under swarm intelligence algorithms [33]. The concept of PSO is to imitate the behavior of birds to acquire food [34]. Each bird modifies its position in the search space to achieve the best solution. Any time a bird finds a food source, it will head toward it [35]. Over the generations (iterations), every particle searches for its own optimal solution and the optimal solution in the swarm. Then, it updates the particle’s position (p) and the rate of change associated with the position (velocity). These updates will affect the personal best (pbest), and global best (gbest) values [36].
The following Equation (1) shows how the velocity is dynamically adjusted for each particle:
v k t + 1 = v k t + c 1 r 1 p b e s t k t p k t + c 2 r 2 ( g b e s t t p k t )
The velocity in the t + 1 th iteration is affected by the three variables listed in Table 3 below [37].
The position equation, Equation (2), can be written as follows:
p k t + 1 = p k t + v k t + 1
Table 4 shows the main parameters used in Equations (1) and (2).
Figure 4 shows how the moving particle looks in the search space, and Figure 5 shows the steps of the PSO algorithm.

3.4. CSE-CIC-IDS2018 Dataset

The CSE-CIC-IDS2018 dataset is a collaboration project proposed in 2018 by CSE [38] and CIC [39]. The dataset is available to download on Amazon Web Services (AWS) [40,41].
The dataset contains ten CSV files representing ten days from the captured network flow, with more than 16.2 million samples. Furthermore, more than 80 features were extracted by the CICFlowMeter tool.
This dataset includes six primary intrusion attack types: distributed DoS (DDoS), DoS, brute force, bot, infiltration, and web attacks, as shown in Table 5 and Table 6 and Figure 6.

4. Methodology

According to the study’s goals, we aimed to build robust and efficient IDSs in the cloud environment. We proposed two models, the first using MLP-BP and the other using MLP-PSO, and used the CSE-CIC-IDS-2018 dataset to train them. The following subsections describe the research design and experimental setup.

4.1. Research Framework and Proposed Methods

This project comprises three phases, and the output from each phase will come as input to the next, as shown in the following Figure 7.
  • Phase 1: Data performed in this phase, including missing, duplicate, infinity, and NaN values. Furthermore, it includes processes such as dropping unnecessary features, feature importance, feature selection, data encoding, and data standardization.
  • Phase 2: In this phase, we will use four scenarios for the training based on DNNs. We will build two models, the first using MLP-BP and the other using MLP-PSO for both binary and multi-class classification.
  • Phase 3: We will use the confusion matrix and equations of performance metrics to evaluate the outcomes of our models.
The flowchart of the proposed methodology is illustrated in Figure 8, and Figure 9 describes the frameworks of our models.

4.2. Data Preparation and Preprocessing

Firstly, we combined the CSV files into one using JupyterLab and Python with the Pandas and Numpy libraries. Furthermore, we cleaned the dataset by removing positive and negative infinity and NaN values. After that, we dropped unnecessary features (“Timestamp”, “Flow ID”, “Src IP”, “Src Port”, and “Dst IP”). In the next step, from the Scikit-learn library, we used the random forest classifier to compute the feature importance, as illustrated in Figure 10.
In contrast, from the Scikit-learn library, we used the SelectFromModel class for feature selection. The outcome of this process was that 24 features were selected, as displayed in Table 7.
Afterward, we made two copies of the dataset. The first copy was for binary classification, and the second was for multi-class classification, containing the following classes: Benign, DDoS, DoS, BruteForce, Bot, Infilteration, and Web.
In the final step, we applied OneHotEncoder to the label column (Y) to convert the categorical data into numerical data, differentiating each class from the others. This process is essential because the neural network requires the data to be numeric. Finally, we applied StandardScaler to the other columns (X); the data had a mean value of 0 and a standard deviation of 1. In other words, after applying StandardScaler, the original distribution of the data remains the same, with the range restricted to [0, 1]. This process is helpful in classification.

4.3. Experimental Setup

Our experiments were applied on AWS, which provides many services, such as the elastic compute cloud (EC2). From EC2, we used an instance size of p3.8xlarge to train our models with GPUs. Accordingly, we used TensorFlow to distribute training across multiple GPUs using the tf.distribute.Strategy API [42]. Furthermore, we used the Pyswarms toolkit package [43] in our proposed MLP-PSO model.
In this subsection, we will summarize our DNN structure and all of the main tools used in this study, as given in the following tables and figures. Figure 11 describes our DNN architecture.
Our DNN parameters are shown in Table 8. Furthermore, Table 9 lists the hardware resources and software environments used in this study. The most important libraries, classes, modules, functions, and algorithms used in our experiments are shown in Table 10.

5. Results

We will use the following equations in Table 11 to evaluate the performance metrics for binary classification, and those in Table 12 to evaluate the performance metrics for multi-class classification [44,45,46].

5.1. MLP-BP (Binary Classification)

Table 13 describes the average accuracy of the proposed models for binary classification using five-fold cross-validation.
Figure 12 and Figure 13 show the training and testing accuracy and the confusion matrix of MLP-BP for binary classification. Table 14 provides the performance metric scores.

5.2. MLP-PSO (Binary Classification)

Figure 14 and Figure 15 show the training and testing accuracy and the confusion matrix of MLP-PSO for binary classification. Table 15 provides the performance metric scores.

5.3. MLP-BP (Multi-Class Classification)

In addition to the binary classification, we show the multi-class classification results with the MLP-BP and MLP-PSO models. Table 16 shows the average accuracy of the proposed models for multi-class classification using five-fold cross-validation.
Table 17 describes the scores of performance metrics for each class of MLP-BP for multi-class classification.
Figure 16 and Figure 17 show the training and testing accuracy and the confusion matrix of MLP-BP for multi-class classification.

5.4. MLP-PSO (Multi-Class Classification)

Table 18 describes the performance metric scores for each class of MLP-PSO for multi-class classification. Figure 18 and Figure 19 show the training and testing accuracy and the confusion matrix of MLP-PSO for multi-class classification.

6. Discussion

The previous results show that the MLP-PSO model performs worse for binary and multi-class classification, with accuracy scores of 96.25% and 95.32%. Conversely, the MLP-BP obtains better accuracy scores (98.97% and 98.41%) for binary and multi-class classification. Furthermore, the MLP-BP obtains better results for FPR, with a value of 0.13% for binary classification and 0.45% for multi-class classification. Furthermore, FNR scores 1.20% and 1.15% for binary and multi-class classification, respectively. Although both types of errors are detrimental to IDSs, FNR (type II error) is more damaging and dangerous since it incorrectly classifies the intrusion attacks as benign. Table 19 summarizes the overall performance scores of the proposed models.
We point out that these results may differ according to the dataset or the strategy. Through our experiments in this study, we find that TensorFlow and Keras have sufficient tools to achieve perfect results without needing an external optimization algorithm. In other situations, it may be helpful to use an optimization algorithm to achieve the best results.
Furthermore, the PSO does not achieve better scores and requires a great deal of training time and hardware resources. Therefore, we suggest building a DNN with TensorFlow and Keras and evaluating the performance before using an external optimization algorithm. After that, according to the outcomes, it can be determined if an optimization algorithm is required.
Additionally, we advise adjusting the learning rate and trying several times to achieve excellent scores and converge toward the minimum in a less noisy manner. TensorFlow and Keras provide several methods to achieve this, such as LearningRateScheduler and ExponentialDecay.
The best performance scores of the proposed MLP-BP and MLP-PSO models compared with those in related works are given in Table 20.

7. Conclusions and Outlooks

In this study, we conducted four experiments based on the MLP-BP and MLP-PSO models for binary and multi-class classification on the CSE-CIC-IDS-2018 dataset. We presented a list of related works and discussed them. Furthermore, we provided an analysis of the CSE-CIC-IDS-2018 dataset and details of our DNN architecture, data preparation, data preprocessing, feature importance, feature selection, and proposed models.
Additionally, we described the details of the hardware resources and software environments used in this study, in addition to the most important libraries, classes, modules, functions, and algorithms used in our experiments.
Finally, we provided and discussed the results. The best accuracy for binary classification was 98.97% and that for multi-class classification was 98.41% based on MLP-BP. Furthermore, we tried to provide some essential tips that we think are useful to researchers in this field. Furthermore, our proposed models showed an improvement in performance metrics scores, which we compared with those in related works.
In the future, we plan to run our IDSs in a cloud computing environment and evaluate the detection efficiency in real time. Furthermore, we propose using another approach or a different metaheuristic optimization algorithm to improve the results.

Author Contributions

Methodology, S.A.; Software, S.A.; Validation, S.E.K.; Writing—original draft, S.A.; Writing—review & editing, S.E.K.; Supervision, S.E.K. All authors have read and agreed to the published version of the manuscript.

Funding

The authors gratefully acknowledge Qassim University, represented by the Deanship of “Scientific Research, on the financial support for this research under the number (COC-2022-1-2-J-29853) during the academic year 1444 AH/2022 AD”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

CCCloud Computing
AIArtificial Intelligence
MLMachine Learning
DLDeep Learning
BPBackpropagation
RFRandom Forest
VMVirtual Machine
AEAutoencoder
DAEDenoising Autoencoder
IDSIntrusion Detection System
ANNArtificial Neural Network
DNNDeep Neural Network
DBNDeep Belief Network
CNNConvolutional Neural Network
TCNTemporal Convolutional Network
U-NetU-shaped Network
MLPMulti-Layer Perceptron
PSOParticle Swarm Optimization
SVMSupport Vector Machine
PCAPrincipal Component Analysis
CSECommunications Security Establishment
CICCanadian Institute for Cybersecurity
AWSAmazon Web Services
EC2Elastic Compute Cloud
IDEInteractive Development Environment
AccAccuracy
ERRError Rate
PPVPositive Predictive Value
TPRTrue Positive Rate
TNRTrue Negative Rate
FPRFalse Positive Rate
FNRFalse Negative Rate
DoSDenial of Service
DDoSDistributed Denial of Service
LSTMLong Short-Term Memory
HIDSHost Intrusion Detection System
NIDSNetwork Intrusion Detection System
DIDSDistributed Intrusion Detection System

References

  1. Saljoughi, A.S.; Mehrvarz, M.; Mirvaziri, H. Attacks and intrusion detection in cloud computing using neural networks and particle swarm optimization algorithms. Emerg. Sci. J. 2017, 1, 179–191. [Google Scholar] [CrossRef]
  2. Riaz, A.; Ahmad, H.F.; Kiani, A.; Qadir, J.; Rasool, R.; Younis, U. Intrusion detection systems in cloud computing: A contemporary review of techniques and solutions. J. Inf. Sci. Eng. 2017, 33, 611–634. [Google Scholar] [CrossRef]
  3. Çavuşoğlu, Ü. A new hybrid approach for intrusion detection using machine learning methods. Appl. Intell. 2019, 49, 2735–2761. [Google Scholar] [CrossRef]
  4. Chong, H.Y.; Yap, H.J.; Tan, S.C.; Yap, K.S.; Wong, S.Y. Advances of metaheuristic algorithms in training neural networks for industrial applications. Soft Comput. 2021, 25, 11209–11233. [Google Scholar] [CrossRef]
  5. Shi, Y. Particle swarm optimization. IEEE Connect. 2004, 2, 8–13. [Google Scholar]
  6. Rehman, M.Z.; Nawi, N.M. The effect of adaptive momentum in improving the accuracy of gradient descent back propagation algorithm on classification problems. In Proceedings of the International Conference on Software Engineering and Computer Systems, Kuantan, Malaysia, 27–29 June 2011; pp. 380–390. [Google Scholar] [CrossRef]
  7. Rezaei, F.; Safavi, H.R. GuASPSO: A new approach to hold a better exploration–exploitation balance in PSO algorithm. Soft Comput. 2020, 24, 4855–4875. [Google Scholar] [CrossRef]
  8. Zhao, F.; Zhang, H.; Peng, J.; Zhuang, X.; Na, S.G. A semi-self-taught network intrusion detection system. Neural Comput. Appl. 2020, 32, 17169–17179. [Google Scholar] [CrossRef]
  9. Liu, L.; Wang, P.; Lin, J.; Liu, L. Intrusion detection of imbalanced network traffic based on machine learning and deep learning. IEEE Access 2020, 9, 7550–7563. [Google Scholar] [CrossRef]
  10. Gamage, S.; Samarabandu, J. Deep learning methods in network intrusion detection: A survey and an objective comparison. J. Netw. Comput. Appl. 2020, 169, 102767. [Google Scholar] [CrossRef]
  11. Kunang, Y.N.; Nurmaini, S.; Stiawan, D.; Suprapto, B.Y. Attack classification of an intrusion detection system using deep learning and hyperparameter optimization. J. Inf. Secur. Appl. 2021, 58, 102804. [Google Scholar] [CrossRef]
  12. Mezina, A.; Burget, R.; Travieso-González, C.M. Network anomaly detection with temporal convolutional network and U-Net model. IEEE Access 2021, 9, 143608–143622. [Google Scholar] [CrossRef]
  13. Al-Fawa’reh, M.; Al-Fayoumi, M.; Nashwan, S.; Fraihat, S. Cyber threat intelligence using PCA-DNN model to detect abnormal network behavior. Egypt. Inform. J. 2022, 23, 173–185. [Google Scholar] [CrossRef]
  14. Elmaaradi, A.; Lyhyaoui, A.; Chairi, I. New security architecture using hybrid IDS for virtual private clouds. In Proceedings of the 2019 Third International Conference on Intelligent Computing in Data Sciences (ICDS), Marrakech, Morocco, 28–30 October 2019; pp. 1–5. [Google Scholar] [CrossRef]
  15. Mehmood, Y.; Shibli, M.A.; Habiba, U.; Masood, R. Intrusion detection system in cloud computing: Challenges and opportunities. In Proceedings of the 2013 2nd National Conference on Information Assurance (NCIA), Rawalpindi, Pakistan, 11–12 December 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 59–66. [Google Scholar] [CrossRef]
  16. Mahajan, V.; Peddoju, S.K. Deployment of intrusion detection system in cloud: A performance-based study. In Proceedings of the 2017 IEEE Trustcom/BigDataSE/ICESS, Sydney, NSW, Australia, 1–4 August 2017; pp. 1103–1108. [Google Scholar] [CrossRef]
  17. Modi, C.; Patel, D.; Borisaniya, B.; Patel, H.; Patel, A.; Rajarajan, M. A survey of intrusion detection techniques in cloud. J. Netw. Comput. Appl. 2013, 36, 42–57. [Google Scholar] [CrossRef]
  18. Elrawy, M.F.; Awad, A.I.; Hamed, H.F. Intrusion detection systems for IoT-based smart environments: A survey. J. Cloud Comput. 2018, 7, 1–20. [Google Scholar] [CrossRef]
  19. Malekian, A.; Chitsaz, N. Concepts, procedures, and applications of artificial neural network models in streamflow forecasting. In Advances in Streamflow Forecasting; Elsevier: Amsterdam, The Netherlands, 2021; pp. 115–147. [Google Scholar] [CrossRef]
  20. Agatonovic-Kustrin, S.; Beresford, R. Basic concepts of artificial neural network (ANN) modeling and its application in pharmaceutical research. J. Pharm. Biomed. Anal. 2000, 22, 717–727. [Google Scholar] [CrossRef]
  21. Kim, D.E.; Gofman, M. Comparison of shallow and deep neural networks for network intrusion detection. In Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 8–10 January 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 204–208. [Google Scholar] [CrossRef]
  22. Gupta, R.; Srivastava, D.; Sahu, M.; Tiwari, S.; Ambasta, R.K.; Kumar, P. Artificial intelligence to deep learning: Machine intelligence approach for drug discovery. Mol. Divers. 2021, 25, 1315–1360. [Google Scholar] [CrossRef] [PubMed]
  23. Ceron, R. AI, Machine Learning and Deep Learning: What’s the Difference. IBM IT Infrastructure Blog. 2019. Available online: https://www.ibm.com/blogs/systems/ai-machine (accessed on 2 February 2023).
  24. Deng, L.; Yu, D. Deep learning: Methods and applications. Found. Trends® Signal Process. 2014, 7, 197–387. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Zhang, K.; Khelifi, A. Multivariate TIME Series Analysis in Climate and Environmental Research; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar] [CrossRef]
  26. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  27. Rojas, R. The backpropagation algorithm. In Neural Networks; Springer: Berlin/Heidelberg, Germany, 1996; pp. 149–182. [Google Scholar] [CrossRef]
  28. Li, L.; Zhao, Y.; Jiang, D.; Zhang, Y.; Wang, F.; Gonzalez, I.; Valentin, E.; Sahli, H. Hybrid deep neural network–Hidden markov model (dnn-hmm) based speech emotion recognition. In Proceedings of the 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 312–317. [Google Scholar] [CrossRef]
  29. Poznyak, T.; Oria, J.I.C.; Poznyak, A. Ozonation and Biodegradation in Environmental Engineering: Dynamic Neural Network Approach; Elsevier: Amsterdam, The Netherlands, 2019. [Google Scholar] [CrossRef]
  30. Camacho Olmedo, M.; Paegelow, M.; Mas, J.F.; Escobar, F. Geomatic approaches for modeling land change scenarios. An introduction. In Geomatic Approaches for Modeling Land Change Scenarios; Springer: Berlin/Heidelberg, Germany, 2018; pp. 1–8. [Google Scholar] [CrossRef]
  31. Aggarwal, C.C. Training deep neural networks. In Neural Networks and Deep Learning; Springer: Berlin/Heidelberg, Germany, 2018; pp. 105–167. [Google Scholar] [CrossRef]
  32. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95–International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
  33. Mavrovouniotis, M.; Li, C.; Yang, S. A survey of swarm intelligence for dynamic optimization: Algorithms and applications. Swarm Evol. Comput. 2017, 33, 1–17. [Google Scholar] [CrossRef]
  34. Carvalho, M.; Ludermir, T.B. An analysis of PSO hybrid algorithms for feed-forward neural networks training. In Proceedings of the 2006 Ninth Brazilian Symposium on Neural Networks (SBRN’06), Ribeirao Preto, Brazil, 23–27 October 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 6–11. [Google Scholar] [CrossRef]
  35. Shelokar, P.; Siarry, P.; Jayaraman, V.K.; Kulkarni, B.D. Particle swarm and ant colony algorithms hybridized for improved continuous optimization. Appl. Math. Comput. 2007, 188, 129–142. [Google Scholar] [CrossRef]
  36. Alam, S.; Dobbie, G.; Riddle, P. An evolutionary particle swarm optimization algorithm for data clustering. In Proceedings of the 2008 IEEE Swarm Intelligence Symposium, St. Louis, MO, USA, 21–23 September 2008; pp. 1–6. [Google Scholar] [CrossRef]
  37. Jain, M.; Saihjpal, V.; Singh, N.; Singh, S.B. An Overview of Variants and Advancements of PSO Algorithm. Appl. Sci. 2022, 12, 8392. [Google Scholar] [CrossRef]
  38. Communications Security Establishment. Government of Canada. Available online: https://www.cse-cst.gc.ca/en (accessed on 2 February 2023).
  39. Canadian Institute for Cybersecurity. University of New Brunswick est.1785. Available online: https://www.unb.ca/cic/ (accessed on 2 February 2023).
  40. CSE-CIC-IDS2018 on AWS. Canadian Institute for Cybersecurity. Available online: https://www.unb.ca/cic/datasets/ids-2018.html (accessed on 2 February 2023).
  41. Registry of Open Data on AWS. A Realistic Cyber Defense Dataset (CSE-CIC-IDS2018). Available online: https://registry.opendata.aws/cse-cic-ids2018 (accessed on 2 February 2023).
  42. TensorFlow. Distributed Training with TensorFlow. Available online: https://www.tensorflow.org/guide/distributed_training (accessed on 2 February 2023).
  43. Miranda, L.J. PySwarms: A research toolkit for Particle Swarm Optimization in Python. J. Open Source Softw. 2018, 3, 433. [Google Scholar] [CrossRef]
  44. Haghighi, S.; Jasemi, M.; Hessabi, S.; Zolanvari, A. PyCM: Multiclass confusion matrix library in Python. J. Open Source Softw. 2018, 3, 729. [Google Scholar] [CrossRef]
  45. Sammut, C.; Webb, G.I. Encyclopedia of Machine Learning; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2010. [Google Scholar] [CrossRef]
  46. Sammut, C.; Webb, G.I. Encyclopedia of Machine Learning and Data Mining; Springer Publishing Company: Berlin/Heidelberg, Germany, 2017. [Google Scholar] [CrossRef]
Figure 1. Classification of cloud-based IDSs.
Figure 1. Classification of cloud-based IDSs.
Applsci 13 02276 g001
Figure 2. Relationship between AI, ML, and DL.
Figure 2. Relationship between AI, ML, and DL.
Applsci 13 02276 g002
Figure 3. MLP: feedforward DNN with BP and two hidden layers.
Figure 3. MLP: feedforward DNN with BP and two hidden layers.
Applsci 13 02276 g003
Figure 4. Moving particle in PSO.
Figure 4. Moving particle in PSO.
Applsci 13 02276 g004
Figure 5. Steps of the PSO algorithm.
Figure 5. Steps of the PSO algorithm.
Applsci 13 02276 g005
Figure 6. Distribution of attack classes in CSE-CIC-IDS2018.
Figure 6. Distribution of attack classes in CSE-CIC-IDS2018.
Applsci 13 02276 g006
Figure 7. The framework of research.
Figure 7. The framework of research.
Applsci 13 02276 g007
Figure 8. General methodology workflow.
Figure 8. General methodology workflow.
Applsci 13 02276 g008
Figure 9. The proposed models’ frameworks: (a) binary classification; (b) multi-class classification.
Figure 9. The proposed models’ frameworks: (a) binary classification; (b) multi-class classification.
Applsci 13 02276 g009
Figure 10. Feature importance ranking of CSE-CIC-IDS2018.
Figure 10. Feature importance ranking of CSE-CIC-IDS2018.
Applsci 13 02276 g010
Figure 11. The architecture of our DNN.
Figure 11. The architecture of our DNN.
Applsci 13 02276 g011
Figure 12. Accuracy of MLP-BP for binary classification.
Figure 12. Accuracy of MLP-BP for binary classification.
Applsci 13 02276 g012
Figure 13. Confusion matrix of MLP-BP for binary classification.
Figure 13. Confusion matrix of MLP-BP for binary classification.
Applsci 13 02276 g013
Figure 14. Accuracy of MLP-PSO for binary classification.
Figure 14. Accuracy of MLP-PSO for binary classification.
Applsci 13 02276 g014
Figure 15. Confusion matrix of MLP-PSO for binary classification.
Figure 15. Confusion matrix of MLP-PSO for binary classification.
Applsci 13 02276 g015
Figure 16. Accuracy of MLP-BP for multi-class classification.
Figure 16. Accuracy of MLP-BP for multi-class classification.
Applsci 13 02276 g016
Figure 17. Confusion matrix of MLP-BP for multi-class classification.
Figure 17. Confusion matrix of MLP-BP for multi-class classification.
Applsci 13 02276 g017
Figure 18. Accuracy of MLP-PSO for multi-class classification.
Figure 18. Accuracy of MLP-PSO for multi-class classification.
Applsci 13 02276 g018
Figure 19. Confusion matrix of MLP-PSO for multi-class classification.
Figure 19. Confusion matrix of MLP-PSO for multi-class classification.
Applsci 13 02276 g019
Table 1. Summary of related works.
Table 1. Summary of related works.
AuthorYearAimMethodEnvironment
[8]2020To understand the nature of an attackDNN AutoencoderPython, Keras, Tensorflow
[9]2020To tackle the class imbalance problemDNNScikit-learn, Tensorflow, Google Colab
[10]2020To identify the capability of IDS modelsDNNSeveral PCs (using GPU on some PCs)
[11]2021To improve detection performanceDNN AutoencoderPython, Keras, Tensorflow-GPU
[12]2021To increase classification accuracyTCN + LSTMPython, Tensorflow
[13]2022To increase security and performanceDNNPython, Numpy
Table 2. Best performance scores on the CSE-CIC-IDS2018 dataset.
Table 2. Best performance scores on the CSE-CIC-IDS2018 dataset.
AuthorAccuracyPrecisionRecallF1-Score
[8]97.90%98.00%98.00%98.00%
[9]96.99%97.46%96.97%97.04%
[10]98.38%97.79%97.74%97.76%
[11]95.79%95.38%95.79%95.11%
[12]97.77%97.94%97.53%97.73%
[13]97.93%99.97%97.42%98.68%
Table 3. Velocity components.
Table 3. Velocity components.
NotationComponent
v k t physical component
c 1 r 1 p b e s t k t p k t cognitive component
c 2 r 2 ( g b e s t t p k t ) social component
Table 4. Parameters of PSO.
Table 4. Parameters of PSO.
ParameterDescription
p k the position of a particle
v k the velocity of a particle
p b e s t k the best solution for the particle
g b e s t k the best solution for the swarm
c 1 cognitive factor
c 2 social factor
r 1 , r 2 random numbers between 0 and 1
t the iteration number
Table 5. The network traffic of CSE-CIC-IDS2018.
Table 5. The network traffic of CSE-CIC-IDS2018.
DayFile NameFeaturesType of AttackCountTotal
1Wednesday-14-02-201880Benign667,6261,048,575
FTP-BruteForce193,360
SSH-Bruteforce187,589
2Thursday-15-02-201880Benign996,0771,048,575
DoS attacks-GoldenEye41,508
DoS attacks-Slowloris10,990
3Friday-16-02-201880Benign446,7721,048,574
DoS attacks-Hulk461,912
DoS attacks-SlowHTTPTest139,890
4Thursday-20-02-201884Benign7,372,5577,948,748
DDoS attacks-LOIC-HTTP576,191
5Wednesday-21-02-201880Benign360,8331,048,575
DDoS attack-HOIC686,012
DDoS attack-LOIC-UDP1730
6Thursday-22-02-201880Benign1,048,2131,048,575
Brute Force-Web249
Brute Force-XSS79
SQL Injection34
7Friday-23-02-201880Benign1,048,0091,048,575
Brute Force-Web362
Brute Force-XSS151
SQL Injection53
8Wednesday-28-02-201880Benign544,200613,071
Infilteration68,871
9Thursday-01-03-201880Benign238,037331,100
Infilteration93,063
10Friday-02-03-201880Benign762,3841,048,575
Botnet286,191
Total of rows (Samples) = 16,232,943; Total of columns (Features) = 84.
Table 6. Primary attack classes of CSE-CIC-IDS2018.
Table 6. Primary attack classes of CSE-CIC-IDS2018.
ClassBenignDDoSDoSBrute ForceBotInfilterationWeb
TypeBenignHOICHulkFTPBotnetInfiltrationWeb
LOIC-UDPGoldenEyeXSS
LOIC-HTTPSlowlorisSSHSQL Injection
SlowHTTPTest
Count13,484,7081,263,933654,300380,949286,191161,934928
Table 7. List of selected features.
Table 7. List of selected features.
RankFeatureRankFeatureRankFeature
1Dst Port9Flow Duration17Fwd IAT Min
2Fwd Seg Size Min10Fwd IAT Mean18Flow Pkts/s
3Init Fwd Win Byts11Fwd IAT Max19Fwd Header Len
4Fwd Pkt Len Max12Flow IAT Max20Init Bwd Win Byts
5TotLen Fwd Pkts13Fwd Seg Size Avg21Bwd Pkts/s
6Subflow Fwd Byts14Flow IAT Min22Pkt Len Max
7Fwd Pkt Len Mean15Fwd Pkts/s23Tot Fwd Pkts
8Fwd IAT Tot16Flow IAT Mean24Subflow Fwd Pkts
Table 8. Our DNN parameters.
Table 8. Our DNN parameters.
ParameterValue
Hidden layers ( h ) 2
n Number of neurons
n f e a t u r e s 24
n c l a s s e s 2 neurons in binary classification
7 neurons in multi-class classification
Loss function“BinaryCrossentropy” for binary classification
“CategoricalCrossentropy” for multi-class classification
Optimization algorithmAdam
Learning rateUsing ExponentialDecay
Batch size2048
Epochs100
Training-testing ratio70% (11,363,060 samples)–30% (4,869,883 samples)
Cross-validation5 folds
PSO iterations50
PSO swarm25
PSO c 1 1.5
PSO c 2 1.5
PSO w 0.9
Table 9. List of hardware resources and software environments.
Table 9. List of hardware resources and software environments.
Hardware/SoftwareDetail
Computing platformPlatform: Amazon Web Services (AWS)
Instance size: p3.8xlarge
OS: Ubuntu 20.04.4 LTS x86_64
CPU: Intel Xeon E5-2686 v4 3.00GHz
GPUs: 4x (NVIDIA Tesla V100 SXM2 16GB)
vCPUs: 32
RAM: 244 (GiB)
Programming languagePython
Web-based IDEJupyterLab
ML frameworkScikit-learn
DL frameworkTensorFlow-GPU
DL APIKeras
Table 10. List of programming tools.
Table 10. List of programming tools.
List 1List 2List 3
mathkeras.layersPyCM
keraskeras.modelsKFold
numpykeras.callbacksEarlyStopping
pandassklearn.metricsGlobalBestPSO
sklearnkeras.optimizersStandardScaler
seabornsklearn.ensembletrain_test_split
datetimesklearn.preprocessingOneHotEncoder
matplotlibsklearn.neural_networkclassification_report
pyswarmssklearn.model_selectionLearningRateScheduler
Tensorflowsklearn.feature_selectionRandomForestClassifier
Table 11. Equations of the performance metrics for binary classification.
Table 11. Equations of the performance metrics for binary classification.
MeasureFormula
AccuracyAcc = (TP + TN)/Total
Misclassification Rate (Error Rate)ERR = 1-Acc
Precision (Positive Predictive Value)PPV = TP/(TP + FP)
Recall (Sensitivity or True Positive Rate)TPR = TP/(TP + FN)
Specificity (True Negative Rate)TNR = TN/(TN + FP)
F1-scoreF1 = 2(PPV * TPR)/(PPV + TPR)
False Positive Rate (Type I Error)FPR = FP/(TN + FP) or FPR = 1-Specificity
False Negative Rate (Type II Error)FNR = FN/(TP + FN) or FNR = 1-Sensitivity
Table 12. Equations of the performance metrics for multi-class classification.
Table 12. Equations of the performance metrics for multi-class classification.
MeasureFormula
Accuracy Acc = T P T o t a l
Precision PPV = T P ( T P + F P )
Recall TPR = T P ( T P + F N )
Table 13. The average accuracy of proposed models for binary classification.
Table 13. The average accuracy of proposed models for binary classification.
ModelFold 1Fold 2Fold 3Fold 4Fold 5Average Accuracy
MLP-BP98.95%98.97%98.99%98.98%98.97%98.97%
MLP-PSO96.23%96.24%96.25%96.26%96.25%96.25%
Table 14. Performance scores for each class of MLP-BP for binary classification.
Table 14. Performance scores for each class of MLP-BP for binary classification.
ClassPrecisionRecallF1-ScoreSamples
Benign0.991.000.994,046,458
Intrusion1.000.940.97823,425
Table 15. Performance scores for each class of MLP-PSO for binary classification.
Table 15. Performance scores for each class of MLP-PSO for binary classification.
ClassPrecisionRecallF1-ScoreSamples
Benign0.970.990.984,046,458
Intrusion0.930.840.88823,425
Table 16. The average accuracy of proposed models for multi-class classification.
Table 16. The average accuracy of proposed models for multi-class classification.
ModelFold 1Fold 2Fold 3Fold 4Fold 5Average Accuracy
MLP-BP98.40%98.39%98.41%98.42%98.41%98.41%
MLP-PSO95.31%95.32%95.32%95.31%95.32%95.32%
Table 17. Performance scores for each class of MLP-BP for multi-class classification.
Table 17. Performance scores for each class of MLP-BP for multi-class classification.
ClassPrecisionRecallF1-ScoreSamples
Benign0.991.000.994,046,458
DDoS1.001.001.00378,558
DoS0.960.900.93196,350
BruteForce0.840.940.89113,902
Bot1.001.001.0085,680
Infilteration0.440.010.0148,649
Web0.950.240.38286
Table 18. Performance scores for each class of MLP-PSO for multi-class classification.
Table 18. Performance scores for each class of MLP-PSO for multi-class classification.
ClassPrecisionRecallF1-ScoreSamples
Benign0.970.990.984,046,458
DDoS0.980.770.86378,558
DoS0.760.860.80196,350
BruteForce0.720.650.68113,902
Bot0.980.980.9885,680
Infilteration0.000.000.0048,649
Web0.000.000.00286
Table 19. Overall performance scores of the proposed models.
Table 19. Overall performance scores of the proposed models.
ModelAccuracyERRPrecisionRecallF1-ScoreFPRFNR
MLP-BP (Binary Classification)98.97%1.03%99.98%98.80%99.38%0.13%1.20%
MLP-PSO (Binary Classification)96.25%3.75%98.75%96.80%97.76%6.48%3.20%
MLP-BP (Multi-class Classification)98.41%1.59%99.55%98.85%99.20%0.45%1.15%
MLP-PSO (Multi-class Classification)95.32%4.68%98.97%96.27%97.60%1.03%3.73%
Table 20. The best performance scores of the proposed models compared with related works.
Table 20. The best performance scores of the proposed models compared with related works.
ModelAccuracyPrecisionRecallF1-Score
[8]97.90%98.00%98.00%98.00%
[9]96.99%97.46%96.97%97.04%
[10]98.38%97.79%97.74%97.76%
[11]95.79%95.38%95.79%95.11%
[12]97.77%97.94%97.53%97.73%
[13]97.93%99.97%97.42%98.68%
Proposed MLP-PSO96.25%98.75%96.80%97.76%
Proposed MLP-BP98.97%99.98%98.80%99.38%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alzughaibi, S.; El Khediri, S. A Cloud Intrusion Detection Systems Based on DNN Using Backpropagation and PSO on the CSE-CIC-IDS2018 Dataset. Appl. Sci. 2023, 13, 2276. https://doi.org/10.3390/app13042276

AMA Style

Alzughaibi S, El Khediri S. A Cloud Intrusion Detection Systems Based on DNN Using Backpropagation and PSO on the CSE-CIC-IDS2018 Dataset. Applied Sciences. 2023; 13(4):2276. https://doi.org/10.3390/app13042276

Chicago/Turabian Style

Alzughaibi, Saud, and Salim El Khediri. 2023. "A Cloud Intrusion Detection Systems Based on DNN Using Backpropagation and PSO on the CSE-CIC-IDS2018 Dataset" Applied Sciences 13, no. 4: 2276. https://doi.org/10.3390/app13042276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop