Novel Machine Learning Approach for DDoS Cloud Detection: Bayesian-Based CNN and Data Fusion Enhancements
<p>Preprocessing of The CICDDoS2019 Dataset Phases.</p> "> Figure 2
<p>CICDDOS 2019 Dataset Class Distribution After Sampling Chart.</p> "> Figure 3
<p>The Heatmap of Correlated Features.</p> "> Figure 4
<p>Cumulative Explained Variance Plot.</p> "> Figure 5
<p>General Structure of BaysCNN Model.</p> "> Figure 6
<p>BaysCNN Model Architecture.</p> "> Figure 7
<p>BaysFusCNN Model Architecture.</p> "> Figure 8
<p>BaysCNN Model Confusion Matrix Chart.</p> "> Figure 9
<p>Rho Weights Distribution for BaysCNN Model.</p> "> Figure 10
<p>Mean Weights Distribution for BaysCNN Model.</p> "> Figure 11
<p>Training Progress for BaysCNN Model.</p> "> Figure 12
<p>Training Accuracy for BaysCNN Model.</p> "> Figure 13
<p>Validation Accuracy for BaysCNN Model.</p> "> Figure 14
<p>Training Loss Accuracy for BaysCNN Model.</p> "> Figure 15
<p>Validation Loss Accuracy for BaysCNN Model.</p> "> Figure 16
<p>BaysFusCNN Model Confusion Matrix Chart.</p> "> Figure 17
<p>BaysFusCNN Model Training Progress and Validation Accuracy Results.</p> "> Figure 18
<p>Training Accuracy for BaysFusCNN Model.</p> "> Figure 19
<p>Validation Accuracy for BaysFusCNN Model.</p> "> Figure 20
<p>Training Loss Accuracy for BaysFusCNN Model.</p> "> Figure 21
<p>Validation Loss Accuracy for BaysFusCNN Model.</p> ">
Abstract
:1. Introduction
2. Research Questions
- How effective is the proposed Bayesian-based CNN (BaysCNN) model in detecting DDoS attacks in cloud computing environments?
- To what extent does the incorporation of the Data Fusion BaysFusCNN approach improve the accuracy, reliability, and key performance metrics (e.g., accuracy, false positive rates) of DDoS detection compared to the BaysCNN model alone and existing methods?
- How do the proposed models compare to existing DDoS detection methods in terms of accuracy and efficiency?
3. Literature Survey
4. Methodology
4.1. Preprocessing
4.1.1. Gathering of Data: Canadian Institute for Cyber-Security Datasets
4.1.2. Sampling Multi-Class Datasets
- Obtain the class distribution of the original dataset.
- Calculate the Shannon Diversity Index (H) using the formula:
- 3.
- Calculate the maximum possible diversity (Hmax) using the formula:
- 4.
- Calculate the Pielou Index (J) by dividing H by Hmax.
4.1.3. Data Cleaning
Removing Categorical Data
Replace Infinities
Flipping Negative Numbers
Remove Constant Features
Remove Quasi-Constant Features
Remove Duplicated Features
4.1.4. Encode Labels Using One-Hot Encoding
4.1.5. Data Scaling
- mydatamydata: This is the feature or variable that we want to scale.
- minDataminData: Represents the minimum value of the feature.
- maxDatamaxData: Represents the maximum value of the feature.
- mydata: The feature or variable that has undergone min-max scaling.
- The multiplication by 2 expands the range of the scaled values.
- Subtracting one shifts the range to be centered around 0.
4.1.6. Implementing PCA
4.1.7. Data Splitting
4.2. BaysCNN Model
- Complex Pattern Recognition: The combination of Bayesian methods and CNN allows for the accurate identification and prediction of complex data patterns within network traffic data.
- Probabilistic Approach: Bayesian ML introduces a probabilistic perspective, enabling us to assess the uncertainties associated with network traffic patterns. This helps improve the model’s robustness in the presence of noise or variations.
4.2.1. Model Architecture Overview
- Input Layer: The model begins with an imageInputLayer of dimensions [41 1 1] corresponding to the size of the input data.
- Convolutional Layers: The input data undergoes convolution via a convolution2dLayer with 64 neurons. This is followed by a batchNormalizationLayer and maxpooling2dLayer. The output of these layers is then passed through a ReLU layer.
- Fully Connected Layers: Following the convolutional layers, the data flows through a fullyConnectedLayer with 200 neurons and two BayesFullyConnectedLayers with 784 output neurons and 392 output neurons.
- Intermediate Layers: The architecture includes additional layers, including ReLU, batchNormalizationLayer, and leakyReLU, to further enhance feature extraction and representation.
- Output Layer: The model concludes with a SoftmaxLayer and a Classification Layer.
4.2.2. Practical Implementation
4.2.3. Developing BaysCNN Model
The Complete Structure
4.2.4. Developing BaysFusCNN Architecture
- Input Layer: The model’s journey begins with an input layer tailored to accommodate data with dimensions of 41 × 1 × 1, designed to seamlessly integrate with the size of the input data.
- Initial Convolutional Layer: The input data are then processed through an initial convolutional layer featuring 16 filters, a padding of 1, and a kernel size of 3. This layer efficiently extracts essential features from the input spectrograms.
- Sequential Processing: The output from the initial convolutional layer undergoes sequential processing, featuring:
- Batch Normalization Layer: To ensure stability and expedited convergence.
- Max Pooling Layer: With pooling size [2 1] and a stride of [2 1], effectively reducing spatial dimensions.
- ReLU Activation Layer: To introduce non-linearity into feature extraction.
- Bayesian Fully Connected Layers: The output from the sequential processing encounters two Bayesian fully connected layers, with output capacities of 784 and 392, respectively. These layers introduce probabilistic weight distributions into the network, enhancing expressiveness and uncertainty modeling. The model parameters, Sigma1 and Sigma2, are tuned to values of 1 and 0.5, respectively.
- Intermediate Processing: The output from the initial Bayesian fully connected layer proceeds through:
- Leaky ReLU Activation Layer: To capture complex patterns.
- Batch Normalization Layer: Ensuring stable training.
- Second Bayesian Fully Connected Layer: Further refine and diversify feature representations.
- Leaky ReLU Activation Layer: Introducing controlled non-linearity.
- Parallel Branch: Simultaneously, the input data undergoes an alternative branch, mirroring the specifications of the initial branch’s convolutional layer, batch normalization layer, max pooling layer, and ReLU activation layer. Like the initial branch, this branch is also subjected to batch normalization.
- Concatenation and Output: The outputs from both branches are concatenated, creating a fusion of features. This fused output proceeds through a softmax layer, culminating in the model’s final output.
- Training and Evaluation: Training the network involves the propagation of training data and predictor data through the network, guided by the options outlined in the ‘options’ variable. Subsequently, the trained network is evaluated using test data to assess its efficacy in classifying input data.
4.2.5. Experimental Setup
Hardware Specifications
Software Specification
- MATLAB: MATLAB version R2023b was utilized for model development, training, and evaluation. It offers a versatile environment for deep learning and data analysis.
- Python 3.7: Python 3.7 was employed for preprocessing the dataset, making use of its rich ecosystem of libraries for data manipulation.
Training Options
Evaluation Metrics
- Accuracy
- 2.
- Recall (Sensitivity)
- 3.
- Precision
- 4.
- Specificity
- 5.
- Fall-out (False Positive Rate)
- 6.
- F1 Score
5. Results
5.1. BaysCNN Model Results
5.1.1. Confusion Matrix Results of BaysCNN Model
- True Positives (TP): These values signify instances correctly classified into their respective classes. In Class 1, 11,223 true positives imply precise recognition of Class 1 instances.
- False Negatives (FN): FN represents instances of a specific class incorrectly assigned to other classes. In Class 1, 316 false negatives indicate misclassified instances that should belong to Class 1.
- False Positives (FP): FP denotes instances inaccurately classified into a particular class when they do not belong. Class 1, for example, records 306 false positives, signifying instances wrongly labeled as Class 1.
- High TP: A high count of true positives demonstrates effective recognition of specific class instances.
- Low FN: Few false negatives indicate infrequent misclassification of class instances as something else.
- Low FP: A minimal number of false positives showcases model precision by reducing the misclassification of non-members into a class.
5.1.2. Rho Weight Distribution of BaysCNN Model
5.1.3. Mean Weight Distribution of BaysCNN Model
5.1.4. Training Progress of BaysCNN Model
5.1.5. Training Accuracy of BaysCNN Model
5.1.6. Training Loss of BaysCNN Model
5.2. Results of the BaysFusCNN Model
- True Positive (TP) Recall: Demonstrates the model’s effectiveness in correctly identifying positive samples within each class. The BaysFusCNN model maintains an average TP Recall of approximately 98.55%, ranging from 97.30% to 99.61% across classes.
- Precision (PP): Measures the model’s accuracy in identifying positive samples while minimizing false positives. The model exhibits an average Precision of approximately 98.57%, with class-specific values ranging from 97.53% to 99.17%.
- Specificity (AN): Evaluates the model’s ability to correctly classify negative samples within each class. The BaysFusCNN model demonstrates an average specificity of 99.88%, varying from 99.81% to 99.92% among classes.
- False Positive Recall (FP Recall): Reflects the rate of false positive predictions by the model. The BaysFusCNN model maintains low FP Recall values, ranging from approximately 0.08% to 0.19% across classes.
- F1 Score: Offers a balanced measure of precision and recall, indicating the model’s overall performance. The average F1 Score stands at 98.56%, with class-specific values spanning from 98.32% to 99.39%.
- Average Performance Metrics: Across all classes, the BaysFusCNN model achieves an average accuracy of approximately 99.79%, an average TP Recall of 98.55%, an average PP Precision of 98.57%, an average Specificity of 99.88%, an average FP Recall of 0.12%, and an average F1 Score of 98.56%. These averages provide a comprehensive overview of the model’s proficiency in classifying diverse network traffic samples.
5.2.1. Confusion Matrix Results of BaysFusCNN
- True Positives (TP): These values indicate instances correctly classified into their respective classes. For example, in Class 1, there were 11,308 instances correctly predicted.
- False Negatives (FN): FN values represent instances that belonged to a specific class but were incorrectly classified as something else. In Class 1, there were 58 instances incorrectly classified.
- False Positives (FP): FP values signify instances wrongly classified into a specific class instead of their true class. For instance, in Class 1, there were 181 instances incorrectly predicted.
5.2.2. Training Progress of BaysFusCNN Model
5.2.3. Training Accuracy of BaysFusCNN Model
5.2.4. Training Loss of BaysFusCNN Model
5.3. Comparing BaysFusCNN & BaysCNN for DDoS Detection in Cloud Environment
- Accuracy: The BaysFusCNN model achieves an average accuracy of 99.79%, a notable improvement over the BaysCNN model’s 99.66%. This increase highlights the advantages of combining multiple classifiers for enhanced network attack detection accuracy.
- True Positive Recall Sensitivity: In most classes, the BaysFusCNN model outperforms the BaysCNN model in True Positive Recall Sensitivity, particularly in classes like DrDoS_LDAP and DrDoS_NTP. This indicates the fusion model’s superior ability to identify previously misclassified instances, a crucial aspect of network security.
- Predicted Positive Precision: The BaysFusCNN model demonstrates superior Predicted Positive Precision in most classes, emphasizing its proficiency in recognizing true positive cases. This reduces false alarms and improves the model’s precision in distinguishing actual threats.
- Actual Negative Specificity: The BaysFusCNN model significantly enhances actual negative Specificity across most classes compared to the BaysCNN model. This reduction in false positives ensures accurate identification of benign traffic, a vital aspect of network security.
- FP Recall: The BaysFusCNN model achieves a lower FP Recall rate (0.12%) compared to the BaysCNN model (0.18%), indicating a reduced rate of false positives and enhanced accuracy in identifying genuine threats.
- F1 Score: The BaysFusCNN model achieves an F1 Score of 98.56%, outperforming the BaysCNN model’s F1 Score of 97.66%. This balanced metric highlights the fusion model’s improved overall performance.
5.4. Comparing BaysFusCNN and Other Similar Models
6. Conclusions
7. Future Work
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Conflicts of Interest
References
- Mell, P.; Grance, T. The NIST Definition of Cloud Computing; NIST Special Publication 800-145; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2011. [Google Scholar]
- Ristenpart, T.; Tromer, E.; Shacham, H.; Savage, S. Hey, you, get off of my cloud: Exploring information leakage in third-party compute clouds. In Proceedings of the 16th ACM Conference on Computer and Communications Security (CCS ‘09), Chicago, IL, USA, 9–13 November 2009. [Google Scholar]
- Mirkovic, J.; Reiher, P. A taxonomy of DDoS attack and DDoS defense mechanisms. ACM SIGCOMM Comput. Commun. Rev. 2004, 34, 39–53. [Google Scholar] [CrossRef]
- Sharafaldin, I.; Lashkari, A.H.; Ghorbani, A.A. Towards a reliable intrusion detection benchmark dataset. Future Gener. Comput. Syst. 2019, 94, 351–362. [Google Scholar] [CrossRef]
- Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press Cambridge: Cambridge, MA, USA, 2016; Volume 1. [Google Scholar]
- Zhang, X.; Wang, Q.; Hu, Y.; Liu, B. DDoS detection in the big data era: A review. IEEE Access 2018, 6, 20666–20677. [Google Scholar]
- Smith, A.; Smith, B. Foundations of Cloud Computing; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
- Johnson, B.; White, T. Cloud Security and Privacy: An Enterprise Perspective on Risks and Compliance; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2018. [Google Scholar]
- Amine, M.; Kishk, M.A.; Alouini, M.-S. Artificial Intelligence for UAV-Enabled Wireless Networks: A Survey. IEEE Open J. Commun. Soc. 2021, 2, 1015–1040. [Google Scholar] [CrossRef]
- Ahmed, M.; Mahmood, A.N.; Hu, J. A survey of network anomaly detection techniques. J. Netw. Comput. Appl. 2016, 60, 19–31. [Google Scholar] [CrossRef]
- Bouzida, Y.; Cuppens, F. Neural networks vs. decision trees for intrusion detection. In Proceedings of the IEEE/IST workshop on monitoring, attack detection and mitigation (MonAM), Tuebingen, Germany, 28–29 September 2006. [Google Scholar]
- Sabeel, U.; Heydari, S.S.; Mohanka, H.; Bendhaou, Y.; Elgazzar, K.; El-Khatib, K. Evaluation of deep learning in detecting unknown network attacks. In Proceedings of the 2019 International Conference on Smart Applications, Communications and Networking, Sharm El Sheik, Egypt, 17–19 December 2019. [Google Scholar]
- Amma, N.G.B.; Subramanian, S. VCDeepFL: Vector Convolutional Deep Feature Learning Approach for Identification of Known and Unknown Denial of Service Attacks. In Proceedings of the TENCON 2018–2018 IEEE Region 10 Conference, Jeju, Korea, 28–31 October 2018; p. 0640. [Google Scholar] [CrossRef]
- Li, Y.; Lu, Y. LSTM-BA: DDoS detection approach combining LSTM and bayes. In Proceedings of the 2019 7th International Conference on Advanced Cloud and Big Data, CBD 2019, IEEE, Suzhou, China, 21–22 September 2019; pp. 180–185. [Google Scholar]
- Gaikwad, D.P.; Thool, R.C. Intrusion detection system using bagging with partial decision treebase classifier. Procedia Comput. Sci. 2015, 49, 92–98. [Google Scholar] [CrossRef]
- Roopak, M.; Tian, G.Y.; Chambers, J. An intrusion detection system against DDoS attacks in IoT networks. In Proceedings of the 2020 10th Annual Computing and Communication Workshop and Conference, CCWC 2020, IEEE, Las Vegas, NV, USA, 6–8 January 2020; pp. 562–567. [Google Scholar]
- Marir, N.; Wang, H.; Feng, G.; Li, B.; Jia, M. Distributed Abnormal Behavior Detection Approach Based on Deep Belief Network and Ensemble SVM Using Spark. IEEE Access 2018, 6, 59657–59671. [Google Scholar] [CrossRef]
- Alotaibi, Y.; Deepa, R.; Shankar, K.; Rajendran, S. Inverse chi-square-based flamingo search optimization with machine learning-based security solution for Internet of Things edge devices. AIMS Math. 2024, 9, 22–37. [Google Scholar] [CrossRef]
- Sharafaldin, I.; Lashkari, A.H.; Hakak, S.; Ghorbani, A.A. Developing Realistic Distributed Denial of Service (DDoS) Attack Dataset and Taxonomy. In Proceedings of the IEEE 53rd International Carnahan Conference on Security Technology, Chennai, India, 1–3 October 2019. [Google Scholar]
- DDoS Evaluation Dataset (CIC-DDoS2019). Available online: https://www.unb.ca/cic/datasets/ddos-2019.html (accessed on 3 August 2021).
- Hauy, P. ddos_feature_selection [Computer Software]. GitHub. 2021. Available online: https://github.com/pedrohauy/ddos_feature_selection (accessed on 3 December 2023).
- Hauy, P.; Silva, E.; Almeida, J.M.; Pras, A. A feature selection approach for DDoS attacks classification in cloud computing environments. In Proceedings of the 2018 International Symposium on Networks, Computers and Communications (ISNCC), IEEE, Rome, Italy, 19–21 June 2018; pp. 1–6. [Google Scholar] [CrossRef]
- Alshammari, R.; Almutairi, A.; Alenezi, A.; Alanazi, H.; Hu, J. CICDDoS2019: A Diverse Dataset for Detection of DDoS Attacks in Cloud Computing Environments. IEEE Access 2020, 8, 129090–129107. [Google Scholar] [CrossRef]
- Pielou, E.C. The Measurement of Diversity in Different Types of Biological Collections. J. Theor. Biol. 1966, 13, 131–144. [Google Scholar] [CrossRef]
- Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
- Gal, Y.; Ghahramani, Z. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In Proceedings of the 33rd International Conference on International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; Volume 48. [Google Scholar]
- Neal, R.M. Bayesian Learning for Neural Networks; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 118. [Google Scholar]
- Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
- MathWorks. Bayesian Fully Connected Layer. Available online: https://www.mathworks.com/help/deeplearning/ug/train-bayesian-neural-network.html (accessed on 24 December 2023).
- Blundell, C.; Cornebise, J.; Kavukcuoglu, K.; Wierstra, D. Weight Uncertainty in Neural Networks. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 6–11 July 2015; Volume 37. [Google Scholar]
Ref. | Year | Description | Dataset | Strengths | Limitations |
---|---|---|---|---|---|
[11] | 2006 | Utilized DT and RNN for intrusion detection on the KDD dataset. | KDD | 94% accuracy in attack detection. | Struggled with detecting 15% of new attack patterns. |
[12] | 2019 | Developed DNN and LSTM models for predicting DoS and DDoS attacks using CICIDS2017. | CICIDS2017 | 98.72% accuracy for DNN and 96.15% for LSTM. | Limited real-time ca- abilities. |
[13] | 2019 | Introduced VCDeepFL technique for identifying DDoS attacks using an outdated dataset. | NSL-KDD | 99.3% accuracy for normal traffic. | Encountered challenges in identifying new and unknown attacks. |
[14] | 2019 | Developed LSTM-BA model, combining LSTM and Bayes techniques for detecting DDoS attacks. | ISCX2012 | Accuracy = 98.15%, Precision = 98.42%, Recall = 97.6%. | Slower attack detection speed. |
[15] | 2015 | Employed Bagging Ensemble framework with Partial Decision Trees for intrusion detection. | KDD Cup ‘99 | 99.71% classification accuracy. | Extended model construction times. |
[16] | 2020 | Introduced a multi-objective optimization approach, combining CNN and LSTM for intrusion detection. | CICIDS2017 | 99.03% accuracy and 99.36% F1 Score. | Lacked information about attack characteristics. |
[17] | 2018 | Proposed hybrid model with DBN and SVM for distributed network intrusion detection. | KDD Cup ‘99, NSL-KDD, UNSW-NB15, CICIDS2017 | High performance in finding suspicious activities. | Increased complexity in model construction. |
[18] | 2024 | Inverse chi square-based flamingo search optimization with machine learning-based security solution for IoT edge devices. | anomaly database | Achieves a commendable 98.25% accuracy in threat recognition for IoT edge devices. Provides an automated security solution, addressing IoT edge device security concerns. | Dataset information lacks details |
Class | Samples |
---|---|
TFTP | 20,082,580 |
DrDoS_SNMP | 5,159,870 |
DrDoS_DNS | 5,071,011 |
DrDoS_MSSQL | 4,522,492 |
DrDoS_NetBIOS | 4,093,279 |
DrDoS_UDP | 3,134,645 |
DrDoS_SSDP | 2,610,611 |
DrDoS_LDAP | 2,179,930 |
Syn | 1,582,289 |
DrDoS_NTP | 1,202,642 |
UDP-lag | 366,461 |
WebDDoS | 439 |
Portmap | 186,960 |
BENIGN | 56,863 |
Class | Samples |
---|---|
TFTP | 44,000 |
DrDoS_SNMP | 44,000 |
DrDoS_DNS | 88,000 |
DrDoS_MSSQL | 44,000 |
DrDoS_NetBIOS | 44,000 |
DrDoS_UDP | 44,000 |
DrDoS_SSDP | 44,000 |
DrDoS_LDAP | 44,000 |
Syn | 44,000 |
DrDoS_NTP | 44,000 |
UDP-lag | 44,000 |
BENIGN | 38,129 |
Component | Specifications |
---|---|
Processor | 2.3 GHz 8-core 9th-generation Intel Core i9 processor with Turbo Boost up to 4.8 GHz |
RAM | 16 GB of DDR4 memory |
GPU | AMD Radeon Pro 5500M with 4 GB of GDDR6 memory |
Storage | 1 TB SSD storage |
Display | 16-inch Retina display with a resolution of 3072 by 1920 pixels |
Ports | Four Thunderbolt 3 (USB-C) ports, a headphone jack, and an SDXC card slot |
Option | Meaning |
---|---|
“Adam” | Optimizer used during training. |
gradientdecayfactor = 0.9000 | Decay factor used for the learning rate. |
squaredgradientdecayfactor = 0.9900 | Decay factor for the second moment estimate used in Adam optimizer. |
Epsilon = 5 × 104 | A small value is added to the denominator to avoid division by zero. |
initiallearnrate = 0.001 | The learning rate used by the optimizer at the start of training. |
learnrateschedule = ‘none’ | The learning rate is not adjusted during training. |
learnratedropfactor = 0.1000 | The learning rate is reduced by 10 |
L2Regularization = 1.0000 × 104 | Strength of the L2 regularization applied to the model. |
gradientthresholdmethod = ‘l2 norm’ | The maximum l2norm of the gradient will be used as the threshold. |
gradientthreshold = Inf | The gradient threshold is set to infinity, meaning there is no threshold. |
maxepochs = 250 | The maximum number of epochs the model can train for. |
minibatchsize = 10,000 | The number of samples used in each batch during training. |
Verbose = 1 | Some output information is displayed in the console during training. |
verbosefrequency = 50 | The output information is displayed after every 50 epochs. |
validationdata = Vdata, vpredictors | The validation dataset is held in a cell array Vdata, vpredictors. |
validationfrequency = 20 | Validation metrics are evaluated every 20 epochs. |
validationpatience = Inf | Training will only stop when the validation loss does not improve any further. |
Shuffle = ‘everyepoch’ | Input batches are shuffled every epoch. |
checkpointpath = “ “ | Training checkpoints are not saved. |
executionenvironment = ‘auto’ | The software selects the best available hardware device. |
sequencelength = ‘longest’ | The input sequences are padded to their longest length. |
sequencepaddingvalue = 0 | The padding value used for the input sequences is zero. |
sequencepaddingdirection = ‘right’ | The padding is executed on the right end of the sequences. |
dispatchinbackground = 0 | The training job is not dispatched in the background. |
resetinputnormalization = 1 | Input normalization is reset before starting training. |
batchnormalizationstatistics = ‘population’ | The full dataset is used when computing batch normalization statistics. |
outputnetwork = ‘lastiteration’ | The trained network is output after the final epoch of training. |
Class | Accuracy | TP Recall | PP Precision | AN Specificity | FP Recall | F1 Score |
---|---|---|---|---|---|---|
Benign | 99.66% | 97.26% | 97.35% | 99.82% | 0.18% | 97.30% |
DrDoS_DNS | 99.84% | 99.67% | 99.20% | 99.86% | 0.14% | 99.43% |
DrDoS_LDAP | 99.54% | 95.01% | 98.55% | 99.89% | 0.11% | 96.74% |
DrDoS_MSSQL | 99.73% | 98.59% | 97.62% | 99.81% | 0.19% | 98.10% |
DrDoS_NTP | 99.51% | 95.67% | 97.45% | 99.80% | 0.19% | 96.55% |
DrDoS_NetBIOS | 99.82% | 100% | 97.54% | 99.80% | 0.19% | 98.76% |
DrDoS_SNMP | 99.67% | 97.90% | 97.53% | 99.80% | 0.19% | 97.71% |
DrDoS_SSDP | 99.75% | 98.58% | 98.03% | 99.84% | 0.15% | 98.31% |
DrDoS_UDP | 99.70% | 98.82% | 97.17% | 99.78% | 0.22% | 97.99% |
Portmap | 99.43% | 95.06% | 97.09% | 99.78% | 0.22% | 96.06% |
Syn | 99.46% | 95.17% | 97.41% | 99.80% | 0.20% | 96.28% |
TFTP | 99.78% | 99.16% | 97.76% | 99.83% | 0.17% | 98.46% |
UDP-lag | 99.70% | 98.64% | 97.23% | 99.78% | 0.22% | 97.93% |
averages | 99.66% | 97.66% | 97.69% | 99.82% | 0.18% | 97.66% |
Class | True Positives TP | False Negatives FN | False Positives FP |
---|---|---|---|
Class 1 | 11,223 | 316 | 306 |
Class 2 | 26,200 | 88 | 212 |
Class 3 | 12,601 | 662 | 186 |
Class 4 | 12,986 | 186 | 317 |
Class 5 | 12,604 | 571 | 330 |
Class 6 | 13,108 | 0 | 330 |
Class 7 | 12,899 | 277 | 327 |
Class 8 | 13,065 | 188 | 262 |
Class 9 | 13,070 | 156 | 380 |
Class 10 | 12,691 | 659 | 381 |
Class 11 | 12,619 | 641 | 335 |
Class 12 | 12,931 | 109 | 296 |
Class 13 | 13,009 | 180 | 371 |
Class | Accuracy | TP Recall | PP Precision | AN Specificity | FP Recall | F1 Score |
---|---|---|---|---|---|---|
Benign | 99.87% | 99.49% | 98.42% | 99.89% | 0.11% | 98.95% |
DrDos_DNS | 99.82% | 99.61% | 99.17% | 99.86% | 0.14% | 99.39% |
DrDos_LDAP | 99.78% | 97.95% | 99.01% | 99.92% | 0.08% | 98.48% |
DrDos_MSSQL | 99.84% | 98.80% | 98.94% | 99.92% | 0.08% | 98.87% |
DrDos_NTP | 99.65% | 97.54% | 97.53% | 99.81% | 0.19% | 97.53% |
DrDos_NetBIOS | 99.80% | 98.70% | 98.61% | 99.89% | 0.11% | 98.66% |
DrDos_SNMP | 99.83% | 99.29% | 98.33% | 99.87% | 0.13% | 98.81% |
DrDos_SSDP | 99.77% | 98.60% | 98.35% | 99.87% | 0.13% | 98.48% |
DrDos_UDP | 99.76% | 97.89% | 98.76% | 99.90% | 0.10% | 98.32% |
Portmap | 99.69% | 97.30% | 98.43% | 99.88% | 0.12% | 97.86% |
Syn | 99.79% | 98.76% | 98.38% | 99.87% | 0.13% | 98.57% |
TFTP | 99.83% | 98.82% | 98.89% | 99.91% | 0.09% | 98.85% |
UDP-lag | 99.79% | 98.41% | 98.62% | 99.89% | 0.11% | 98.51% |
Averages | 99.79% | 98.55% | 98.57% | 99.88% | 0.12% | 98.56% |
Class | True Positives TP | False Negatives FN | False Positives FP |
---|---|---|---|
Class 1 | 11,308 | 58 | 181 |
Class 2 | 26,175 | 103 | 219 |
Class 3 | 12,879 | 269 | 129 |
Class 4 | 12,955 | 157 | 139 |
Class 5 | 12,823 | 324 | 325 |
Class 6 | 13,279 | 174 | 187 |
Class 7 | 13,048 | 93 | 221 |
Class 8 | 13,073 | 185 | 219 |
Class 9 | 13,063 | 282 | 164 |
Class 10 | 12,795 | 355 | 204 |
Class 11 | 12,974 | 163 | 213 |
Class 12 | 13,208 | 158 | 148 |
Class 13 | 12,929 | 209 | 181 |
Metric | BaysCNN Model (%) | BaysFusCNN Model (%) |
---|---|---|
Accuracy | 99.66% | 99.79% |
TP Recall | 97.66% | 98.55% |
PP Precision | 97.69% | 98.57% |
AN Specificity | 99.82% | 99.88% |
FP Recall | 0.18% | 0.12% |
F1 Score | 97.66% | 98.56% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
AlSaleh, I.; Al-Samawi, A.; Nissirat, L. Novel Machine Learning Approach for DDoS Cloud Detection: Bayesian-Based CNN and Data Fusion Enhancements. Sensors 2024, 24, 1418. https://doi.org/10.3390/s24051418
AlSaleh I, Al-Samawi A, Nissirat L. Novel Machine Learning Approach for DDoS Cloud Detection: Bayesian-Based CNN and Data Fusion Enhancements. Sensors. 2024; 24(5):1418. https://doi.org/10.3390/s24051418
Chicago/Turabian StyleAlSaleh, Ibtihal, Aida Al-Samawi, and Liyth Nissirat. 2024. "Novel Machine Learning Approach for DDoS Cloud Detection: Bayesian-Based CNN and Data Fusion Enhancements" Sensors 24, no. 5: 1418. https://doi.org/10.3390/s24051418