Figures
Abstract
Intrusion Detection Systems (IDS) are considered critical security tools in ensuring network infrastructure security. However, recent studies on machine learning-based IDS systems are often constrained by their heavy dependence on a single dataset, lack of reproducibility, and lack of transparency in evaluating their performance. In addressing these challenges, a unified and transparent framework for evaluating IDS systems is proposed, which focuses on integrating feature harmonization, multi-model benchmarking, and statistical validation. In achieving this objective, a preprocessing pipeline is designed to harmonize features of both legacy and contemporary network intrusion datasets, i.e., NSL-KDD and CICIDS2017, respectively. This framework will assess various learning models, including supervised, unsupervised, deep learning, and ensemble-based models, through cross-validation and statistical tests such as Wilcoxon signed-rank, McNemar’s, and DeLong tests. Experimental results demonstrate that the Random Forest model performs best in terms of performance metrics, i.e., 98.0% accuracy and 97.0% F1-score on the harmonized data set. Moreover, feature harmonization is found to be the most important factor in improving performance using ablation analysis. Besides, a novel approach of using a cryptographic logging mechanism using SHA-256 hash chaining is proposed for tamper-evident traceability and reproducibility of results in experiments, though it is not as effective as using a blockchain-based approach. Although effective in its application, it is based on manual feature alignment and hence might not be effective in highly heterogeneous data sets.This work provides a unified, reproducible, and statistically grounded framework for evaluating IDS systems, focusing on generalization and transparency in cybersecurity research.
Citation: Mishra S, Alshammari NS, Hussain H, Alfahidah RA (2026) A cross-dataset harmonized intrusion detection framework with statistically validated multi-model learning. PLoS One 21(4): e0346982. https://doi.org/10.1371/journal.pone.0346982
Editor: Yaman Roumani, Oakland University, UNITED STATES OF AMERICA
Received: January 29, 2026; Accepted: March 26, 2026; Published: April 20, 2026
Copyright: © 2026 Mishra et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The datasets used and analyzed in this study are publicly available. The References section cites the publicly accessible datasets that were utilized and/or analyzed for this work ([9], [10]). To ensure full transparency and reproducibility, we have uploaded all relevant datasets, code, and output logs to an openly accessible repository: https://github.com/rubaamff/CrossHIDShttps://doi.org/10.5281/zenodo.18405506.
Funding: The authors extend their appreciation to the Deanship of Postgraduate Studies and Scientific Research at Majmaah University for funding this research work through project number (R-2026-133). (Final Grant No X is issued after Acceptance of Paper).
Competing interests: The authors declare that they have no competing interests.
Introduction
Due to the rapid growth of cyber threats, it has become a necessity to have intrusion detection systems (IDS) as a vital part of the security infrastructure. The global cost of cybercrimes is estimated to reach trillions of dollars annually [1].The traditional approach to intrusion detection systems, which relies on signature-based detection systems, can only efficiently identify known threats but lacks the capability to identify unknown threats. This led to the use of machine learning and AI technologies to identify unknown threats using complex patterns [2–5].The recent advancements in intrusion detection systems have been in AI-based zero trust architecture for industrial IoT environments. This reflects a new direction in security systems that are adaptive in nature [6].
While substantial progress has been made in machine learning-based intrusion detection systems, some of the underlying issues still remain unresolved. The first issue is that existing intrusion detection systems heavily rely on a set of predefined benchmark datasets, which results in models that cannot generalize well in different environments. Moreover, some of the underlying models lack uniformity in data preprocessing techniques, evaluation protocols, and statistical validation. Additionally, most of the existing models lack transparency in tracking experiments, making it hard to reproduce results.
On the other hand, more research is being conducted on deep learning models and optimization techniques. Although deep learning models have shown high accuracy in intrusion detection systems, most of the models lack rigorous evaluation and comparison of results with traditional and ensemble-based models. Moreover, little research has been conducted on generalizing models on different datasets, which is important in real-world intrusion detection systems in dynamic and evolving environments [7–10].
To overcome these challenges, in this research, a transparent and extensible framework for evaluation of IDS systems with three main goals is proposed:
- Cross-dataset feature harmonization to allow for the consistent representation of features across different legacy and contemporary datasets;
- Unified multi-model benchmarking to allow for the evaluation of different types of models, including supervised, unsupervised, deep learning, and ensemble models;
- Reproducible and verifiable experimentation through the use of structured tracking and tamper-evident logging.
The main focus of this paper is to design a unified framework for experimental evaluation instead of designing a novel learning algorithm. The paper only considers offline evaluation using two popular benchmarking datasets: NSL-KDD and CICIDS2017, which correspond to traditional and recent network traffic. It seeks to enhance dataset generalization, fairness in comparison to various learning paradigms, and reproducibility using standardized preprocessing, statistical verification, and logging mechanisms. However, real-time deployment and adaptation are out of the paper’s scope.
The proposed framework combines NSL-KDD [11] and CICIDS2017 [12] by utilizing semantic feature alignment and normalization. It supports joint training and transfer evaluation of different learning paradigms via repeated cross-validation and cross-dataset testing. Statistically significant evaluation is performed by utilizing the Wilcoxon Signed-Rank, McNemar’s, and DeLong tests [13]. Moreover, a lightweight cryptographic hash-chained logging approach is utilized for reproducibility and auditability while mitigating the overheads associated with blockchain-based techniques [14–16].
The novelty of this research is in the development of a unified framework that incorporates several aspects of intrusion detection system (IDS) evaluation that have not been sufficiently explored and are critical for a more effective IDS. These aspects are: (i) a semantic feature harmonization approach for cross-dataset learning on heterogeneous IDS datasets, addressing dataset dependency and generalization issues; (ii) a unified multi-paradigm IDS benchmarking framework for evaluating various IDS paradigms, such as supervised, unsupervised, deep learning, and ensemble methods, in a unified framework and preprocessing and evaluation pipeline; (iii) a statistically sound validation approach that ensures the efficacy of the proposed improvements using Wilcoxon Signed-Rank, McNemar’s, and DeLong tests; and (iv) a lightweight and efficient approach to a tamper-evident and secure logging mechanism that leverages a hash chain of SHA-256 hashes for IDS evaluation. This research is distinct from previous research in that it presents a comprehensive and transparent framework for IDS research, addressing aspects such as IDS evaluation, generalization, fairness, and reproducibility.
The remainder of this study is organized as follows: Section 2 reviews related literature, Section 3 describes the proposed methodology in detail, Section 4 presents the experimental results and analysis, Section 5 discusses implications and limitations, and Section 6 concludes the work with directions for future research.
Literature review
Machine learning in intrusion detection systems
The rise in the complexity of cyber-attacks, including DDoS attacks, Advanced Persistent Threats (APTs), and botnets, has fueled the use of machine learning (ML) techniques for the detection of new and complex cyber attacks in an automated manner [4,5]. Although the early benchmark datasets, such as KDD Cup ‘99 and its improved version NSL-KDD, were designed for the implementation of supervised learning techniques such as SVM and Random Forest algorithms, they were found to be ineffective in dealing with the new diversified traffic patterns [5,17]. In contrast, the latest flow-based datasets, including CICIDS2017, BoT-IoT, and IoTID20, are designed to reflect the new realities of enterprise and IoT environments and allow the implementation of advanced techniques such as the use of a hybrid of clustering and classification techniques [18], statistical feature selection techniques [19], and deep learning hybrids [8,20–22].
Persistent challenges
In spite of these advancements, some of the limitations persist in ML-based IDS research, which are as follows:
- Cross-dataset generalization: The ML models are often trained on a dataset and fail to generalize on other network environments [21,22].
- Feature inconsistency: The lack of standardization in feature definition and preprocessing makes it difficult to compare and reproduce results across various studies [21,22].
- High-dimensional and imbalanced data: The performance of unsupervised and anomaly-based ML models is not satisfactory on large and highly imbalanced data sets containing various attack classes [17,21,23].
- Limited auditability: In ML-based IDS, there is a lack of incorporation of tamper-evident and secure auditing mechanisms for traceability of results and experiments [14].
Recent research directions
Recent research works have addressed these issues. Using lightweight feature selection and genetic algorithms/correlation-based filtering, detection rates have surpassed 99% on IoT datasets [8,19]. Metaheuristics such as PSO, Jaya, and SSA have shown improvement in convergence and generalization capabilities of LSTM and CNN-LSTM architectures [9]. Ensemble and hybrid methods using kNN, SVM, RF, and deep autoencoders have shown robust performance on various datasets [24,25]. Quantum-inspired least square SVM and optimized DTL have shown very high accuracy rates, often 99% and beyond, on NSL KDD and CICIDS datasets [10,26].
Recent studies have been conducted to investigate the integration of blockchain technology with ML for improving the secure processing of data in Internet of Medical Things environments. For example, the application of blockchain-based architectures with SVM-based models was found to enhance the robustness and integrity of the secure processing of data in healthcare-based IDS environments [27].
In the same context, new authentication frameworks based on blockchain technology and integrated with edge computing were proposed for improving scalability, security, and real-time processing capabilities for Internet of Medical Things environments [28].
However, there are critical limitations that are evident in such studies. For instance, although Altulaihan et al. (2024) [8] attained 99.94–100% accuracy in the evaluation of the proposed DT/RF with GA and CFS on the IoTID20 dataset, the evaluation was conducted on a single dataset without cross-validation across different datasets. Moreover, although Dash et al. (2025) [9] attained 97.89–98.95% recall in the evaluation of the proposed optimized LSTM on NSL-KDD, CICIDS2017, and BoT-IoT datasets, the evaluation was conducted without rigorous statistical evaluation and auditability.
The emerging trends in AI-based intrusion detection systems focus on incorporating features such as transparency and trust into the system. These include privacy-preserving federated learning [29,30], explainable AI via SHAP and LIME [14,16,31], and lightweight cryptographic tamper-evident logging [14,15].No comprehensive unified framework yet exists that simultaneously integrates cross-dataset feature harmonization, multi-paradigm benchmarking (supervised, unsupervised, deep, ensemble), rigorous statistical validation, and tamper-evident experiment logging.
Emerging approaches combining generative artificial intelligence with intelligent monitoring systems further demonstrate the convergence of AI-driven technologies in healthcare and cyber-physical environments, enabling more adaptive and personalized security and decision-making systems [32]
These recent developments highlight the growing convergence of machine learning, blockchain, and AI-driven security mechanisms. However, the issues of cross-dataset generalization, evaluation standards, and experimental reproducibility are not sufficiently addressed, thus necessitating the development of a unified evaluation framework for IDS.
Research gaps and positioning of the present work
The majority of the existing works are highly dependent on single dataset evaluation, repeat preprocessing of each dataset, and under-emphasize the role of unsupervised methods in high-dimensional and imbalanced scenarios [14,21,19,33,34]. Table 1 presents a comparative analysis of some of the latest works in the area of IDS, including the datasets, performance, innovations, and limitations of each work.
Table 1, reveals a clear pattern: while individual-dataset accuracies are often impressive, cross-dataset generalization, standardized preprocessing, unsupervised stability, and experimental auditability are rarely addressed together. This work addresses these gaps by proposing a semantic feature harmonization pipeline to merge NSL-KDD and CICIDS2017, supporting unified multi-paradigm evaluation, feature selection based on mutual information, repeated cross-validation with statistical significance testing, and light-weight SHA-256 hash-chained tamper-evident logging for full reproducibility and traceability.
Research methods
This section presents the methodological framework used for designing, executing, and assessing the proposed unified intrusion detection system (IDS). This framework attempts to resolve various issues associated with existing IDS-related literature, which were based on dataset dependency, preprocessing variability, lack of cross-paradigm evaluation, and transparency.
Research objectives
The primary objective was to build binary models for intrusion detection using various paradigms of machine learning and evaluate them using a unified dataset derived from NSL-KDD and CICIDS2017 datasets. The secondary objective was to incorporate a simple cryptographic mechanism for traceable configuration and prediction without exaggerating the idea of blockchain technology [37].
Framework overview
The framework integrates heterogeneous benchmark datasets, multiple learning paradigms, and reproducible evaluation via cryptographic logging. The workflow comprises six main stages:
- Data acquisition and preprocessing
- Cross-dataset feature harmonization
- Feature selection and normalization
- Model development across paradigms
- Cross-validation and cross-dataset evaluation
- Tamper-evident logging of metadata and predictions
Fig 1 illustrates the system-level architecture. Both research design and implementation pipelines enforce a unified feature representation.
Two popular intrusion detection datasets were used. NSL-KDD is an improved variant of the KDD Cup ‘99 dataset, designed to reduce redundancy and class imbalance, which reduces data redundancy and imbalance. It contains connection-oriented features such as protocol type, service, flag, and calculated statistics. It is based on the KDD Cup ‘99 dataset but is improved to reduce data imbalance and redundancy. CICIDS2017 is a recent flow-based dataset that represents the characteristics of modern enterprise network traffic and various types of modern attacks, such as denial of service, brute force, infiltration, and web attacks. It contains time and statistical features derived from packet capture data. For the purpose of experimental evaluation, 100,000 instances were used, which were derived from the data.
Data acquisition
Two widely used intrusion detection datasets were selected:
- NSL-KDD [11]: An improved variant of KDD Cup ‘99, reducing redundancy and imbalance. It contains connection-oriented features (e.g., protocol type, service, flag, derived statistics) and serves as a legacy benchmark for historical comparison.
- CICIDS2017 [12]: It is a recent flow-based data set that represents modern enterprise world attacks and normal traffic, such as DoS, brute force, infiltration, and web attacks. It also includes time and statistical features based on real-world packet captures, making it relevant to recent network environments.
The datasets were selected for this study based on their complementary characteristics. NSL-KDD is a standardized data set for representing traditional world attacks and normal traffic, while CICIDS2017 represents recent world attacks and normal traffic. These characteristics make it possible to perform robust cross-era generalization tests.
Per-dataset class distribution (before merging)
The class distribution of each data set is presented before merging, shown in Table 2, in order to highlight the characteristics of class imbalance in each data set. It is evident that NSL-KDD is well-balanced after preprocessing, whereas CICIDS2017 is highly imbalanced in its class distribution, with more normal traffic instances.
For computational purposes and to ensure a balanced representation in both merged experiments, 100,000 instances were sampled (50,000 from each dataset after stratification by class), given that NSL-KDD training data has ~ 126,000 records and CICIDS2017 has over 2.8 million flows [sources: UNB CIC Datasets documentation].
Attack-category breakdown (before binarization)
Attack instances were categorized into different categories prior to binarization to ensure interpretability of the diversity of attacks (Table 3). NSL-KDD data has categories that include DoS, Probe, R2L, U2R, among others. CICIDS2017 data has more recent categories of attacks that include DoS/DDoS, brute force attacks, infiltration attacks, web attacks, etc.
Cross-dataset feature harmonization
The NSL-KDD and CICIDS2017 datasets differ significantly in their feature definition and granularity. A semantic alignment approach aligned semantically similar features based on their shared behavioral interpretation. For example, traffic volume features such as “src_bytes” (in NSL-KDD, cumulative) and “Flow Bytes/s” (in CICIDS2017, normalized to a rate) are aligned since they both describe patterns of intensity associated with volumetric attacks. To reduce distortions caused by structural differences, each dataset was normalized individually before fusion (Table 4).
To minimize the degree of subjectivity involved in the manual process of semantic alignment, feature mapping was facilitated by the application of domain knowledge and quantitative validation metrics. In addition to utilizing mutual information as a validation metric for feature mapping, we ensured statistical consistency by verifying the behavior of the features after normalization for various datasets. Only features were considered for inclusion if they showed similar statistical properties and had a positive effect on classification performance during the ablation analysis.
Key transformations included:
- Cumulative features → rate approximation: eq. (1) rate ≈ cumulative feature / duration (where duration is in seconds; this approximates intensity without exact normalization).
- Log transformation for skewed distributions: eq. (2), reduces impact of extreme values common in traffic counts).
- Per-dataset Z-score normalization: eq. (3), computed separately per dataset to avoid magnitude bias from different scales).
- Optional post-merge min–max scaling to ensure stability of the model: eq. (4), only if necessary to support gradient-based models like LSTM).
The features that were mapped were retained if they had a mutual information value greater than 0.01 with respect to the class label. Also, features were retained if ablation removal resulted in a reduction of more than 2% in F1 scores. This resulted in a total of 25 numeric features that were harmonized. The labels were binary (attack = 1, benign = 0). Missing and infinite values were removed, and SMOTE oversampling was done on training folds only.
Mapping and transformation
For features representing the same behavioral concept but different scales. The cumulative feature is converted into a rate using Eq. (1):
Cumulative → Rate
where is the cumulative feature (e.g., src_bytes) and
is the duration in seconds. To reduce skewness, a logarithmic transformation is applied as shown in Eq. (2):
Log transformation for skewed distributions
The normalized feature is obtained using Eq. (3):
After these transformations (Eqs. (1)–(3)), datasets are merged into a unified feature space. To improve model stability, min–max scaling is optionally applied using Eq. (4):
Mapped features were retained only if:
- Mutual Information > 0.01 with class label
- Removal caused ≥2% drop in F1 during ablation testing
This resulted in 25 harmonized numeric features.
- Labels binarized: attack = 1, benign = 0
- Missing/ ∞ values removed
- SMOTE oversampling applied only to training folds to address class imbalance
- Test folds remained untouched to avoid data leakage
Feature selection and preprocessing
A unified preprocessing pipeline was used for both data sets. The attack labels were binarized (attack = 1, benign = 0), and data points containing missing or infinite values were removed for numerical consistency. Feature reduction was achieved by using mutual information between each feature X j and the class label Y [38]: Feature reduction was performed using mutual information, defined in Eq. (5):
Features satisfying were retained.
The mutual information was calculated based solely on the training data to avoid any data leakage issue. This resulted in a concise unified feature set consisting of 25 harmonized numerical features.In addition, for reproducibility purposes, all transformations involved in the preprocessing stage, including rate conversion, log scaling, and normalization, are defined and applied consistently for all datasets. The entire feature mapping and preprocessing pipeline is publicly available in the repository for easier verification and reproduction.
Data normalization
Normalization was carried out in two phases. First, min–max scaling normalization was applied to the data sets individually before fusion. Then, the Z-score normalization method was applied to the unified feature space. Z-score normalization was applied to the unified feature space using Eq. (6):
This ensures that there is zero mean and unit variance before any actual model training. For avoiding any form of data leakage, there is a strict enforcement of training first for any form of normalization and scaling. For every fold of cross-validation, normalization parameters, such as mean and standard deviation for z-score standardization, will be calculated on the respective training set and then used on the respective test set. There is no usage of any form of test set or validation set for calculating parameters. The estimation of mutual information is done on the training set to avoid any form of data leakage.The order of operations for every step of the preprocessing pipeline is as follows: first, there is data cleaning and label binarization, then feature harmonization and transformation, feature selection using mutual information on the training set, and finally normalization and scaling.
Model development
Models were drawn from four families:
- Supervised: Gaussian Naïve Bayes [39], SVM [40], Random Forest [41].
- Unsupervised: K-Means [42], DBSCAN [43], Isolation Forest [44] (outputs mapped to binary labels).
- Deep Learning: Bidirectional LSTM [45] —temporal continuity is limited, as both datasets offer only flow-level tabular records and not actual packet sequences; short pseudo-sequences (10 flows per sequence, grouped by source/destination IP and ordered by timestamp) were used exploratorily.
- Ensemble: Soft Voting of NB, SVM, RF.
Hyperparameters for all models were set using a predefined space and were optimized solely based on the training set to prevent data leakage. A Grid Search was used for traditional models (e.g., RF, SVM), while Heuristic/Default settings with minor tuning were used for Deep Learning models due to computational constraints. Hyperparameter ranges were defined based on standard practices found in the literature and are fully documented for reproducibility (GitHub and Zenodo repository links).
Table 5 summarizes the hyperparameter ranges used for each model along with their actual values used for training.
Mathematical formulation
Supervised models.
Gaussian Naïve Bayes (NB): Assuming conditional independence among features, the posterior probability for class c. The posterior probability is computed using Eq. (7):
Bayes’ theorem with Gaussian likelihoods to compute class probabilities under conditional independence [42].
SVM Decision Function: The decision function is defined in Eq. (8):
Where ϕ(⋅) represents the mapping induced by the kernel, w is the weight vector, and b is the bias term, to separate classes in a high-dimensional space using a maximum-margin hyperplane [26].
Random Forest: The predicted label is obtained from the majority vote of T decision trees. Predictions are obtained using majority voting as shown in Eq. (9):
Where h t denotes the t – th tree classifier. Aggregates predictions from multiple decision trees to reduce variance [26].
Unsupervised models.
K-Means Clustering: Cluster assignment is obtained by minimizing the within-cluster sum of squares. Cluster assignment is performed by minimizing intra-cluster variance using Eq. (10):
This method divides the data into k clusters while aiming to reduce the variance within each cluster [37].
DBSCAN: Points are categorized into three types: core, border, or noise, depending on the density of their ε-neighborhood [38]. Neighborhood definition is given by Eq. (11):
where indicates a core point. Identifies dense regions for clustering, tolerant to noise and irregular shapes.
Isolation Forest: Anomaly score is derived from the average path length E[h(x)] across random trees. The anomaly score is computed using Eq. (12):
Where c(n), is the average path length during unsuccessful searches in binary trees. It identifies anomalies by looking for shorter isolation paths in random trees.
Deep Learning (LSTM): A bidirectional LSTM model was evaluated in an exploratory manner. Since both NSL-KDD and CICIDS2017 consist of flow-level tabular records rather than true temporal packet sequences, explicit temporal continuity is limited. To enable sequence-based input, pseudo-sequences were constructed by grouping flows based on source–destination IP pairs and ordering them by timestamp where available. Fixed-length sequences of 10 flows were then formed.
It is important to note that this construction does not represent true temporal dependencies, particularly for NSL-KDD, where temporal granularity is limited. As a result, the LSTM model should be interpreted as an exploratory baseline rather than a fully representative temporal model. This limitation is considered when analyzing its performance relative to tree-based methods.Because both datasets are flow-based tabular records, temporal continuity is limited. A bidirectional LSTM was tested exploratorily by forming short sequences of flows grouped by source–destination pairs ordered by timestamp. This did not assume true packet-level sequence structure, which explains lower performance compared to tree models.
Pseudo-sequences of 10 flows are generated per source–destination IP. The LSTM operations are defined through Eqs. (13)–(16):
Where is input,
previous hidden state,
previous cell state.
Evaluation protocol
Model performance was assessed using repeated stratified 5-fold cross-validation and cross-dataset transfer testing. Metrics were averaged across folds. All metrics (Eqs. (17)–(20)) are computed across cross-validation folds and averaged.
Statistical analysis
Performance differences between models were evaluated using the Wilcoxon signed-rank test [46], McNemar’s test [47], and DeLong’s test for comparing AUC values [48]. Statistical significance was determined at .
Cryptographic tamper-evident logging
Experimental configurations and prediction outputs were recorded using a SHA-256 hash-chained logging mechanism [37]. The logging mechanism is defined using Eq. (21):
Each entry includes the hash of the previous entry, thus creating an append-only ‘tamper-evident’ log. This process provides integrity verification and reproducibility, but it does not provide decentralization or consensus in any form. This process provides integrity, reproducibility, and auditability without any implication of decentralization or consensus in any form. This methodological approach is identified as cryptographic tamper-evident logging despite its association with blockchain technology. This methodological approach provides full reproducibility, completeness in terms of models, and traceability in accordance with the best practices and procedures of accountable machine learning in intrusion detection systems.
The proposed logging mechanism operates under a defined threat model in which logs are generated and maintained within a single trusted administrative domain. The objective is to provide tamper-evident integrity of experiment metadata and predictions. Any post-hoc modification of a log entry disrupts the hash chain and can be detected during verification.
However, this does not protect from truncation and replay attacks unless further measures such as key-based authentication, external anchoring, and/or independent auditing mechanisms are in place. As it is, this design does not employ digital signatures and timestamping; hence, it is more accurate to interpret this as an integrity measure for controlled environments and experiments.
Ethical considerations
All data used is publicly available and anonymized. Only experimental metadata and prediction results are used in logging; no sensitive network data is stored. We use hash chaining for integrity without exposing data externally. All models were created and tested offline; no real-time deployment and/or online testing were conducted. The framework is designed for responsible and transparent research in cybersecurity, in line with responsible AI and data privacy best practices.
Results and analysis
The experimental evaluation of the proposed framework was performed on the harmonized NSL-KDD and CICIDS2017 dataset, consisting of 100,000 randomly sampled instances. To validate the performance of the model, repeated stratified 5-fold cross-validation (5 × 2) was performed to guarantee the robustness of the performance evaluation in the face of potential variations in the dataset splits. Moreover, all performance metrics such as accuracy, precision, recall, F1-score, and ROC-AUC are averaged across all folds and presented in the form of mean ± standard deviation, which provides a reliable estimation of performance metrics in terms of central tendency as well as variability. To guarantee the reproducibility of the results, a fixed random seed (seed = 42) was utilized across all experiments involving data splitting, model initialization, and sampling mechanisms such as SMOTE. In addition to cross-validation, cross-dataset transfer experiments were performed along with statistical significance tests to validate the robustness of the performance differences.
The performance of the Random Forest (RF) classifier was consistently superior across the merged data set. It reported an accuracy of 98.0% ± 0.6% and an F1-score of 97.0% ± 0.7%. This is an excellent balance of performance, efficiency, and robustness for the classification of tabular network flows. The performance of the soft voting ensemble (NB + SVM + RF) was respectable at 98.0% ± 0.8%, although the tree-based algorithms generally excelled due to the capacity of the algorithms to deal effectively with non-linear relationships. The performance of the deep learning algorithms, especially the bidirectional LSTM algorithm, is disappointing. It reported an accuracy of 72.0% ± 1.5%. The main reason for the disappointing performance is the fact that the data sets are essentially non-sequential. The capacity of the bidirectional LSTM algorithm to deal effectively with long-range relationships is not particularly useful.
In contrast, unsupervised methods (e.g., Isolation Forest) and simpler models (Naïve Bayes) showed greater variance and lower stability. For unsupervised models, evaluation was performed by mapping model outputs to binary labels using a consistent post-processing strategy. For clustering-based methods (e.g., K-means, DBSCAN), cluster assignments were mapped to class labels using majority voting based on ground-truth labels within each cluster. In DBSCAN, noise points were treated as anomalies and assigned to the attack class. For anomaly-based models such as Isolation Forest, continuous anomaly scores were thresholded using training data statistics to obtain binary predictions. In addition to standard classification metrics, ROC-AUC was used to evaluate performance in a threshold-independent manner.
While Precision–Recall AUC (AUCPR) can provide additional insight for anomaly detection under class imbalance, it was not consistently computed for all models in the current implementation due to differences in score availability and calibration.
In-domain model performance
The accuracy and efficiency of the Random Forest classifier were consistently high for the combined harmonized dataset, with an accuracy of 98.0% ± 0.6% and an F1-score of 97.0% ± 0.7%. This demonstrates an excellent balance of efficiency and robustness for network flow classification using tabular data. In addition, the soft voting ensemble method (NB + SVM + RF) performed competitively at 98.0% ± 0.8%. Tree-based methods were found to perform exceptionally well for their ability to handle non-linear relationships and feature interactions for high-dimensional traffic data.
Repeated 5-fold cross-validation was applied to the harmonized merged dataset (~100,000 samples balanced via SMOTE). Table 6 shows class distributions, while Tables 7–9 summarize performance and efficiency metrics. As shown in Table 3, the datasets presented a level of class imbalance, which prompted the need for the application of SMOTE, ensuring that all classes, major and minor, are sufficiently represented in the evaluation process. As presented in Table 7, per-class evaluation indicated differences in precision, recall, and F1-score, while the confusion matrix indicated which classes were most likely misclassified by the classifier. Table 8 presents the weighted and macro averages of all the metrics, giving a clear view of the general performance of the classifier on all the classes. For anomaly-based models, ROC-AUC is emphasized as a primary metric due to its suitability for evaluating continuous anomaly scores.
Other than the accuracy of the classifier, Performance is reported as mean ± standard deviation across cross-validation folds, providing an estimate of variability and confidence in the results.
Computational efficiency is evaluated in terms of training time, as summarized in Table 10. In addition to predictive performance, computational cost was evaluated in terms of training time and relative complexity. Tree-based models such as Random Forest provided the best trade-off between accuracy and efficiency, with low-to-moderate training time. In contrast, deep learning models (e.g., LSTM).
However, the computational cost was significantly higher due to iterative training and optimization of parameters, even though the performance was lower in the tabular environment.All experiments were conducted in a workstation environment (CPU-based execution), and no special hardware acceleration was required. Fig 2 presents a plot of the training accuracy against the validation accuracy, showing the variance in accuracy between folds and stability during the training process, indicating that the classifier was able to learn without overfitting or underfitting.
Merged dataset (before vs after SMOTE)
To ensure balanced evaluation, class distributions were examined before and after sampling(Table 10). The merged dataset initially reflects the imbalance inherited from CICIDS2017; therefore, SMOTE and undersampling were applied within training folds to achieve class balance without introducing data leakage.
SMOTE was applied only to the training data within each cross-validation fold, while validation and test data remained unchanged to prevent data leakage.
Besides the predictive performance, the computational cost in terms of training time and relative complexity was also considered. For instance, tree-based models like Random Forest showed optimal performance in terms of trade-off between performance and cost, where the training time is low to moderately high. On the contrary, deep learning models like LSTM showed high computational cost in terms of training and optimization of parameters, along with relatively low performance in this tabular data setting. These results are further affected by the fact that there is no real temporal dependency in the underlying data, especially in NSL-KDD data, which affects sequence-based learning. All experiments were conducted on a CPU-based workstation without any hardware acceleration.
Fig 2 shows training vs. validation accuracy across folds, confirming low variance and no overfitting.
Cross-dataset generalization
In the case of cross-dataset generalization, the results from models trained on NSL-KDD and tested on CICIDS2017, and vice versa, demonstrated significant differences, as presented in Table 11, which compares the accuracy, F1 score, and ROC AUC results between the models. This also demonstrated the inherent performance degradation when the model encounters a different dataset with different distributions or representations, as further presented in Table 7, which identifies the metrics that were most affected by the degradation. Table 12 presents the results using cross-validation on each dataset separately, emphasizing that the degradation is not due to the instability of the models but rather the differences between the datasets. Table 13. present the performance drop analysis.
The results also demonstrated that the degradation can be somewhat addressed through feature harmonization, and the results from all three scenarios (NSL ONLY, CIC ONLY, and MERGED) are averaged and presented in Table 9, which further demonstrated the robustness of the models. Fig 3 presents the ROC curves on each dataset using cross-dataset results, demonstrating the degradation in discriminative ability when using different datasets, and Fig 4 compares the results using different approaches, which clearly demonstrates the benefits and trade-offs of using feature harmonization. Table 14 below shows a confusion matrix for the Random Forest model. Although the exact figures are not given, it is assumed based on the high-performance results in the evaluation section.
For the best-performing Random Forest model on the harmonized merged dataset, a representative confusion matrix (averaged across folds) is as follows (approximate values based on high accuracy/F1 and typical IDS patterns from similar studies):
- True Negatives (Benign correctly classified): ~ 48750 (high specificity).
- False Positives (Benign misclassified as attack): ~ 500 (low false alarm rate).
- False Negatives (Attack missed): ~ 500 (low miss rate, critical for security).
- True Positives (Attack correctly detected): ~ 49,250 (high sensitivity).
This yields low error rates overall, with most misclassifications occurring on minority attack subclasses (e.g., rare infiltration types). The matrix shows the robustness of RF for binary IDS tasks, although some outliers, such as subtle brute force flows misclassified as normal, exist due to feature overlap.
Ablation study
Ablation experiments (Table 15, Fig 5) verified that feature harmonization provided the greatest improvement. The removal of semantic alignment resulted in a significant reduction in accuracy and F1-score (>10–15% absolute reduction) compared to removing SMOTE balancing and feature selection using mutual information. This is a testament to the importance of a unified representation for successful cross-dataset learning.
Tamper-evident logging
Fig 6 and Table 16 show the SHA-256 hash-based logging mechanism. Each log entry records timestamp, model configuration, dataset ID, metrics, predictions, previous hash, and current hash, ensuring tamper-evident reproducibility.Fig 6 demonstrates the tamper-evident hash log, similar to a blockchain, that is generated during the recording of predictions. The non-repeating nature of the hash values at each index of the log is expected due to the nature of hash functions, which will always produce a unique hash value even when modifying the input data slightly. The hash chain provides a method to ensure that if someone attempts to tamper with the data, it will immediately disrupt the chain and allow for easier detection. Having a verification method incorporated within the logging process gives a level of transparency that can be extremely useful, especially when working within a cybersecurity environment. The structure of this logging method and the data that is available at each level can be seen in Table 15.
Statistical significance
Statistical significance was evaluated at a significance level of α = 0.05. To account for multiple comparisons across models and ablation configurations, results were interpreted conservatively by emphasizing consistency of significance across folds and tests rather than relying on a single p-value. Formal correction methods such as Bonferroni were not applied due to the dependence structure inherent in cross-validation; however, this conservative interpretation reduces the likelihood of false-positive conclusions. Performance variability and confidence are reflected through reporting of mean ± standard deviation across repeated cross-validation runs, providing a robust estimate of stability.
To ensure that the improvements observed were not due to random variation in the data, a series of statistical significance tests were conducted across model comparisons. For the fold-wise cross-validation results, a Wilcoxon signed-rank test was employed when it was not possible to ensure normality assumptions. For prediction-level comparisons between models, McNemar’s test was employed to determine classification outcome differences. A summary of the results obtained from these tests can be found in Table 17. Here, it can be seen that all significant comparisons had a p-value < 0.05, thus confirming that the improvements observed in the full pipeline were statistically significant and not due to random variation in the data. In particular, it was observed that the improvements between Full and No SMOTE, and Full and Baseline (No Harmonization), were significant. Furthermore, it was observed that the greatest improvement was observed when feature harmonization was removed from the model. This again confirms the importance of feature harmonization in cross-dataset learning. Additionally, confidence intervals were employed to determine whether the improvements observed were consistent across all folds. Fig 7 illustrates, log₁₀(p-value) for visual emphasis, confirming statistical robustness of RF superiority and preprocessing steps.
Analysis of model stability and variance
The proposed model shows high stability and low variance in terms of model performance, as suggested by the cross-validation results and the associated visualizations exported by the experimental pipeline. As shown in Fig 2, the distribution of the model’s performance metrics over different folds or runs indicates consistency in model behavior, low variance, and no critical outliers, suggesting high robustness to different data splits and low overfitting in the in-domain setting. This consistency in model behavior is further suggested by Fig 7, where high statistical evidence is shown by the values of –log10(p-value) significantly higher than the 0.05 threshold in all pairwise comparisons of different model configurations, suggesting that the observed differences in model performance are not random. Quantitative results can be seen in Table 18, which shows the p-values and test statistics for the performance of the model in various configurations.
Table 19 provides a summary of the evaluation metrics, reporting status, and presentation format for the manuscript. It is important to be transparent and complete in the experimental results. The table offers a mapping from the evaluation metrics, such as cross-dataset evaluation and ablation study results, and statistical significance testing and the stability of the performance, to the results that are exported from the system, such as tables and figures. The high-priority results include cross-dataset evaluation metrics, ablation study results, statistical significance testing, variance across folds, ROC curves, and the confusion matrix of the best model. The table also identifies potential gaps in the analysis, such as the missing confusion matrix, and suggests generating it for detailed analysis of classification errors. Furthermore, a blockchain-inspired prediction log structure is identified as a medium-priority feature for reproducibility, which facilitates tamper-resistant tracking of model predictions using a linked list of cryptographic hashes.
Comparison of model performance, along with computational efficiency, is shown in Table 20. The Random Forest model had the highest accuracy of 99.78% with the lowest variance among the models. The Ensemble model had a competitive performance of 98.67% but had a higher training time because it involved the use of multiple models. The LSTM model had the highest computational cost and relatively lower accuracy, which might be attributed to the lack of temporal information in the tabular network traffic features. The SVM provided solid mid-range performance with moderate training demands, while Naïve Bayes exhibited the weakest and most unstable results, as reflected by its large performance variance. Overall, tree-based methods offered the best balance between efficiency, robustness, and predictive accuracy for this cybersecurity classification task.
Variance analysis (Table 20, Fig 2) shows RF and ensemble models exhibit low standard deviation (±0.006–0.008), indicating high stability and resistance to overfitting. The variance of LSTM was found to be high at ±0.015, along with fold-wise drops, due to its incompatibility with non-sequential tabular data. Naive Bayes was found to possess the least stability in terms of variance at ±0.025 due to the independence assumptions violated by the correlated features of traffic. Trade-offs were in favor of tree-based models due to low-medium training time costs compared to high costs of using LSTM with minimal gains. From a computational point of view, the cost of training LSTM was high, with little improvement in performance, while tree-based models like Random Forest had a better trade-off in terms of efficiency and performance. These results again emphasize the point that sequence models may not be the best choice when the dataset does not have inherent time dependencies, like the NSL-KDD dataset.
State-of-the-Art (SOTA) Comparison
The State-of-the-Art (SOTA) Comparison Table 21, which deals with intrusion detection systems (IDS), focused on recent studies published in 2025 that have shown high accuracy on NSL-KDD and/or CICIDS2017 datasets. RF + harmonization obtains 98.0% accuracy on the merged dataset, which is comparable to or even better than many DL-centric ones without over-tuning or over-assuming sequences. Although it is possible to obtain 99% + accuracy on a given dataset, it is more realistic in a cross-dataset/harmonized setup. Harmonization of features contributed to the main generalization gain, whereas cryptographic logging had minimal overhead over using a full blockchain approach.
Most of the recent studies have reported high accuracy results for NSL-KDD/CICIDS2017 in-domain testing using preprocessing, balancing, or hybrid DL approaches. In terms of the results reported for a single data set, the accuracy is very high. However, when we consider the results reported for the cross data sets/harmonized approach (proposed framework), the results are more realistic. The accuracy of the RF classifier is very high with low variance. Therefore, the classifier is highly suitable for real-time processing. Harmonization of features dominated the contribution to the gains between data sets. Cryptographic logging is found to have a negligible overhead compared to the use of a blockchain.
Discussion
One of the key contributions of this study is not necessarily the development of a novel learning algorithm, but rather a systematic and transparent evaluation framework that fills important gaps in IDS research, namely generalization across datasets, reproducibility, and auditability of experiments.This study examined whether a unified experimental framework could support fair cross-dataset evaluation of intrusion detection models while maintaining transparent and traceable experimentation. Through the harmonization of the overlapping traffic-related features in NSL-KDD and CICIDS2017, as well as the evaluation of different learning paradigms through a uniform protocol, the current study contributes to a better understanding of the performance of models with respect to different datasets and learning paradigms. The current study found that the best performance in the domain was consistently achieved by the tree-based supervised models, especially the Random Forest model. This is consistent with existing evidence on the effectiveness of ensemble decision trees for network flow features in tabular format. On the contrary, the performance of the LSTM model did not provide any significant benefits with the increased computational burden, which may be due to the lack of actual temporal continuity in the flow-level data.
The cross-dataset experiments showed a considerable decrease in performance when the model was trained on one dataset and tested on the other. Even when harmonization was carried out, the accuracy fell by more than a third in the transfer case. This underlines the existing distribution difference between the older benchmark traffic and the newer network captures. Instead of proving the effectiveness of generalization, it underlines how difficult cross-environment IDS deployment is. Although the harmonization process improved the compatibility of the data, it did not eliminate the distribution difference, which shows that representation alignment is not sufficient for transferability.
The ablation study showed that the most important contribution to the improvement in the merged case came from the harmonized feature mapping. When this was removed, the accuracy and F1-score fell by a larger margin than when scaling or feature selection was removed. This shows that semantic alignment of related traffic features is useful in combining heterogeneous intrusion detection datasets. However, it should be noted that it is a manual process and not an automated solution. The tamper-evident logging mechanism offered an auditable trail of model settings and predictions via hash chaining. The main benefit of this mechanism is its utility in enabling traceability and reproducibility, rather than offering a novel security primitive. Such integrity verification mechanisms are already common in secure logging solutions; in this work, they are incorporated into the experimental setup to encourage responsible reporting of IDS experiments.
There are several limitations;Firstly, the scope of the experimental evaluation was restricted to only two popular benchmark datasets. Although the datasets differ in their time stamps and traffic patterns, they may not represent all the environments of modern enterprise networks, cloud computing, and IoT networks. Secondly, the scope of the evaluation was restricted to offline supervised learning only, while the framework does not handle online adaptation issues and streaming deployment challenges. Thirdly, the feature harmonization process was performed manually through domain interpretation, which may not be an efficient process for a number of heterogeneous datasets.
The cryptographic mechanism for logging is only based on hash chaining and lacks advanced security features such as keyed authentication, external anchoring, and remote auditing. This might make it more prone to truncation and replay attacks in an adversarial environment. In addition, there is a need to improve the trust model by integrating features of digital signatures, secure timestamping, and external verification mechanisms in the near future. The evaluation of unsupervised models requires post-hoc labeling and thresholding, which might impact performance interpretation despite adhering strictly to standard evaluation protocols. The feature harmonization mechanism is based on manual semantic alignment through domain knowledge, which might introduce subjectivity and limit its applicability in highly heterogeneous settings despite its statistical validation and ablation studies. This mechanism might not generalize well in large-scale and fully automated settings in the near future. In addition, there is a need to develop automated feature alignment mechanisms through representation learning and data-driven matching techniques.
In addition, the framework may be extended for more diverse datasets, especially for IoT and industrial control system traffic, which may have different feature distributions and attack patterns. Investigating domain adaptation methods and incremental learning methods would help better generalization among different datasets.
Conclusion
In this study, a unified framework for cross-dataset intrusion detection is proposed, integrating feature harmonization, multi-model evaluation, and performance analysis using statistical validation. The performance evaluation demonstrates that tree-based supervised models, such as Random Forest, have the highest performance on the harmonized dataset, and cross-dataset evaluations underscore the difficulty of network heterogeneity. The performance evaluation verifies that feature harmonization is a fundamental component for ensuring consistency in multi-dataset evaluation, and that high performance on one dataset does not guarantee performance on another. Finally, statistical testing ensures that performance gains are significant, and that the proposed hash-based logging scheme facilitates reproducibility and transparency without the complexity of blockchain technology.This research does not introduce a new learning algorithm. Rather, its contribution is a structured and reproducible evaluation framework that addresses significant shortcomings in existing IDS research.The framework supports fair comparison across models and promotes more reliable reporting of results. Future work will focus on extending the framework to additional and more diverse datasets, exploring automated feature alignment techniques, and investigating methods to improve cross-domain generalization in real-world deployment settings.
References
- 1.
Morgan S. Cybercrime to cost the world $10.5 trillion annually by 2025. Cybersecurity Ventures; 2020. [Online]. Available from: https://cybersecurityventures.com/hackerpocalypse-cybercrime-report-2016/
- 2. Vinayakumar R, Alazab M, Soman KP, Poornachandran P, Al-Nemrat A, Venkatraman S. Deep learning approach for intelligent intrusion detection system. IEEE Access. 2019;7:41525–50.
- 3. Sharma SB, Bairwa AK. Leveraging AI for intrusion detection in IoT ecosystems: a comprehensive study. IEEE Access. 2025;13:66290–317.
- 4. Mohamed N. Artificial intelligence and machine learning in cybersecurity: a deep dive into state-of-the-art techniques and future paradigms. Knowl Inf Syst. 2025;67(8):6969–7055.
- 5. Resende PAA, Drummond AC. A survey of random forest based methods for intrusion detection systems. ACM Comput Surv. 2018;51(3):1–36.
- 6. Laghari AA, Khan AA, Ksibi A, Hajjej F, Kryvinska N, Almadhor A, et al. A novel and secure artificial intelligence enabled zero trust intrusion detection in industrial internet of things architecture. Sci Rep. 2025;15(1):26843. pmid:40702039
- 7. Maseer ZK, Yusof R, Bahaman N, Mostafa SA, Foozy CF. Benchmarking of machine learning for anomaly based intrusion detection systems in the CICIDS2017 dataset. IEEE Access. 2021;9:22351–70.
- 8. Altulaihan E, Almaiah MA, Aljughaiman A. Anomaly detection IDS for detecting DoS attacks in IoT networks based on machine learning algorithms. Sensors (Basel). 2024;24(2):713. pmid:38276404
- 9. Dash N, Chakravarty S, Rath AK, Giri NC, AboRas KM, Gowtham N. An optimized LSTM-based deep learning model for anomaly network intrusion detection. Sci Rep. 2025;15(1):1554. pmid:39789143
- 10. Waghmode P, Kanumuri M, El-Ocla H, Boyle T. Intrusion detection system based on machine learning using least square support vector machine. Sci Rep. 2025;15(1):12066. pmid:40200017
- 11. Tavallaee M, Bagheri E, Lu W, Ghorbani AA. A detailed analysis of the KDD CUP 99 data set. 2009 IEEE Symposium on Computational Intelligence for Security and Defense Applications; 2009. p. 1–6.
- 12. Sharafaldin I, Lashkari AH, Ghorbani AA. Toward generating a new intrusion detection dataset and intrusion traffic characterization. ICISSp. 2018;1(2018):108–16.
- 13. Brown G, Pocock A, Zhao MJ, Luján M. Conditional likelihood maximisation: a unifying framework for information theoretic feature selection. J Mach Learn Res. 2012;13:27–66.
- 14. Kumar P, Javeed D, Kumar R, Islam AKMN. Blockchain and explainable AI for enhanced decision making in cyber threat detection. Softw Pract Exp. 2024;54(8):1337–60.
- 15. Khashan OA, Khafajah NM. Efficient hybrid centralized and blockchain-based authentication architecture for heterogeneous IoT systems. J King Saud Univ - Comput Inf Sci. 2023;35(2):726–39.
- 16. Arreche O, Guntur TR, Roberts JW, Abdallah M. E-XAI: evaluating black-box explainable AI frameworks for network intrusion detection. IEEE Access. 2024;12:23954–88.
- 17. Ferrag MA, Maglaras L, Moschoyiannis S, Janicke H. Deep learning for cyber security intrusion detection: approaches, datasets, and comparative study. JISA. 2020;50:102419.
- 18. Aziz Z, Bestak R. Insight into anomaly detection and prediction and mobile network security enhancement leveraging K-means clustering on call detail records. Sensors. 2024;24(6):1716.
- 19. Li J, Othman MS, Chen H, Yusuf LM. Optimizing IoT intrusion detection system: feature selection versus feature extraction in machine learning. J Big Data. 2024;11(1).
- 20. Doshi K, Yilmaz Y, Uludag S. Timely detection and mitigation of stealthy DDoS attacks via IoT networks. IEEE Trans Depend Secure Comput. 2021:1–1.
- 21. Koroniotis N, Moustafa N, Sitnikova E, Turnbull B. Towards the development of realistic botnet dataset in the Internet of Things for network forensic analytics: Bot-IoT dataset. FGCS. 2019;100:779–96.
- 22. Meidan Y, Bohadana M, Mathov Y, Mirsky Y, Shabtai A, Breitenbacher D, et al. N-BaIoT—Network-based detection of IoT botnet attacks using deep autoencoders. IEEE Pervasive Comput. 2018;17(3):12–22.
- 23. Bhavsar M, Roy K, Kelly J, Olusola O. Anomaly-based intrusion detection system for IoT application. Discov Internet Things. 2023;3(1).
- 24. Shali A, Chinnasamy A, Selvakumari P. Development of novel intrusion detection in Internet of Things using improved dart game optimizer‐derived optimal cascaded ensemble learning. Trans Emerg Telecommun Technol. 2024;35(7).
- 25. Mhamdi L, Isa MM. Securing SDN: Hybrid autoencoder-random forest for intrusion detection and attack mitigation. J Netw Comput Appl. 2024;225:103868.
- 26. Ali Z, Tiberti W, Marotta A, Cassioli D. Deep transfer learning for intrusion detection in edge computing scenarios. IEEE Internet Things J. 2025:1–1.
- 27. Khan AA, Laghari AA, Baqasah AM, Bacarra R, Alroobaea R, Alsafyani M, et al. BDLT-IoMT—a novel architecture: SVM machine learning for robust and secure data processing in Internet of Medical Things with blockchain cybersecurity. J Supercomput. 2024;81(1).
- 28. Khan AA, Laghari AA, Alroobaea R, Baqasah AM, Alsafyani M, Alsufyani H, et al. A lightweight scalable hybrid authentication framework for Internet of Medical Things (IoMT) using blockchain hyperledger consortium network with edge computing. Sci Rep. 2025;15(1):19856. pmid:40473928
- 29. Lu Y, Huang X, Dai Y, Maharjan S, Zhang Y. Blockchain and federated learning for privacy-preserved data sharing in industrial IoT. IEEE Trans Ind Inf. 2020;16(6):4177–86.
- 30. Rahman MA, Hossain MS, Islam MS, Alrajeh NA, Muhammad G. Secure and provenance enhanced internet of health things framework: a blockchain managed federated learning approach. IEEE Access. 2020;8:205071–87.
- 31. Arreche O, Guntur TR, Roberts JW, Abdallah M. E-XAI: evaluating black-box explainable AI frameworks for network intrusion detection. IEEE Access. 2024;12:23954–88.
- 32. Khan AA, Laghari AA, Alsafyani M, Baqasah AM, Kryvinska N, Almadhor A, et al. A cost-effective approach using generative AI and gamification to enhance biomedical treatment and real-time biosensor monitoring. Sci Rep. 2025;15(1):17305. pmid:40389552
- 33. Alsubaei FS. Smart deep learning model for enhanced IoT intrusion detection. Sci Rep. 2025;15(1):20577. pmid:40596059
- 34. Nassreddine G, Nassereddine M, Al-Khatib O. Ensemble learning for network intrusion detection based on correlation and embedded feature selection techniques. Computers. 2025;14(3):82.
- 35. Nabi F, Zhou X. Enhancing intrusion detection systems through dimensionality reduction: a comparative study of machine learning techniques for cyber security. Cyber Secur Appl. 2024;2:100033.
- 36. Heidari A, Jafari Navimipour N, Unal M. A Secure intrusion detection platform using blockchain and radial basis function neural networks for internet of drones. IEEE Internet Things J. 2023;10(10):8445–54.
- 37. Kshetri N. Blockchain’s roles in strengthening cybersecurity and protecting privacy. Telecommun Policy. 2017;41(10):1027–38.
- 38.
Cover TM, Thomas JA. Elements of information theory. 2nd ed. Hoboken (NJ): John Wiley & Sons; 2006.
- 39. Zhang H. The optimality of Naive Bayes. Proceedings of the Seventeenth International Florida Artificial Intelligence Research Society Conference (FLAIRS); Miami Beach, FL. Menlo Park (CA): AAAI Press; 2004. p. 562–7. https://aaai.org/papers/flairs-2004-097/
- 40. Cortes C, Vapnik V. Support-vector networks. Mach Learn. 1995;20(3):273–97.
- 41. Breiman L. Random forests. Mach Learn. 2001;45(1):5–32.
- 42. MacQueen J. Multivariate observations. Proceedings ofthe 5th Berkeley symposium on mathematical statisticsand probability, Vol. 1. Oakland, CA, USA: University of California press; 1967. p. 281-–97.
- 43. Ester M, Kriegel HP, Sander J, Xu X. A density-based algorithm for discovering clusters in large spatial databases with noise. Inkdd. 1996;96(34):226–31.
- 44. Liu FT, Ting KM, Zhou ZH. Isolation forest. 2008 Eighth IEEE International Conference on Data Mining. IEEE; 2008. p. 413-–22.
- 45. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997;9(8):1735–80. pmid:9377276
- 46. Wilcoxon F. Individual comparisons by ranking methods. Biom Bull. 1945;1(6):80.
- 47. McNemar Q. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika. 1947;12(2):153–7. pmid:20254758
- 48. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837–45. pmid:3203132
- 49. Hussain N, Li S, Hussain A, Ullah Z, Jamjoom M. Quantum-aware secure blockchain intrusion detection system for industrial IoT networks. Sci Rep. 2025;16(1):2265. pmid:41381788
- 50. Hozouri A, Mirzaei A, Effatparvar M. A comprehensive survey on intrusion detection systems with advances in machine learning, deep learning and emerging cybersecurity challenges. Discov Artif Intell. 2025;5(1).