Figures
Abstract
Outlier detection is essential for identifying unusual patterns or observations that significantly deviate from the normal behavior of a dataset. With the rapid growth of data science, the prevalence of anomalies and outliers has increased, which can disrupt system modeling and parameter estimation, leading to inaccurate results. Recently, deep learning-based outlier detection methods have gained significant attention, but their performance is often limited by challenges in parameter selection and the nearest neighbor search. To overcome these limitations, we propose a three-stage Efficient Outlier Detection Approach (named EODA), that not only detects outliers with high accuracy but also emphasizes dataset characteristics. In the first stage, we apply a feature selection algorithm based on the Boruta method and Random Forest to reduce the data size by selecting the most relevant attributes and calculating the highest Z-score of shadow features. In the second stage, we improve the K-nearest neighbors algorithm to enhance the accuracy of nearest neighbor identification in the clustering phase. Finally, the third stage efficiently identifies the most significant outliers within clustered datasets. We evaluate the proposed EODA algorithm across eight UCI machine-learning repository datasets. The results demonstrate the effectiveness of our EODA approach, achieving a Precision of 63.07%, Recall of 82.49%, and an F1-Score of 64.53%, outperforming the existing techniques in the field.
Citation: Kumar S, Varshney S, Jain U, Johri P, Almazyad AS, Mohamed AW, et al. (2025) EODA: A three-stage efficient outlier detection approach using Boruta-RF feature selection and enhanced KNN-based clustering algorithm. PLoS One 20(5): e0322738. https://doi.org/10.1371/journal.pone.0322738
Editor: Razieh Sheikhpour, Ardakan University, IRANISLAMIC REPUBLIC OF
Received: October 25, 2024; Accepted: March 26, 2025; Published: May 30, 2025
Copyright: © 2025 Kumar et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data and codes supporting the findings of this study are publicly available at: https://github.com/Sunildangi107/Oulier_Detection_Dataset.
Funding: The research is funded by Researchers Supporting Program number (RSPD2025R809), King Saud University, Riyadh, Saudi Arabia. The funders provided financial support for this research and facilitated access to key resources. However, they had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
1. Introduction
An outcome that significantly differs from the other datasets and seems to be unsuitable is referred to as an anomaly. Hawkins defines an outlier [1] as: “A deviation from the usual pattern that raises the possibility that an unusual technique was used to create the object is known as an outlier”. Data mining includes outlier detection as a key component and has recently gained popularity. The goal of identifying outliers is to locate data objects that stand out from the majority. This task aims to identify outliers or unusual findings [2]. Current applications of the outlier detection technique include network resilience analysis, detection of intrusions, clustering categorization, medical and health care, assessment of stock networks, and the protection of credit card scams, among others [3–4]. The last few years have seen the development of a wide range of recognition of outlier techniques, including clustering-based models, distance-based models, density-based models, probabilistic models, etc. [5]. Several deep recognitions of outlier strategies have also been presented [6–7] in addition to these shallow systems.
Outlier detection techniques are frequently used in a variety of data mining applications, including data cleaning, recognition of credit card fraud, intrusion detection for networks, web service QoS forecasting, DRAM failure estimation, and even recently trendy government legislation [8]. In recent years, the study on outlier detection has drawn the attention of researchers. Recently, several Outlier detection algorithms have been proposed based on different classifications, such as supervised, unsupervised, and semi-supervised techniques. When a training dataset already has prior knowledge about the category that every single instance falls under—normal or abnormal—supervised outlier detection is concerned. When training data are assumed to be entirely normal instances, one class support vector machines (OCSVM) or support vector data descriptions (SVDD) [9,10] carry out a hypersphere within the normal data and use the created hypersphere to identify an unidentified sample as an inlier or outlier. In numerous practical scenarios, the trained outlier detection problem presents a challenging case because obtaining label details for the whole training dataset can be costly, time-consuming, and subjective.
Unsupervised machine learning outlier detection, which does not require knowledge of the class distribution beforehand, can be broadly divided into four categories: distribution-based [11], distance-based [12], density-based [13], and clustering-based [14] methods. While anomalies are not consistent with the algorithm, the distribution-based technique believes that every data point is the result of a specific statistical model. The supervised learning-based approaches now in use have two major limitations, the first of which is the time-consuming requirement for labeled training datasets in real-time applications. Second, it fails to identify new categories of unusual events. A method based on unsupervised learning, however, does not need a labeled training dataset. It can identify the types or categories of unusual events. According to the classification, the difficulty in identifying the local density issue is a barrier to the distance-based strategy. Furthermore, it is still difficult to identify low-density patterns using density-based methods.
This paper proposes an effective method for identifying outliers that explicitly addresses the limitations of unsupervised methods. This proposed method eliminates irrelevant attributes present in the datasets by examining relevant features of the attributes and using K-nearest neighbors (KNN) search; the proposed approach uses unsupervised learning to create class-based clusters. The primary goal of the proposed study is to use data reduction techniques to remove irrelevant attributes present in the dataset and reduce the size of the dataset in its first stages. Each attribute in the dataset contributes significantly to the outlier detection process. Therefore, removing irrelevant attributes from the dataset significantly impacts how well the proposed method performs. The KNN search algorithm is used in the proposed approach to group similar class clusters based on their distance from each other. The proposed strategy also computes the degree of each point and finds outliers in the clustered dataset. The originality of this research resides in the method of training the dataset by eliminating redundant attributes, creating class-based clusters, and then identifying outliers within these clusters. The emphasis is on preparing the data by removing extraneous features, which ultimately enhances the accuracy and effectiveness of outlier recognition.
1.1. Our motivation
Outlier detection, particularly in large datasets, can be incredibly challenging due to the complexity of the data. As datasets grow, they often contain a mix of both useful and irrelevant features, making it difficult to accurately identify patterns that truly stand out. These irrelevant characteristics can negatively affect the precision and efficiency of outlier detection methods, leading to misleading or erroneous results. Another major challenge is accurately identifying the nearest neighbors, especially in high-dimensional data. Many existing outlier detection methods struggle with this, and if the nearest neighbors are chosen incorrectly, it can cause the entire clustering process to break down, leading to incorrect identification of outliers. Moreover, many current approaches don’t compute the Z-score and the degree of each point simultaneously, which is essential for understanding how far an observation deviates from the norm. Without this, it is harder to determine the true significance of potential outliers. Given these challenges, our work is motivated by a clear need to improve outlier detection by addressing the issue of irrelevant data, ensuring more accurate neighbor selection, and providing a more comprehensive analysis of each data point.
1.2. Our contributions
In this paper, we introduce EODA (Efficient Outlier Detection Approach) to tackle the mentioned challenges and provide a more reliable way of identifying outliers in complex datasets. The summary of our main contributions is as follows:
- Proposing a three-stage outlier detection method: One of the key innovations in this work is the integration of Boruta-Random Forest (RF) feature selection, KNN clustering, and outlier detection, allowing us to not only identify outliers but also understand the structure of the clusters they deviate from. The proposed EODA algorithm combines feature selection, improved KNN, and simultaneous calculation of both Z-scores and data point degrees. This three-stage approach enables us to detect the most significant outliers within clustered datasets, making the process more effective and efficient.
- Reducing irrelevant data using Boruta-RF feature selection: In the first phase of the proposed EODA algorithm, we streamline the dataset by using Boruta-RF to identify the most relevant features. By focusing only on the attributes that matter, we remove noise and redundancy, which significantly improves both the accuracy and the speed of the outlier detection process.
- Enhancing nearest neighbor accuracy in clustering: In the second phase, we address the problem of nearest neighbor selection. We improve the traditional KNN algorithm by using the Euclidean distance metric to find the closest neighbors more precisely. This ensures that we form accurate clusters, which is crucial for identifying outliers with confidence.
- Proving the effectiveness across real-world datasets: Finally, we tested our method on eight UCI datasets and compared its performance against five existing outlier detection algorithms including LOF [15], CBLOF [16], LDCOF [17], DBSCAN [18], and H-DBSCAN [19]. The results demonstrate that the proposed EODA algorithm consistently outperforms the existing methods in terms of precision, recall, and F1 score.
1.3. Organization
The remainder of the paper is divided into five sections: Section 2 describes the related works of the current unsupervised outlier identification system, which summarises the research on density-based techniques that have been proposed over the years. Section 3 discusses the proposed framework of outlier detection to address the issues of data reduction, searching for the nearest neighbor, and computing outlier detection along with the degree of each point. Section 4 presents the results of outlier detection over eight UCI datasets and discusses each obtained result. Section 5 focuses on the conclusion of the research paper and its future direction.
2. Related work
The recent papers on outlier detection are reviewed in this section. Several published studies on outlier detection have had an impact on our research. This section explores various topics such as the classification of outlier methods, different types of techniques used to identify outliers based on cluster-based techniques, and density-based outlier detection techniques. The efficiency of the models for identifying outliers may be impacted by the outliers. As a result, the outliers in the dataset must be eliminated to improve the performance of such models. A comparative analysis of distance-based methods for outliers’ detection is shown in Fig 1.
Existing techniques for outlier detection fall into the following categories: The following techniques are used: statistical, ensemble, distance-based, density-based, clustering-based, learning-based. Statistics-based methods are employed to draw out information from the underlying data. These methods deal with the issues presented by certain distribution points that do not fit the model. These methods, however, may not be effective when dealing with high-dimensional data adaptation caused by data points connected to several unknown distributions. An observation is considered an outlier if its immediate neighborhood has a low population density.
Distance-based outliers are identified by taking into consideration the distance to their neighbors. An adequate distance and the right number of neighbors are identified by analyzing the observation’s distance from its neighbors to predetermined threshold values. Two points are referred to as neighbors to detect outliers if their distances fall outside of a predetermined range. A parameter that is that depends on the choice of neighboring points affects the determination of an outlier. The specifications for detecting anomalies in the dataset are greatly influenced by the parameter [11]. If an object meets following conditions for a particular dataset X, it is referred to as an outlier:
where n represents the number of objects in the data collection, and stand for the threshold values.
A distance-based strategy utilizing KNN for outlier detection is developed [20] to solve the drawback of outlier ranking. By calculating the distance to each item’s kth nearest neighbor, this method determines the Z-score for each object. The proposed approach has demonstrated impressive effectiveness in outlier detection while preserving minimal computational complexity.
2.1. Density based outlier detection
Breunig et al. [21] developed the idea of the local outlier factor (LOF) to combine with the idea of relative density in identifying outliers. By calculating the degree of each point in relation to its neighbors, LOF evaluates KNN outlier detection and distance-based determination of outlier’s algorithms [22–29] and [30]. To assess whether a data point is abnormal, LOF considers the density of nearby points in the nearby region. Data points with higher LOF scores have a higher chance of being outliers. To find the top-n local outliers, Jin et al. [31] presented an approach since outliers frequently take a small portion of a dataset. As they are more likely to reflect the most significant outliers in the dataset, this model concentrates on choosing data points with the highest LOF scores. Additionally, several techniques [32–35] have been developed in the literature already in existence to find local outliers in static datasets. These techniques seek to detect and examine outliers in particular local dataset regions.
Tang et al. [36] proposed a method based on the connectivity-based outlier factor (COF) method, which seeks to ascertain the relative connectivity amongst data points. To determine the level of outlier-ness for each data point, this method applies the KNN methodology. The COF approach offers a trustworthy way to find outliers in a dataset by taking into consideration the closeness and connection of points. The local distance-based outlier factor (LDOF) approach has been proposed [37]. This variant calculates the local density of the data points by averaging the distances between each data point and its k-nearest neighbors. The outlier factor is then generated by estimating the relative density using the KNN method. The LDOF approach is an efficient way to find outliers in a dataset which takes advantage of both local density and relative density. A more economical variation of LOF has been published by Erich Schubert et al. [35], which uses the k-distance measure to assess the density of data points rather than the achievable distance employed in conventional LOF. However, this adjustment might provide inliers with lesser reliable outcomes. This method entails calculating the distances between data points to evaluate their density and specify the degree of irregularity for each data point. The precision of identifying inliers may be compromised by using this method, but it allows for an effective assessment of the density and irregularity of data points.
The LOF algorithm is a well-known technique for outlier detection and was first introduced by Angiulli and Pizzuti in 2002 [38]. Each data point’s neighborhood density is calculated using the LOF technique. The LOF technique provides useful insights about a data point’s local density in comparison to its surrounding data points by assessing the density of a data point concerning its neighboring points. Based on how far outliers deviate from the regional patterns of densities, this method has shown to be useful for detecting anomalies.
This work uses the LOF estimator, developed by Dragoljub Pokrajac et al. [23], to calculate the local density deviation of a given sample with nearby data points. It measures how much the density of a particular sample varies from the nearby data points, giving important information on the anomalous status of the sample. According to our method, an outlier is a sample point with a density that is noticeably lower than the sample points nearby. This description implies that the outlier is an observation that differs from the predicted density patterns displayed by its neighbors. The measure is expressed as per the following:
where the measurement of the distance between a data point’s k nearest neighbors, or the set , and that point’s
can be generally referred to as the phrase
. It estimates the separation between
and the edge of the data point’s immediate neighborhood and provides important details about the local density features of the data point.
The COF algorithm, which was developed by Tang et al. [39], aims to identify underlying data patterns. To determine how many neighbors are closest, the COF algorithm uses the set-based nearest (SBN) method. An in-depth analysis of the local density features in the dataset becomes feasible using the SBN route within COF to find and calculate its nearest neighbors for each data point. Additionally, the average chain distance of the experiment site is taken into consideration when using the SBN technique to compute the relative density. By comparing the test point’s density to those of its surrounding data points, this calculation aids in determining the test point’s local density characteristics. The relative density distributions are examined in the LOF and COF techniques to identify outliers. The LOF algorithm has been significantly enhanced by Jin et al. [40], who have also unveiled INFLO, a brand-new density-based outlier detection technique. The core concepts of LOF are expanded upon in INFLO, which also incorporates enhancements to increase the precision and efficacy of outlier detection using density analysis. The influenced space, which is created by merging an object’s neighboring data points and reverse neighboring data points, is used in the INFLO method to estimate relative density. This extensive method considers the influence of both the object’s neighbors and its reverse neighbors, allowing for a more accurate calculation of the relative density. Each data point’s connectivity pattern with its neighbors is the focus of the COF algorithm. The COF algorithm offers useful insights into the connections and interactions among data points by focusing on the connectivity element, allowing for a more in-depth examination of outlier detection based on density characteristics. This connection pattern is measured using a score represented by the following ratio:
where is the average chaining distance between point p and its k-1 nearest neighbors in the set
, which is defined as
, where
stands for the edges used in the calculation. The average distance between point p and its k-1 nearest neighbors is measured by the
, which is a key parameter in some outlier detection algorithms as well as analysis.
2.2. Clustering-based outlier detection
The primary objective of clustering algorithms, on the other hand, is to uncover patterns that classify the fundamental information and group it into discrete clusters, distinguishing between normal clusters and outliers that vary from typical clusters. Clustering-based unsupervised outlier detection techniques have drawn a lot of interest and have grown in popularity [41]. These methods are beneficial as they create detection models without the need for labelled training data, enabling them to be particularly helpful in circumstances when labeled data is scarce or unavailable. The benefit of clustering-based techniques in the outlier detection problem is their applicability in situations where acquiring labeled data or perfectly regular data is difficult.
2.2.1. K-means clustering method.
Partitional and hierarchical methods can be generally used to categorize clustering methods. Partitioning techniques, like distance-based methods, use distance measures (such as Euclidean or city block distances) to compare the similarity of data points and group them into clusters according to predetermined standards. While identifying clusters, density-based algorithms evaluate the density of data regions. By repeatedly integrating or dividing clusters, hierarchical clustering creates a hierarchy of clusters. Depending on the data properties and desired level of control during the clustering process, one may choose between partitional and hierarchical clustering [24,25]. The Euclidean distance, which is used in Eq (6), is a popular option for methods of clustering because it accurately expresses geographic relationships between data points.
The 2-dimensional samples can be used to express by Equation (7) using the Euclidean distance: and
as
Additionally, as shown in Equation (8), new centroids are identified by modifying at every repetition. The K-means technique can be terminated if the constituents of
and
remain unchanged during subsequent iterations.
2.2.2. K-Nearest neighbor method.
The KNN algorithm, a popular regression and classification technique, is employed to identify the closest neighbors for a given sample, considering relevant sample measurements. For a given dataset D consisting of k clusters with similar member labels (p), the KNN algorithm can be utilized to determine the nearest neighbors for a test sample from a test set (TS). To achieve this, the distance between the test sample
and the cluster members
is calculated using Equation (9), and it is denoted as d_ji. Next, the values of
are sorted in ascending order to create
as shown in Equation (10). The average of the K members from
is then calculated as 𝑎𝑣𝑒𝑖, which represents the best value of K according to the model described in Equation (11). The outlier detection within
is based on the cluster with the least 𝑎𝑣𝑒 value, as determined by Equation (12).
Although KNN is frequently used as a classification approach, in this case, it has been employed to determine distance. Support vector machine (SVM) and RF have been applied on two-labeled and multi-labeled datasets, respectively, to assess the efficacy of the methods. A linear kernel function with a 2-box restriction was used by SVM in this work. Instead of using just one decision tree, the RF approach used many single decision trees, which resulted in more accurate results.
An extensively researched technique for identifying outliers is the CBLOF (clustering-based LOF) algorithm [42]. It has been introduced by He et al. in 2003 [42]. By taking into consideration the properties of data clusters, the algorithm concentrates on locating local outliers. Using a clustering method like k-means or DBSCAN, the CBLOF algorithm divides the data into clusters. Then, a score is given to each cluster depending on the size and average distance between the data points within the cluster. Each data point’s LOF, which reflects how out of the ordinary it is compared to its nearby cluster, is determined. The LOF is determined by comparing the density of the data point to the densities of its neighbors. The CBLOF algorithm offers a computationally efficient approach for outlier detection, as it only considers the size of clusters and does not require calculating the exact density of each cluster. It has been applied in various domains, such as fraud detection, intrusion detection, and network traffic analysis, where identifying local anomalies within clusters is crucial. Further research and literature reviews on the CBLOF algorithm can provide more in-depth insights into its theoretical foundations, variations, and experimental evaluations.
In the LDCOF (local density cluster-based outlier factor) algorithm, the data points are grouped into clusters based on their similarities using clustering techniques such as k-means, DBSCAN, or hierarchical clustering. Each cluster’s local density is calculated by considering the distances between the data points within the cluster. To determine the outlier factor for each data point, the LDCOF algorithm assesses the density of the data point concerning its neighboring points and the density of its corresponding cluster [43–47]. A higher outlier factor indicates that a data point has a lower density compared to its neighbor’s and the density of its cluster, suggesting its potential outlier status. By combining local density and clustering information, the LDCOF algorithm aims to provide a more comprehensive and accurate outlier detection approach. It can be applied to various domains and datasets where understanding local density patterns and cluster structures is crucial for identifying outliers. However, further literature or references regarding LDCOF are needed for a more H-DBSCAN (hierarchical density-based spatial clustering of applications with noise) is an extension of the DBSCAN algorithm that incorporates a hierarchical approach to clustering. While there may not be a specific algorithm known as “H-DBSCAN” in the literature, there are variations and related works that incorporate hierarchical clustering with density-based techniques. The density-based clustering and hierarchical approaches hybrid HDBSCAN algorithm were introduced in [48–51]. It is possible to find clusters of various densities and sizes using HDBSCAN, which builds a hierarchical cluster tree. It deals with the shortcomings of the conventional DBSCAN method.
Several well-known outlier detection methods have been assessed in this section, and it has been noticed that many of them rely on the same model parameter, k (neighborhood size). Determining whether k is used simultaneously with the distance-based or density-based strategy is essential since it has a significant impact on the efficiency of outlier detection. We describe a unique outlier detection technique that makes use of the idea of natural neighbors to get around these restrictions. A summary of comparison analysis of the existing research for outliers’ detection methods is discussed in Table 1.
2.3. Problem definition
The purpose of this research is to create an efficient feature selection and clustering approach for outlier detection. Outliers, which dramatically differ from the rest of the data, might offer insightful information or hint at problems with the quality of the data. However, because the data is so complicated and has a high degree of dimensionality, finding outliers in large datasets can be difficult. The primary issue to deal with is feature selection: Find the most relevant characteristics for outlier detection by developing a feature selection strategy [52–55]. The subset of attributes that can most effectively distinguish between typical data and outliers should be automatically identified by the system.
- Clustering Techniques: The work requires exploring and using clustering methods to group comparable data points together. These algorithms examine the underlying structures and patterns in the data to find clusters and separate outliers from typical data points.
- Outlier Detection: Create an outlier detection system that makes use of the selected attributes and clustering outcomes to locate and highlight probable outliers. For each data point, the approach should assess the degree of abnormality by considering both the feature values and the distance from cluster borders.
This paper proposes a stable and effective strategy for outlier detection in big and complicated datasets by combining feature selection and clustering. The proposed EODA approach has the potential to be used in a variety of fields such as banking, healthcare, fraud detection, and detection of anomalies in industrial systems. In Section 3, the proposed EODA model is introduced to accomplish the research paper’s objectives. Using feature selection and clustering approaches, the EODA model was created to overcome the difficulties associated with the existing outlier detection techniques.
3. Proposed EODA outlier detection algorithm
To achieve high accuracy and efficiency when mining outliers in high-dimensional datasets, there are multiple challenges to overcome. When used to mine for outliers in these datasets, existing techniques frequently have issues with the effectiveness of the outlier method. To address this problem, we provide a data reduction strategy that concentrates on reducing the dimensions and data points before executing outlier identification methods. In this section, the process of eliminating irrelevant attributes from the dataset is outlined in Section 3.1. The technique explores cluster-exhibiting dimensions in the dataset, enabling us to locate dense clusters and their positions in each dimension. The objective is to distinguish between dimensions that matter and those that don’t for outliers. Performing this, we can locate a subspace with dense points in irrelevant dimensions. By reducing the number of sparse regions, the efficiency of data clustering has also been improved using this attribute relevance analysis idea. In the methodology section, Fig 2 provides a description of the proposed scheme architecture for finding outliers in huge datasets. This scheme works in three stages as follows:
- Stage 1: Feature Selection: The objective of this stage is to select the high-dimensional dataset’s most significant and instructive features. The significance of each feature can be assessed using feature selection approaches like mutual information, correlation analysis, or recursive feature reduction. We can decrease dimensionality to raise the precision of outlier detection by choosing effective features that accurately depict the fundamental trends in the data.
- Stage 2: Clustering: Clustering methods are used to combine data points that are associated after the relevant attributes have been recognized. You can use a variety of clustering methods, including k-means, hierarchical clustering, and density-based clustering. The idea is to divide the dataset into discrete clusters, where each cluster stands for a collection of data points with a common set of attributes. Clustering makes it easy to spot outliers that do not fit into any cluster by identifying the regular behaviors that occur within the data.
- Stage 3: Outlier Detection: The next phase is to find outliers within the dataset after collecting the clusters. Considering the difference or differing characteristics between each data point and its associated cluster will help with this. Potential outliers can be data points with a large divergence from the specified cluster centroid, a large distance from it, or low density in the grouping region. To further detect outliers based on how they differ from the data’s usual distribution, statistical techniques like the Z-score or Mahalanobis distance can be computed.
The proposed approach follows a structured three-stage process to enhance outlier detection. In Stage 1: Feature Selection, the method identifies the most significant and informative features from high-dimensional datasets. Techniques such as mutual information, correlation analysis, or recursive feature reduction are utilized to retain features that capture key trends and variations. This step reduces dimensionality, improves computational efficiency, and enhances the precision of outlier detection. In Stage 2: Clustering, the selected attributes are used to group data points into clusters using k-means, hierarchical clustering, or density-based clustering. These clusters represent groups of data points with similar characteristics, allowing for identifying regular patterns within the data. In Stage 3: Outlier Detection, the deviations of individual data points from their respective clusters are analyzed. Outliers are identified based on significant divergence from cluster centroids, large distances from cluster boundaries, or low-density regions. Statistical techniques such as the z-score or Mahalanobis distance further refine this process, enabling precise and reliable detection of anomalies within the dataset.
3.1. Elimination of irrelevant attributes
The process of developing predictive models needs to incorporate the feature selection process since it aids in determining which features are most useful for forecasting the target variable. Outliers in the dataset, however, may affect the importance ratings of the features, which may then impact the selection procedure. The Boruta algorithm is a feature selection approach created to solve the drawbacks of conventional techniques that depend upon statistical hypotheses that might not hold in datasets from the real world.
The first phase of the Boruta method generates “shadow” properties for every variable. The original features are permuted arbitrarily to produce these shadow qualities, which eliminates biases and correlations between them. To train an RF classifier, which determines the Z-scores indicating the significance of each feature, this expanded information system is subsequently utilized. The technique is iteratively performed to produce statistically significant outcomes, and the maximum Z-score amongst the shadow characteristics (abbreviated as MZSA) is established for each iteration.
The Z-scores for each of the actual attributes are analyzed and compared to the MZSA. Any attribute that has a Z-score below the MZSA is labeled “Unimportant” and is then deleted from the database. On the other hand, each attribute that has a Z-score greater than the MZSA is labeled as “Important.” This procedure is repeated until all attributes are classified as “Important” or “Unimportant,” or until a predetermined number of RF iterations have been completed. The final result of the Boruta method then returns the features that have been selected after the shadow characteristics have been eliminated.
Algorithm 1. Feature Selection Method for Removing Irrelevant Attributes
Inputs: Feature Matrix and Target Matrix of Original Data, maximum number of iterations
Output: Removal of the irrelevant features and selecting the relevant features from datasets
1. INITIALIZATION:
2. x⇓Original Features
3. y⇓Shadow Features
4. R⇓Origin
5. M⇓ Feature Matrix of Original Data
6. N⇓ Target Matrix of Original Data
7. n⇓maximum number of iterations
8. S⇓New Matrix
9. D⇓Dataset
10. ⇓Z_Score of Original Features
11. ⇓Maximum Z_Score of Shadow Features
12. BEGIN:
13. for (i = 0; i < n; i++)
14. P⇓Generate(Matrix of Shadow Features)
15. S⇓Fit_Model(P,R)
16. if ()
17. D⇓ Assign (x)
18. else
19. Removal of the original features from the dataset
20. end if
21. End for
22. Compute average gain of the selected original features.
23. END
The proposed method assigns each attribute a relevance score depending on how closely it relates to the target variable. Then, it resolves redundancy across each of the selected features and identifies the relevant attributes above a predetermined level. The result is a dataset that has been condensed to just include the crucial elements and is prepared for additional simulation or assessment.
The novelty of this algorithm lies in its integration of advanced feature selection techniques, such as mutual information or recursive feature elimination, with the specific aim of optimizing outlier detection. By focusing on identifying and removing irrelevant attributes, the algorithm significantly reduces the dimensionality of the dataset while preserving the core information required for accurate analysis. This step is particularly innovative in the context of outlier detection, as it ensures that the detection process is both computationally efficient and precise.
3.2. Searching the nearest neighbors
The KNN algorithm is a supervised learning algorithm commonly used for classification and regression tasks. It operates by identifying the k nearest neighbors to a given data point in the feature space and utilizing their labels or values to make predictions. While the KNN algorithm itself does not explicitly handle outliers, it can be employed as a tool for outlier identification and handling within a dataset. This can be achieved by using the KNN algorithm to calculate the distances between each data point and its k nearest neighbors. If a data point is significantly farther away from its k nearest neighbors compared to the rest of the data points, it can be classified as an outlier. These outliers can then be managed by either removing them from the dataset or treating them differently, depending on the specific analysis requirements. Additionally, the KNN algorithm can be utilized for imputing missing values in a dataset. This involves using the KNN algorithm to identify the k nearest neighbors for each data point with missing values and then using the average (for numerical data) or mode (for categorical data) of the corresponding values from the nearest neighbors to impute the missing values.
The algorithm improves upon the traditional KNN search by incorporating enhancements tailored for clustered datasets. Its novelty lies in dynamically adjusting the neighborhood search based on the clustering structure, which mitigates the impact of high-dimensional noise and improves the identification of nearest neighbors. This approach allows for more accurate clustering and subsequently enhances the precision of outlier identification in the next stage.
Algorithm 2. KNN Search Algorithm for Cluster Dataset
Inputs: Dataset as X = {,
,
, ………,
} where
,
,
, ………,
are the data points of clusters to identify the class, Threshold Value as Thresh_k, Nearest Neighbors C: {
,
,
, ………,
}
Output: Class of Clusters as where
, x = 1,2,3,……
1. INITIALIZATION:
2. K⇓3
3. m⇓len (C)
4. BEGIN:
5. for (j = 1; j<=m; j++)
6. Compute Euclidian_Distance (j, })
7. End for
8. for (K = 3; K<=Thresh_k; K++)
9. for (= 1;
<=m;
++)
10. if (K_nearest < k && Euclidian_Dist >= Euclidian_Dist_Ordered)
11. Take the data points from x
12. End if
13. End for
14. End for
15. Compute the class of the x:
16. END
3.3. Detection of outlier instances
Data points that greatly depart from the rest of the data and display anomalous behavior are those that need to be identified to be classified as outlier cases. Detecting abnormalities or pointing out problems with data quality, outliers offer insightful information. The LOF and KNN algorithms, which evaluate distance or density between data points, are examples of distance-based methods. The term “outlier” refers to a point with exceptionally high or low density.
The proposed algorithm combines feature selection, clustering, and outlier detection techniques to identify and classify outlier instances in the dataset. This code represents an outlier detection algorithm. It takes as input a set of clusters C obtained by applying another algorithm (Algorithm 2). For each cluster i in C, it calculates the density of instances within that cluster by summing the distances of each instance to all other instances in the cluster. Then it calculates an outlier score for each instance in the cluster by multiplying its density by its distance to the nearest higher density instance. Alternatively, it can use the inverse of distance instead of distance in the outlier score calculation. It then merges each cluster of probable outliers and computes the set of non-outlier instances N_pts by subtracting the probable outlier instances from the original data set D. Finally, it sorts the outlier scores in descending order and generates the top-N outliers.
The proposed outlier detection algorithm exhibits novelty in its multi-faceted approach to isolating outliers. By combining statistical methods, such as z-scores or Mahalanobis distance, with cluster-based anomaly analysis, the algorithm addresses both local and global variations in the dataset. Its ability to analyze deviations within clusters while simultaneously considering inter-cluster relationships provides a comprehensive framework for outlier detection. This integration of techniques distinguishes it from existing methods and underscores its innovative contribution to the field. The combination of the Boruta-RF feature selection method, enhanced KNN clustering, and the proposed outlier detection approach, is rooted in the complementary strengths of each component. Boruta-RF effectively identifies the most relevant features, improving the accuracy of the subsequent clustering process by reducing dimensionality and eliminating irrelevant noise. Clustering then groups data points with similar characteristics, which facilitates the detection of outliers based on their deviation from the typical patterns. This unified framework highlights how each stage contributes to the overall performance, demonstrating that the combination of these techniques leads to more accurate and robust outlier detection.
Algorithm 3. Proposed EODA Algorithm for Detection of Outlier Instances
Inputs: Dataset as D, Set of Clusters as C
Output: Instances of Outliers, Set of Outlier Clusters
1. INITIALIZATION:
2. C⇓{,
,
, ………,
}// Set of Clusters after applying algo. 2
3. BEGIN:
4. for (i=1; i<=; i++)
5. //
is the instance distance to other instance distance present within clusters
6. Compute the nearest distance from the higher-density instances
7. Calculate outliers’ instances as:
⇓
or
8. End for
9. Merging each cluster of probable outliers.
10. Compute ⇓ D -
.
11. Compute ⇓ Sort (
).
12. Generate Top-N outliers.
13. END
3.4. Computational complexity analysis
The computational complexity of the proposed EODA is derived by analysing its three main algorithms: feature selection, KNN-based clustering, and outlier detection. Algorithm 1, the feature selection method, processes the dataset by iterating up to a predefined maximum number of iterations . During each iteration, shadow features are generated,
scores are calculated, and features are evaluated for relevance. These operations yield a complexity of
, where
represents the number of features. This step is efficient for reducing dimensionality while retaining essential attributes.
In Algorithm 2, the KNN-based search algorithm computes distances between data points within clusters and identifies the nearest neighbours. The Euclidean distance computation for clusters and
instances dominate the process, with a complexity of
, including sorting operations for nearest neighbour identification. This complexity reflects the effort to balance accuracy and scalability for clustering.
Finally, Algorithm 3 calculates density values for clustered data points and identifies outliers based on their scores. For each cluster, density and outlier scores are computed with a complexity of , where
is the average number of points per cluster. Sorting outlier scores to identify the top instances introduces an additional complexity of
. Thus, the overall complexity of this stage is
.
Combining the complexities of all three stages, the overall computational cost of EODA is approximately . This complexity indicates that EODA is efficient for datasets of moderate size but may face scalability challenges with extremely large datasets due to the dependence on the number of instances and clusters.
4. Results and discussion
The purpose of this experimental research work is to assess how well the proposed EODA algorithm performs. To demonstrate the enhanced performance of our machine learning approach in outlier detection, we performed a comparison study employing eight selected datasets from the UCI Machine Learning Repository, openly available in [56] (website: http://archive.ics.uci.edu/ml/datasets). In the following, the effectiveness of our work is assessed using these real-world datasets as useful benchmarks. In our studies, we examined the results that were obtained from our proposed method with those five well-known unsupervised outlier detection methods: LOF [15], CBLOF [16], LDCOF [17], DBSCAN [18], and H-DBSCAN [19]. Based on their applicability to unsupervised learning techniques for outlier detection and their shared parameter, k, these algorithms were chosen.
The Python programming language was used to implement the EODA algorithm. The tests were carried out on a computer with a 3.40 GHz Intel(R) Core (TM) i7-4770 CPU, 6 GB of RAM, and a RAM frequency of 799.0 MHz. Table 2 displays the performance measurements that were taken into consideration and offers perspectives on multiple dimensions of the outlier detection method’s performance, including its capacity to accurately detect outliers, manage false positives, and strike a balance between precision and recall. Depending on the specific objectives and specifications of the identifying outlier’s task, the best metric needs to be chosen. Other often-used metrics include precision, recall, and F1-Score. Precision is defined as the percentage of outliers among all samples that were accurately recognized to be outliers. The fraction of accurately detected outliers among all real outliers is known as recall, also known as the true positive rate or sensitivity. Table 2 provides the performance metrics used for the evaluation of the performance of the proposed method.
The efficiency of the proposed outlier detection method has been assessed in this section using various parameters. First, the data reduction is required to be performed on the given dataset and then classified into clusters by employing the KNN method. By setting the different parameter values for the proposed method to compute the degree of each data point for identifying the outliers, precision, recall, and F1-score are secondarily calculated for analyzing the performance on 8 randomly chosen datasets which include Diab, Audiology, Hepatitis, Cred, Heart, Lymp, Grem, and Monks. A thorough case study is also used to demonstrate the logic of the proposed technique. Additionally, we have compared our EODA method with the other existing state-of-the-art techniques to demonstrate the effectiveness of the proposed method in terms of outlier identification. In this study, precision (P), recall (R), and F1-Score are used as assessment measures to evaluate the performance of the EODA method. These measures enable us to evaluate the effectiveness and performance of the EODA algorithm in outlier detection objectively. Since most outlier detection techniques produce an outlier factor value for each sample, these values can be ordered downward and given sequence numbers beginning at 1. The initial position can be indicated by setting the currently selected sequence ID as t = 1. By considering the value of the outlier factor with t, we can demonstrate if a sample qualifies as an outlier. It is important to keep in consideration that selecting t too small or too large could lead to the misclassification of normal points as outliers or missing the significance of true outliers. We use the parameters P(t) and R(t) to determine the ideal value for t to solve this problem. We can determine the relevant precision (P) and recall (R) values by adjusting t. For a given t, the precision (P(t)) is the proportion of correctly detected outliers to all samples that have been categorized as outliers. The recall (R(t)) shows the percentage of all the true outliers that have been correctly recognized as outliers.
Let OS(t) be the collection of outliers identified by a certain threshold t, which may be considered as a function of t. The real set of outliers is referred to as OS. Additionally, P(t) and R(t) stand for the ratios of outliers that were successfully identified at the specified threshold t to the total number of true outliers (P(t)) and all true outliers (R(t)), respectively. The EODA technique removes the dense regions of each dataset using a data reduction approach. This approach becomes particularly significant when dealing with datasets that have many attributes, as it results in a higher number of attribute reductions. In comparison, for the Grem and Audiology datasets, the EODA approach demonstrates competitive performance slightly superior to that of CBLOF and LDCOF.
When evaluating outlier detection techniques using precision, the number of true outliers among the top-n ranked instances is considered. Consequently, the precision value can be low for datasets with a small proportion of outliers, such as Grem and Audiology. Conversely, the precision value might be high if the proportion of inliers is relatively low. To provide a comprehensive assessment of the overall performance of outlier detection techniques, a new approach called average precision (AP) has been introduced. Average precision takes into account the performance across a wide range of possible choices for n. Instead of evaluating at a single value of n, the value of precision is averaged over the ranks of all outlier points. This enables a more comprehensive evaluation of the effectiveness of outlier detection techniques. Experimental results of the proposed EODA method against existing outlier detection techniques on eight datasets are shown in Tables 3–10. The obtained results demonstrate the effectiveness of the proposed EODA algorithm compared to other existing methods across various datasets.
The proposed EODA incorporates several controllable hyperparameters across its three stages, significantly influencing its performance and adaptability. In the Boruta-RF feature selection stage, the primary hyperparameters include the number of decision trees (), which controls the accuracy and stability of feature importance estimation, and the maximum number of iterations (
), which determines the extent of exploration for feature relevance. Additionally, the threshold for Z-score comparison affects the sensitivity of feature selection by controlling the trade-off between retaining relevant features and removing noise. In the KNN clustering stage, key hyperparameters include K, the number of nearest neighbours, which determines the neighbourhood size and influences the clustering accuracy. Smaller values of K capture local structures but may lead to overfitting, while larger values provide smoother clustering results at the risk of underfitting. The distance metric (e.g., Euclidean or Manhattan) also plays a critical role, with the choice depending on the dataset’s characteristics, and the threshold for neighbourhood consideration (
) defines the maximum number of neighbours to evaluate, balancing local density and global spread.
The density function’s parameters are key hyperparameters in the outlier detection stage. The density function aggregates instance distances within a cluster, and its sensitivity can be adjusted through scaling or normalization to highlight local or global density variations. The distance weighting in the outlier score computation, which combines density and distance, also significantly impacts the detection of anomalies, offering flexibility to emphasize either highly dense or sparse regions. Furthermore, the number of outliers () selected from the sorted outlier scores is a critical parameter, directly influencing the number of anomalies detected. These hyperparameters can be tuned using methods such as grid search or cross-validation to ensure optimal performance, allowing EODA to adapt effectively to diverse datasets and anomaly detection tasks. This detailed discussion provides clarity on the controllable aspects of the proposed method and enhances its reproducibility.
A comparative study of the proposed EODA algorithm with existing algorithms in terms of the average precision, recall, and F1-Score is discussed in Tables 11, 12, and 13, respectively. The results show that the EODA algorithm consistently outperforms other outlier detection methods. In terms of precision (Table 11), which measures how accurately the algorithm identifies true outliers, EODA comes out on top with an average of 63.07%, slightly ahead of LDCOF at 61.14%. This means EODA does a better job of minimizing false positives. When it comes to recall (Table 12), which indicates how well the algorithm catches actual outliers, EODA performs strongly with an average score of 82.49%, just shy of LDCOF’s 83.83%. While LDCOF has a slight edge here, EODA remains highly reliable. The F1 score (Table 13), which balances precision and recall, highlights EODA’s strength, achieving the highest average score of 64.53%. This shows that EODA not only accurately detects outliers but also maintains consistency across different datasets. These results highlight the proposed method’s ability to handle outliers more effectively than existing techniques. Furthermore, the EODA algorithm provides a more balanced approach for detecting outliers compared to other algorithms.
To evaluate the effectiveness of different outlier detection techniques, we conducted a thorough analysis using eight diverse UCI datasets. The comparison focused on key performance metrics (precision, recall, and F1-Score) presented through bar charts and boxplots to visualize the performance distribution across the algorithms. Figs 3 and 4 highlight the comparative performance of the proposed EODA method against other methods. Fig 3 focuses on precision, demonstrating how EODA consistently surpasses existing methods, showing higher precision rates across all datasets. This indicates that EODA is more effective at minimizing false positives and accurately identifying true outliers. Fig 4 presents the recall rates, which measure how effectively the methods catch actual outliers. In Fig 5, rectangular and notched boxplots provide a more detailed data-specific comparison to visualize the distribution of performance metrics across each technique and each dataset. The notched boxplots reveal that EODA not only performs better on average but also maintains less variability in its performance, indicating robustness and stability across different types of datasets. The analysis shows that EODA provides a more balanced and effective approach to outlier detection.
5. Conclusions
In this paper, a three-sage efficient outlier detection approach, named EODA, has been presented. The proposed EODA algorithm follows a structured three-stage process to address the challenges of the existing techniques. In the first stage, the Boruta method combined with a random forest ensemble classifier is used for feature selection, to extract all relevant features from the dataset. This ensures that only the most important attributes are considered for subsequent analysis, reducing noise and improving the quality of the data. In the second stage, the selected features are clustered using the KNN algorithm, which groups the data based on similarities. Finally, the third stage introduces an efficient outlier detection mechanism that identifies the most significant outliers within the clustered data. This stage also computes the degree of outliers for each data point, ensuring a precise and accurate identification of anomalies.
The effectiveness of the proposed EODA algorithm has been evaluated on eight UCI datasets through several key performance metrics including precision, recall, detection rate, accuracy, and F1-Score. These metrics demonstrate the method’s ability to not only detect outliers but also to do so with a high level of accuracy and reliability across different datasets. Compared to existing methods, EODA shows promising results, making it a robust tool for outlier detection in various domains.
Future research can concentrate on feature selection and clustering strategies to improve the efficiency and scalability of outlier detection. This may entail creating complicated feature selection algorithms that take a variety of data types into account and incorporate domain-specific information. Anomaly detection accuracy can also be greatly increased by analyzing advanced clustering techniques that may identify complicated data structures and take noise and outliers into consideration. The progress of outlier detection approaches made possible by these research initiatives will allow for the more robust and accurate detection of anomalies across a range of application domains. As another future work, we will prioritize addressing the current limitations by expanding our evaluation to include real-world and high-dimensional datasets, ensuring broader validation and practical applicability of the proposed approach.
References
- 1.
Hawkins DM. Identification of Outliers. 1980. Vol 11. London: Chapman and Hall.
- 2. Yang J, Chen Y, Rahardja S. Neighborhood representative for improving outlier detectors. Inform Sci. 2023;625:192–205.
- 3. Djenouri Y, Belhadi A, Lin JC-W, Djenouri D, Cano A. A survey on urban traffic anomalies detection algorithms. IEEE Access. 2019;7:12192–205.
- 4. Shokouhifar M, Fanian F, Kuchaki Rafsanjani M, Hosseinzadeh M, Mirjalili S. AI-driven cluster-based routing protocols in WSNs: a survey of fuzzy heuristics, metaheuristics, and machine learning models. Comput Sci Rev. 2024;54:100684.
- 5. Shen L., Li Z., Kwok J. Timeseries anomaly detection using temporal hierarchical one-class network Adv Neural Inform Process Syst 2020;33:13016–13026.
- 6. Pang G, Shen C, Cao L, Hengel AVD. Deep learning for anomaly detection. ACM Comput Surv. 2021;54(2):1–38.
- 7. Ruff L, Kauffmann JR, Vandermeulen RA, Montavon G, Samek W, Kloft M, et al. A unifying review of deep and shallow anomaly detection. Proc IEEE. 2021;109(5):756–95.
- 8.
Sattler F, Müller KR, Wiegand T, Samek W. On the byzantine robustness of clustered federated learning. In ICASSP IEEE International Conference on Acoustics, Speech and Signal Processing. 2020. IEEE; 8861–8865.
- 9. Zhang P, Li T, Wang G, Wang D, Lai P, Zhang F. A multi-source information fusion model for outlier detection. Inform Fus. 2023;93:192–208.
- 10. Ke W, Wei J, Xiong N, Hou Q. GSS: a group similarity system based on unsupervised outlier detection for big data computing. Inform Sci. 2023;620:1–15.
- 11. Abhaya A, Patra BK. An efficient method for autoencoder based outlier detection. Expert Syst Appl. 2023;213:118904.
- 12. Song Y, Lin H, Li Z. Outlier detection in a multiset-valued information system based on rough set theory and granular computing. Inform Sci. 2024;657:119950.
- 13. Du X, Chen J, Yu J, Li S, Tan Q. Generative adversarial nets for unsupervised outlier detection. Expert Syst Appl. 2024;236:121161.
- 14. Zhao J, Liu K, Wang W, Liu Y. Adaptive fuzzy clustering based anomaly data detection in energy system of steel industry. Inform Sci. 2014;259:335–45.
- 15. Yang J, Rahardja S, Fränti P. Mean-shift outlier detection and filtering. Patt Recogn. 2021;115:107874.
- 16. He Z, Xu X, Deng S. Discovering cluster-based local outliers. Patt Recogn Lett. 2003;24(9–10):1641–50.
- 17.
Amer M, Goldstein M. Nearest-neighbor and clustering based anomaly detection algorithms for rapidminer. In Proc. of the 3rd RapidMiner Community Meeting and Conference (RCOMM 2012). 2012;1–12.
- 18. Chen H, Liang M, Liu W, Wang W, Liu PX. An approach to boundary detection for 3D point clouds based on DBSCAN clustering. Patt Recogn. 2022;124:108431.
- 19. Garg S, Kaur K, Batra S, Kaddoum G, Kumar N, Boukerche A. A multi-stage anomaly detection scheme for augmenting the security in IoT-enabled applications. Fut Gener Comput Syst. 2020;104:105–18.
- 20. Wang X, Pan Y, Li M, Chen J. A novel data-driven optimization framework for unsupervised and multivariate early-warning threshold modification in risk assessment of deep excavations. Expert Syst Appl. 2024;238:121872.
- 21.
Breunig MM, Kriegel HP, Ng RT, Sander J. LOF: identifying density-based local outliers. In Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data. 2000;93–104.
- 22. Li Z, Zhao Y, Hu X, Botta N, Ionescu C, Chen G. Ecod: unsupervised outlier detection using empirical cumulative distribution functions. IEEE Trans Knowl Data Eng. 2022.
- 23.
Pokrajac D, Lazarevic A, Latecki LJ. Incremental local outlier detection for data streams. In 2007 IEEE Symposium on Computational Intelligence and Data Mining. 2007:504–15. https://doi.org/10.1109/cidm.2007.368917
- 24. Salehi M, Leckie C, Bezdek JC, Vaithianathan T, Zhang X. Fast memory efficient local outlier detection in data streams. IEEE Trans Knowl Data Eng. 2016;28(12):3246–60.
- 25.
Na GS, Kim D, Yu H. Dilof: effective and memory efficient local outlier detection in data streams. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2018;1993–2002.
- 26.
Schubert E, Zimek A, Kriegel HP. Generalized outlier detection with flexible kernel density estimates. In Proceedings of the 2014 SIAM International Conference on Data Mining. 2014. Society for Industrial and Applied Mathematics; 542–550.
- 27. Hilal W, Gadsden SA, Yawney J. Financial fraud: a review of anomaly detection techniques and recent advances. Expert Syst Appl. 2022;193:116429.
- 28. Tang B, He H. A local density-based approach for outlier detection. Neurocomputing. 2017;241:171–80.
- 29. Kumar S, Jailia M, Varshney S. An efficient approach for highway lane detection based on the Hough transform and Kalman filter. Innov Infrastruct Solut. 2022;7(5).
- 30. Sánchez Pedroche D, García Herrero J, Molina López JM. Context learning from a ship trajectory cluster for anomaly detection. Neurocomputing. 2024;563:126920.
- 31.
Jin W. Tung A. K. Han J. Mining top-n local outliers in large databases. In Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2001;293–298.
- 32. Olszewski D, Iwanowski M, Graniszewski W. Dimensionality reduction for detection of anomalies in the IoT traffic data. Future Gener Comput Syst. 2024;151:137–51.
- 33. Kumar S, Jailia M, Varshney S, Pathak N, Urooj S, Elmunim NA. Robust vehicle detection based on improved you look only once. Comput Mater Continua. 2023;74(2).
- 34. Wang G, Chen Y. Robust feature matching using guided local outlier factor. Patt Recogn. 2021;117:107986.
- 35. Schubert E, Zimek A, Kriegel H-P. Local outlier detection reconsidered: a generalized view on locality with applications to spatial, video, and network outlier detection. Data Min Knowl Disc. 2012;28(1):190–237.
- 36.
Tang J, Chen Z, Fu AWC, Cheung DW. Enhancing effectiveness of outlier detections for low density patterns. In Advances in Knowledge Discovery and Data Mining: 6th Pacific-Asia Conference, PAKDD 2002 Taipei, Taiwan, May 6–8, 2002 Proceedings, vol 6;535–548. Berlin, Heidelberg: Springer. 2002. 6th Pacific-Asia Conference, PAKDD 2002.
- 37. López-Oriona Á, Vilar JA. Outlier detection for multivariate time series: a functional data approach. Knowl-Based Syst. 2021;233:107527.
- 38.
Angiulli F, Pizzuti C. Fast outlier detection in high dimensional spaces. In European Conference on Principles of Data Mining and Knowledge Discovery. 2002. Berlin, Heidelberg: Springer Berlin Heidelberg; 15–27.
- 39.
Tang J, Chen Z, Fu AWC, Cheung DW. Enhancing effectiveness of outlier detections for low density patterns. In Advances in Knowledge Discovery and Data Mining: 6th Pacific-Asia Conference, PAKDD 2002 Taipei, Taiwan, May 6–8, 2002 Proceedings, 2002:6. Springer Berlin Heidelberg;535–548.
- 40.
Jin W, Tung AK, Han J, Wang W. Ranking outliers using symmetric neighborhood relationship. In Advances in Knowledge Discovery and Data Mining: 10th Pacific-Asia Conference, PAKDD 2006, Singapore, April 9–12, 2006. Proceedings, 2006;10. Springer Berlin Heidelberg, pp. 577–593.
- 41. Zhang L, Lin J, Karim R. Adaptive kernel density-based anomaly detection for nonlinear systems. Knowl-Based Syst. 2018;139:50–63.
- 42. He Z, Xu X, Deng S. Discovering cluster-based local outliers. Patt Recogn Lett. 2003;24(9–10):1641–50.
- 43.
Amer M, Goldstein M. Nearest-neighbor and clustering based anomaly detection algorithms for RapidMiner. In Proc. of the 3rd RapidMiner Community Meeting and Conference (RCOMM 2012). 2012;1–12.
- 44. Mahesh Kumar K, Rama Mohan Reddy A. A fast DBSCAN clustering algorithm by accelerating neighbor searching using Groups method. Patt Recogn. 2016;58:39–48.
- 45. Lv Y, Ma T, Tang M, Cao J, Tian Y, Al-Dhelaan A, et al. An efficient and scalable density-based clustering algorithm for datasets with complex structures. Neurocomputing. 2016;171:9–22.
- 46. Kumar S, Pandey A, Varshney S. Exploring the impact of deep learning models on lane detection through semantic segmentation. SN Comput Sci. 2024;5(1).
- 47. Kumar S, Singh SK, Varshney S, Singh S, Kumar P, Kim B-G, et al. Fusion of Deep Sort and Yolov5 for effective vehicle detection and tracking scheme in real-time traffic management sustainable system. Sustainability. 2023;15(24):16869.
- 48.
McInnes L, Healy J. Accelerated Hierarchical Density Based Clustering. In 2017 IEEE International Conference on Data Mining Workshops (ICDMW). 2017:33–42. https://doi.org/10.1109/icdmw.2017.12
- 49.
Smith J. Nouretdinov I. Craddock R. Offer C. Gammerman A. Anomaly detection of trajectories with kernel density estimation by conformal prediction. In Artificial Intelligence Applications and Innovations: AIAI 2014 Workshops: CoPA, MHDW, IIVC, and MT4BD, Rhodes, Greece, September 19-21, 2014. 2014. Springer Berlin Heidelberg; 271–280.
- 50. Li J, Zhang J, Pang N, Qin X. Weighted outlier detection of high-dimensional categorical data using feature grouping. IEEE Trans Syst Man Cybern Syst. 2020;50(11):4295–308.
- 51. Huang L, Ran J, Wang W, Yang T, Xiang Y. A multi-channel anomaly detection method with feature selection and multi-scale analysis. Comput Netw. 2021;185:107645.
- 52. Saberi-Movahed F, Rostami M, Berahmand K, Karami S, Tiwari P, Oussalah M, et al. Dual regularized unsupervised feature selection based on matrix factorization and minimum redundancy with application in gene selection. Knowl-Based Syst. 2022;256:109884.
- 53.
Saberi-Movahed F, Berahman K, Sheikhpour R, Li Y, Pan S. Nonnegative Matrix Factorization in Dimensionality Reduction: A Survey. 2024. arXiv preprint arXiv:2405.03615. https://arxiv.org/abs/2405.03615
- 54. Sheikhpour R, Berahmand K, Mohammadi M, Khosravi H. Sparse feature selection using hypergraph Laplacian-based semi-supervised discriminant analysis. Patt Recogn. 2025;157:110882.
- 55. Shokouhifar M, Hasanvand M, Moharamkhani E, Werner F. ensemble heuristic–metaheuristic feature fusion learning for heart disease diagnosis using tabular data. Algorithms. 2024;17(1):34.
- 56. UCI Machine Learning Repository, http://archive.ics.uci.edu/ml/datasets