Next Article in Journal
Deep Link-Prediction Based on the Local Structure of Bipartite Networks
Previous Article in Journal
Weighted Similarity and Core-User-Core-Item Based Recommendations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cluster-Based Improved Isolation Forest

School of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
*
Author to whom correspondence should be addressed.
Submission received: 12 April 2022 / Revised: 24 April 2022 / Accepted: 24 April 2022 / Published: 27 April 2022

Abstract

:
Outlier detection is an important research direction in the field of data mining. Aiming at the problem of unstable detection results and low efficiency caused by randomly dividing features of the data set in the Isolation Forest algorithm in outlier detection, an algorithm CIIF (Cluster-based Improved Isolation Forest) that combines clustering and Isolation Forest is proposed. CIIF first uses the k-means method to cluster the data set, selects a specific cluster to construct a selection matrix based on the results of the clustering, and implements the selection mechanism of the algorithm through the selection matrix; then builds multiple isolation trees. Finally, the outliers are calculated according to the average search length of each sample in different isolation trees, and the Top-n objects with the highest outlier scores are regarded as outliers. Through comparative experiments with six algorithms in eleven real data sets, the results show that the CIIF algorithm has better performance. Compared to the Isolation Forest algorithm, the average AUC (Area under the Curve of ROC) value of our proposed CIIF algorithm is improved by 7%.

1. Introduction

Outlier detection is an important research direction in the field of data mining, which aims to uncover the unusual data present in a dataset [1,2]. The most widespread definition of an outlier is that proposed by Hawkins [3]. Outliers are those data objects that deviate from most of the data set, raising the suspicion that these deviations are not generated by random factors, but by a completely different mechanism. The main reasons for outliers are anomalies in the data itself and errors caused by the collection of data.
Isolation Forest is an unsupervised detection method specially designed based on the isolation of outliers [4]. The method isolates outliers by splitting the data space through a random hyper plane, reflecting the characteristic that outliers are easily isolated. With high accuracy and low computational complexity, this method is widely used in the industry. However, the Isolation Forest uses a completely random selection of features and feature values when constructing isolation trees, and the overly random selection leads to a possible invalid selection of feature values, resulting in divided features as interference features and affecting the detection results.
In response to the limitations of the Isolation Forest method, this paper proposes an algorithm CIIF that combines clustering and Isolation Forest. First, the proposed method clusters the dataset using the k-means method [5] and constructs a selection matrix based on the results of the clustering. Then, the process of isolation trees construction splits the sample set using a selection matrix, which can effectively avoid the error caused by the defects of the traditional Isolation Forest. Finally, outliers are calculated based on the average search length of each sample in each decision tree, and the n samples with the highest outliers are listed as outliers.
Our main contributions are summarized as follows:
  • The proposed method introduces a pre-selection mechanism to improve the shortcomings of Isolation Forest which are the unstable detection results and the low efficiency caused by randomly dividing features of the dataset.
  • The proposed method uses the k-means algorithm to obtain the distribution of the dataset, which is used to construct a selection matrix to implement a pre-selection mechanism.
  • The proposed method introduces the parameter selection degree I to control the influence of the pre-selection mechanism on the method and avoid overfitting.
The methods for outlier detection can be classified into distribution-based methods, nearest-neighbor-based methods, clustering-based methods, neural network-based methods, classification-based methods, and isolation-based methods.
The distribution-based outlier detection algorithm is one of the first proposed algorithms, whose main idea is to assume that the data distribution of a dataset fits a statistical model and define outliers as those points that are in the low probability region [6]. Classic representative models include the Gaussian distribution model [7,8,9,10], etc.
The outlier detection algorithm based on nearest neighbors is to detect outliers based on the relationship between all data and their nearest neighbors. This class of methods can be divided into two categories: distance-based methods [11,12,13] and density-based methods [14,15,16,17]. Classic representative algorithms are the KNN (K-Nearest-Neighbor) algorithm [18] based on distance and the LOF (Local Outlier Factor) algorithm based on density [19].
The clustering-based outlier detection algorithm [20] is an unsupervised algorithm whose main idea is to detect outliers by analyzing the relationship between data points and clusters, which has good results for most data sets [21,22,23]. The disadvantage of the clustering-based outlier detection algorithm is that the main purpose of the algorithm is to obtain the distribution characteristics of the dataset, and the detection efficiency for outlier points is not optimal, and the model needs to be adjusted according to the actual application, so it cannot be flexibly applied to different datasets. The DBSCAN (Density-Based Spatial Clustering of Applications with Noise) algorithm is the representative of this type of method [24,25,26].
The classification-based outlier detection algorithm trains a classifier from a labeled dataset and uses this classifier to detect outliers [27]. The algorithm also has a disadvantage. When the amount of data in the training dataset is insufficient, the efficiency and accuracy of the trained classifier will fall as expected.
With the development of deep learning techniques, neural network-based methods [28,29,30] have also advanced. This type of method has high detection accuracy and good performance on different types of datasets. However, the models for this type of approach are usually more complex and require a lot of time to train the model. Some of the popular methods are as follows: Autoencoder Ensemble [31,32,33], GAN (Generative Adversarial Network)-based model [34,35,36], Graph neural network [37,38,39,40], etc.
The isolation-based outlier detection method defines data that can easily be isolated as outliers [41,42,43]. Isolation Forest is the representative of this type of method, which constructs multiple isolation trees by splitting the sample space through hyper planes. These isolation trees are completely random in the selection of attributes and split values each time during the construction process. These isolation trees constitute the Isolation Forest. The Isolation Forest algorithm defines those points that are easily isolated as outliers, which tend to be the leaf nodes closest to the root node in the isolation trees. These outliers are too different from other samples in the sample space and are far from the distribution center of the sample. Therefore, we can locate potential outliers by calculating the average finding length of sample points in the entire forest.

2. Materials and Methods

The Isolation Forest algorithm can cause the constructed isolation tree to fail to accurately reflect the difference between normal and outlier points due to the random selection of split values in the process of constructing the isolation tree, which finally affects the detection results.
As shown in Figure 1, the blue asterisk indicates normal data, and the red asterisk indicates an outlier, due to the random selection of the split value, normal data may be more likely to be isolated than an outlier.
To improve the shortcomings of the IF, CIIF introduces a pre-selection mechanism. The main idea is to select a suitable cluster based on the data distribution of the dataset, and preferentially select the boundary and center of that cluster as the split values, which is shown in Figure 2:
As shown in Figure 2, the outliers are isolated more accurately in CIIF.
CIIF is divided into two phases, the training phase, and the evaluation phase. The training phase is divided into two steps, firstly, the construction of the selection matrix, and then the construction of the Isolation Forest.

2.1. Training Phase

2.1.1. Selection Matrix

The CIIF algorithm is an unsupervised outlier detection algorithm that analyzes the distribution of the dataset through the k-means clustering algorithm, divides the dataset into k clusters, and selects the appropriate cluster as the selection cluster Cs.
Definition 1.
selection cluster Cs
Let the dataset X be divided into k clusters C1, C2, …, Ck by the k-means clustering method, and each cluster is scored as follows:
S c o r e c i = j = 1 m i d i s t c i x j n i
where dist(ci,xj) is the Euclidean distance, n1, n2, … nk are the amount of data contained in each cluster, and c1, c2, … ck are the cluster centers of each cluster.
Define the cluster with the lowest score as the selection cluster Cs.
The choice of selection cluster Cs directly affects the performance of the whole algorithm. Different clusters as the selection cluster will lead to a large difference in the results of the algorithm. Clusters with large data-to-data differences as the selection cluster can seriously degrade the performance of the algorithm, and clusters with larger amounts of data are more suitable as the selection cluster than those with smaller amounts of data. Therefore, it is necessary to score each cluster to determine the best choice of selection cluster.
The selection matrix S is built based on the selection cluster.
Definition 2.
selection matrix S
Let the dimension of the dataset be d, and define the selection matrix as:
S = S 1 , S 2 , S d
Define the maximum value of the selection cluster Cs in dimension d as Max(cd), the minimum value as Min(cd), and the average value as Mean(cd), then:
s d = { M a x ( c ) , M i n ( c ) , M e a n ( c ) }
The construction process of matrix S is shown in Algorithm 1.
Algorithm 1.Get-S(D, k)
 Input: D-input data, k-number of clusters
 Output: selection matrix S
  • InitializeS
  • Ck-means(D, k)//Cluster the dataset, return the result C = {C1,C2,…,Ck}
  • fori = 1 to k do
  •   scoreiscoreiScore(Ci);//Score each cluster
  • end for
  • s ← argmin(scorei)//Get the serial number of the selection cluster
  • n ← size(Cs);//Cs = {x1,x2,…,xn}, xn = {nn1,nn2,…,nnd}, d-number of dimensions
  • /* Construct the selection matrix S*/
  • forI = 1 to d do
  •   SamΦ
  •   for j = 1 to n do
  •     Sam ← Sam∪nij
  •   end for
  •   si ← {Min(Sam), Max(Sam), Mean(Sam)}
  •   SSsi;
  • end for
  • returnS
The selection matrix S is the set of data boundaries and means of the selection cluster Cs in each attribute, which reflects the distribution characteristics of the selection cluster. The CIIF algorithm implements the selection mechanism for split value selected by the selection matrix when constructing the forest. In this process, the selection of the split values will be prioritized from the optional points of the selection matrix S in that attribute.

2.1.2. Isolation Forest

Set the selection degree I to control the degree of influence of the selection matrix S on the algorithm.
Definition 3.
selection degree I
The selection degree I is defined as the maximum number of times that split value can be selected by the selection matrix S in each attribute.
The degree of selection is a parameter that controls the randomness of the algorithm and is determined artificially. The larger the value of I, the more the forest is influenced by the selection cluster Cs and the lower the randomness; the smaller the value of I, the less the forest is influenced by the selection cluster Cs, the greater the randomness, and the closer it is to the original Isolation Forest algorithm; when the selection degree I is 0, the algorithm is the original Isolation Forest algorithm at this time.
Definition 4.
discriminant matrix J
Define the discriminant matrix J as the record of the number of split values decided by each dimension according to the selection matrix S during the construction of the Isolation Forest by CIIF:
J ( d ) = i
where d represents the dimension and i is the record value. Equation (4) indicates that CIIF performs i times split value selection for d dimensions in constructing the Isolation Forest.
The isolation tree is the core of the whole CIIF algorithm. To construct the isolation tree, we first select a subsample from the sample space, use the subsample as the root node of the isolation tree, then randomly select an attribute, choose a value from the candidate values of the selection matrix S in the range of the subsample in the selected attribute, use the value as the split value, and update the record of the selected attribute in the discriminant matrix; If the candidate values of the selection matrix in the selected attribute are not in the range of the subsample or the record of the selected attribute in the discriminant matrix J is greater than the selection degree I, then randomly select a value as the split value in the range of the subsample.
The subsample space is divided into two subspaces according to the split value, and the data with value less than the split value in the selected attribute are grouped in the left subspace, and the data with value greater than the split value are grouped in the right subspace, and the two subspaces are the two subtrees of the root node. The above process is repeated recursively for both subtrees until the leaf nodes contain only one data, or all the data in the leaf nodes have the same value; or the height of the tree exceeds the limit, at which point the isolation tree construction is completed. The construction process of the isolation tree is shown in Algorithm 2.
Algorithm 2. iTree(D, l, L, S, J)
Input: D-input data, l-current tree height, L-height limit, S-selection matrix, J-discriminant matrix
Output: an iTree
  • get selection degree I
  • ifl > Lor|x| <= 1 then
  •   return exNode
  • else
  •   let Q be a list of attributes in D
  •   randomly select an attribute qQ
  •   let k be the serial number of q in D
  •   s ← {x | xS(k, : ),min <= x <= max}
  •   if sΦ and J(k) < I then
  •     randomly select a split point p from s
  •     J(k) ← J(k) + 1
  •   else
  •     randomly select a split point p from max and min values of attribute q in D
  •   end if
  •   Dlfilter(D, q < p)
  •   Drfilter(D, q > =p)
  •   return inNode{LeftiTree(Dl, l + 1, L, S, J),
  •     RightiTree(Dr, l + 1, L, S, J),
  •     SplitAttq,
  •     SplitValuep}
  • end if
Construct multiple isolation trees to form an isolation forest, the construction process of an isolation forest is shown in Algorithm 3.
Algorithm 3.iForest(D, t, X)
Input: D-input data, t-number of isolation trees, X-subsampling size
 Output: a set of iTrees
  • InitializeForest
  • InitializeJ
  • set height limit L = ceiling(log2X)
  • SGet-S(D, k) //get the selection matrix
  • forI = 1 to t do
  •   D’Sample(D,X)
  •   ForestForestiTree(D’, 0, L, S, J)
  • end for
  • returnForest

2.2. Evaluation Phase

After the training phase, the proposed method will calculate the outlier scores of all data points in the isolation forest with the following outlier score calculation formula:
s c o r e ( x , n ) = 2 E ( h ( x ) ) c ( n )
where h(x) is the path length of sample x from the root node to the leaf node where it is located, E(h(x)) is the expectation of path length h(x) in an isolated forest, and c(n) is the average of the path lengths of all data points, calculated as follows:
c ( n ) = 2 H ( n 1 ) 2 ( n 1 ) n
where H(i) is the Harmonic series, which can be calculated as ln(i) + γ, and γ is the Euler’s constant, which is approximately equal to 0.5772156649.
When E(h(x)) tends to 0, the outlier score tends to 1, and the data point x is judged to be an outlier. On the contrary, if the score tends to 0, the data point x will be judged as a normal point. When the score tends to 0.5, it is not possible to determine whether the data point x is an outlier.
This algorithm has two stages, the first stage is to construct the selection matrix and the second stage is the improved Isolation Forest algorithm. The first stage clusters the data set by the k-means algorithm, and the selection matrix is constructed according to the clustering results. The computational complexity of computing the Euclidean distance of the data set is O(n2), the computational complexity of the k-means algorithm is O(n), and the computational complexity of constructing the selection matrix is O(n), so the computational complexity of the first stage is O(n2). The second stage is the improved Isolation Forest algorithm with linear computational complexity. Thus, the computational complexity of the whole improved algorithm is O(n2).

3. Results

3.1. Subsection

To verify the effectiveness of the algorithm, experiments were conducted on 11 different publicly available real datasets from UCI and ODDS [44,45,46,47,48,49,50,51,52,53,54], and the AUC value was used as the Accuracy Metric of the algorithm.
The specific attributes of datasets are shown in Table 1. The breastw dataset is the Wisconsin breast cancer diagnosis dataset, which is a high-dimensional dataset publicly available at UCI and contains diagnostic data for malignant and benign tumors. The diagnostic data for malignant tumors are labeled as outliers. The annthyroid dataset is a thyroid disease dataset, which is divided into two categories: noisy and normal, and the noisy data are labeled as outliers. The arrhythmia dataset, which is a cardiac arrhythmia dataset, divides the data into multiple categories, the eight categories with less data are labeled as outliers. The pima dataset is an Indian diabetes dataset, divided into two categories: abnormal and normal; the abnormal data are labeled as outliers. The vertebral dataset is a genomic dataset with six dimensions, classifying data into normal and abnormal categories, the abnormal data are labeled as outliers. The wine dataset is a dataset of results of chemical analyses of wines made from three different grapes from the same region of Italy, which identified the number of 13 components contained in the three wines, the data for one of the wines are labeled as outliers. The ionosphere dataset is a binary dataset with 34 dimensions, classifying the data into bad and good classes, removing an invalid attribute, the bad class data are labeled as outliers. The shuttle dataset is the flight data of the aircraft; the data are divided into two categories, the data of the category with smaller number are labeled as outliers. The cardio dataset is the fetal heart rate measurements on the ECG that have been processed by a professional physician. The data are divided into three categories: normal, suspicious, and pathological, with the suspicious category discarded and the pathological category labeled as outliers.

3.2. Evaluation Metric

For a binary classification algorithm, data samples can be classified into four categories based on the classification results and true labels: True Positive (TP), False Positive (FP), True Negative (TN), and False Negative (FN), as shown in Table 2.
Area under the Curve of ROC (AUC) is the value of the area between the Receiver Operating Characteristic (ROC) curve and the horizontal coordinate. The ROC curve is a curve on a two-dimensional plane with the horizontal coordinate of the false-positive rate (FPR) and the vertical coordinate of the true-positive rate (TPR). The formula for calculating FPR and TPR is as follows:
TPR = TP TP + FT
FPR = FP TN + FP
AUC is calculated as:
AUC = i p r a n k i p M ( M + 1 ) 2 M N
where ip denotes a positive sample, rank is the sample serial number, M is the number of positive samples, and N is the number of negative samples. the AUC value is generally between 1 and 0.5, and the closer the AUC value is to 1, the better the performance of the algorithm. If the AUC value is below 0.5, the algorithm is not applicable to the detection dataset.

3.3. Experimental Results

Six typical outlier detection algorithms are used as comparison algorithms with the proposed CIIF to compare the AUC values and computational times on 11 datasets. The six comparison algorithms are Isolation Forest, LOF, KNN, COF (Connectivity-based Outlier Factor) [55], FastABOD (Fast Angle-Based Outlier Detection) [56], and LDOF (Local Distance-based Outlier Factor) [57].
Table 3 shows the AUC values of each algorithm on the 11 datasets and highlights the best AUC value with the second-highest AUC value on each dataset. By comparing the AUC values of each algorithm, we can see that the CIIF performs well on the annthyroid, arrhythmia, speech, vertebral, wine, and ionosphere datasets, and has a significant improvement compared to the Isolation Forest and other comparison algorithms; outperforms other comparative algorithms on thyroid and shuttle dataset, with less difference compared to the Isolation Forest; outperforms the original Isolation Forest algorithm and other comparison algorithms on the breastw dataset, and differs less from the COF; outperforms the Isolation Forest and other comparison algorithms on the pima dataset, and differs less from FastABOD. The performance on the cardio dataset is slightly worse than the original Isolation Forest algorithm and COF algorithm, and less different from the KNN algorithm.
Figure 3 shows the ROC curves of CIIF and other comparison algorithms on 11 datasets. The ROC curves of the proposed algorithm on the annthyroid, pima, thyroid, vertebral, wine, and cardio datasets are above the other algorithms; The ROC curves of the proposed algorithm on the shuttle, breast, and pima datasets nearly overlap with those of the Isolation Forest and are higher than those of other algorithms. On the speech dataset, CIIF does not work as well as LDOF. On the ionosphere dataset, CIIF does not work as well as LDOF, KNN and LOF. The results of comparing six state-of-the-art algorithms on eleven real-world datasets show that CIIF achieves the highest Area under ROC Curve (AUC) on nine datasets. Thus, the CIIF outperforms the IF and the other comparison algorithms in overall performance.
As shown in Figure 4, the proposed algorithm has a higher computational time than LOF, KNN, COF, FastABOD, LDOF on datasets with smaller datasets. From Figure 5, the difference between the computational time of LOF, KNN, COF, LDOF, and the CIIF is not significant on the datasets with larger data volume such as shuttle data set, and the computational time of FastABOD is even much higher than the proposed algorithm. The computational time of the Isolation Forest is smaller than that of the CIIF on each dataset, but the CIIF has higher AUC values and better detection results on most of the datasets.
The experimental results show that CIIF has good detection performance on most of the datasets; the computational time on the datasets with less data is slightly higher than other algorithms, but within the acceptable range; the computational time on the datasets with larger data is smaller or not much different compared to other algorithms. Therefore, the CIIF is effective and feasible.

3.4. Parameter Analysis

The effect of selection degree I, number of subsampling X, on the proposed algorithm was analyzed experimentally on annthyroid, arrhythmia, pima, ionosphere, shuttle, and cardio dataset.

3.4.1. Effect of Selectivity I

Experiments were conducted with different I on six datasets. The range of I was set to integers from 1 to 10, because the proposed algorithm is no different from the Isolation Forest when I is less than 1, and the AUC values tend to be smooth when I is greater than 10. The experimental results are shown in Figure 6.
From the results, the influence of different I on each dataset is small. When I = 2, the AUC values reach the optimal value on each dataset and then decrease. Therefore, in the CIIF, the value of the selection degree I is generally set to 2, so that the CIIF can achieve the optimal performance on most dataset.

3.4.2. Effect of Selectivity Subsampling X

Experiments were conducted on the datasets with large data such as annthyroid, speech, and shuttle, to explore the effect of subsampling X on the results. Because the subsampling number is too large for datasets with smaller data, they will no longer be subsampled, but use all the data to construct isolation trees, which eventually leads to all isolation trees in the isolation forest being constructed from the same set of samples.
From Figure 7, the detection performance of the CIIF on the shuttle dataset increases with the subsampling X and reaches the best at subsampling of 256. The detection performance on the annthyroid and speech datasets starts to level off at subsampling of 256, before which there are large fluctuations in detection performance. When the subsampling is too small, the detection performance of the CIIF is poor and unstable; when the subsampling is too large, the sample set will contain too many normal samples, leading to a certain degradation in performance and leading to greater time cost. Therefore, the best comprehensive performance of the proposed algorithm is achieved when the subsampling is 256.
IF takes a completely random selection of attributes and split values in the training process, ignoring the distribution characteristics of the dataset itself, so, as a result, the constructed isolation forest cannot accurately reflect the isolation of each sample, resulting in a decrease in detection accuracy. CIIF takes account of the distribution characteristics of the dataset and performs a heuristic training process based on these characteristics, resulting in better performance.

4. Conclusions

In this paper, we propose an improved isolation forest algorithm, which constructs a selection matrix to realize the pre-selection mechanism of attribute values for isolation forest division by clustering and analyzing the data distribution of the dataset, avoiding the problem of low accuracy caused by too much randomness of the Isolation Forest. Experiments on 11 datasets on UCI and ODDS verified the effectiveness of the algorithm. In the experiments, it was found that the performance of the k-means algorithm is too greatly affected by the given k value, and there is no less lossy way to select the appropriate k value for a data set, and the result of the clustering algorithm directly affects the performance of the CIIF, so the next step is to study the effect of different clustering algorithms on the CIIF and the improvement of k-means algorithm.

Author Contributions

Conceptualization, C.S. and X.D.; methodology, C.S. and X.D.; software, J.C.; validation, X.D., J.Y. and J.C.; formal analysis, C.S.; investigation, C.S.; resources, J.Y.; data curation, J.Y.; writing—original draft preparation, J.C.; writing—review and editing, C.S.; visualization, J.Y.; supervision, X.D.; project administration, C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China under grants 61862060, 61462079, 61562086, and 61562078.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Boukerche, A.; Zheng, L.; Alfandi, O. Outlier detection: Methods, models, and classification. ACM Comput. Surv. 2020, 53, 1–37. [Google Scholar] [CrossRef]
  2. Wang, H.; Bah, M.J.; Hammad, M. Progress in outlier detection techniques: A survey. IEEE Access 2019, 7, 107964–108000. [Google Scholar] [CrossRef]
  3. Hawkins, D.M. Identification of Outliers; Chapman and Hall: London, UK, 1980. [Google Scholar]
  4. Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar]
  5. Krishna, K.; Murty, M.N. Genetic K-means algorithm. IEEE Trans. Syst. Man Cybern. Part B 1999, 29, 433–439. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Ben-Gal, I. Outlier detection. In Data Mining and Knowledge Discovery Handbook; Springer: Boston, MA, USA, 2005; pp. 131–146. [Google Scholar]
  7. Wang, H.; Li, H.; Fang, J.; Wang, H. Robust Gaussian Kalman filter with outlier detectio. IEEE Signal Process. Lett. 2018, 25, 1236–1240. [Google Scholar] [CrossRef]
  8. Liao, W.; Guo, Y.; Chen, X.; Li, P. A unified unsupervised gaussian mixture variational autoencoder for high dimensional outlier detection. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 1208–1217. [Google Scholar]
  9. Wang, B.; Mao, Z. Outlier detection based on Gaussian process with application to industrial processes. Appl. Soft Comput. 2019, 76, 505–516. [Google Scholar] [CrossRef]
  10. Dwivedi, R.K.; Kumar, R.; Buyya, R. Gaussian distribution-based machine learning scheme for anomaly detection in healthcare sensor cloud. Int. J. Cloud Appl. Comput. 2021, 11, 52–72. [Google Scholar] [CrossRef]
  11. Pang, G.; Cao, L.; Chen, L.; Liu, H. Learning representations of ultrahigh-dimensional data for random distance-based outlier detection. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 2041–2050. [Google Scholar]
  12. Ahn, J.; Lee, M.H.; Lee, J.A. Distance-based outlier detection for high dimension, low sample size data. J. Appl. Stat. 2019, 46, 13–29. [Google Scholar] [CrossRef]
  13. Wahid, A.; Rao, A.C.S. A distance-based outlier detection using particle swarm optimization technique. In Information and Communication Technology for Competitive Strategies; Springer: Singapore, 2019; pp. 633–643. [Google Scholar]
  14. Su, S.; Xiao, L.; Ruan, L.; Gu, F.; Li, S.; Wang, Z.; Xu, R. An efficient density-based local outlier detection approach for scattered data. IEEE Access 2018, 7, 1006–1020. [Google Scholar] [CrossRef]
  15. Boddy, A.J.; Hurst, W.; Mackay, M.; El Rhalibi, A. Density-based outlier detection for safeguarding electronic patient record systems. IEEE Access 2019, 7, 40285–40294. [Google Scholar] [CrossRef]
  16. Lin, C.H.; Hsu, K.C.; Johnson, K.R.; Luby, M.; Fann, Y.C. Applying density-based outlier identifications using multiple datasets for validation of stroke clinical outcomes. Int. J. Med. Inform. 2019, 132, 103988. [Google Scholar] [CrossRef]
  17. Riahi-Madvar, M.; Azirani, A.A.; Nasersharif, B.; Raahemi, B. A new density-based subspace selection method using mutual information for high dimensional outlier detection. Knowl.-Based Syst. 2021, 216, 106733. [Google Scholar] [CrossRef]
  18. Brito, M.R.; Chávez, E.L.; Quiroz, A.J.; Yukich, J.E. Connectivity of the mutual k-nearest-neighbor graph in clustering and outlier detection. Stat. Probab. Lett. 1997, 35, 33–42. [Google Scholar] [CrossRef]
  19. Breunig, M.M.; Kriegel, H.P.; Ng, R.T.; Sander, J. LOF: Identifying density-based local outliers. In Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, TX, USA, 15–18 May 2000; pp. 93–104. [Google Scholar]
  20. Elahi, M.; Li, K.; Nisar, W.; Lv, X.; Wang, H. Efficient clustering-based outlier detection algorithm for dynamic data stream. In Proceedings of the 2008 Fifth International Conference on Fuzzy Systems and Knowledge Discovery, Jinan, China, 18–20 October 2008; Volume 5, pp. 298–304. [Google Scholar]
  21. Ariyaluran Habeeb, R.A.; Nasaruddin, F.; Gani, A.; Amanullah, M.A.; Abaker Targio Hashem, I.; Ahmed, E.; Imran, M. Clustering-based real-time anomaly detection—A breakthrough in big data technologies. Trans. Emerg. Telecommun. Technol. 2019, 32, 367–378. [Google Scholar] [CrossRef]
  22. Pu, G.; Wang, L.; Shen, J.; Dong, F. A hybrid unsupervised clustering-based anomaly detection method. Tsinghua Sci. Technol. 2020, 26, 146–153. [Google Scholar] [CrossRef]
  23. Li, J.; Izakian, H.; Pedrycz, W.; Jamal, I. Clustering-based anomaly detection in multivariate time series data. Appl. Soft Comput. 2021, 100, 106919. [Google Scholar] [CrossRef]
  24. Ijaz, M.F.; Alfian, G.; Syafrudin, M.; Rhee, J. Hybrid prediction model for type 2 diabetes and hypertension using DBSCAN-based outlier detection, synthetic minority over sampling technique (SMOTE), and random forest. Appl. Sci. 2018, 8, 1325. [Google Scholar] [CrossRef] [Green Version]
  25. Sheridan, K.; Puranik, T.G.; Mangortey, E.; Pinon-Fischer, O.J.; Kirby, M.; Mavris, D.N. An application of dbscan clustering for flight anomaly detection during the approach phase. In Proceedings of the AIAA Scitech 2020 Forum, Orlando, FL, USA, 6–10 January 2020; p. 1851. [Google Scholar]
  26. Jin, C.H.; Na, H.J.; Piao, M.; Pok, G.; Ryu, K.H. A novel DBSCAN-based defect pattern detection and classification framework for wafer bin map. IEEE Trans. Semicond. Manuf. 2019, 32, 286–292. [Google Scholar] [CrossRef]
  27. Bergman, L.; Hoshen, Y. Classification-based anomaly detection for general data. arXiv 2020, arXiv:2005.02359. [Google Scholar]
  28. Kieu, T.; Yang, B.; Jensen, C.S. Outlier detection for multidimensional time series using deep neural networks. In Proceedings of the 2018 19th IEEE International Conference on Mobile Data Management (MDM), Aalborg, Denmark, 25–28 June 2018; pp. 125–134. [Google Scholar]
  29. Zhao, T.; Schranz, M.E. Network-based microsynteny analysis identifies major differences and genomic outliers in mammalian and angiosperm genomes. Proc. Natl. Acad. Sci. USA 2019, 116, 2165–2174. [Google Scholar] [CrossRef] [Green Version]
  30. Tang, Z.; Chen, Z.; Bao, Y.; Li, H. Convolutional neural network-based data anomaly detection method using multiple information for structural health monitoring. Struct. Control. Health Monit. 2019, 26, e2296. [Google Scholar] [CrossRef] [Green Version]
  31. Chen, J.; Sathe, S.; Aggarwal, C.; Turaga, D. Outlier detection with autoencoder ensembles. In Proceedings of the 2017 SIAM International Conference on Data Mining, Houston, TX, USA, 27–29 April 2017; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2017; pp. 90–98. [Google Scholar]
  32. Kieu, T.; Yang, B.; Guo, C.; Jensen, C.S. Outlier Detection for Time Series with Recurrent Autoencoder Ensembles. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 2725–2732. [Google Scholar]
  33. Sarvari, H.; Domeniconi, C.; Prenkaj, B.; Stilo, G. Unsupervised boosting-based autoencoder ensembles for outlier detection. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Cham, Switzerland, 2021; pp. 91–103. [Google Scholar]
  34. Zenati, H.; Foo, C.S.; Lecouat, B.; Manek, G.; Chandrasekhar, V.R. Efficient gan-based anomaly detection. arXiv 2018, arXiv:1802.06222. [Google Scholar]
  35. Liu, Y.; Li, Z.; Zhou, C.; Jiang, Y.; Sun, J.; Wang, M.; He, X. Generative adversarial active learning for unsupervised outlier detection. IEEE Trans. Knowl. Data Eng. 2019, 32, 1517–1528. [Google Scholar] [CrossRef] [Green Version]
  36. Ibrahim, B.I.; Nicolae, D.C.; Khan, A.; Ali, S.I.; Khattak, A. VAE-GAN based zero-shot outlier detection. In Proceedings of the 2020 4th International Symposium on Computer Science and Intelligent Control, Newcastle upon Tyne, UK, 17–19 November 2020; pp. 1–5. [Google Scholar]
  37. Chaudhary, A.; Mittal, H.; Arora, A. Anomaly detection using graph neural networks. In Proceedings of the 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon), Faridabad, India, 14–16 February 2019; pp. 346–350. [Google Scholar]
  38. Protogerou, A.; Papadopoulos, S.; Drosou, A.; Tzovaras, D.; Refanidis, I. A graph neural network method for distributed anomaly detection in IoT. Evol. Syst. 2021, 12, 19–36. [Google Scholar] [CrossRef]
  39. Deng, A.; Hooi, B. Graph neural network-based anomaly detection in multivariate time series. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually. 2–9 February 2021; Volume 35, pp. 4027–4035. [Google Scholar]
  40. Ma, X.; Wu, J.; Xue, S.; Yang, J.; Zhou, C.; Sheng, Q.Z.; Xiong, H.; Akoglu, L. A comprehensive survey on graph anomaly detection with deep learning. IEEE Trans. Knowl. Data Eng. 2021, 8, 58. [Google Scholar] [CrossRef]
  41. Hariri, S.; Kind, M.C.; Brunner, R.J. Extended isolation forest. IEEE Trans. Knowl. Data Eng. 2019, 33, 1479–1489. [Google Scholar] [CrossRef] [Green Version]
  42. Karczmarek, P.; Kiersztyn, A.; Pedrycz, W.; Al, E. K-Means-based isolation forest. Knowl.-Based Syst. 2020, 195, 105659. [Google Scholar] [CrossRef]
  43. Staerman, G.; Mozharovskyi, P.; Clémençon, S.; D’Alché-Buc, F. Functional isolation forest. In Proceedings of the 11th Asian Conference on Machine Learning, PMLR, Nagoya, Japan, 17–19 November 2019; pp. 332–347. [Google Scholar]
  44. Mangasarian, O.L.; Wolberg, W.H. Cancer Diagnosis Via Linear Programming. SIAM News. 1990. Available online: http://archive.ics.uci.edu/ml/datasets/Breast+Cancer+Wisconsin+%28Original%29 (accessed on 11 April 2022).
  45. Ting, K.M.; Tan, S.C.; Liu, F.T. Mass: A New Ranking Measure for Anomaly Detection. IEEE Transactions on Knowledge and Data Engineering. 2009. Available online: http://odds.cs.stonybrook.edu/annthyroid-dataset/ (accessed on 11 April 2022).
  46. Dua, D.; Graff, C. Arrhythmia Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml/datasets/Arrhythmia (accessed on 11 April 2022).
  47. Ting, K.M.; Tan, S.C.; Liu, F.T. Mass: A New Ranking Measure for Anomaly Detection. IEEE Transactions on Knowledge and Data Engineering. 2009. Available online: http://odds.cs.stonybrook.edu/pima-indians-diabetes-dataset/ (accessed on 11 April 2022).
  48. Learing Outlier Ensembles: The Best of Both Worlds—Supervised and Unsupervised. Barbora Micenkova, Brian McWilliams, and Ira Assent, KDD ODD2 Workshop. 2014. Available online: http://odds.cs.stonybrook.edu/speech-dataset/ (accessed on 11 April 2022).
  49. Dua, D.; Graff, C. Thyroid Disease Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: https://archive.ics.uci.edu/ml/datasets/Thyroid+Disease (accessed on 11 April 2022).
  50. Dua, D.; Graff, C. Vertebral Column Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: https://archive.ics.uci.edu/ml/datasets/Vertebral+Column (accessed on 11 April 2022).
  51. Dua, D.; Graff, C. Wine Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: https://archive.ics.uci.edu/ml/datasets/Wine (accessed on 11 April 2022).
  52. Dua, D.; Graff, C. Ionosphere Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml/datasets/Ionosphere (accessed on 11 April 2022).
  53. Dua, D.; Graff, C. Shuttle Landing Control Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml/datasets/Shuttle+Landing+Control (accessed on 11 April 2022).
  54. Dua, D.; Graff, C. Cardiotocography Data Set. In UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml/datasets/Cardiotocography (accessed on 11 April 2022).
  55. Tang, J.; Chen, Z.; Fu, A.W.C.; Cheung, D.W. Enhancing effectiveness of outlier detections for low density patterns. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2002; pp. 535–548. [Google Scholar]
  56. Kriegel, H.P.; Schubert, M.; Zimek, A. Angle-based outlier detection in high-dimensional data. In Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 444–452. [Google Scholar]
  57. Zhang, K.; Hutter, M.; Jin, H. A new local distance-based outlier detection approach for scattered real-world data. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2009; pp. 813–822. [Google Scholar]
Figure 1. The principle of IF, (a) normal data; (b) outlier. The blue asterisk indicates normal data, and the red asterisk indicates an outlier.
Figure 1. The principle of IF, (a) normal data; (b) outlier. The blue asterisk indicates normal data, and the red asterisk indicates an outlier.
Entropy 24 00611 g001
Figure 2. The principle of CIIF, (a) normal data; (b) outlier. The blue asterisk indicates normal data and the red asterisk indicates an outlier, the data in the selection cluster are indicated by the purple dots.
Figure 2. The principle of CIIF, (a) normal data; (b) outlier. The blue asterisk indicates normal data and the red asterisk indicates an outlier, the data in the selection cluster are indicated by the purple dots.
Entropy 24 00611 g002
Figure 3. ROC curve of several algorithms in several datasets: (a) breast; (b) annthyroid; (c) arrhythmia; (d) pima; (e) speech; (f) thyroid; (g) vertebral; (h) wine; (i) ionosphere; (j) shuttle; (k) cardio.
Figure 3. ROC curve of several algorithms in several datasets: (a) breast; (b) annthyroid; (c) arrhythmia; (d) pima; (e) speech; (f) thyroid; (g) vertebral; (h) wine; (i) ionosphere; (j) shuttle; (k) cardio.
Entropy 24 00611 g003aEntropy 24 00611 g003b
Figure 4. Computational times(s).
Figure 4. Computational times(s).
Entropy 24 00611 g004
Figure 5. Computational time of shuttle(s).
Figure 5. Computational time of shuttle(s).
Entropy 24 00611 g005
Figure 6. AUC impact of I.
Figure 6. AUC impact of I.
Entropy 24 00611 g006
Figure 7. AUC impact of subsampling X.
Figure 7. AUC impact of subsampling X.
Entropy 24 00611 g007
Table 1. Testing dataset.
Table 1. Testing dataset.
DatasetData VolumeDimensionNumber of OutliersOutlier Ratio %
breastw683923934.9927
annthyroid720065347.4167
arrhythmia4522746614.6018
pima768926834.8958
speech3686400611.6549
thyroid37726932.4655
vertebral24063012.5
wine12913107.7519
ionosphere3513312635.8974
shuttle49,097935117.1511
cardio1822211769.6122
Table 2. Confusion matrix of classification results.
Table 2. Confusion matrix of classification results.
ActualForecast
PositiveNegative
PositiveTPFN
NegativeFPTN
Table 3. AUC results on 11 real-world datasets AUC of several algorithms. The highlighted data indicate the best and second best values on each data set.
Table 3. AUC results on 11 real-world datasets AUC of several algorithms. The highlighted data indicate the best and second best values on each data set.
DataAUC
CIIFIFLOFKNNCOFFastABODLDOF
breastw0.99220.98760.24210.98810.12730.62200.6394
annthyroid0.90730.82120.69580.69380.65230.21530.7377
arrhythmia0.83920.80350.50920.50920.72290.25620.5092
pima0.85020.80640.44910.82750.48590.25800.5221
speech0.60480.45390.54670.48210.57470.26620.6592
thyroid0.97990.97490.68360.94810.61210.16420.7098
vertebral0.69110.36590.48460.32380.48050.62700.5281
wine0.94030.78290.40080.84620.23190.56470.4496
ionosphere0.86240.85270.86430.87930.85290.17380.8831
shuttle0.99830.99680.51840.63390.55340.41720.5208
cardio0.93050.90420.61280.91610.57960.47590.5798
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shao, C.; Du, X.; Yu, J.; Chen, J. Cluster-Based Improved Isolation Forest. Entropy 2022, 24, 611. https://0-doi-org.brum.beds.ac.uk/10.3390/e24050611

AMA Style

Shao C, Du X, Yu J, Chen J. Cluster-Based Improved Isolation Forest. Entropy. 2022; 24(5):611. https://0-doi-org.brum.beds.ac.uk/10.3390/e24050611

Chicago/Turabian Style

Shao, Chen, Xusheng Du, Jiong Yu, and Jiaying Chen. 2022. "Cluster-Based Improved Isolation Forest" Entropy 24, no. 5: 611. https://0-doi-org.brum.beds.ac.uk/10.3390/e24050611

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop