Next Article in Journal
Temporal Statistical Relationship between Regional Cerebral Oxygen Saturation (rSO2) and Brain Tissue Oxygen Tension (PbtO2) in Moderate-to-Severe Traumatic Brain Injury: A Canadian High Resolution-TBI (CAHR-TBI) Cohort Study
Next Article in Special Issue
A Deep Learning Method of Human Identification from Radar Signal for Daily Sleep Health Monitoring
Previous Article in Journal
Cell-Tissue Interaction: The Biomimetic Approach to Design Tissue Engineered Biomaterials
Previous Article in Special Issue
Sparse Logistic Regression-Based EEG Channel Optimization Algorithm for Improved Universality across Participants
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Binary Sand Cat Swarm Optimization for Feature Selection in High-Dimensional Biomedical Data

1
Department of Computer Engineering, Istanbul Aydin University, Istanbul 34295, Turkey
2
Department of Medical and Molecular Genetics, Indiana University School of Medicine, Indianapolis, IN 46202, USA
Submission received: 28 August 2023 / Revised: 19 September 2023 / Accepted: 21 September 2023 / Published: 25 September 2023

Abstract

:
Recent breakthroughs are making a significant contribution to big data in biomedicine which are anticipated to assist in disease diagnosis and patient care management. To obtain relevant information from this data, effective administration and analysis are required. One of the major challenges associated with biomedical data analysis is the so-called “curse of dimensionality”. For this issue, a new version of Binary Sand Cat Swarm Optimization (called PILC-BSCSO), incorporating a pinhole-imaging-based learning strategy and crossover operator, is presented for selecting the most informative features. First, the crossover operator is used to strengthen the search capability of BSCSO. Second, the pinhole-imaging learning strategy is utilized to effectively increase exploration capacity while avoiding premature convergence. The Support Vector Machine (SVM) classifier with a linear kernel is used to assess classification accuracy. The experimental results show that the PILC-BSCSO algorithm beats 11 cutting-edge techniques in terms of classification accuracy and the number of selected features using three public medical datasets. Moreover, PILC-BSCSO achieves a classification accuracy of 100% for colon cancer, which is difficult to classify accurately, based on just 10 genes. A real Liver Hepatocellular Carcinoma (TCGA-HCC) data set was also used to further evaluate the effectiveness of the PILC-BSCSO approach. PILC-BSCSO identifies a subset of five marker genes, including prognostic biomarkers HMMR, CHST4, and COL15A1, that have excellent predictive potential for liver cancer using TCGA data.

1. Introduction

Enormously large, rapidly growing collections of biomedical and clinical data pose significant challenges to their analysis and interpretation. Health data are large-scale, multimodal, and high-dimensional. The promise of Big Data in healthcare is based on the ability to discover patterns and transform massive volumes of data into meaningful information for precision, diagnosis, treatment, and decision-makers. Biomedical datasets, encompassing genomics, proteomics, clinical attributes, imaging, and more, often present researchers with a staggering number of variables. While this wealth of data holds the potential to unveil crucial insights into disease mechanisms and patient profiles, it simultaneously poses formidable challenges, giving rise to the ‘curse of dimensionality’.
In biomedical data analysis, the ‘curse of dimensionality’ arises from the combination of high-dimensional feature spaces, sparsity, computational demands, risk of overfitting, and the need to capture complex biological phenomena. Addressing this challenge requires innovative feature selection techniques and dimensionality reduction methods. This difficulty in navigating high-dimensional biomedical data has led to a growing interest among researchers in the biomedical domain, inspiring the development of new robust algorithms that are best suited to appropriately evaluate this big data [1]. The task of extracting meaningful information and identifying key aspects within these vast datasets has become a focal point of exploration and innovation within the field of biomedical research.
Feature selection is a powerful data mining approach for shrinking the dimensionality of feature space. It is broadly known that feature selection is an NP-hard task, and therefore determining the optimal or near-optimal feature set is a challenging task [2,3].
Feature selection’s primary role is to identify and retain the most informative and relevant attributes while discarding redundant or noisy variables. Doing so not only mitigates the computational burden associated with high dimensionality but also enhances the interpretability and generalization of analytical models. In the context of disease diagnosis, feature selection serves as a compass guiding researchers and clinicians toward the most discriminating biomarkers or attributes associated with specific diseases. This precision enables the development of diagnostic models that are not only accurate but also clinically interpretable. Such models, informed by selected features, provide the foundation for early disease detection and stratification, facilitating timely interventions and improved patient outcomes. Moreover, feature selection plays a pivotal role in patient care management. In the era of personalized medicine, where treatment strategies are tailored to individual patients, the identification of relevant biomarkers and clinical attributes is paramount. Feature selection aids in constructing predictive models that inform treatment decisions, predict patient responses, and gauge disease prognosis. By focusing on the most influential factors, healthcare providers can optimize treatment plans, minimize adverse effects, and maximize therapeutic efficacy.
There are three popular feature selection methods: filter-based, wrapper-based, and hybrid approaches. Filter techniques assess the importance of features based on their correlation with the dependent variable using statistical methods and are significantly quicker than wrapper approaches, whereas wrapper methods assess the utility of a subset of features by training a model on it and can provide the most effective subset of features. Nature-inspired optimization algorithms (NIOAs) are used as search techniques in wrapper methods to identify informative features. A hybrid feature selection method combines filters and wrappers approaches. Hybrid approaches are still in their fancy and further research is needed to develop a more effective feature selection methodology [4]. In the literature, various feature selection strategies have been offered. Some of them are a hybrid of minimum redundancy maximum relevance (mRMR) and a mutated binary Aquila optimizer (MBAO) [5], a hybrid of mutual information maximization (MIM) and moth flame optimization algorithm (MFOA) [6], binary coral reefs optimization with simulated annealing and tournament selection strategy (BCROSAT) [1], an improved binary clonal flower pollination algorithm (IBCFPA) [7], an improved shuffled frog leaping algorithm (ISFLA) [8], a hybrid of mRMR with a combination of binary black hole algorithm and binary dragonfly optimization algorithm (DBH) [9], a hybrid of symmetrical uncertainty (SU) and reference set harmony search algorithm (RSHSA) [10], “Technique for Order Preference by Similarity to Ideal Solution” (TOPSIS) filtering and binary Jaya algorithm [11], a hybrid of information gain (IG) and modified krill herd algorithm (MKHA) [12], and hybrid of mRMR and binary Coot with simulated annealing and crossover operator (mRMR BCOOT-CSA) [13]. Difficulty in parameter tuning, lack of interpretability, risk of premature convergence, and limited adaptability are some limitations of the above approaches. Nevertheless, recognizing that no single solution can entirely alleviate the dimensionality curse within the original dataset, these limitations have motivated numerous researchers to propose new algorithms with the aim of achieving improved performance.
The sand cat swarm optimization (SCSO) algorithm [14] is a new NIOA, that has been utilized to solve various optimization problems such as engineering problems [15,16], power transformer fault diagnosis [17], and feature selection [2,18]. Low solution precision and early convergence are two main drawbacks of most existing SCSO variations [15]. This paper puts forward an improved version of binary SCSO (PILC-BSCSO) by incorporating crossover and opposition-based learning for feature selection challenges of high-dimensional medical data. This is the main innovation of this paper and shows promise in finding the best feature subset.
The key contributions of this paper are as follows:
  • A novel gene selection approach is proposed based on an enhanced binary sand cat swarm optimization for high-dimensional biomedical data.
  • A pinhole-imaging opposition-based learning (PIOBL) scheme is employed to boost the exploration and convergence characteristics of the BSCSO.
  • The Crossover operator is fused with BSCSO to improve the search performance of the original BSCSO.
  • An initial population strategy based on the Differential Expression (DE) analysis is conducted to identify differentially expressed genes (DEGs), which makes the proposed algorithm, called PILC-BSCSO, obtain higher classification accuracy with a better-initialized population.
  • The suggested PILC-BSCSO approach is compared to 11 state-of-the-art methods on three benchmark microarray datasets and outperforms them all.
  • The efficiency of the PILC-BSCSO approach was further assessed using a real Liver Hepatocellular Carcinoma (TCGA-HCC) data set, and PILC-BSCSO selects a subset of five marker genes while offering the best accuracy.

2. Materials and Methods

2.1. Sand Cat Swarm Optimization

The SCSO Algorithm is a new nature-inspired optimization algorithm proposed by Seyyedabbasi [14], which simulates the behavior of sand cats in hunting. These animals utilize their acute hearing to detect low-frequency disturbances. Therefore, they may sense prey movement underground. They also have an unusual ability to dig swiftly if the prey is underground. In SCSO, the population consists of N sand cat individuals (solutions) with D dimensions, thus the population vector contains an N × D dimensional matrix. The X ( t ) demonstrates the position vector of each sand cat in searching space at iteration t .
The sound cat has a sensitivity range of (2, 0) kHz in perceiving low-frequency noises. It starts at 2 kHz and decreases linearly till it approaches 0 kHz. The sensitivity level is known as r g in SCSO, which is calculated as follows:
r g = s M s M × t T
where s M is taken to be 2. t is the current iteration number, while T is the maximum number of iterations. Meanwhile, the R parameter determines the trade-off between the exploration and exploitation phases and is computed as follows:
R = ( ( 2 × r g ) × r a n d 0 ,   1 ) r g
where r a n d 0 ,   1 produces a random number between 0 and 1. The r parameter, which specifies the sensitivity range of each potential solution, is determined as follows:
r = r g × r a n d 0 ,   1
The sand cat’s next location is decided by the value of R , which runs between −1 and 1. When R 1 , the SCSO approach concentrates on exploitation and guiding the sand cat to hunt the prey (4–5). Otherwise, the algorithm concentrates on exploration and forces the sand cats to look for food (6–8).
In SCSO the mathematical expression of attacking the prey (exploitation) is as follows:
X r a n d = r a n d 0 ,   1 × X b e s t X ( t )
X t + 1 = X b e s t r a n d 0 ,   1 X r a n d cos   ( θ )
where X r a n d calculates the distance between the best position X b e s t and current position X ( t ) in the related iteration t . X t + 1 demonstrates the position update for the corresponding search agent, i.e., X . Moreover, the sand cats’ precise sensitivity is supposed to be circular, hence the direction of each movement is decided by a random angle θ based on a roulette wheel selection.
In SCSO, the mathematical expression of searching for prey (exploration), is as follows:
c p = f l o o r N r a n d ( 0 ,   1 ) + 1
X C a n d i d a t e t = X ( c p , : )
X t + 1 = r × ( X C a n d i d a t e t r a n d 0 ,   1 × X ( t ) )
where X C a n d i d a t e t indicates a random candidate position. The pseudo-code of the SCSO algorithm is shown in Algorithm 1.
Algorithm 1: Pseudo-code of the SCSO algorithm.
1.Determine the number of population  N , and maximum number of iteration T
2.Initialize the sand cat population X i i = 1 ,   2 ,   ,   N
3.While  t T do
4. Calculate the fitness function of each sand cat based on the objective function
5. Determine X b e s t
6. Calculate r g when s M = 2
7. For  i = 1  to  N  do
8.  Calculate R and r
9.  For  j = 1  to  D  do
10.   Randomly selected 0 θ 360 using Roulette wheel selection
11.   if ( ( 1 < = R ) & & ( R < = 1 ) )  then
12.     X r a n d = r a n d 0 ,   1 × X b e s t , j X ( i , j )
13.     X i , j = X b e s t , j r a n d 0 ,   1 X r a n d cos   ( θ )
    //update position using (5)
14.    else
15.     c p = f l o o r N r a n d ( 0 ,   1 ) + 1
16.     X C a n d i d a t e = X ( c p , : )
17.     X i , j = r × ( X C a n d i d a t e , j r a n d 0 ,   1 × X ( i , j ) ) //update position using (8)
18.    End if
19.   End for
20.  End for
21.  t = t + 1
22.End while
23.Return  X b e s t

2.2. Binary Sand Cat Swarm Optimization for Feature Selectıon

In the context of feature selection, each feature can be thought of as a binary decision–either included in the final subset or not. This binary choice can be represented using a binary vector of size D , where D is the total number of features in the dataset. Each element of the vector corresponds to a feature, and is set to 1 if the feature is selected and 0 if not.
The SCSO method is applied in a continuous space, whereas the feature selection problem is applied in a discrete space. Before the SCSO algorithm can be used for the feature selection issue, the continuous space must be transformed into the discrete space. The transfer functions are used for this conversion. Seyyedabbasi [18] presented the first binary version of the SCSO method, which employed a V-shaped transfer function. The transfer function determines the probability that the binary solution element changes from 0 to 1. Also, Qtaish et al. [2] introduced a memory-based BSCSO (BMSCSO) method that incorporates a memory-based approach into the BSCSO position-updating process, employing an S-shaped transfer function to pick the most relevant subset of features.

2.3. Pinhole Imaging Opposition-Based Learning

Various techniques, including mutation [5], Lévy flight [19], and opposition-based learning (OBL) [20], have been used in the literature to increase NIOA’s exploration capabilities. OBL broadens the search range by computing the inverse of the existing viable solution and locating candidate solutions in more ideal places. OBL is a subset of pinhole-imaging opposition-based learning (PIOBL) [21]. Pinhole imaging is a general physical phenomenon in which a light source flows through a tiny hole in a plate, forming an inverted actual picture on the opposite side of the plate. Figure 1 depicts the basic PIOBL concept.
The coordinate x-axis’ upper and lower bounds are labeled a and b in the picture. A tiny aperture screen is installed at the base point O . Once the light source via the small aperture receives a reversed image p of height h at the imaging screen, the projection of p on the x-axis is X   b e s t (the newly created reverse solution), whereas the projection of p whose height is h, on the x-axis is X b e s t (the current global optimal solution). The geometric connection of the line subdivisions in the figure allows us to deduce:
( a + b ) 2 X b e s t X b e s t ( a + b ) / 2 = h h
Substituting h / h = K into the foregoing equation produces the expression for X b e s t :
X b e s t = ( a + b ) 2 + ( a + b ) 2 K X b e s t K
When the method is solving a high-dimensional complex function, X b e s t can be computed using the following equation:
X b e s t , j = ( a j + b j ) 2 + ( a j + b j ) 2 K X b e s t , j K
where X b e s t , j is the inverse solution of X b e s t , j , and X b e s t , j demonstrates the optimal solution in the j t h dimension. a j and b j are the minimum and maximum values in the j t h dimension and the scale factor K = 0.05 .

2.4. Single Point Crossover

Crossover is a genetic operator that mixes two parents’ genetic information to produce new offspring. After selecting a random cut point on parents to create offspring, all data in the parents’ string after that point is swapped between the two parents.

2.5. The Proposed Algorithm

A modified binary SCSO (called PILC-BSCSO) with pinhole-imaging-based learning and crossover operator is proposed as a novel wrapper feature selection to find the optimal gene subset with the highest accuracy.
The crossover operator is a fundamental mechanism in BSCSO, facilitating the exchange of genetic information to create diverse offspring. This diversity enhances the algorithm’s search capabilities, allowing it to effectively explore a wider range of feature combinations and identify feature subsets with improved predictive power for biomedical data analysis.
The pinhole-imaging-based learning strategy provides a localized focus as well as adaptability and balance in the BSCSO process. It strategically narrows the focus when needed for in-depth exploration and widens it to exploit promising regions. This intelligent strategy not only enhances the algorithm’s ability to navigate the vast solution space but also safeguards against premature convergence, ultimately contributing to its effectiveness in feature selection for high-dimensional biomedical data analysis.
The detailed implementation of the proposed algorithm is elaborated upon in the following steps:
Step 1. First, a Limma differential expression analysis of microarray data is conducted as a preprocessing step to identify DEGs, and the genes with an adjusted p-value lower than 0.05 are selected. Then, the shrink dataset (GEGs) is used as the input for the proposed PILC-BSCSO algorithm where the Cohen’s kappa score of the support vector machine (SVM) [22,23,24] with the linear kernel is utilized as the fitness function.
Step 2. Population initialization is performed, and each sand cat individual is encoded as a binary vector with an initial value of 1.
Step 3. Binary SCSO is used to further select the optimal subset of genes from a provided pool of DEGs. Each individual within the sand cat population undergoes fitness value computation, enabling the identification of the individual with the most optimal fitness—a role granted to the best individual. After this process, the updating of the solution is performed using (5) and (8). The transfer function affects the efficiency of binary optimization techniques. There are several transfer functions accessible in the literature; nevertheless, selecting one is not an easy process [25]. We are using a hyperbolic tangent sigmoid (tansig) transfer function to convert the continuous SCSO algorithm to a binary version with the following equations:
T f X ( i , j ) = 2 1 + e 2 X ( i , j ) 1
X ( i , j ) = 1 ,     T f ( X i , j ) > r a n d ( 0 ,   1 ) 0 ,     o t h e r w i s e
Step 4. Low solution accuracy and early convergence are two main drawbacks in the majority of current SCSO versions. Therefore, PIOBL and crossover mechanisms are utilized to effectively boost the exploration ability of SCSO. The process of updating individuals after step 3 is continued using either the crossover operators or the PIOBL strategy according to random probability. The individual updating procedure is repeated until the stop criteria are met. The comprehensive sequence of steps involved in the PILC-BSCSO algorithm is depicted in Figure 2, while the precise algorithmic details are provided in Algorithm 2.
Algorithm 2: Pseudo-code of the proposed PILC-BSCSO algorithm for feature selection.
1.Load Microarray dataset
2.Extracting DEG lists using Limma and obtaining shrinking dataset with D features
3.//Perform PILC-BSCSO algorithm
4. Determine   the   number   of   population   N ,   and   maximum   number   of   iterations   T
5. Initialize   the   sand   cat   population   X i i = 1 ,   2 ,   ,   N with the binary value 1
6.While  t T do
7.    Calculate the fitness function of each sand cat using SVM with a 10-fold CV
8.     Determine   X b e s t
9.     Calculate   r g when s M = 1
10.    For i = 1   to   N  do
11.         Calculate   R   and   r
12.        For j = 1   to   D  do
13.             Randomly   selected   0 θ 360 using Roulette wheel selection
14.            if  ( ( 0 < R ) & & ( R < 1 ) )  then
15.                Update the search agent position using Equation (5)
16.            else
17.                Update the search agent position using Equation (8)
18.            End if
19             T f = 2 1 + e 2 X ( i , j ) 1
20            if t f > r a n d ( 0 ,   1 )  then  X ( i , j ) = 1  else  X ( i , j ) = 0
21.        End for//j
22.        if r a n d 0 ,   1 < 0.5  then
23.            //Perform crossover operator
24.             [ q 1 ,   q 2 ] = Crossover ( X b e s t , X ( i , : ) )
25.            Calculate the fitness values of p 1 ,   p 2 using SVM
26.            if fitness value of q 1 is better than fitness values of q 2 and X b e s t  then
27.                 X b e s t = q 1
28.            else if the fitness value of q 2 is better than the fitness value of X b e s t  then
29.                 X b e s t = q 2
30.             End if
31.        else
32.            //Perform PIOBL operator
33.             Calculate   q 1 = 1 2 + 1 2 K X ( i ,   : ) K when k = 0.05
34.             X b e s t = q 1   A N D   X b e s t
35.            Calculate the fitness values of X b e s t using SVM
36.            if the fitness value of X b e s t is better than the fitness values of X b e s t  then
37.                 X b e s t = X b e s t
38.                 X ( i , : ) = X b e s t
            End if
        End if
39.    End for//i
40.     t = t + 1
41.End while
42.Return  X b e s t

3. Results

3.1. Experimental Setup

The proposed method is a two-step procedure. In the first step, Z-score normalization and DEG analysis are performed as a preprocessing step to scale and identify genes whose expression levels differ significantly between the two experimental conditions. In the second step, the proposed approach is applied to gain an optimal subset of genes. The effectiveness of our proposed gene selection approach was examined on three binary-class microarray cancer datasets and one real The Cancer Genome Atlas Liver Hepatocellular Carcinoma (TCGA-LIHC) dataset. Table 1 describes the characteristics of the datasets. In this study, we employed an SVM classifier with a linear kernel as a fitness function to explore the prediction ability of gene subsets. Tuning parameter ‘C’ was held constant at a value of 1 (default value).
To avoid bias, we subjected each subset of potential candidate genes to rigorous validation and analysis, employing a repeated 10-fold cross-validation approach with three repetitions. To show stability, the proposed methodology was executed independently multiple times on distinct datasets, with subsequent reporting of the averaged outcomes. For the implementation of algorithms, the R programming language was used. Specifically, the ‘limma’ package was harnessed for the analysis of DEGs, while the construction of the SVM classifier was carried out using the ‘e1071’ package. The “rmcfs” package was used for Monte Carlo Feature Selection (MCFS) [26], while the “praznik” package was employed for feature ranking using Minimum Redundancy Maximum Relevance (mRMR) [27]. Particle Swarm Optimization (PSO) and Genetic Algorithms (GA) optimization techniques were implemented using the Weka platform. R code of PILC-BSCSO is available at https://github.com/nazpashaei/PILC-BSCSO, accessed on 27 August 2023.
Computational experiments were conducted on an AMD Ryzen 7 5700U processor operating at 1.80 GHz, ×64 architecture, and bolstered by 16 GB of RAM. For four optimization algorithms, we configured the algorithm parameters, setting the number of populations at 100 and the maximum number of iterations at 50.

3.2. Experimental Results on Three Benchmark Microarray Datasets

The results of this study reveal significant insights into the performance and effectiveness of the proposed approach. The investigation of Differentially Expressed Genes (DEGs) led to the identification of distinct gene sets across different datasets. Specifically, there are 358 DEGs with an adjusted p-value of 0.05 in the colon, 328 with a p-value of 0.05 in the CNS, and 154 with a p-value of 0.05 and |LogFC| > 0.68 in the Breast datasets, respectively. To evaluate the potential of these gene sets for classification tasks, the LOOCV (Leave-One-Out Cross-Validation) classification accuracy was assessed using an SVM classifier. mRMR and MCFS feature ranking algorithms with various cut-offs were utilized to compare with DEG performance. The mRMR is an entropy-based feature selection method that calculates the mutual information (MI) between a group of features and a class variable. Features with high MI values with respect to the class variable and low MI values with respect to other selected features are considered more informative and less redundant. The MCFS method evaluates the feature importance by creating numerous decision trees. Each decision tree is trained on a subset of the data with a random feature subset. The importance of each feature is determined by how much it contributes to the quality of the decision trees.
The outcomes, detailed in Table 2, provided an initial assessment of the DEGs’ predictive power compared to MRMR and MCFS. Table 2 reveals that MCFS with cutoffs of 100, 200, and 300 consistently demonstrates better classification accuracy on three datasets. Notably, the 300-cutoff threshold outperforms DEGs in terms of classification accuracy.
Visual representations further enhanced our understanding of the data. The volcano plot (Figure 3) depicted the distribution of Log2(fold-change) against the significance (p-value) of the identified DEGs, with cut-off values indicated by vertical and horizontal dotted lines. The comparison of the proposed PILC-BSCSO method with the basic BSCSO technique, PSO, and GA (Table 3) showcases their respective performance in 10 separate runs. Strikingly, PILC-BSCSO consistently outperformed all three swarm optimization algorithms (BSCSO, GA, and PSO) in terms of classification accuracy across all datasets. A nuanced observation was made for the colon and breast datasets, where BSCSO exhibited a slight advantage over PILC-BSCSO in terms of the average number of selected genes. Table 3 also shows the statistical test results, where a p-value < 0.05 indicates that the PILC-BCSO methodology produces statistically different results than other techniques.
The convergence behavior of PILC-BSCSO and the basic BSCSO methods was examined, and the results are depicted in Figure 4. This visualization showcases the trajectories of their convergence across four distinct datasets, all derived from the same random seed. Significantly, PILC-BSCSO exhibited more favorable convergence trends in terms of fitness value (Cohen’s kappa) compared to conventional BSCSO, which tended to converge to local optima. It is worth noting that PILC-BSCSO may take longer (two and a half times) to converge than the traditional BSCSO approach.
Figure 5 offered a visual representation of the gene expression profiles for the best subset of discriminative genes identified by the proposed method for each dataset, represented through a heatmap.
To comprehensively assess the proposed method’s efficacy, comparisons were made against 11 state-of-the-art approaches. The average results, summarized in Table 4 and Figure 6, demonstrated that PILC-BSCSO consistently achieved superb classification accuracy while selecting a reasonable number of genes, outperforming 11 competing techniques across all three datasets. These findings collectively underscore the effectiveness of the proposed PILC-BSCSO approach in identifying significant gene subsets and its potential for robust classification tasks across diverse datasets. PILC-BSCSO’s superior performance can be attributed to several factors: enhanced exploration and exploitation, population initialization, and fitness function evaluation. PILC-BSCSO leverages the Pinhole-Imaging Opposition-Based Learning (PIOBL) scheme and the crossover operator to enhance both the exploration and exploitation phases. This allows it to effectively explore a wide solution space while also exploiting promising regions more efficiently, leading to improved solutions. The algorithm also uses an initial population strategy based on differential expression analysis. This strategy provides a better-initialized population, guiding the optimization process toward more promising solutions from the start. It also employs repeated 10-fold cross-validation with three repetitions contributing to more stable and reliable results, especially when dealing with unbalanced datasets. Additionally, utilizing the kappa measure of SVM further enhances the appropriateness of the evaluation metric for accurately assessing model performance in the context of class imbalance. This approach ensures a robust evaluation framework that is well-suited for the challenges posed by the dataset at hand.

3.3. Experimental Results on Liver Hepatocellular Carcinoma TCGA

To demonstrate the effectiveness of the proposed method, it was applied to data on HCC sourced from TCGA. HCC, a devastating malignancy ranked as the third leading cause of global cancer-related deaths, often evades early detection, resulting in diagnosis at advanced stages. Therefore, the development of innovative treatment targets is of paramount importance to enhance patient survival outcomes.
The RNA-Seq data encompassed 371 samples from HCC patients and 50 control samples, all derived from the TCGA-liver hepatocellular carcinoma (LIHC) dataset, comprising a total of 421 samples and 56,602 genes. Following data acquisition, various preprocessing steps were executed, including the removal of genes with low counts, conversion of counts to DGEList format, quality control, and normalization to mitigate batch effects. Subsequently, 1656 genes with |LogFC| > 2 were identified as DEGs out of the initial 14,899 genes, based on an adjusted p-value threshold of 0.05 (as depicted in Figure 3). The dataset was partitioned into training (75%) and testing (25%) sets, with the latter serving as an independent dataset to validate the PILC-BSCSO results.
Figure 4 illustrates the convergence behavior of the TCGA-LIHC training dataset comprising 1546 DEGs and 317 samples for both BSCSO and PILC-BSCSO. The experimental results on the test data (104 samples) reveal that PILC-BSCSO outperforms BSCSO in terms of classification accuracy, achieving an average of 98.87% ± 1.2, compared to BSCSO’s 97.6% ± 3. PILC-BSCSO demonstrates superior efficiency in feature selection, with an average selection of 8 ± 2.6 genes, in contrast to BSCSO’s average of 73 ± 20.2 genes, for the achievement of higher classification accuracy.
Figure 5 portrays the expression patterns of the best subset of identified genes, including ANGPTL6 [28], HMMR [29], CHST4 [30], COL15A1 [31], and PZP [32], utilizing the proposed approach. These genes exhibit remarkable classification accuracy and an Area Under the Curve (AUC) of 100% in the test data.
Furthermore, Kaplan–Meier survival analyses were conducted to evaluate the prognostic potential of these genes. Among the five genes in the subset, HMMR, CHST4, and COL15A1 emerged as potential independent biomarkers (Figure 7), signifying a robust and statistically significant association with patient survival in HCC.
Figure 8 depicts the tissue-wise expression patterns of the identified best subset of genes associated with LIHC. From this figure, it can be observed that the identified subset of five genes (ANGPTL6, HMMR, CHST4, COL15A1, and PZP) has discriminative gene expression patterns. These genes can potentially serve as diagnostic or prognostic biomarkers, aiding in the early detection or risk assessment of LIHC.

4. Discussion

Due to the rapid technological improvement in medical research, a vast volume of biomedical data is regularly created from various biomedical equipment and investigations these days. The effective analysis of this biomedical data, such as identifying the key biological and diagnostic features, is a difficult challenge. Here, a new feature selection method based on the BSCSO algorithm was proposed. Pinhole-imaging-based learning strategy and crossover operator are combined with BSCSO to design the PILC-BSCSO algorithm which is capable of efficiently addressing feature selection problems for high-dimensional biomedical data. Experimental results on three benchmark datasets reveal that the suggested PILC-BSCSO-SVM method can achieve a superior classification accuracy with a lower number of features simultaneously when compared to the 11 most recent state-of-the-art methods. In the context of HCC analysis, the PILC-BSCSO algorithm demonstrated outstanding performance. It successfully pinpointed a subset of target genes, including HMMR, CHST4, and COL15A1, that function as both prognostic and diagnostic biomarkers. The proposed approach holds promise for enhancing HCC diagnosis and patient outcome prediction.
While the PILC-BSCSO algorithm shows promise, it is important to acknowledge potential limitations, including the need for further validation in larger and more diverse datasets such as single-cell data to ensure its generalizability. Although PILC-BSCSO demonstrates impressive feature selection and classification accuracy, the algorithm’s output may lack interpretability, particularly when dealing with a very large number of genes. Identifying the biological relevance of the selected genes or understanding the underlying biological mechanisms contributing to high classification accuracy may require additional post-processing and domain expertise. Enhancing the algorithm’s interpretability and providing insights into the biological significance of the selected genes could be an area for further improvement. The robustness of PILC-BSCSO in selecting biologically informative genes can indeed be a potential concern, as it is for many feature selection algorithms.
In future work, other transfer functions, such as X-shaped and U-shaped, might be used to determine how they affect the suggested approach. Additionally, we believe that the incorporation of Protein–Protein interaction networks will improve the algorithm’s capacity for biomarker identification. Furthermore, the suggested PILC-BSCSO may be evaluated to address various optimization issues, including clustering, task scheduling in fog computing, image segmentation, sentiment analysis, and more. PILC-BSCSO can be adapted to tackle clustering tasks by modifying its objective function and fitness evaluation criteria. Instead of feature selection, the algorithm could be tailored to group similar data points together while maximizing the dissimilarity between clusters. The algorithm’s optimization capabilities can help identify meaningful cluster centroids or representative data points, contributing to improved clustering accuracy and robustness. By defining a suitable objective function, PILC-BSCSO may be applied to task scheduling in fog computing. The method can efficiently schedule jobs to fog nodes, minimizing execution time and resource usage while optimizing overall system performance. In image processing, PILC-BSCSO can be adapted for image segmentation tasks. The objective function can be designed to identify optimal segmentation boundaries within an image. The algorithm’s optimization capabilities can help automate the process of partitioning an image into distinct regions or objects based on various image attributes, such as intensity, color, or texture. PILC-BSCSO can contribute to sentiment analysis by optimizing feature selection for sentiment classification tasks. The algorithm can identify the most informative features from text or data sources, enhancing the performance of sentiment analysis models. In each of these applications, the key lies in customizing the objective function, fitness evaluation criteria, and problem-specific parameters to align with the optimization goals. PILC-BSCSO’s adaptability and optimization capabilities make it a versatile tool for addressing a wide range of optimization challenges beyond gene selection, enhancing performance and efficiency in diverse domains.
In summary, PILC-BSCSO holds the potential to significantly impact the field of biomedicine by providing an advanced gene selection approach that enhances disease diagnosis and prognosis, and its versatility extends to broader applications in various domains, including healthcare, bioinformatics, and beyond.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

All relevant data are within the paper.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Yan, C.; Ma, J.; Luo, H.; Patel, A. Hybrid binary Coral Reefs Optimization algorithm with Simulated Annealing for Feature Selection in high-dimensional biomedical datasets. Chemom. Intell. Lab. Syst. 2019, 184, 102–111. [Google Scholar] [CrossRef]
  2. Qtaish, A.; Albashish, D.; Braik, M.; Alshammari, M.T.; Alreshidi, A.; Alreshidi, E.J. Memory-Based Sand Cat Swarm Optimization for Feature Selection in Medical Diagnosis. Electronics 2023, 12, 2042. [Google Scholar] [CrossRef]
  3. Pashaei, E.; Ozen, M.; Aydin, N. Biomarker discovery based on BBHA and AdaboostM1 on microarray data for cancer classification. In Proceedings of the 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS), Orlando, FL, USA, 16–20 August 2016. [Google Scholar]
  4. Pashaei, E.; Pashaei, E. Gene Selection for Cancer Classification using a New Hybrid of Binary Black Hole Algorithm. In Proceedings of the 28th IEEE Conference on Signal Processing and Communications Applications (SIU2020), Gaziantep, Turkey, 5–7 October 2020. [Google Scholar]
  5. Pashaei, E. Mutation-based Binary Aquila optimizer for gene selection in cancer classification. Comput. Biol. Chem. 2022, 101, 107767. [Google Scholar] [CrossRef]
  6. Dabba, A.; Tari, A.; Meftali, S.; Mokhtari, R. Gene selection and classification of microarray data method based on mutual information and moth flame algorithm. Expert Syst. Appl. 2021, 166, 114012. [Google Scholar] [CrossRef]
  7. Yan, C.; Ma, J.; Luo, H.; Zhang, G.; Luo, J. A Novel Feature Selection Method for High-Dimensional Biomedical Data Based on an Improved Binary Clonal Flower Pollination Algorithm. Hum. Hered. 2019, 84, 34–46. [Google Scholar] [CrossRef]
  8. Hu, B.; Dai, Y.; Su, Y.; Moore, P.; Zhang, X.; Mao, C.; Chen, J.; Xu, L. Feature Selection for Optimized High-Dimensional Biomedical Data Using an Improved Shuffled Frog Leaping Algorithm. IEEE/ACM Trans. Comput. Biol. Bioinform. 2018, 15, 1765–1773. [Google Scholar] [CrossRef] [PubMed]
  9. Pashaei, E.; Pashaei, E. Gene selection using hybrid dragonfly black hole algorithm: A case study on RNA-seq COVID-19 data. Anal. Biochem. 2021, 627, 114242. [Google Scholar] [CrossRef] [PubMed]
  10. Shreem, S.S.; Ahmad Nazri, M.Z.; Abdullah, S.; Sani, N.S. Hybrid Symmetrical Uncertainty and Reference Set Harmony Search Algorithm for Gene Selection Problem. Mathematics 2022, 10, 374. [Google Scholar] [CrossRef]
  11. Chaudhuri, A.; Sahu, T.P. A hybrid feature selection method based on Binary Jaya algorithm for micro-array data classification. Comput. Electr. Eng. 2021, 90, 106963. [Google Scholar] [CrossRef]
  12. Zhang, G.; Hou, J.; Wang, J.; Yan, C.; Luo, J. Feature Selection for Microarray Data Classification Using Hybrid Information Gain and a Modified Binary Krill Herd Algorithm. Interdiscip. Sci. Comput. Life Sci. 2020, 12, 288–301. [Google Scholar] [CrossRef]
  13. Pashaei, E.; Pashaei, E. Hybrid binary COOT algorithm with simulated annealing for feature selection in high-dimensional microarray data. Neural Comput. Appl. 2023, 35, 353–374. [Google Scholar] [CrossRef]
  14. Seyyedabbasi, A.; Kiani, F. Sand Cat swarm optimization: A nature-inspired algorithm to solve global optimization problems. Eng. Comput. 2022, 39, 2627–2651. [Google Scholar] [CrossRef]
  15. Kiani, F.; Anka, F.A.; Erenel, F. PSCSO: Enhanced sand cat swarm optimization inspired by the political system to solve complex problems. In Advances in Engineering Software; Elsevier: Amsterdam, The Netherlands, 2023; Volume 178, p. 103423. [Google Scholar]
  16. Yu, Y.; Li, Y.; Li, J.; Gu, X.; Royel, S. Nonlinear Characterization of the MRE Isolator Using Binary-Coded Discrete CSO and ELM. Int. J. Struct. Stab. Dyn. 2017, 18, 1840007. [Google Scholar] [CrossRef]
  17. Lu, W.; Shi, C.; Fu, H.; Xu, Y. A Power Transformer Fault Diagnosis Method Based on Improved Sand Cat Swarm Optimization Algorithm and Bidirectional Gated Recurrent Unit. Electronics 2023, 12, 672. [Google Scholar] [CrossRef]
  18. Zhao, W.; Zhang, Z.; Seyyedabbasi, A. Binary Sand Cat Swarm Optimization Algorithm for Wrapper Feature Selection on Biological Data. Biomimetics 2023, 8, 310. [Google Scholar] [CrossRef]
  19. Pashaei, E.; Pashaei, E. Training Feedforward Neural Network Using Enhanced Black Hole Algorithm: A Case Study on COVID-19 Related ACE2 Gene Expression Classification. Arab. J. Sci. Eng. 2021, 46, 3807–3828. [Google Scholar] [CrossRef] [PubMed]
  20. Yao, J.; Sha, Y.; Chen, Y.; Zhang, G.; Hu, X.; Bai, G.; Liu, J. IHSSAO: An Improved Hybrid Salp Swarm Algorithm and Aquila Optimizer for UAV Path Planning in Complex Terrain. Appl. Sci. 2022, 12, 5634. [Google Scholar] [CrossRef]
  21. Long, W.; Jiao, J.; Liang, X.; Wu, T.; Xu, M.; Cai, S. Pinhole-imaging-based learning butterfly optimization algorithm for global optimization and feature selection. Appl. Soft Comput. 2021, 103, 107146. [Google Scholar] [CrossRef]
  22. Shukla, A.K.; Singh, P.; Vardhan, M. A new hybrid wrapper TLBO and SA with SVM approach for gene expression data. Inf. Sci. 2019, 503, 238–254. [Google Scholar] [CrossRef]
  23. Yu, Y.; Rashidi, M.; Samali, B.; Yousefi, A.M.; Wang, W. Multi-Image-Feature-Based Hierarchical Concrete Crack Identification Framework Using Optimized SVM Multi-Classifiers and D–S Fusion Algorithm for Bridge Structures. Remote Sens. 2021, 13, 240. [Google Scholar] [CrossRef]
  24. Pashaei, E.; Yilmaz, A.; Aydin, N. A combined SVM and Markov model approach for splice site identification. In Proceedings of the 6th International Conference on Computer and Knowledge Engineering (ICCKE 2016), Mashhad, Iran, 20 October 2016. [Google Scholar]
  25. Pashaei, E.; Pashaei, E. Hybrid binary arithmetic optimization algorithm with simulated annealing for feature selection in high-dimensional biomedical data. J. Supercomput. 2022, 78, 15598–15637. [Google Scholar] [CrossRef]
  26. Dramiński, M.; Koronacki, J. rmcfs: An R Package for Monte Carlo Feature Selection and Interdependency Discovery. J. Stat. Softw. 2018, 85, 1–28. [Google Scholar] [CrossRef]
  27. Kursa, M.B. Praznik: High performance information-based feature selection. SoftwareX 2021, 16, 100819. [Google Scholar] [CrossRef]
  28. Bai, Y.; Lu, D.; Qu, D.; Li, Y.; Zhao, N.; Cui, G.; Li, X.; Sun, X.; Liu, Y.; Wei, M.; et al. The Role of ANGPTL Gene Family Members in Hepatocellular Carcinoma. Dis. Markers 2022, 2022, 1844352. [Google Scholar] [CrossRef]
  29. Lu, D.; Bai, X.; Zou, Q.; Gan, Z.; Lv, Y. Identification of the association between HMMR expression and progression of hepatocellular carcinoma via construction of a co-expression network. Oncol. Lett. 2020, 20, 2645–2654. [Google Scholar] [CrossRef]
  30. Zhang, L.; Fan, Y.; Wang, X.; Yang, M.; Wu, X.; Huang, W.; Lan, J.; Liao, L.; Huang, W.; Yuan, L.; et al. Carbohydrate Sulfotransferase 4 Inhibits the Progression of Hepatitis B Virus-Related Hepatocellular Carcinoma and Is a Potential Prognostic Marker in Several Tumors. Front. Oncol. 2020, 10, 554331. [Google Scholar] [CrossRef] [PubMed]
  31. Yao, T.; Hu, W.; Chen, J.; Shen, L.; Yu, Y.; Tang, Z.; Zang, G.; Zhang, Y.; Chen, X. Collagen XV mediated the epithelial-mesenchymal transition to inhibit hepatocellular carcinoma metastasis. J. Gastrointest. Oncol. 2022, 13, 2472–2484. [Google Scholar] [CrossRef]
  32. Wu, M.; Lan, H.; Ye, Z.; Wang, Y. Hypermethylation of the PZP gene is associated with hepatocellular carcinoma cell proliferation, invasion and migration. FEBS Open Bio 2021, 11, 826–832. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Principle of pinhole imaging opposition-based learning.
Figure 1. Principle of pinhole imaging opposition-based learning.
Bioengineering 10 01123 g001
Figure 2. Flow chart of the proposed PILC-BSCSO algorithm for gene selection.
Figure 2. Flow chart of the proposed PILC-BSCSO algorithm for gene selection.
Bioengineering 10 01123 g002
Figure 3. Volcano plot of the DEGs identified by limma for each dataset.
Figure 3. Volcano plot of the DEGs identified by limma for each dataset.
Bioengineering 10 01123 g003
Figure 4. The convergence behavior of BSCSO and PILC-BSCSO for three microarray datasets.
Figure 4. The convergence behavior of BSCSO and PILC-BSCSO for three microarray datasets.
Bioengineering 10 01123 g004
Figure 5. The gene expression level of the best subset of genes with the highest accuracy is shown as a heatmap.
Figure 5. The gene expression level of the best subset of genes with the highest accuracy is shown as a heatmap.
Bioengineering 10 01123 g005
Figure 6. Comparing the performance of the suggested methodology to approaches from the literature.
Figure 6. Comparing the performance of the suggested methodology to approaches from the literature.
Bioengineering 10 01123 g006
Figure 7. Kaplan–Meier analysis of the survival rates of the high- and low-expression groups of HMMR, CHST4, and COL15A1.
Figure 7. Kaplan–Meier analysis of the survival rates of the high- and low-expression groups of HMMR, CHST4, and COL15A1.
Bioengineering 10 01123 g007
Figure 8. Tissue-wise expression patterns of the identified best subset of genes associated with LIHC. Red color indicates tumors, and gray indicates normal samples.
Figure 8. Tissue-wise expression patterns of the identified best subset of genes associated with LIHC. Red color indicates tumors, and gray indicates normal samples.
Bioengineering 10 01123 g008
Table 1. Characteristics of Gene Expression Datasets.
Table 1. Characteristics of Gene Expression Datasets.
Dataset
Name
No. of SamplesNo. of FeaturesNo. of ClassesDistribution of Class Label
Colon cancer 62 2000 2 40, 22
CNS607129239, 21
Breast97 24,481251, 46
TCGA-LIHC 42156,6022371, 50
Table 2. The LOOCV classification accuracy of identified DEGs, mRMR, and MCFS with an SVM classifier.
Table 2. The LOOCV classification accuracy of identified DEGs, mRMR, and MCFS with an SVM classifier.
Dataset NameAll FeaturesDEGsmRMR
(50)
mRMR
(100)
mRMR
(200)
mRMR
(300)
MCFS
(50)
MCFS
(100)
MCFS
(200)
MCFS
(300)
Colon cancer 83.8785.4880.6483.8783.8780.6479.0388.7085.48388.70
CNS68.3390606333783368.3381.660.8591.6693.33
Breast67.0175.2576.2878.3578.3579.3872.1676.2887.6289.69
Table 3. Comparison between BSCSO and PILC-BSCSO in terms of classification accuracy and number of selected genes.
Table 3. Comparison between BSCSO and PILC-BSCSO in terms of classification accuracy and number of selected genes.
DatasetMetricsAccuracy #Genes
BSCSOGAPSOPILC-BSCSOBSCSOGAPSOPILC-BSCSO
ColonAVG97.6391.31194.3599.228.33133.670.415
best10093.5498.3810061135010
worst93.818583.8796.991458923
STDEV2.353.3494.471.3481.96612.9115.515.244
t-test (p-value)0.01950.00660.0519 0.01591.2259 × 10⁻⁵0.0022
CNSAVG99.3498.33299.1610033.25100.573.416.25
best10010010010014455413
worst98.499598.333100591449022
STDEV0.7552.0410.914018.7642.91413.294.0311
t-test (p-value)0.071980.06220.0755 0.04790.008080.00118
BreastAVG91.81991.0696.296.3811.4625826.4
best97.92695.871001005565215
worst88.753384.5393.8193.9816666540
STDEV3.8084.362.612.54.7224.325.0912.30
t-test (p-value)0.000970.0082460.6330 0.007300.000360.00047
Note: ‘#’ represents number of selected genes.
Table 4. Comparing the performance of the suggested methodology to approaches from the literature.
Table 4. Comparing the performance of the suggested methodology to approaches from the literature.
MethodsHigh Dimensional Biomedical Datasets
Colon CancerCNSBreast
∣#G∣ACC∣#G∣ACC∣#G∣ACC
PILC-BSCSO1599.2216.2510026.496.38
BMSCSO [2]997.8093.33----
mRMR-MBAO [5]16.1195.7421.3788.5723.5889.12
SU-RSHSA [10]7.5993.1713.1589.3618.3180.40
mRMR-DBH [9]1297.0239.7597.191490.21
IBCFPA [7]25.9092.1625.284.82--
MIM-MFO [6]24.2599.191785.0022.5084.11
BCROSAT [1]20.592.3121.4082.00--
ISFLA [8]37.189.5641.177.46--
TOPSIS-Jaya [11]18.9097.768.796.22--
IG-MBKH [12]17.1096.4714.7090.34--
mRMR-BCOOT-CSA [13]8.7594.75793.221595.54
Note: ‘#’ represents number of selected genes.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pashaei, E. An Efficient Binary Sand Cat Swarm Optimization for Feature Selection in High-Dimensional Biomedical Data. Bioengineering 2023, 10, 1123. https://0-doi-org.brum.beds.ac.uk/10.3390/bioengineering10101123

AMA Style

Pashaei E. An Efficient Binary Sand Cat Swarm Optimization for Feature Selection in High-Dimensional Biomedical Data. Bioengineering. 2023; 10(10):1123. https://0-doi-org.brum.beds.ac.uk/10.3390/bioengineering10101123

Chicago/Turabian Style

Pashaei, Elnaz. 2023. "An Efficient Binary Sand Cat Swarm Optimization for Feature Selection in High-Dimensional Biomedical Data" Bioengineering 10, no. 10: 1123. https://0-doi-org.brum.beds.ac.uk/10.3390/bioengineering10101123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop