Next Article in Journal
Differences in Dynamics of Lung Computed Tomography Patterns between Survivors and Deceased Adult Patients with COVID-19
Next Article in Special Issue
The Role of 3D CT Imaging in the Accurate Diagnosis of Lung Function in Coronavirus Patients
Previous Article in Journal
The Role of Bronchoscopy in the Diagnosis and Management of Patients with SARS-Cov-2 Infection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Brain Cancer Prediction Based on Novel Interpretable Ensemble Gene Selection Algorithm and Classifier

1
College of Computer Science and Engineering, Taibah University, Yanbu 46411, Saudi Arabia
2
Department of Computer Science, Faculty of Science, Abdelmalek Essaadi University, Tetouan 93000, Morocco
3
Math and Computer Science Department, Amran University, Amran 891-6162, Yemen
4
Math and Computer Science Department, Menoufia University, Menoufia 32511, Egypt
5
Department of Genomic Medicine, Faculty of Science, Cairo University, Cairo 12613, Egypt
6
Department of Mathematics, College of Computer Sciences and Mathematics, Tikrit University, Tikrit 34001, Iraq
7
Computers Engineering and Control Systems Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt
*
Authors to whom correspondence should be addressed.
Submission received: 23 August 2021 / Revised: 12 October 2021 / Accepted: 13 October 2021 / Published: 19 October 2021
(This article belongs to the Special Issue Computer-Assisted Functional Diagnostics)

Abstract

:
The growth of abnormal cells in the brain causes human brain tumors. Identifying the type of tumor is crucial for the prognosis and treatment of the patient. Data from cancer microarrays typically include fewer samples with many gene expression levels as features, reflecting the curse of dimensionality and making classifying data from microarrays challenging. In most of the examined studies, cancer classification (Malignant and benign) accuracy was examined without disclosing biological information related to the classification process. A new approach was proposed to bridge the gap between cancer classification and the interpretation of the biological studies of the genes implicated in cancer. This study aims to develop a new hybrid model for cancer classification (by using feature selection mRMRe as a key step to improve the performance of classification methods and a distributed hyperparameter optimization for gradient boosting ensemble methods). To evaluate the proposed method, NB, RF, and SVM classifiers have been chosen. In terms of the AUC, sensitivity, and specificity, the optimized CatBoost classifier performed better than the optimized XGBoost in cross-validation 5, 6, 8, and 10. With an accuracy of 0.91 ± 0.12 , the optimized CatBoost classifier is more accurate than the CatBoost classifier without optimization, which is 0.81 ±   0.24 . By using hybrid algorithms, SVM, RF, and NB automatically become more accurate. Furthermore, in terms of accuracy, SVM and RF ( 0.97 ± 0.08 ) achieve equivalent and higher classification accuracy than NB ( 0.91 ± 0.12 ). The findings of relevant biomedical studies confirm the findings of the selected genes.

1. Introduction

Brain cancer is the leading cause of death in women under the age of 20 and men under 40 [1,2]. Moreover, the prevalence of malignant brain tumors is rising [3], severely impacting society and human health [4]. Primary brain tumors arise from the brain cells themselves, while secondary brain tumors arise from malignant cells outside of the brain and spread there [5]. According to studies, brain tumors are extremely heterogeneous, posing a major challenge for classification, segmentation, diagnosis, and prognosis [6].
Microarray-based gene expression profiling has proven useful for cancer detection, prognosis, and treatment [7]. In addition, in recent years, DNA microarray technology has significantly impacted the information that we have bout what causes cancer [8,9].
Cancer microarray data typically contain a small number of samples with many gene expression levels as features, which leads to the curse of dimensionality, making the classification of microarray data a difficult task. The bioinformatics community uses a variety of approaches to classify microarray data using machine learning systems. Most of the examined studies on cancer classification using microarray data sets take cancer classification accuracy into account without disclosing any biological information of the cancer classification process. Few studies have investigated the biological interpretation of microarray data sets in addition to the model classification accuracy. This research aims to bridge the gap between cancer classification and biological interpretation by improving accuracy performance, and the selected significant genes agree with the findings of relevant biomedical studies.
This paper proposes a hybrid model based on three different machine learning techniques, including the commonly used ensemble classification methods of gradient boosting [10], an extremely efficient ML algorithm that produces a strong learner in the form of an ensemble of weak learners/models. Furthermore, to optimize the hyperparameters of machine learning algorithms, distributed hyperparameter optimization [11] is one of the most efficient methods (per function evaluation) that is utilized for parameter optimization. Furthermore, minimum redundancy maximum relevance (mrmr) [12] is a particularly fast feature selection method that can be used to find a set of relevant and complementary features. Our model was assessed using three different machine learning classifiers: random forest (RF), naive Bayes (NB), and support vector machines (SVM).
The experiments show that the proposed model substantially reduces the number of genes required for classification and improves classifier accuracy. Additionally, the proposed hybrid model’s selected genes (features) are biologically interpreted, and the biological interpretation coincides with the findings of relevant biomedical studies.
The main contributions of this work are:
  • The proposal of a novel ensemble classifier to ensure that the genes selected in our model are biologically interpreted. On top of that, the results are also satisfactory and in line with pertinent biomedical studies.
  • The identification of relevant and non-redundant genes for the biological context by ensemble mRMRs, allowing for enhanced biological interpretations.
  • The analysis of a brain cancer microarray dataset on high-dimensional data using Catboost and XGboost.
  • The optimization of the hyperparameters of the two classifiers using the hyperboot optimizer.
  • The outperformance of Catboost compared to XGboost with regard to the AUC, sensitivity, specificity, and accuracy.
The remainder of the paper is structured as follows: Section 2 briefly reviews related work, while Section 3 reviews some background research that was used in the proposed methodology. Section 4 explains the proposed model, while Section 5 discusses the experimental design, findings, and discussion. Finally, Section 6 concludes the whole paper by summarizing the contributions of the paper.

2. Literature Review

In recent years, multiple pieces of research have used a wide range of machine learning methods for the classification, diagnosis, and treatment of cancer disease. BU-Net was developed by Rehman [13] and others to segment and classify brain tumor regions. Their model with a modified encoder–decoder architecture was proposed for segmenting brain tumors. Tests of the proposed BU-Net architecture were conducted on BraTS 2017 and 2018. BU-Net proved to be a significant improvement over the baseline U-Net architecture and other existing segmentation models. Radiology mMRI imaging sub-regions can affect tumor localization, so researchers suggested a deep learning method [14] that considers the uncertainty of the tumor location. Then, they classified the tumor segments into subtypes using a conventional 3D convolutional neural network (CNN). Performance was measured by widely used measures such as the dice score coefficient, the Hausdorff distance at percentile 95 (HD95), classification accuracy, and mean square error. According to the findings, the suggested method can accurately segment tumors and predict survival rates.
Bashir et al. [15] used a fusion of five classifiers: naive Bayes, decision tree using Gini index, decision tree using information gain, support vector machines, and memory-based learner, to diagnose breast cancer. The weighted vote-based ensemble technique was then used to make the final prediction. Several preprocessing and feature selection methods were also used on four breast cancer datasets to improve the prediction accuracy. The proposed ensemble classifiers achieved remarkable results, with average accuracy, precision, and recall of 85.23%, 86.18%, and 76.68%. However, small datasets were used to test the performance of the model. Applying the proposed ensemble on a large dataset with many features may lead to computational instability.
To overcome the dimensionality problem, Kumar et al. [16] introduced the ANOVA, Kruskal–Wallis, and Friedman tests as examples of statistical methods (tests) based on MapReduce to select relevant features. The MapReduce-based proximal support vector machine (mrPSVM) classifier was also applied to classify the microarray data after feature selection. The Hadoop framework was used to implement these algorithms. Using the microarray datasets of different dimensions, a comparative study of these feature selection methodologies was performed. The experimental results showed that an ensemble of the mrPSVM classifier and various feature selection methods produced higher accuracy than other models. Thus, the proposed model successfully handled big data, but it could only interpret biological microarray data. Finally, a two-phase hybrid model for cancer classification was proposed (iBPSO) by Jain et al. [17] that consisted of correlation-based feature selection (CFS) and improved-binary particle swarm optimization. The proposed model uses the naive Bayes classifier to select a low-dimensional collection of prognostic genes to identify biological samples of binary and multi-class cancers. The model was evaluated and tested on eleven benchmark microarray datasets. The findings of the experiments showed that the model outperformed several well-known approaches in terms of classification accuracy and the number of selected genes.
Pradana et al. [18] introduce an approach that used binary particle swarm optimization (BPSO) as a feature selection and C4.5 decision tree as a classifier to investigate cancer diagnosis based on microarray data. The decision tree rule model requires discretization, which is accomplished by the use of K-Means. Applying BPSO and decision tree showed that the model could successfully find the most significant features and increase the accuracy. The model achieved accuracies of 54% and 99%, respectively, for C4.5 and BPSO. Shukla et al. [19] also proposed a new filter-based gene selection approachto identify highly important genes in microarray gene expression datasets. The proposed approach was evaluated using well-known classification techniques such as support vector machine, naive Bayes, k-nearest neighbor, and decision tree on the diffuse large B cell lymphoma (DLBCL) dataset. Experiments confirmed that the proposed model could successfully compete, with excellent predictions concerning the accuracy, precision, sensitivity, F-measure, and ROC value.
Sampathkumar et al. [20] developed a novel cuckoo search with a crossover algorithm that could accurately classify a variety of cancer subtypes. The model was tested on benchmark cancer gene expression, and the results show that CSC outperformed CS and other well-known methods. Kilicarslan et al. [21] used the relief algorithm for dimension reduction and feature ranking. The most important features were then used by support vector machines (SVM) and convolutional neural networks (CNN) [22] for prediction. The experimental results show that the proposed approach could improve the accuracy of SVM and CNN classification methods. Finally, Lee et al. [23] suggested a novel multivariate feature ranking approach to increase gene selection efficiency in microarray data. The proposed method created a new feature ranking method by embedding the formal concept of relevance into a Markov blanket (MB). The results show that the model performs well in high-dimensional microarray data classification.
Given all that has been mentioned so far, all of the algorithms listed above address two major issues: brain cancer classifiers and the curse of dimensionality. However, there is no biological interpretation of the microarray data set discussed in the literature. To the best of our knowledge, this paper is the first paper to provide a consolidated biological interpretation of the results of the proposed work. Table 1 summarizes some of the previous research methods for microarray cancer classification to conclude and review related work.

3. Materials and Methods

This section discusses ensemble classification and how to set the hyperparameter values and the minimum redundancy maximum relevance (mRMR) feature selection. The three main components of the proposed method are ensemble classification, hyperparameter optimization, and minimum redundancy maximum relevance (mRMR). First, the ensemble methods use multiple learning algorithms to achieve better predictive performance than if they were used by themselves. Secondly, we have hyperparameter optimization, in which the parameters of the classifier are tuned to find the optimal setting. The third one is minimum redundancy maximum relevance (mRMR), which is a filter-type feature selection method that obtains the best feature set by minimizing the similarities between features and classified variables and by maximizing their correlations

3.1. Ensemble Classification

Ensemble learning methods use multiple machine learning algorithms to generate weak predictive results based on the features extracted from various data projections. Following this, the results are fused with various voting mechanisms to achieve better results than any constituent algorithm alone [32,33]. Figure 1 depicts the basic concept of a typical ensemble classification model [34], which consists of two steps: (1) generating classification results using multiple weak classifiers and (2) integrating multiple results into a consistency function to obtain the results with voting schemes.
The gradient boosting method, which constructs the solution stagewise and solves the overfitting problem by optimizing the loss functions, is one of the most commonly used ensemble classification methods. The main concept of a gradient boosting model is depicted in Figure 2.
XGBoost is an efficient gradient-boosted decision tree algorithm [35]. Gradient boosting is a technique introduced by XGB. The new models are fitted to residuals from previous models, and the combined results are minimized using gradient descent. Yandex released CatBoost [36] in 2017 and demonstrated that it was faster in terms of prediction making, that it was more accurate, and that it was easier to use for categorical data across a series of GBDT tasks based on their benchmark. As a better gradient boosting algorithm, Catboost introduces ordered boosting. Table 2 illustrates the two gradient boosting algorithms.

3.2. Hyperparameter Optimization

Hyperopt is a Python library that implements sequential model-based optimization (SMBO) [9]. Hyperopt provides algorithms and software infrastructure to conduct hyperparameter optimization for machine learning algorithms. Hyperopt has an optimization interface that separates a configuration space from an evaluation function that assigns real-valued loss values to points in the configuration space. It works by treating the search of hyperparameters as an optimization task.
Hyperparameter search spaces are typically large multi-dimensional spaces. Hyperopt outperforms grid and random searches, particularly as the search space grows. Within the framework of our proposed model, Hyperopt is used to optimize the settings for the XGBoost and CatBoost hyperparameters. It aims to identify the optimal genes for microarray data analysis and to improve the classification of cancer microarrays.

3.3. Minimum Redundancy Maximum Relevance (mRMR) for Feature Selection

Minimum redundancy maximum relevance (mRMR) is a filter-type feature selection method that obtains the best feature set by maximizing the correlation between the features and the classified variables and by minimizing the correlation between features. The classic function enables the collection of appropriate and non-redundant features with ease [37]. In set S, the maximally important and minimally redundant gene i * is given by:
i * = a r g m a X i S   R S Q S , i
Ensemble (mRMR) feature selection implements two ensemble approaches: exhaustive and bootstrap ensemble mRMR. The exhaustive variant of the mRMR heuristic extends it by starting multiple feature selection procedures, with the k > 1 being the most relevant feature. Following that, k   m R M R solutions are generated in parallel, with the first selected feature being guaranteed to be different. The bootstrap variant resamples the original dataset (with replacement) to generate k bootstraps. Finally, classical mRMR feature selection is performed in parallel for each bootstrapped dataset, resulting in k mRMR solutions.

4. The Proposed Hybrid Model

This section describes the model (as shown in Figure 3) used for brain cancer classification. The Hyperopt optimizer is used to estimate the optimal values of the CatBoost hyperparameters.
The main process of the model can be defined as:
(i)
Preprocessing the dataset (brain cancer microarray). This step is vital to
  • Avoid features in greater numeric ranges dominating those in smaller ranges;
  • Avoid numerical difficulties during calculation;
  • Ensure that each feature is scaled to the range [0, 1].
(ii)
The data were partitioned into two sets: The training set is used for the training. The testing set is used to test final model 3, initializing CatBoost with specific solution parameters. Table 3 describes the parameter initialization of the classifiers
(iii)
CatBoost is used as a feature selector with 8-fold cross-validations (8 cross-validations of different levels of importance for every gene index). CatBoost calculates the means for each fold.
(iv)
By setting a threshold, irrelevant features are then removed. Suppose the score of a gene is above the threshold. In that case, the gene will be selected (as seen in Appendix A, the optimal threshold that offers the maximum accuracy is: 0.84). The genes are shuffled, and unique genes are kept.
(v)
The importance value of each gene is registered using a voting process. For example, the gene with index 1 in fold 0 receives an importance value of 1 if the same gene is present in the next fold; then, the gene importance is +1, and so on, for all of the 8-fold cross-validations. After this was applied, voting is conducted 50 genes (six of the filtered genes are genes with an importance >8.

5. Results and Discussion

In this section, we present the findings of the various experiments to judge the performance of the proposed hybrid model. A PC with the following features was used to test the proposed hybrid model: Intel(R) Core (TM) i5-7500 CPU with a 32-bit operating system, 4 GB RAM, and the Windows 7 operating system as well as the NumPy, SciPy, Pandas, Keras, and Matplotlib frameworks and Python 2.7 programming language.

5.1. Datasets

Molecular profiles from 28 patient samples were analyzed (data set A: medulloblastomas, CNS AT/RTs, renal and extrarenal rhabdoid tumors, supratentorial PNETs, and normal human cerebella). In addition, an analysis of frozen specimen RNA with oligonucleotide microarrays containing probes for 1070 genes was conducted. Gene expression data are available in the Supplementary Information [30].
k-fold tests were used in the majority of previous experiments. k-fold verification works to find crossed validity by randomly dividing data into k subsets of (approximately) equal size and k-times. As a result, it will run several times, with one subset serving as a test group and the other subset, k-1, serving as a training group (see Figure 3). The mean of the k-fold results can then be averaged to present a single evaluation. In our experiments, we used eight-fold cross-validation to evaluate the outcomes of the proposed hybrid model, and the results are represented as an average standard deviation. Furthermore, the total number of iterations in all of the experiments was 30. In our experiments, we used three evaluation methods [38]: specificity (Spec.), sensitivity (Sen.), and area under the curve (AUC) (AUC).
Sensitivity, as calculated by S = T P / T P + F N , is the likelihood that a diseased person is recognized as diseased through the test, where TN is the true negatives number, and TP is the true positives number, and N is the false negative. Specificity is the likelihood that a person without the illness is defined by the ((TEST) as non-diseased (or healthy). It is described as T N R = T N /   T N + F P , where FP implies the number of false positives, and TN is the number of true negatives. The AUC shows the area under the receiver operating characteristics (ROC) curve, calculated as A U C = ( 1 + T P R F P R ) / 2 .

5.2. Experiment 1: Comparing Performance of Optimized (CatBoost and XGBoost) with the Proposed Hybrid Model

This experiment compares optimized (CatBoost and XGBoost) classifiers in the proposed hybrid model to achieve the best hyperparameter of the two classifiers. The Hyperopt optimizer is used. Table 3 lists the hyperparameters settings values of XGBoost and CatBoost and the range of each parameter. The experimental results are shown in Table 4 and Figure 4 and Figure 5. The most striking results are better results for the optimized CatBoost classifier than the optimized XGBoost with cross-validation 5, 6, 8, and 10 in the AUC, Sen, and Spec results.
From the data in Figure 4 and Figure 5, it is apparent that CatBoost classifier has the best accuracy, 0.97   ±   0.08 , with 8-fold cross-validation and the best A U C = 0.97   ±   0.08 , S e n = 0.94   ± 0.17 , and S p e c = 1.00   ±   0.00 compared to the XGBoost classifier, which had an accuracy of 0.80   ±   0.21 , and where the A U C = 0.80   ±   0.21 , S e n = 0.80   ±   0.21 , and S p e c = 0.80   ±   0.21 . All in all, these results point to CatBoost having a higher performance than XGBoost in the hybrid model we developed. Table 5 shows the training vs. testing performance with 8-fold cross-validation in 28 samples.

5.3. Experiment 2: Comparing Performance of CatBoost and Optimized CatBoost Classifier

In this experiment, CatBoost and optimized CatBoost were compared with the original brain cancer microarray data (1070 feature, 28 samples). Table 6 and Figure 6 and Figure 7 show the classification report of the CatBoost and optimized CatBoost classifier with the brain cancer microarray data. The threshold value was 0.84, which is the threshold value with the highest accuracy (see Appendix A) with the final optimal genes (features) selected with our proposed hybrid model. The number of genes selected in each stage of feature selection is as follows:
  • Number of non-zero genes importance (every fold).
  • (588, 576, 590, 599, 594, 579, 585, 584).
  • The number of genes selected by embedded SVM (with Redundant), 980 genes.
  • The number of genes selected by embedded SVM (Unique), 671 genes.
  • The final number of genes after we applied voting was 50 genes.
Based on the performance metrics in Table 6, optimized CatBoost had higher performance than the CatBoost classifier without optimization. Figure 6 and Figure 7 show the accuracy curve of CatBoost and the optimized CatBoost classifier. The optimized CatBoost classifier had an accuracy of 0.91 ± 0.12, which is higher than that of the CatBoost classifier without optimization, which had an accuracy of 0.81 ± 0.24.

5.4. Experiment 3: Comparison of Hybrid Proposed Model Performance by Different Classification

The classic learning algorithms random forest (RF), naive Bayes (NB), and support vector machines (SVM) were used to assess the gene classification accuracy of selected optimal genes by the proposed hybrid model. These learning algorithms were applied to the newly collected dataset, which only included the best genes, and the overall accuracy was calculated. In Figure 8, Figure 9 and Figure 10, the learning accuracy of three classifiers is illustrated using the newly generated gene (feature) set. The proposed hybrid model increased the accuracy of the SVM, RF, and NB classifiers while the accuracy is weighted on brain data set; on the other hand, SVM and RF, with accuracies of (0.97 ± 0.08), achieve equal and higher classification accuracy than the NB (0.91 ± 0.12) classifier.

5.5. Biological Interpretation

A subset of genes (features) from the brain cancer data set was biologically interpreted to demonstrate the proposed model’s efficacy in improving critical items such as classification accuracy and for selecting genes with important biological backgrounds. A few classes of important genes derived from microarray technologies were used to diagnose and to provide the prognostic purposes of brain cancer after using the proposed hybrid model’s biological portrait.
The proposed hybrid model aims to determine crucial gene subsets with the maximum amount of accuracy needed to treat a brain cancer patient. In this segment, the selected group of probe sets could be studied by using the web tool DAVID (Database for Annotation, Integrated Discovery, and Visualization) https://david.ncifcrf.gov/list.jsp (accessed on 18 June 2020) [31,39]. Table 7 shows the gene name and gene ID from the Entrez probe set. GO Research Tools: http://www.geneontology.org/GO.tools.microarray, (accessed on 18 June 2020) are generally considered to be the most inclusive and fastest-growing public repository for grouping functionally related genes. Following that, it can be shown that the proposed approach is the most effective way to pick a large group of genes for brain cancer pathway detection and prognosis.
In Figure 11, heat maps representing the frequency of selected features over the cross-validation analysis are used to evaluate the consistency of the selected features over time and to identify genes that are differentially expressed between the two disease classes (cancer class and healthy class).
Figure 12 shows the correlation among the selected feature using our proposed model with the Catboost classifier. The correlation coefficient has values between −1 to 1. A value closer to 0 implies a weaker correlation (exact 0 implying no correlation). A value closer to 1 implies a stronger positive correlation, and a value closer to −1 implies a stronger negative correlation. We compared the correlation between features and removed one of two features that correlate to >= 0.5, that correlate to =0.5 in the first threshold, and to 0.5 + 0.01 in the second fold of the threshold.

6. Conclusions

Brain disorders are becoming a major issue, particularly malignant brain tumors, which significantly impact people’s lives. The brain cancer microarray data have proven to be a complicated classification task due to the small number of samples that have a large number of gene expression levels as features. As part of brain cancer microarray data analysis, the present study proposed an effective and powerful technique for the selection of significant and relevant genes with biomedical relevance. Three distinct techniques were used for classification and prediction (feature selection, optimization, and classification). We used the same dataset and three different algorithms to evaluate the performance of the proposed model (NB, RF, and SVM). The experimental results demonstrate that the proposed hybrid model significantly improves critical items such as classification accuracy and that the selected genes have an important biological background.
Furthermore, selecting optimal genes (features) with biological significance can assist biological researchers in brain cancer treatment. The major contributions of this paper are: (a) The application of Catboost and XGboost on high-dimensional microarray data to create a cancer microarray dataset; (b) the use of the hyperboot optimizer to optimize the hyperparameters of the two classifiers, and the outperformance of the Catboost on XGboost in terms of AUC, Sen, Spec, and accuracy; (c) the selection of genes that are non-redundant and relevant to the biological context using ensemble mRMR, which leads to more detailed biological interpretations. Later, the output of the gene subset was combined with the Catboost-selected features. Then, a voting process was applied to obtain unique, informative genes (features) with high relevance and minimum redundancy; (d) the selected genes in our proposed model were biologically interpreted, and the results agree with the findings of relevant biomedical studies. Developing robustness should be a priority for future work.

Author Contributions

Conceptualization: A.M.A., M.Q.; Data curation, H.F., A.K.K.; Formal analysis, A.M.A., H.F.; Investigation, A.M.A., H.F., A.K.K.; Methodology, A.M.A., M.Q., S.A. and M.E.; Software, A.M.A., N.K.H.; Supervision, M.E.; Validation, A.M.A., M.Q. and M.E.; Visualization, H.F., N.K.H.; Writing–original draft, A.M.A., H.F., M.A.; Writing–review & editing, A.M.A., M.A. and M.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Threshold (28: 1068)SVMRandom ForestNaive BayesCatBoost
AccuracySpecSENAUCAccuracySpecSENAUCAccuracySpecSENAUCAccuracySpecSENAUC
0.50.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.081.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.510.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.081.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.520.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.91 ± 0.170.88 ± 0.330.94 ± 0.170.91 ± 0.171.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.530.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.111.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.540.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.081.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.550.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.93 ± 0.131.00 ± 0.000.88 ± 0.220.94 ± 0.111.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.560.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.081.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.570.91 ± 0.171.00 ± 0.000.81 ± 0.350.91 ± 0.170.94 ± 0.110.94 ± 0.170.94 ± 0.170.94 ± 0.111.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.580.93 ± 0.131.00 ± 0.000.88 ± 0.220.94 ± 0.110.88 ± 0.120.94 ± 0.170.81 ± 0.240.88 ± 0.120.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.590.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.121.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.60.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.121.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.610.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.86 ± 0.190.88 ± 0.220.88 ± 0.220.88 ± 0.180.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.90 ± 0.141.00 ± 0.000.81 ± 0.240.91 ± 0.12
0.620.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.94 ± 0.170.88 ± 0.220.91 ± 0.120.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.90 ± 0.141.00 ± 0.000.81 ± 0.240.91 ± 0.12
0.630.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.83 ± 0.180.88 ± 0.220.81 ± 0.240.84 ± 0.170.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.640.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.650.94± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.081.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.91 ± 0.120.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.660.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.670.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.94 ± 0.170.94 ± 0.170.94 ± 0.170.94 ± 0.170.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.94 ± 0.11
0.680.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.94 ± 0.170.94 ± 0.170.94 ± 0.170.94 ± 0.170.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.690.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.91 ± 0.170.94 ± 0.170.88 ± 0.220.91 ± 0.170.97 ± 0.080.94 ± 0.171.00 ± 0.000.97 ± 0.080.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.70.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.91 ± 0.170.94 ± 0.170.88 ± 0.220.91 ± 0.170.91 ± 0.120.81 ± 0.241.00 ± 0.000.91 ± 0.120.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.710.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.91 ± 0.170.94 ± 0.170.88 ± 0.220.91 ± 0.170.93 ± 0.130.88 ± 0.221.00 ± 0.000.94 ± 0.110.91 ± 0.120.94 ± 0.170.81 ± 0.240.91 ± 0.12
0.720.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.88 ± 0.220.94 ± 0.170.91 ± 0.120.91 ± 0.121.00 ± 0.000.81 ± 0.240.91 ± 0.12
0.730.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.94 ± 0.170.94 ± 0.170.94 ± 0.170.94 ± 0.170.90 ± 0.140.88 ± 0.220.94 ± 0.170.91 ± 0.120.91 ± 0.121.00 ± 0.000.81 ± 0.240.91 ± 0.12
0.740.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.91 ± 0.170.94 ± 0.170.88 ± 0.220.91 ± 0.170.93 ± 0.130.88 ± 0.221.00 ± 0.000.94 ± 0.110.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.750.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.88 ± 0.220.94 ± 0.170.91 ± 0.120.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.760.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.94 ± 0.170.94 ± 0.170.94 ± 0.170.94 ± 0.170.80 ± 0.210.75 ± 0.350.88 ± 0.220.81 ± 0.210.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.770.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.80 ± 0.210.75 ± 0.350.88 ± 0.220.81 ± 0.210.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.780.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.86 ± 0.190.88 ± 0.220.88 ± 0.220.88 ± 0.180.83 ± 0.180.75 ± 0.350.94 ± 0.170.84 ± 0.170.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.791.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.90 ± 0.190.88 ± 0.220.94 ± 0.170.91 ± 0.170.83 ± 0.180.75 ± 0.350.94 ± 0.170.84 ± 0.170.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.81.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.86 ± 0.140.75 ± 0.350.94 ± 0.170.84 ± 0.170.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.11
0.811.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.83 ± 0.190.81 ± 0.240.94 ± 0.170.88 ± 0.180.90 ± 0.140.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.821.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.86 ± 0.190.81 ± 0.240.94 ± 0.170.88 ± 0.180.90 ± 0.140.94 ± 0.170.88 ± 0.220.91 ± 0.12
0.831.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.11
0.841.00 ± 0.001.00 ± 0.001.00 ± 0.001.00 ± 0.000.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.850.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.860.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.11
0.870.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.11
0.880.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.11
0.890.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.94 ± 0.111.00 ± 0.000.88 ± 0.220.94 ± 0.110.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.90.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.910.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.90 ± 0.140.81 ± 0.241.00 ± 0.000.91 ± 0.120.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.920.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.930.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.90 ± 0.140.94 ± 0.170.88 ± 0.220.91 ± 0.120.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.940.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.950.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.960.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.11
0.970.94 ± 0.110.94 ± 0.170.94 ± 0.170.94 ± 0.110.93 ± 0.130.94 ± 0.170.94 ± 0.170.94 ± 0.110.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.980.94 ± 0.110.94 ± 0.170.94 ± 0.170.94 ± 0.110.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08
0.990.94 ± 0.110.94 ± 0.170.94 ± 0.170.94 ± 0.110.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.080.83 ± 0.180.69 ± 0.351.00 ± 0.000.84 ± 0.170.97 ± 0.081.00 ± 0.000.94 ± 0.170.97 ± 0.08

References

  1. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer statistics, 2021. CA A Cancer J. Clin. 2021, 66, 7–30. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Rehman, M.U.; Cho, S.; Kim, J.; Chong, K.T. BrainSeg-Net: Brain Tumor MR Image Segmentation via Enhanced Encoder–Decoder Network. Diagnostics 2021, 11, 169. [Google Scholar] [CrossRef]
  3. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; Larochelle, H. Brain tumor segmentation with Deep Neural Networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Isensee, F.; Jäger, P.F.; Full, P.M.; Vollmuth, P.; Maier-Hein, K.H. nnU-net for brain tumor segmentation. In International MICCAI Brainlesion Workshop; Springer: Cham, Switzerland, 2020; pp. 118–132. [Google Scholar]
  5. Zeineldin, R.A.; Karar, M.E.; Coburger, J.; Wirtz, C.R.; Burgert, O. DeepSeg: Deep neural network framework for automatic brain tumor segmentation using magnetic resonance FLAIR images. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 909–920. [Google Scholar] [CrossRef]
  6. Perrin, S.L.; Samuel, M.S.; Koszyca, B.; Brown, M.P.; Ebert, L.M.; Oksdath, M.; Gomez, G.A. Glioblastoma heterogeneity and the tumour microenvironment: Implications for preclinical research and development of new treatments. Biochem. Soc. Trans. 2019, 47, 625–638. [Google Scholar] [CrossRef]
  7. Golub, T.R.; Slonim, D.K.; Tamayo, P.; Huard, C.; Gaasenbeek, M.; Mesirov, J.P.; Coller, H.; Loh, M.L.; Downing, J.R.; Caligiuri, M.A.; et al. Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring. Science 1999, 286, 531–537. [Google Scholar] [CrossRef] [Green Version]
  8. Leung, Y.F.; Cavalieri, D. Fundamentals of cDNA microarray data analysis. Trends Genet. 2003, 19, 649–659. [Google Scholar] [CrossRef] [PubMed]
  9. Flores, M.; Hsiao, T.-H.; Chiu, Y.-C.; Chuang, E.Y.; Huang, Y.; Chen, Y. Gene Regulation, Modulation, and Their Applications in Gene Expression Data Analysis. Adv. Bioinform. 2013, 2013, 1–11. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  11. Bergstra, J.; Komer, B.; Eliasmith, C.; Yamins, D.; Cox, D. Hyperopt: A Python library for model selection and hyperparameter optimization. Comput. Sci. Discov. 2015, 8, 014008. [Google Scholar] [CrossRef]
  12. De Jay, N.; Papillon-Cavanagh, S.; Olsen, C.; El-Hachem, N.; Bontempi, G.; Haibe-Kains, B. mRMRe: An R package for parallelized mRMR ensemble feature selection. Bioinformatics 2013, 29, 2365–2368. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Rehman, M.U.; Cho, S.; Kim, J.H.; Chong, K.T. BU-Net: Brain Tumor Segmentation Using Modified U-Net Architecture. Electronics 2020, 9, 2203. [Google Scholar] [CrossRef]
  14. Pei, L.; Vidyaratne, L.; Rahman, M.; Iftekharuddin, K.M. Context aware deep learning for brain tumor segmentation, subtype classification, and survival prediction using radiology images. Sci. Rep. 2020, 10, 1–11. [Google Scholar] [CrossRef] [PubMed]
  15. Bashir, S.; Qamar, U.; Khan, F.H. Heterogeneous classifiers fusion for dynamic breast cancer diagnosis using weighted vote based ensemble. Qual. Quant. 2015, 49, 2061–2076. [Google Scholar] [CrossRef]
  16. Kumar, M.; Rath, S. Classification of microarray using MapReduce based proximal support vector machine classifier. Knowl.-Based Syst. 2015, 89, 584–602. [Google Scholar] [CrossRef]
  17. Jain, I.; Jain, V.K.; Jain, R. Correlation feature selection based improved-Binary Particle Swarm Optimization for gene selection and cancer classification. Appl. Soft Comput. 2018, 62, 203–215. [Google Scholar] [CrossRef]
  18. Pradana, A.C.; Aditsania, A. Implementing binary particle swarm optimization and C4.5 decision tree for cancer detection based on microarray data classification. J. Physics: Conf. Ser. 2019, 1192, 012014. [Google Scholar] [CrossRef]
  19. Shukla, A.K.; Tripathi, D. Detecting biomarkers from microarray data using distributed correlation based gene selection. Genes Genom. 2020, 42, 449–465. [Google Scholar] [CrossRef] [PubMed]
  20. Sampathkumar, A.; Rastogi, R.; Arukonda, S.; Shankar, A.; Kautish, S.; Sivaram, M. An efficient hybrid methodology for detection of cancer-causing gene using CSC for micro array data. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 4743–4751. [Google Scholar] [CrossRef]
  21. Kilicarslan, S.; Adem, K.; Celik, M. Diagnosis and classification of cancer using hybrid model based on ReliefF and convolutional neural network. Med. Hypotheses 2020, 137, 109577. [Google Scholar] [CrossRef] [PubMed]
  22. Naser, M.A.; Deen, M.J. Brain tumor segmentation and grading of lower-grade glioma using deep learning in MRI images. Comput. Biol. Med. 2020, 121, 103758. [Google Scholar] [CrossRef] [PubMed]
  23. Lee, J.; Choi, I.Y.; Jun, C.-H. An efficient multivariate feature ranking method for gene selection in high-dimensional microarray data. Expert Syst. Appl. 2021, 166, 113971. [Google Scholar] [CrossRef]
  24. Zexuan Zhu, Y.S. Markov Blanket-Embedded Genetic Algorithm for Gene Selection. Available online: http://csse.szu.edu.cn/staff/zhuzx/Datasets.html (accessed on 12 May 2020).
  25. Ramey, J. Datamicroarray. Available online: https://github.com/ramhiser/datamicroarray/blob/master/inst/data_scripts/singh-2002/1-download.r (accessed on 18 October 2021).
  26. Smith, A. GEO DataSets. (N. L. Medicine, Editor). Available online: https://0-www-ncbi-nlm-nih-gov.brum.beds.ac.uk/gds (accessed on 21 March 2021).
  27. Street, N. Datasets.php. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 7 June 2021).
  28. Jamil, L.S. Data analysis based on data mining algorithms using weka workbench. Int. J. Eng. Sci. Res. Technol. 2016. [Google Scholar] [CrossRef]
  29. Statnikov, A.; Tsamardinos, I.; Dosbayev, Y.; Aliferis, C.F. GEMS: A system for automated cancer diagnosis and biomarker discovery from microarray gene expression data. Int. J. Med Inform. 2005, 74, 491–503. [Google Scholar] [CrossRef] [PubMed]
  30. Scott, L.; Pomeroy, P.; Tamayo, P.; Michelle, G. Prediction of central nervous system embryonal tumour outcome based on gene expression: Letters to nature. Nature 2002, 415, 436–442. [Google Scholar]
  31. Dennis, G.; Sherman, B.T.; Hosack, D.A.; Yang, J.; Gao, W.; Lane, H.C.; Lempicki, R.A. DAVID: Database for annotation, visualization, and integrated discovery. Genome Biol. 2003, 4, 1–11. [Google Scholar] [CrossRef] [Green Version]
  32. Qu, Y.; Li, R.; Deng, A.; Shang, C.; Shen, Q. Non-unique decision differential entropy-based feature selection. Neurocomputing 2019, 393, 187–193. [Google Scholar] [CrossRef]
  33. Zhou, Z.H. Ensemble Methods: Foundations and Algorithms; CRC Press: Boca Raton, FL, USA, 2012. [Google Scholar]
  34. Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
  35. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, New York, NY, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  36. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
  37. Ding, C.; Peng, H. Minimum redundancy feature selection from microarraygene expression data. J. Bioinform. Comput. Biol. 2005, 3, 185–205. [Google Scholar] [CrossRef] [PubMed]
  38. Nema, S.; Dudhane, A.; Murala, S.; Naidu, S. RescueNet: An unpaired GAN for brain tumor segmentation. Biomed. Signal Process. Control. 2020, 55, 101641. [Google Scholar] [CrossRef]
  39. Hosack, D.A.; Dennis, G., Jr.; Sherman, B.T.; Lane, H.C.; Lempicki, R.A. Identifying biological themes within lists of genes with EASE. Genome Biol. 2003, 4, R70. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The ensemble classification framework.
Figure 1. The ensemble classification framework.
Diagnostics 11 01936 g001
Figure 2. The gradient boosting framework.
Figure 2. The gradient boosting framework.
Diagnostics 11 01936 g002
Figure 3. The proposed hybrid model.
Figure 3. The proposed hybrid model.
Diagnostics 11 01936 g003
Figure 4. Accuracy curve obtained using the hybrid model with optimized CatBoost classifier.
Figure 4. Accuracy curve obtained using the hybrid model with optimized CatBoost classifier.
Diagnostics 11 01936 g004
Figure 5. Accuracy curve obtained using the hybrid model with optimized XGBoost classifier.
Figure 5. Accuracy curve obtained using the hybrid model with optimized XGBoost classifier.
Diagnostics 11 01936 g005
Figure 6. Accuracy curve obtained using the hybrid model with CatBoost classifier.
Figure 6. Accuracy curve obtained using the hybrid model with CatBoost classifier.
Diagnostics 11 01936 g006
Figure 7. Accuracy curve obtained using the hybrid model with optimized CatBoost classifier.
Figure 7. Accuracy curve obtained using the hybrid model with optimized CatBoost classifier.
Diagnostics 11 01936 g007
Figure 8. Accuracy curve of random forest classifier.
Figure 8. Accuracy curve of random forest classifier.
Diagnostics 11 01936 g008
Figure 9. Accuracy curve of SVM classifier.
Figure 9. Accuracy curve of SVM classifier.
Diagnostics 11 01936 g009
Figure 10. Accuracy curve of NB classifier.
Figure 10. Accuracy curve of NB classifier.
Diagnostics 11 01936 g010
Figure 11. (a) Hierarchical clustering dendrogram maps of the genes selected in the proposed hybrid model. (b) Heat maps of the genes selected in the proposed hybrid model.
Figure 11. (a) Hierarchical clustering dendrogram maps of the genes selected in the proposed hybrid model. (b) Heat maps of the genes selected in the proposed hybrid model.
Diagnostics 11 01936 g011aDiagnostics 11 01936 g011b
Figure 12. The correlation among the selected genes with the proposed hybrid mode.
Figure 12. The correlation among the selected genes with the proposed hybrid mode.
Diagnostics 11 01936 g012
Table 1. Review of previous studies on the cancer microarray data classification.
Table 1. Review of previous studies on the cancer microarray data classification.
AuthorMethodRemarkLimitationsDataset
Bashir, S., Qamar, U., and Khan, F. H. (2015) [15].(Naïve Bayes, DT-Gini, DT-IG, MBL
and SVM)
-
A weighted vote-based ensemble fusion of heterogeneous classifiers was introduced for dynamic breast cancer diagnosis.
-
Small datasets were used to test the performance of the model.
-
A small set of features are used to test the proposed model.
-
UCI [24]
-
Wisconsin Clinical [25]
Kumar, M., and Rath, S. K. (2015) [16].(MrPSVM)
-
Data on microarrays are classified using proximal support vector machines (mrPSVMs) based on MapReduce.
-
Large-scale results were managed using Hadoop.
-
No biological interpretation of the microarray data set was discussed.
-
Kent Ridge Bio-medical Data Set Repository [26]
-
National center of Biotechnology Information (NCBI GEO) [27]
Jain, I., Jain, V. K., and Jain, R. (2018) [17](CFS) and (iBPSO)
-
Hybrid feature selection is proposed for gene selection and cancer classification that combines correlation-based and binary particle swarm optimization.
-
Tested on eleven benchmark microarray datasets
Biological information on the cancer classification process is not discussed.
-
Kent Ridge Bio-medical Data Set Repository [28]
Pradana, A. C., and Aditsania, A. (2019, March) [18](BPSO)
Decision Tree C4.5)
-
Introduced binary particle swarm optimization and C4. 5 decision tree for cancer detection based on microarray data classification.
-
Used Random Forest Ranking (RFR) as filtering methods to order genes
By using filtering methods, some important features may not be included.
There is no interpretation of the results.
-
Kent Ridge Bio-medical Data Set Repository [29]
Shukla, A. K., and Tripathi, D. (2020) [19]Spearman’s Correlation (SC) and distributed FS
-
Introduced a new filter-based method for gene selection that can select the highly relevant genes for distinguishing tissues from the gene expression dataset
-
Biological information is not addressed.
-
DLBCL [30,31]
Table 2. XGBoost and CatBoost classifiers.
Table 2. XGBoost and CatBoost classifiers.
Base
Classifiers
Ways to Prevent
Overfitting
The Loss Function for Binary Classification
XGBoostRegression trees
Row Subsampling
Shrinkage parameter
Column subsampling
Regularization term in the objective function
L = y   l o g   p x + 1 y   l o g   1 p x , p x = 1 1 + exp   F x
CatBoostClassification trees
Row Subsampling
Shrinkage parameter
Column subsampling
L i = y i log p i 1 y i l o g 1 p i
Table 3. Parameter initialization.
Table 3. Parameter initialization.
XGBoost ClassifierCatBoost Classifier
HyperparametersRangeHyperparametersRange
iterations[1, 500]n_estimators[50, 900]
depth[1, 16]max_depth[1, 12]
subsample[0.5, 1]m_child_weight[1, 6]
rsm[0.75, 1.0]gamma[0.5, 1]
learning_rate[−3.0, −0.7]subsample[0.5, 1]
l2_leaf_reg[1, 10]learning_rate[log(0.001), log(0.3)]
random_strength[1 × 10−9, 10]colsample_bytree[0.5, 1]
bagging_temperature[0.0, 1.0]
scale_pos_weight[0.01, 1.0]
Table 4. Comparing performance of optimized (XGboost and Catboost) in terms of AUC, Sen., and Spec.
Table 4. Comparing performance of optimized (XGboost and Catboost) in terms of AUC, Sen., and Spec.
Cross Validation (CV)XGBoost ClassifierCatBoost Classifier
AUCSenSpecAUCSenSpec
Cv = 50.80 ± 0.160.80 ± 0.160.80 ± 0.160.87 ± 0.070.80 ± 0.160.93 ± 0.13
Cv = 60.83 ± 0.130.75 ± 0.260.93 ± 0.190.89 ± 0.110.86 ± 0.200.92 ± 0.19
Cv = 80.81 ± 0.160.81 ± 0.240.81 ± 0.240.91 ± 0.120.88± 0.210.93 ± 0.16
Cv = 100.75 ± 0.290.75 ± 0.330.75 ± 0.400.88 ± 0.170.85 ± 0.230.90 ± 0.03
Table 5. Training vs. testing performance of optimized Catboost with 8-fold cross-validation.
Table 5. Training vs. testing performance of optimized Catboost with 8-fold cross-validation.
Fold NumberOptimized CatBoost Classifier
Train AccuracyTest Accuracy
11.000.750
21.001.00
31.000.750
41.001.00
51.001.00
61.001.00
71.001.00
81.001.00
Table 6. Classification report for CatBoost and optimized CatBoost.
Table 6. Classification report for CatBoost and optimized CatBoost.
Brain Cancer DatasetCatBoostOptimized CatBoost
PrecisionRecallf1-ScoreSupportPrecisionRecallf1-ScoreSupport
0.00.850.790.81140.920.790.8514
1.00.800.860.83140.810.930.8714
accuracy 0.8214 0.8214
Macro avg0.820.820.82280.820.820.8228
Weighted avg0.820.820.82280.820.820.8228
Table 7. Gene accession number and gene description of the selected genes of brain cancer by the proposed hybrid model.
Table 7. Gene accession number and gene description of the selected genes of brain cancer by the proposed hybrid model.
Prob setGene IDGene NameDiagnostic MarkerPrognostic MarkerOverexpressionDown
Expression
1860_at1860Tumor protein p53 binding protein 2(TP53BP2) Unfavorable+
286_at286Histone cluster 2 H2A family member a4(HIST2H2AA4)Yes +
31667_r_at31667_rNuclear receptor subfamily 2 group E member 3(NR2E3)Yes +
33242_at33242TSR2, ribosome maturation factor(TSR2)Yes +
34088_at34088Neurexophilin 4(NXPH4) Unfavorable+
37055_at37055ETS variant 1(ETV1) Unfavorable+
37701_at37701Regulator of G-protein signaling 2(RGS2) Unfavorable+
40388_at40388DLG associated protein 1(DLGAP1) Unfavorable
41098_at41098Dishevelled associated activator of morphogenesis 2(DAAM2) +
1972_s_at1972_sMicrotubule associated protein 2(MAP2) Unfavorable+
32647_at32647Vesicle transport through interaction with t-SNAREs 1B(VTI1B)Yes +
36073_at36073Necdin, MAGE family member(NDN) Unfavorable+
37360_at37360Lymphocyte antigen 6 complex, locus E(LY6E)Yes +
38420_at38420Collagen type V alpha 2 chain(COL5A2) Unfavorable+
39673_i_at39673_iExtracellular matrix protein 2(ECM2) Unfavorable +
41387_r_at41387_rLysine demethylase 6B(KDM6B) Unfavorable+
41407_at41407MicroRNA 1236(MIR1236) Unfavorable+
41725_at41725Casein kinase 1 gamma 2(CSNK1G2) Unfavorable+
41732_at41732BolA family member 2(BOLA2) Favorable+
103_at103Thrombospondin 4(THBS4) Unfavorable +
1230_g_at1230_gMyotubularin related protein 11(MTMR11)Yes +
1396_at1396Insulin like growth factor binding protein 5(IGFBP5) Unfavorable+
32988_at32988Chloride voltage-gated channel Ka(CLCNKA) Unfavorable+
33854_at33854ATPase H+ transporting V1 subunit D(ATP6V1D) Unfavorable+
37209_g_at37209_gPhosphoserine phosphatase(PSPH) Unfavorable+
35297_at35297NADH:ubiquinone oxidoreductase subunit AB1(NDUFAB1) Unfavorable+
36155_at36155SPARC/osteonectin, cwcv and kazal-like domains proteoglycan 2(SPOCK2) Favorable+
36534_at36534DIX domain containing 1(DIXDC1) Unfavorable+
36617_at36617Inhibitor of DNA binding 1, HLH protein(ID1) Unfavorable +
38440_s_at38440_sArmadillo repeat containing, X-linked 6(ARMCX6) Unfavorable+
39315_at39315Angiopoietin 1(ANGPT1) Unfavorable+
39364_s_at39364_sProtein phosphatase 1 regulatory subunit 3C(PPP1R3C) Unfavorable
39512_s_at39512_sInositol polyphosphate-4-phosphatase type I A(INPP4A) +
39850_at39850Ankyrin 2(ANK2) Unfavorable+
755_at755Inositol 1,4,5-trisphosphate receptor type 1(ITPR1)
31386_at31386Immunoglobulin kappa variable 1/OR2-118 (IGKV1OR2-118) (pseudogene) Unfavorable+
33580_r_at33580_rGalanin receptor 3(GALR3) +
34193_at34193Cell adhesion molecule L1 like(CHL1) Unfavorable+
35349_at35349COP9 signalosome subunit 3(COPS3) Unfavorable +
35719_at35719PH domain and leucine rich repeat protein phosphatase 1(PHLPP1) Unfavorable+
38967_at38967Chromosome 14 open reading frame 2(C14orf2) Unfavorable+
39329_at39329Actinin alpha 1(ACTN1)yesUnfavorable+
41530_at41530Acetyl-CoA acyltransferase 2(ACAA2) Favorable+
38397_at38397DNA polymerase delta 4, accessory subunit(POLD4) Unfavorable
39008_at39008Ceruloplasmin(CP)
40767_at40767Tissue factor pathway inhibitor(TFPI) Unfavorable+
41214_at41214Ribosomal protein S4, Y-linked 1(RPS4Y1) Unfavorable+
31342_at31342Polypeptide N-acetylgalactosaminyltransferase 2(GALNT2) Unfavorable+
32109_at32109FXYD domain-containing ion transport regulator 1(FXYD1)yesUnfavorable+
32458_f_at32458_fProline rich protein BstNI subfamily 4(PRB4) Unfavorable+
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Almars, A.M.; Alwateer, M.; Qaraad, M.; Amjad, S.; Fathi, H.; Kelany, A.K.; Hussein, N.K.; Elhosseini, M. Brain Cancer Prediction Based on Novel Interpretable Ensemble Gene Selection Algorithm and Classifier. Diagnostics 2021, 11, 1936. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics11101936

AMA Style

Almars AM, Alwateer M, Qaraad M, Amjad S, Fathi H, Kelany AK, Hussein NK, Elhosseini M. Brain Cancer Prediction Based on Novel Interpretable Ensemble Gene Selection Algorithm and Classifier. Diagnostics. 2021; 11(10):1936. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics11101936

Chicago/Turabian Style

Almars, Abdulqader M., Majed Alwateer, Mohammed Qaraad, Souad Amjad, Hanaa Fathi, Ayda K. Kelany, Nazar K. Hussein, and Mostafa Elhosseini. 2021. "Brain Cancer Prediction Based on Novel Interpretable Ensemble Gene Selection Algorithm and Classifier" Diagnostics 11, no. 10: 1936. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics11101936

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop