Next Article in Journal
Switching Coefficients or Automatic Variable Selection: An Application in Forecasting Commodity Returns
Previous Article in Journal
High-Resolution Gridded Air Temperature Data for the Urban Environment: The Milan Data Set
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable Ensemble Machine Learning for Breast Cancer Diagnosis Based on Ultrasound Image Texture Features

1
Department of Electrical and Computer Engineering, University of Minnesota, Minneapolis, MN 55414, USA
2
Department Computer Science and Engineering, University of Minnesota, Minneapolis, MN 55414, USA
3
Division of Interventional Radiology, Department of Radiology, University of Cincinnati, Cincinnati, OH 45221, USA
*
Author to whom correspondence should be addressed.
Submission received: 17 January 2022 / Revised: 6 February 2022 / Accepted: 11 February 2022 / Published: 13 February 2022
(This article belongs to the Section Forecasting in Computer Science)

Abstract

:
Image classification is widely used to build predictive models for breast cancer diagnosis. Most existing approaches overwhelmingly rely on deep convolutional networks to build such diagnosis pipelines. These model architectures, although remarkable in performance, are black-box systems that provide minimal insight into the inner logic behind their predictions. This is a major drawback as the explainability of prediction is vital for applications such as cancer diagnosis. In this paper, we address this issue by proposing an explainable machine learning pipeline for breast cancer diagnosis based on ultrasound images. We extract first- and second-order texture features of the ultrasound images and use them to build a probabilistic ensemble of decision tree classifiers. Each decision tree learns to classify the input ultrasound image by learning a set of robust decision thresholds for texture features of the image. The decision path of the model predictions can then be interpreted by decomposing the learned decision trees. Our results show that our proposed framework achieves high predictive performance while being explainable.

1. Introduction

Ultrasound imaging is an effective method for breast cancer diagnosis [1,2] that, compared to alternative modalities, is more accessible and less costly. Several recent studies have explored building data-driven automated breast cancer diagnosis machine learning pipelines to detect the malignancy of tumors observed in ultrasound images [3,4,5,6]. These studies dominantly rely on deep convolutional neural network architectures to classify tumor images. Convolutional neural networks essentially learn to map the input image pixel information to a lower-dimensional feature space through a series of hidden layers.
Although notable in prediction performance, convolutional networks are largely black-box machine learning models that provide little to no insight into the logic behind their predictions [7,8]. In general, humans tend to be unwilling to rely on procedures that are not interpretable, explainable and transparent, especially for making critical predictions such as cancer diagnosis [8,9]. However, the explainability of machine learning models for cancer diagnosis falls short of the increasing demand for interpretable and reliable artificial intelligence [10].
In a recent study, Moon et al. [3] adopted standard deep convolutional neural network architectures (including VGG, ResNet, and DenseNet) to classify breast ultrasound images and detect the malignancy of tumors. The study reports a high predictive performance for these standard architectures. In a similar study, Masud et al. [11] evaluated pretrained convolutional models for ultrasound image classification. Other studies, such as [12,13], focused on semantic segmentation of the breast tumor from ultrasound images.
In this paper, we propose an explainable machine learning pipeline for probabilistic breast cancer diagnosis based on ultrasound images. We formulate this as a binary classification problem. First, through a comprehensive texture analysis, we extract first- and second-order texture features of the region of interest in the ultrasound image. We then use these features to train an ensemble of decision trees. Each decision tree in the model learns to classify the input image through a set of test conjunctions where each test compares a texture feature with a robust numerical decision threshold. We show that our proposed pipeline achieves a high predictive performance that is comparable to the existing black box convolution neural network architectures. More importantly, we demonstrate that our proposed model can be probed to accurately track and explain the decision path behind its prediction.

2. Materials and Methods

2.1. Data

We use a public dataset of breast ultrasound images [2]. In this dataset, a total number of 780 images are obtained from 600 female patients (age of 25–75 years old). This includes 133 normal cases with no mass, 210 cases with a benign mass, and 487 cases with a malignant mass. Images are obtained using a LOGIQ E9 ultrasound and LOGIQ E9 Agile ultrasound systems. These instruments produce DICOM images with 1280 × 1024 resolution using 1–5 MHz transducers on an ML6-15-D Matrix linear probe. The raw DICOM images are cropped, preprocessed, and converted to PNG format with an average resolution of 500 × 500 pixels. For each case with a mass, a ground-truth binary mask of the region of interest (ROI) is manually created. The dataset is split into an 80 % train set and a 20 % test set. The model performance metrics are reported based on a k-fold cross validation (k = 5).

2.2. Texture Analysis

Texture features are important quantifiable metrics to characterize and describe a region of interest in an image [14,15,16]. Texture feature analysis is typically measured using first-order and second-order statistical metrics [17].

2.2.1. First-Order Statistics’ Texture Features

First-order texture features are computed based on first-order statistics of the one-dimensional gray level histogram of the image. Therefore, a first-order texture feature does not take into account the pixel neighborhood information [17,18]. In this study, we compute eight common first-order statics of the ROI pixels: Mean, Variance, RMS, Energy, Entropy, Kurtosis, Skewness, and Uniformity. These statistical metrics described in Table A1.

2.2.2. Second-Order Statistics’ Texture Features

The second-order statistical texture features are computed based on the gray-level co-occurrence matrix (GLCM). GLCM elements are an estimation of the probability of transition from one gray level to another along a certain pixel distance and direction [19,20,21]. We measure five common statistics based on the GLCM computed for the ROI: Contrast, Dissimilarity, Homogeneity, Energy, and Correlation. A description of these statistics is included in Table A2.

2.3. Decision Tree Models

A decision tree (DT) is a predictive model that consists of a set of test conjunctions, where each test compares a feature of data with a numerical threshold [22]. Decision tree classification models are learned by recursively partitioning the feature space to discover a set of robust decision rules [23,24]. One major advantage of decision tree modeling is their interpretability. The set decision rules learned in a decision tree model can be directly used to explain the logic behind the prediction of the model. Explainable predictions, as opposed to black-box predictions, can be used more reliably in applications such as medical diagnosis.

2.3.1. Gradient Boosting Decision Tree

Gradient boosting decision tree (GBDT) is an ensemble of sequential decision tree models. GBDT is frequently used in a variety of machine learning tasks due to its accuracy and efficiency. At each boosting iteration, the ensemble learns a decision tree to predict the residual errors [22,25]. Specifically, for a classification problem, let { ( X i , y i ) } i = 1 N be a dataset where for each entry, X and y, correspond to feature vector and the class label the entry belongs to. The goal is to approximate the function F ^ ( X ) = y that learns to map feature vectors to their corresponding class label under an arbitrary differentiable loss function L ( y i , F ( X i ) ) , cross-entropy loss, in our case.
With gradient boosting, at first, the model is initialized with a constant,
F 0 ( x ) = min λ i = 1 n L ( y i , λ )
Next, at each boosting iteration m, for each entry, the residuals are computed as
r i , m = L ( y , F ) / F | F = F m 1
A decision tree with J terminal nodes is fitted to the residuals where
λ j , m = min λ j L ( y , F m 1 ( x ) + λ )
Finally, the boosted model is updated as:
F m ( x ) = F m 1 ( x ) + ν J λ j , m I ( x )
Through grading boosting, GBDT combines multiple “weak” learner classifiers into an ensemble of a strong classification model.

2.3.2. LightGBM

LightGBM is an open-source GBDT framework [25]. LightGBM is based on a gradient-based one-side sampling to filter data instances and an exclusive feature bundling to encode features into less dense space. Specifically, LightGBM discretizes continuous features using a histogram-based algorithm for a faster training process and reduced memory consumption. In addition, LightGBM uses a leaf-wise strategy of growing decision trees by discovering a leaf with the highest gain of variance. This enables LightGBM to achieve state-of-the-art performance in a variety of applications [26,27,28].

2.4. Machine Learning Diagnosis Pipeline

In this paper, we propose an interpretable machine learning pipeline for breast cancer diagnosis based on ultrasound images. We formulate this as a binary classification problem where class labels belong to {benign, malignant}. As shown in Figure 1, the pipeline takes as input the ultrasound image of the mass and a binary mask of the ROI. We use this binary mask to crop and compute the GLCM of the ROI. This yields a total number of 60 texture features (5 GLCM statistics ×   3 distances ×   4 angles). These texture features are denoted as {statistic}_d{distance}_a{angle} (e.g., energy_d3_a135 refers to the GLCM energy computed within a distance of three pixels and along a 135 direction).
We then use a LightGBM classification model with a gradient boosting decision tree strategy, 10 leaves per tree, and a maximum feature bin size of 512. The classification model is trained by minimizing a binary log loss, with a learning rate of 0.05, and for a total number of 500 boosting iterations. For a given ultrasound image input, the model outputs the probability of the mass in ROI being benign or malignant. Importantly, the decision tree ensemble can be decomposed to explain how the model comes up with a prediction. The learned model is a set of decision trees with multiple test conjunctions that compare the texture features of the ROI with numerical thresholds inferred from the data.

3. Results

In this section, we first summarize the texture analysis results and then evaluate the performance of our purposed pipeline. Lastly, we highlight how our pipeline can be used as an explainable machine learning framework to understand the logic behind each of its diagnostic predictions.

3.1. Texture Analysis and Statistical Analysis

We perform texture analysis by computing the ROI first- and second-order statistics (see Table A1 and Table A2 for mathematical descriptions). A standard t-test is used to compare each texture feature between the two groups of benign and malignant masses.
Most first-order statistic texture features are significantly different for the two sets of benign and malignant masses ( p < 0.001 ). Mean ( p = 0.43 ) and RMS ( p = 0.28 ) first-order statistics, however, are not significantly different across the groups. Figure 2 demonstrates the first-order texture features comparison. All t-test results of this comparison are thoroughly reported in Table A1.
Second-order statistics based on GLCM are computed across three distances { 1 , 3 , 5 } pixels and four angles { 0 , 45 , 90 , 135 } . All GLCM features are significantly different between benign and malignant groups ( p < 0.001 ). See Table A2, Table A3, Table A4, Table A5, Table A6 and Table A7 for the complete t-test results. Interestingly, the difference between the two groups is consistent for various distances and angles. This further indicates that GLCM features are consistent for different orientations of the ROI. Figure 3 exemplifies the persistence of the difference in each GLCM feature across all angles for d = 3 pixels.

3.2. Machine Learning Diagnosis Pipeline Performance

In this section, we compare multiple variations of our framework. The comparison is made using common standard statistical metrics for evaluating classification performance on the test set. Specifically, we report: Accuracy, Precision, Recall, F1-score, and Area under the ROC Curve (AUC) (see [29] for mathematical descriptions of these metrics).
Using a LightGBM algorithm with 500 iterations of gradient boosting, our pipeline reaches its best performance with 0.91 accuracy, 0.94 precision, 0.93 recall, and 0.93 F1-score (Table 1). With a smaller number of boosting iterations, the accuracy slightly drops down. We compare our model with a simple DT and a random forest (RF) with 100 estimators and a maximum depth of two. Using LightGBM significantly improves the classification performance which further emphasizes the importance of gradient boosting approaches in our pipeline.
We also compare our model’s performance with the standard convolutional neural network architectures from Moon et al. [3], which is also using a similar dataset to ours [2]. Our pipeline, based on decision tree ensembles, achieves comparable results to the convolutional network models (Table 2) while being explainable. Our model can be probed to trace the logic behind its predictions while the convolutional models do not provide any insight into the process behind their predictions. In comparison, our model achieves higher precision, recall, and F1-score. Note that the F1-score is more suitable to assess models for their classification performance on imbalanced datasets.
To further identify the most important texture features for the model, we quantify the feature importance with SHAP values [30,31]. The SHAP values are Shapley values from coalitional game theory and correspond to the magnitude of each feature’s attribution on the output of the model. Intuitively, SHAP value of each feature singles out its importance on the model’s prediction. With our dataset, the most important features are: GLCM correlation within three pixels along the 90 direction, GLCM energy within a 5-pixel distance along the 90 direction, GLCM energy within a 3-pixel distance along the 90 direction, and GLCM correlation within a 5-pixel distance along the 90 direction (see Figure 4). Interestingly, all top four important features are statistics measured in the 90 direction.

3.3. Explainability of the Machine Learning Diagnosis Pipeline

In this section, we give an overview of how the explainability of our pipeline enables tracking down the logic behind its predictions. First, we evaluate a benign case from the test set (Figure 5). Given the ultrasound image and the cropped ROI, the learned model predicts the mass in ROI is benign with a likelihood of 0.97. We can further decompose the learned model to understand the decision path that leads to this prediction. The learned model is an ensemble of multiple decision trees ( T = 500 ).
In this example (Figure 5), two decision trees from the ensemble are depicted. In Tree Classifier 1, the first nodes split the data based on c o r r e l a t i o n _ d 3 _ a 90 . This texture feature is measured as 0.36 in the cropped ROI, which, in comparison with the learned split threshold 0.78, determines the outcome of the split at this node. The histogram of nodes also compares the split threshold with the distribution of the texture feature for both malignant and benign cases in the train set. At the next node e n e r g y _ d 5 _ a 0 = 0.23 < 0.26 , which leads to the final node d i s s i m i l a r i t y _ d 1 _ a 0 = 4.09 < 5.88 . Based on this decision path, the model predicts the ROI that belongs to the benign class. Other decision trees in the ensemble can also be interpreted in a similar way.
Next, we evaluate a malignant case from the test set (Figure 6). For this input, the model predicts that the mass in ROI is malignant with a likelihood of 0.93. In Tree Classifier 1, the model’s decision path begins with a split based on c o r r e l a t i o n _ d 3 _ a 90 = 0.86 > 0.78 . This is followed by the final node e n e r g y _ d 5 _ a 90 = 0.31 > 0.30 , which predicts that the ROI mass is malignant.

4. Conclusions

Ultrasound imaging is an accessible and cost-effective imaging modality to diagnose breast cancer. Most recent work on building machine learning models for breast cancer diagnosis depends on convolutional neural network architectures. Although accurate in performance, convolutional networks are black-box models and cannot be interpreted in terms of the logic behind their predictions.
In this paper, we proposed a novel explainable machine learning pipeline for breast cancer diagnosis based on ultrasound images. Our pipeline uses texture analysis of the ultrasound images as its basis to learn an ensemble of decision trees to predict the likelihood of malignancy of breast tumors. Importantly, our model can be decomposed into its underlying decision trees to fully interpret the decision path behind its outputs by following test conjunctions in each decision tree.
We believe our work is a step towards building more practical and comprehensible machine learning tools for cancer diagnosis by increasing the transparency of the prediction process. An important limitation of our work is relying on manually extracted ROIs. This could be a drawback for scaling to larger datasets without ground truth ROI masks. An interesting future work is to combine convolutional networks with decision trees. Finally, we hope our approach in this work to inspire future research on data-driven medical diagnosis to devote more attention into increasing the explainability of their solutions.

Author Contributions

Conceptualization, A.R., Y.J. and A.K.; methodology, A.R. and Y.J.; software, A.R.; validation, A.R. and Y.J.; formal analysis, A.R. and Y.J.; investigation, A.R., Y.J. and A.K.; resources, A.R. and Y.J.; data curation, A.R.; writing—original draft preparation, A.R.; writing—review and editing, A.R., Y.J. and A.K.; visualization, A.R.; supervision, A.R. and Y.J.; project administration, A.R., Y.J. and A.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The breast ultrasound image dataset used in the paper is publicly available at https://bit.ly/3I441wN (accessed on 10 February 2022). Ultrasound image texture analysis from DICOM images is performed by a costume-written MATLAB code available at https://github.com/arezaz/ultrasound-texture-analysis (accessed on 10 February 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

This section gives an overview of the first- and second-order GLCM texture feature statistics along with a detailed report of the t-test of each statistic between the two groups of benign and malignant masses.
Table A1. First-order Texture Features. The metrics are computed for a set of N pixels inside of the ROI denoted as X. Benign and malignant groups are compared using a t-test.
Table A1. First-order Texture Features. The metrics are computed for a set of N pixels inside of the ROI denoted as X. Benign and malignant groups are compared using a t-test.
FeatureDescription μ benign μ malignant p-Value
energy i = 1 N ( X ( i ) + c ) 2 6.66 × 10 7 1.17 × 10 8 <0.001
entropy i = 1 N p ( i ) log 2 p ( i ) + ϵ 1.982.23<0.001
kurtosis 1 N i = 1 N ( X ( i ) X ¯ ) 4 1 N i = 1 N ( X ( i ) X ¯ ) 2 2 6.504.27<0.001
mean 1 N i = 1 N X ( i ) 6.40 × 10 1 6.59 × 10 1 0.43
rms 1 N i = 1 N ( X ( i ) + c ) 2 7.26 × 10 1 7.51 × 10 1 0.28
skewness 1 N i = 1 N ( X ( i ) X ¯ ) 3 1 N i = 1 N ( X ( i ) X ¯ ) 2 3 1.450.99<0.001
uniformity i = 1 N p ( i ) 2 3.40 × 10 1 2.69 × 10 1 <0.001
variance 1 N i = 1 N ( X ( i ) X ¯ ) 2 1.04 × 10 3 1.22 × 10 3 <0.001
Table A2. Second-order Texture Features. The metrics are computed based on the gray-level co-occurrence matrix for a set of N gray level of pixels inside of the ROI.
Table A2. Second-order Texture Features. The metrics are computed based on the gray-level co-occurrence matrix for a set of N gray level of pixels inside of the ROI.
GLCM FeatureDescription
contrast i = 1 N j = 1 N ( i j ) 2 p ( i , j )
energy i = 1 N ( X ( i ) + c ) 2
correlation i = 1 N j = 1 N p ( i , j ) i j μ x μ y σ x ( i ) σ y ( j )
dissimilarity i = 1 N j = 1 N | i j | p ( i , j )
energy i = 1 N j = 1 N p ( i , j ) 2
energy i = 1 N j = 1 N p ( i , j ) 2
homogeneity i = 1 N j = 1 N p ( i , j ) 1 + | i j |
Table A3. The gray level co-occurrence matrix contrast feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
Table A3. The gray level co-occurrence matrix contrast feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
GLCM Contrast
DistanceAngle μ benign μ malignant p-Value
1184.86114.20<0.001
45°433.20238.06<0.001
90°304.20154.26<0.001
135°438.01238.07<0.001
3730.45493.19<0.001
45°1032.13582.21<0.001
90°1229.20660.83<0.001
135°1049.76582.42<0.001
51136.95830.26<0.001
45°1768.761132.49<0.001
90°1681.631040.57<0.001
135°1801.471130.44<0.001
Table A4. The gray level co-occurrence matrix correlation feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
Table A4. The gray level co-occurrence matrix correlation feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
GLCM Correlation
Distance Angle μ benign μ malignant p-Value
10.940.97<0.001
45°0.870.94<0.001
90°0.910.96<0.001
135°0.860.94<0.001
30.770.87<0.001
45°0.680.85<0.001
90°0.630.83<0.001
135°0.680.85<0.001
50.640.78<0.001
45°0.460.70<0.001
90°0.470.73<0.001
135°0.440.70<0.001
Table A5. The gray level co-occurrence matrix dissimilarity feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
Table A5. The gray level co-occurrence matrix dissimilarity feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
GLCM Dissimilarity
Distance Angle μ benign μ malignant p-Value
14.843.41<0.001
45°8.675.75<0.001
90°7.094.58<0.001
135°8.695.74<0.001
312.108.84<0.001
45°15.3310.07<0.001
90°17.0410.97<0.001
135°15.4010.06<0.001
516.7612.76<0.001
45°23.1015.95<0.001
90°22.0614.98<0.001
135°23.1815.90<0.001
Table A6. The gray level co-occurrence matrix energy feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
Table A6. The gray level co-occurrence matrix energy feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
GLCM Energy
Distance Angle μ benign μ malignant p-Value
10.300.38<0.001
45°0.280.37<0.001
90°0.290.38<0.001
135°0.280.37<0.001
30.260.36<0.001
45°0.240.35<0.001
90°0.250.35<0.001
135°0.240.35<0.001
50.240.34<0.001
45°0.190.32<0.001
90°0.220.33<0.001
135°0.190.32<0.001
Table A7. The gray level co-occurrence matrix homogeneity feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
Table A7. The gray level co-occurrence matrix homogeneity feature measured across three pixel distances { 1 , 3 , 5 } and four angles { 0 , 45 , 90 , 135 } .
GLCM Homogeneity
Distance Angle μ benign μ malignant p-Value
10.490.56<0.001
45°0.410.49<0.001
90°0.430.51<0.001
135°0.410.50<0.001
30.370.45<0.001
45°0.330.44<0.001
90°0.330.43<0.001
135°0.330.44<0.001
50.330.41<0.001
45°0.260.32<0.001
90°0.300.40<0.001
135°0.270.39<0.001

References

  1. Kuhl, C.K.; Schrading, S.; Leutner, C.C.; Morakkabati-Spitz, N.; Wardelmann, E.; Fimmers, R.; Kuhn, W.; Schild, H.H. Mammography, breast ultrasound, and magnetic resonance imaging for surveillance of women at high familial risk for breast cancer. J. Clin. Oncol. 2005, 23, 8469–8476. [Google Scholar] [CrossRef] [PubMed]
  2. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief 2020, 28, 104863. [Google Scholar] [CrossRef] [PubMed]
  3. Moon, W.K.; Lee, Y.W.; Ke, H.H.; Lee, S.H.; Huang, C.S.; Chang, R.F. Computer-aided diagnosis of breast ultrasound images using ensemble learning from convolutional neural networks. Comput. Methods Programs Biomed. 2020, 190, 105361. [Google Scholar] [CrossRef] [PubMed]
  4. Samulski, M.; Hupse, R.; Boetes, C.; Mus, R.D.; den Heeten, G.J.; Karssemeijer, N. Using computer-aided detection in mammography as a decision support. Eur. Radiol. 2010, 20, 2323–2330. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Sahiner, B.; Chan, H.P.; Roubidoux, M.A.; Hadjiiski, L.M.; Helvie, M.A.; Paramagul, C.; Bailey, J.; Nees, A.V.; Blane, C. Malignant and benign breast masses on 3D US volumetric images: Effect of computer-aided diagnosis on radiologist accuracy. Radiology 2007, 242, 716–724. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Jiménez-Gaona, Y.; Rodríguez-Álvarez, M.J.; Lakshminarayanan, V. Deep-Learning-Based Computer-Aided Systems for Breast Cancer Imaging: A Critical Review. Appl. Sci. 2020, 10, 8298. [Google Scholar] [CrossRef]
  7. Castelvecchi, D. Can we open the black box of AI? Nat. News 2016, 538, 20. [Google Scholar] [CrossRef] [Green Version]
  8. Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef] [Green Version]
  9. Zhu, J.; Liapis, A.; Risi, S.; Bidarra, R.; Youngblood, G.M. Explainable AI for designers: A human-centered perspective on mixed-initiative co-creation. In Proceedings of the 2018 IEEE Conference on Computational Intelligence and Games (CIG), Maastricht, The Netherlands, 14–17 August 2018; pp. 1–8. [Google Scholar]
  10. Preece, A.; Harborne, D.; Braines, D.; Tomsett, R.; Chakraborty, S. Stakeholders in explainable AI. arXiv 2018, arXiv:1810.00184. [Google Scholar]
  11. Masud, M.; Rashed, A.E.E.; Hossain, M.S. Convolutional neural network-based models for diagnosis of breast cancer. Neural Comput. Appl. 2020, 1–12. [Google Scholar] [CrossRef]
  12. Byra, M.; Jarosik, P.; Szubert, A.; Galperin, M.; Ojeda-Fournier, H.; Olson, L.; O’Boyle, M.; Comstock, C.; Andre, M. Breast mass segmentation in ultrasound with selective kernel U-Net convolutional neural network. Biomed. Signal Process. Control 2020, 61, 102027. [Google Scholar] [CrossRef] [PubMed]
  13. Irfan, R.; Almazroi, A.A.; Rauf, H.T.; Damaševičius, R.; Nasr, E.A.; Abdelgawad, A.E. Dilated semantic segmentation for breast ultrasonic lesion detection using parallel feature fusion. Diagnostics 2021, 11, 1212. [Google Scholar] [CrossRef] [PubMed]
  14. Tuceryan, M.; Jain, A.K. Texture analysis. Handb. Pattern Recognit. Comput. Vis. 1993, 235–276. [Google Scholar]
  15. Materka, A.; Strzelecki, M. Texture analysis methods—A review. Tech. Univ. Lodz Inst. Electron. COST B11 Rep. Bruss. 1998, 10, 4968. [Google Scholar]
  16. Varghese, B.A.; Cen, S.Y.; Hwang, D.H.; Duddalwar, V.A. Texture analysis of imaging: What radiologists need to know. Am. J. Roentgenol. 2019, 212, 520–528. [Google Scholar] [CrossRef]
  17. Srinivasan, G.; Shobha, G. Statistical texture analysis. World Acad. Sci. Eng. Technol. 2008, 36, 1264–1269. [Google Scholar]
  18. Kim, N.D.; Amin, V.; Wilson, D.; Rouse, G.; Udpa, S. Ultrasound image texture analysis for characterizing intramuscular fat content of live beef cattle. Ultrason. Imaging 1998, 20, 191–205. [Google Scholar] [CrossRef]
  19. Sebastian V, B.; Unnikrishnan, A.; Balakrishnan, K. Gray level co-occurrence matrices: Generalisation and some new features. arXiv 2012, arXiv:1205.4831. [Google Scholar]
  20. Iqbal, F.; Pallewatte, A.S.; Wansapura, J.P. Texture analysis of ultrasound images of chronic kidney disease. In Proceedings of the 2017 Seventeenth International Conference on Advances in ICT for Emerging Regions (ICTer), Colombo, Sri Lanka, 6–9 September 2017; pp. 1–5. [Google Scholar]
  21. Xu, S.S.D.; Chang, C.C.; Su, C.T.; Phu, P.Q. Classification of liver diseases based on ultrasound image texture features. Appl. Sci. 2019, 9, 342. [Google Scholar] [CrossRef] [Green Version]
  22. Sharma, H.; Kumar, S. A survey on decision tree algorithms of classification in data mining. Int. J. Sci. Res. (IJSR) 2016, 5, 2094–2097. [Google Scholar]
  23. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An introduction to decision tree modeling. J. Chemom. A J. Chemom. Soc. 2004, 18, 275–285. [Google Scholar] [CrossRef]
  24. Safavian, S.R.; Landgrebe, D. A survey of decision tree classifier methodology. IEEE Trans. Syst. Man Cybern. 1991, 21, 660–674. [Google Scholar] [CrossRef] [Green Version]
  25. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3146–3154. [Google Scholar]
  26. Rezazadeh, A. A Generalized Flow for B2B Sales Predictive Modeling: An Azure Machine-Learning Approach. Forecasting 2020, 2, 267–283. [Google Scholar] [CrossRef]
  27. Chen, C.; Zhang, Q.; Ma, Q.; Yu, B. LightGBM-PPI: Predicting protein-protein interactions through LightGBM with multi-information fusion. Chemom. Intell. Lab. Syst. 2019, 191, 54–64. [Google Scholar] [CrossRef]
  28. Sun, X.; Liu, M.; Sima, Z. A novel cryptocurrency price trend forecasting model based on LightGBM. Financ. Res. Lett. 2020, 32, 101084. [Google Scholar] [CrossRef]
  29. Hossin, M.; Sulaiman, M.N. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1. [Google Scholar]
  30. Lundberg, S.M.; Erion, G.G.; Lee, S.I. Consistent individualized feature attribution for tree ensembles. arXiv 2018, arXiv:1802.03888. [Google Scholar]
  31. Sundararajan, M.; Najmi, A. The many Shapley values for model explanation. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 9269–9278. [Google Scholar]
Figure 1. Our proposed pipeline uses the ultrasound image and ROI mask to extract GLCM texture features and learns a boosted decision tree model to make probabilistic diagnosis based on explainable decision trees.
Figure 1. Our proposed pipeline uses the ultrasound image and ROI mask to extract GLCM texture features and learns a boosted decision tree model to make probabilistic diagnosis based on explainable decision trees.
Forecasting 04 00015 g001
Figure 2. First-order texture features. All measured statistics, except Mean and RMS, are significantly different between the two groups of benign and malignant masses. Refer to Table A1 for a description of the metrics and t-test results.
Figure 2. First-order texture features. All measured statistics, except Mean and RMS, are significantly different between the two groups of benign and malignant masses. Refer to Table A1 for a description of the metrics and t-test results.
Forecasting 04 00015 g002
Figure 3. Second-order GLCM texture features. All measure statistics are significantly different between the benign and malignant groups. This significant difference is consistent across various pixel distances ( d = { 1 , 3 , 5 } , only d = 3 results are demonstrated here) and angles ( θ = { 0 , 45 , 90 , 135 } ). Refer to Table A2, Table A3, Table A4, Table A5, Table A6 and Table A7 for the deception of each metric and detailed t-test results.
Figure 3. Second-order GLCM texture features. All measure statistics are significantly different between the benign and malignant groups. This significant difference is consistent across various pixel distances ( d = { 1 , 3 , 5 } , only d = 3 results are demonstrated here) and angles ( θ = { 0 , 45 , 90 , 135 } ). Refer to Table A2, Table A3, Table A4, Table A5, Table A6 and Table A7 for the deception of each metric and detailed t-test results.
Forecasting 04 00015 g003
Figure 4. Feature importance. SHAP value quantifies the contribution of each feature to the prediction of each class. The features are denoted as {statistic}_d{distance}_a{angle}.
Figure 4. Feature importance. SHAP value quantifies the contribution of each feature to the prediction of each class. The features are denoted as {statistic}_d{distance}_a{angle}.
Forecasting 04 00015 g004
Figure 5. Qualitative results of a benign case. Our pipeline learns to infer a probabilistic diagnosis of the breast ultrasound images. The learned ensemble can be probed to obtain explainable decision paths in a set of learned decision trees. In each learned tree classifier, the orange arrows highlight the decision path. The model learns to compare the texture features obtained from the input image (orange numbers at the bottom of each dashed box) with the learned thresholds (black triangle on each histogram) at each node of the decision tree.
Figure 5. Qualitative results of a benign case. Our pipeline learns to infer a probabilistic diagnosis of the breast ultrasound images. The learned ensemble can be probed to obtain explainable decision paths in a set of learned decision trees. In each learned tree classifier, the orange arrows highlight the decision path. The model learns to compare the texture features obtained from the input image (orange numbers at the bottom of each dashed box) with the learned thresholds (black triangle on each histogram) at each node of the decision tree.
Forecasting 04 00015 g005
Figure 6. Qualitative results of a malignant case. Refer to the Figure 5 caption for more details.
Figure 6. Qualitative results of a malignant case. Refer to the Figure 5 caption for more details.
Forecasting 04 00015 g006
Table 1. Model evaluation. Standard classification performance metrics measured on the test set. The best performance is achieved with the LightGBM model with 500 gradient boosting iterations.
Table 1. Model evaluation. Standard classification performance metrics measured on the test set. The best performance is achieved with the LightGBM model with 500 gradient boosting iterations.
ModelPrecisionRecallF1-ScoreAUCAccuracy
DT0.850.820.830.820.86
RF0.870.840.860.850.88
LightGBM-100.900.800.830.790.87
LightGBM-500.880.870.880.900.87
LightGBM-1000.930.910.920.920.90
LightGBM-5000.940.930.930.930.91
Table 2. Model comparison with convolutional architectures. Standard classification performance metrics measured on the test set. Our explainable model based on decision trees achieves high predictive performance that is comparable to existing black box convolutional neural network architectures.
Table 2. Model comparison with convolutional architectures. Standard classification performance metrics measured on the test set. Our explainable model based on decision trees achieves high predictive performance that is comparable to existing black box convolutional neural network architectures.
ModelPrecisionRecallF1-ScoreAUCAccuracy
VGG0.750.760.760.870.85
ResNet0.890.890.890.960.91
DenseNet0.900.920.910.970.94
Decision Tree Ensemble (ours)0.940.930.930.930.91
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rezazadeh, A.; Jafarian, Y.; Kord, A. Explainable Ensemble Machine Learning for Breast Cancer Diagnosis Based on Ultrasound Image Texture Features. Forecasting 2022, 4, 262-274. https://0-doi-org.brum.beds.ac.uk/10.3390/forecast4010015

AMA Style

Rezazadeh A, Jafarian Y, Kord A. Explainable Ensemble Machine Learning for Breast Cancer Diagnosis Based on Ultrasound Image Texture Features. Forecasting. 2022; 4(1):262-274. https://0-doi-org.brum.beds.ac.uk/10.3390/forecast4010015

Chicago/Turabian Style

Rezazadeh, Alireza, Yasamin Jafarian, and Ali Kord. 2022. "Explainable Ensemble Machine Learning for Breast Cancer Diagnosis Based on Ultrasound Image Texture Features" Forecasting 4, no. 1: 262-274. https://0-doi-org.brum.beds.ac.uk/10.3390/forecast4010015

Article Metrics

Back to TopTop