Next Article in Journal
Estimating System State through Similarity Analysis of Signal Patterns
Previous Article in Journal
Low Power Wide Area Network, Cognitive Radio and the Internet of Things: Potentials for Integration
Previous Article in Special Issue
Evaluation of Hemodialysis Arteriovenous Bruit by Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Breast Tumor Classification in Ultrasound Images Using Combined Deep and Handcrafted Features

1
Department of Computer Engineering, German Jordanian University, Amman 11180, Jordan
2
Chair for Computer Aided Medical Procedure, Technical University of Munich, 85748 Munich, Germany
3
Department of Diagnostic Radiology, The University of Jordan Hospital, Queen Rania Street, Amman 11942, Jordan
*
Author to whom correspondence should be addressed.
Submission received: 23 October 2020 / Revised: 19 November 2020 / Accepted: 22 November 2020 / Published: 30 November 2020
(This article belongs to the Special Issue Artificial Intelligence in Medical Sensors)

Abstract

:
This study aims to enable effective breast ultrasound image classification by combining deep features with conventional handcrafted features to classify the tumors. In particular, the deep features are extracted from a pre-trained convolutional neural network model, namely the VGG19 model, at six different extraction levels. The deep features extracted at each level are analyzed using a features selection algorithm to identify the deep feature combination that achieves the highest classification performance. Furthermore, the extracted deep features are combined with handcrafted texture and morphological features and processed using features selection to investigate the possibility of improving the classification performance. The cross-validation analysis, which is performed using 380 breast ultrasound images, shows that the best combination of deep features is obtained using a feature set, denoted by C O N V features that include convolution features extracted from all convolution blocks of the VGG19 model. In particular, the C O N V features achieved mean accuracy, sensitivity, and specificity values of 94.2 % , 93.3 % , and 94.9 % , respectively. The analysis also shows that the performance of the C O N V features degrades substantially when the features selection algorithm is not applied. The classification performance of the C O N V features is improved by combining these features with handcrafted morphological features to achieve mean accuracy, sensitivity, and specificity values of 96.1 % , 95.7 % , and 96.3 % , respectively. Furthermore, the cross-validation analysis demonstrates that the C O N V features and the combined C O N V and morphological features outperform the handcrafted texture and morphological features as well as the fine-tuned VGG19 model. The generalization performance of the C O N V features and the combined C O N V and morphological features is demonstrated by performing the training using the 380 breast ultrasound images and the testing using another dataset that includes 163 images. The results suggest that the combined C O N V and morphological features can achieve effective breast ultrasound image classifications that increase the capability of detecting malignant tumors and reduce the potential of misclassifying benign tumors.

1. Introduction

Breast cancer is the most common cancer in females and a major cause of cancer-related deaths in women worldwide [1]. Ultrasound imaging is one of the widely used modalities for breast cancer diagnosis [2,3]. However, breast ultrasound (BUS) imaging is considered operator-dependent, and hence the reading of BUS images is a subjective task that requires well-trained and experienced radiologists [3,4]. Several studies, such as [5,6,7,8,9], proposed computer-aided diagnosis (CAD) systems to analyze BUS images with the goal of achieving objective computer-based classifications of breast tumors. In general, conventional CAD systems employ handcrafted features, such as the morphological features [10,11,12,13] and texture features [11,14,15], to characterize the tumor. Hence, the performance of conventional CAD systems depends mainly on the capability of the handcrafted features to capture the differences between benign and malignant tumors [16].
Several recent studies investigated the possibility of employing deep learning technology, and particularly convolutional neural networks (CNNs), to improve the performance of CAD systems [17,18]. However, a major challenge that restricts the full exploitation of CNNs for developing accurate CAD systems is the limited size of publicly available, well-annotated BUS image datasets that are crucial to achieve effective learning process [17,19]. Hence, the development of CNN models using limited BUS image datasets has become an important research challenge that was investigated in many previous studies [8,9,16,20,21,22,23,24,25]. A large group of these studies were essentially focused on the use of pre-trained CNNs to classify BUS images using two main strategies, namely the fine-tuning strategy and the deep features extraction strategy. In the fine-tuning strategy, the last fully connected layer of the pre-trained CNN model is modified to match the classes targeted by the BUS image classification problem and the parameters of the modified pre-trained CNN model are fine-tuned using the available BUS images. An example of this strategy is the study by Han et al. [9] who employed a modified GoogLeNet CNN model [26] that is pre-trained on the ImageNet dataset and fine-tuned using a BUS image dataset to classify the BUS images with a classification accuracy of 90 % . In the study by Xiao et al. [23], three ImageNet pre-trained CNN models, namely Xception [27], InceptionV3 [28], and ResNet50 [29], were fine-tuned using a BUS image dataset to enable the classification of BUS images. Moreover, the classification performance achieved using these three fine-tuned CNN models was compared with the classification performance obtained using a dedicated CNN model that was developed specifically for the BUS image classification problem and trained using the same BUS image dataset. The results reported in [23] showed that the fine-tuned InceptionV3 CNN model achieved the highest classification performance with accuracy value of 85 % . Tanaka et al. [24] compared the classification performance obtained using the ImageNet pre-trained VGG19 [30] and ResNet152 [29] CNN models, which were fine-tuned using a BUS image dataset, with the performance achieved by combining these two fine-tuned CNN models. The process of combining the two fine-tuned CNN models is achieved by averaging the class likelihoods achieved individually by each model. The results provided in [24] showed that the combined VGG19 and ResNet152 models achieved classification accuracy of 89 % , which is higher than the accuracy obtained using each model alone. Despite the effective classification performance obtained using the fine-tuning strategy, this strategy involves several challenges, such as the selection of the parameters of the pre-trained model that are included in the fine-tuning process, configuring the learning rates, and ensuring that the available BUS images are sufficient to perform the fine-tuning process.
The second strategy, which involves the use of pre-trained CNNs to extract deep features from the BUS images and classify the extracted features using a computer classifier, was studied by several research groups. For instance, Byra et al. [21] employed the individual convolution layers of the ImageNet pre-trained VGG19 CNN model to extract deep features from the BUS image. The deep features were analyzed using Fisher linear discriminant analysis [31] to classify the BUS images with accuracy values as high as 80 % . Antropova et al. [25] used the max-pooling layers and the first fully connected layer of the ImageNet pre-trained VGG19 CNN model to extract deep features for quantifying the BUS images. The extracted deep features were classified using a support vector machine (SVM) classifier [32]. The results reported in [25] showed that the deep features extracted from the max pooling layers were able to achieve the highest classification performance with area under the curve (AUC) of 0.87. In a recent study, Byra et al. [22] worked to improve the method proposed in [25] by adding a matching layer to interface the gray-scale BUS image with the three channels of the input layer of the ImageNet pre-trained VGG19 CNN model. Moreover, they compared the classification performance obtained by using the pre-trained VGG19 CNN model as a deep features extractor with the performance achieved by fine-tuning the pre-trained VGG19 model using a dataset of BUS images. The results reported in [22] showed that the matching layer was able to improve the BUS image classification performance. Furthermore, the results provided in [22] indicated that fine-tuning the pre-trained VGG19 CNN model using the BUS image dataset obtained the highest performance with classification accuracy of 88.7 % , which is higher than the accuracies achieved by using the pre-trained VGG19 CNN model as a deep features extractor. A major drawback in the previous studies that employed the deep features extraction strategy is that the extracted deep features were not processed using feature selection algorithms to select the deep feature combinations that optimize the classification performance. In fact, the use of features selection algorithms is crucial to identify the most effective features and eliminate irrelevant features that might degrade the classification performance.
The main goal of the current study is to improve the classification performance obtained using the deep features extraction strategy. In addition, the current study aims to investigate the possibility of combining the deep features with conventional handcrafted features to achieve higher classification performance. The pre-trained CNN model considered in the current study for extracting the deep features is the ImageNet pre-trained VGG19 model since this model was commonly used in several previous studies [21,22,25] that employed the deep features extraction strategy. In fact, the main contributions of our study can be summarized as follows:
  • Employ a two-phase optimization procedure to select the best deep feature combination that maximizes the BUS image classification performance as well as identify the components of the pre-trained VGG19 model that correspond to the optimized deep feature combination. In the first phase, we extracted deep features from the pre-trained VGG19 model at six different extraction levels. In the second phase, the deep features extracted at each extraction level are processed using a features selection algorithm and classified using a SVM classifier to identify the feature combinations that maximize the BUS image classification performance at that level. Furthermore, the deep feature combinations that are identified at all extraction levels are analyzed to find the best deep feature combination that enables the highest classification performance across all levels.
  • Investigate the possibility of improving the BUS image classification performance by combining the deep features extracted from the pre-trained VGG19 model with the handcrafted features that were introduced in previous studies. In fact, the handcrafted features considered in the current study are the texture and morphological features. The features selection algorithm is used to select the best combination of deep and handcrafted features that maximizes the classification performance.
  • Perform cross-validation analysis using a dataset that includes 380 BUS images to evaluate the classification performance obtained using the optimized combinations of deep features and combined deep and handcrafted features. Moreover, the classification performance obtained using the optimized combinations of deep features and combined deep and handcrafted features was compared with the results achieved using the optimized combinations of handcrafted features as well as the fine-tuned VGG19 model. Furthermore, the cross-validation analysis investigates the effect of classifying the deep features without applying the features selection algorithm.
  • Evaluate the generalization performance of the optimized combination of deep features and the optimized combination of combined deep and handcrafted features. In particular, the 380 breast ultrasound images were used to train two SVM classifiers that employ the optimized combination of deep features and the optimized combination of combined deep and handcrafted features. The performance of the trained classifiers were evaluated using another dataset that includes 163 BUS images.
The remainder of the paper is organized as follows. Section 2 presents the BUS image datasets that are employed in the current study, the two-phase optimization procedure that is used to selected the best combination of deep features, and the handcrafted texture and morphological features that were combined with the deep features. Moreover, Section 2 introduces the procedures employed to evaluate the performance of the deep features, the combined deep features and handcrafted texture and morphological features, the handcrafted texture and morphological features, and the fine-tuned VGG19 model. The results are provided and discussed in Section 3. Finally, the conclusions are presented in Section 4.

2. Materials and Methods

2.1. BUS Image Datasets

Two BUS image datasets were employed in the current study. The first dataset, denoted by Dataset 1, was collected during the period between 2015 and 2019 at the Jordan University Hospital, Amman, Jordan. The dataset is composed of 380 BUS images (217 benign tumors and 163 malignant tumors). BUS image acquisition was carried out during routine breast cancer screening procedures using an Acuson S2000 system (Siemens AG, Munich, Germany) equipped with a 14L5 linear ultrasound transducer that has a bandwidth of 5–14 MHz. During imaging, the radiologist was free to adjust the parameters of the ultrasound imaging system, including the gain, depth, and focal length, to achieve the best view of the tumor. All acquired BUS images were resampled to have a uniform resolution of 0.1 mm × 0.1 mm. The tumors were labeled as benign or malignant based on biopsy findings. The mean ± standard deviation diameter of the tumors is 14.5 ± 5.8 mm. All participating patients were females. The mean ± standard deviation age of the patients was 47.8 ± 11.9 years. For each patient, one BUS image was acquired. The tumor in each BUS image was outlined by asking a radiologist (fourth author) with more than 15 years of experience to segment the tumor for three times and the gold standard segmentation was taken as the mean of the three manual segmentations. The study protocol was approved by the Ethics Committee at the Jordan University Hospital. Furthermore, each participating patient was asked to sign informed consent to the study protocol.
The second BUS dataset, denoted by Dataset 2, was provided by the authors of [33]. The dataset was acquired in 2012 at the UDIAT Diagnostic Centre of the Parc Tauli Corporation, Sabadell, Spain. The dataset is composed of 163 BUS images (110 benign tumors and 53 malignant tumors). As described in [33], the tumors in the BUS images were outlined by experienced radiologists. Detailed description of the dataset is provided in [33].

2.2. BUS Image Classification Using the Deep Features

A two-phase procedure is used to identify the best combination of deep features, which are extracted using the ImageNet pre-trained VGG19 model that maximizes the BUS image classification performance as well as the components of the pre-trained VGG19 model that correspond to the best deep feature combination. This two-phase procedure is applied using the BUS images included in Dataset 1. The first phase of this procedure, which is described in Section 2.2.1, aims to extract deep features from the pre-trained VGG19 model at six different extraction levels. The second phase, which is described in Section 2.2.2, aims to process the deep features that are extracted at each level using features selection and classification to identify the feature combinations that maximize the classification performance at that extraction level. The deep feature combination that achieves the highest classification performance across all six deep features extraction levels is denoted as the best-performing combination of deep features.

2.2.1. Deep Features Extraction

Each BUS image in Dataset 1 is processed to define a region of interest (ROI) around the tumor. As suggested in [22], the ROI is obtained by computing the smallest box that contains the tumor and adding 30 pixels to each side of the box to include the surrounding tissue. For each image, the pre-trained VGG19 model [30] is used to extract deep features from the ROI that includes the tumor, as illustrated in Figure 1. As shown in the figure, the VGG19 model is composed of five convolution blocks, where each block includes a group of convolution layers followed by a max-pooling layer, and three fully connected layers. The first two convolution blocks comprise two convolution layers that use 64 kernels with a size of 3 × 3 . The third convolution block is composed of four convolution layers that employ 256 kernels with a size of 3 × 3 . The last two convolution blocks include four convolution layers that use 512 kernels with a size of 3 × 3 . The first two fully connected layers (FC6 and FC7) have 4096 units and the last fully connected layer (FC8) includes 1000 units that correspond to the 1000 classes of the ImageNet dataset [34].
The deep features extracted using the VGG19 model can be grouped into two main categories: convolution features and fully connected features. The convolution features are extracted from the convolution layers based on the method described in [35]. In particular, consider a convolution layer, denoted by C L , with a size of W × H × N , where W, H, and N are the width, height, and number of convolution kernels of the layer. The convolution layer, C L , can be processed to extract N convolution features, denoted by C L m a x , using the maximum pooling operation as follows [35]:
C L m a x = [ C L m a x ( 1 ) , C L m a x ( 2 ) , . . . , C L m a x ( N ) ] , where C L m a x ( k ) = m a x C L ( . , . , k ) , k = 1 , 2 , 3 , . . . , N
In addition, the C L can be processed to extract N convolution features, denoted by C L a v g , using the average pooling operation, as expressed below [35]:
C L a v g = [ C L a v g ( 1 ) , C L a v g ( 2 ) , . . . , C L a v g ( N ) ] , where C L a v g ( k ) = 1 W × H C L ( . , . , k ) , k = 1 , 2 , 3 , . . . , N
In the current study, a three-level convolution features extraction approach is employed. In the first level, denoted as convolution features extraction level 1 (CF Extraction Level 1), the convolution features, i.e., C L a v g and C L a v g , extracted from each convolution layer are concatenated to form a feature set with a size of 2 N and this set is normalized using l 2 -normalization [36]. As illustrated in Figure 1, eighteen feature sets were extracted at CF Extraction Level 1, where each set includes between 128 and 1024 convolution features. In the second level, denoted as convolution features extraction level 2 (CF Extraction Level 2), the feature sets extracted from all layers of each convolution block of the VGG19 model are concatenated to form a feature set and this set is normalized using l 2 -normalization. As shown in Figure 1, five feature sets were extracted at CF Extraction Level 2, where each set includes between 256 and 4096 convolution features. In the third level, called convolution features extraction level 3 (CF Extraction Level 3), the feature sets extracted from all convolution blocks of the VGG19 model are concatenated to form a feature set and this set is normalized using l 2 -normalization. As shown in Figure 1, the feature set extracted at CF Extraction Level 3 is composed of 11,008 convolution features.
The VGG19 model is also used to extract fully connected features from the ROI that includes the tumor. Particularity, a two-level fully connected features extraction approach is employed. In the first level, denoted as fully connected features extraction level 1 (FCF Extraction Level 1), the activations of each fully connected layer are extracted and normalized using l 2 -normalization to form a feature set. As shown in Figure 1, two fully connected feature sets were extracted at FCF Extraction Level 1, where each set includes 4096 fully connected features. In the second level, called fully connected features extraction level 2 (FCF Extraction Level 2), the two feature sets extracted at FCF Extraction Level 1 are concatenated and normalized using l 2 -normalization to form one feature set. As shown in Figure 1, the feature set extracted at FCF Extraction Level 2 is composed of 8192 fully connected features.
In addition to the five deep features extraction levels described above, an additional deep features extraction level, called combined convolution and fully connected features extraction (Combined CF FCF Extraction), is used to extract deep features from all convolution blocks and all fully connected layers of the VGG19 model. The features extracted at Combined CF FCF Extraction are concatenated and normalized using l 2 -normalization to form one feature set. As shown in Figure 1, the feature set computed at Combined CF FCF Extraction includes 19,200 convolution and fully connected features. The six deep features extraction levels that are employed in the current study are summarized in Table 1.

2.2.2. Deep Features Selection and Classification

The use of the extracted features to directly classify the BUS images in Dataset 1 can limit the classification performance as these features might include redundant and irrelevant information [37]. To overcome this limitation, the extracted features are processed using features selection to identify the relevant and effective features that can achieve high classification performance. Despite the fact that the exhaustive search of all possible feature combinations can identify the optimal feature combination, this search approach requires long processing times and extensive computational resources, particularly when a large number of features is considered [37]. In the current study, features selection is performed using a two-phase heuristic procedure [7] that is based on the features selection algorithm described in [14]. In the first phase, the extracted features are ranked using the minimum redundancy maximum relevance (mRMR) features selection algorithm [38]. After ranking the features, M feature groups are formed such that the m t h feature group is composed of the m top-ranked features, where M is the total number of features and m = { 1 , 2 , 3 , , M } . Moreover, the classification performance of the M feature groups is evaluated using the Matthews correlation coefficient (MCC) metric [39]. The MCC is used as it provides an effective classification performance metric for imbalanced data [40]. Then, the smallest feature group that achieves the highest classification performance is selected as the candidate feature group. In the second phase, the candidate feature group obtained by the first phase is refined using a backward elimination procedure to achieve a compact subset of features that maximizes the classification performance. Assume that the size of the candidate feature group obtained by the first phase is equal to n, then the aim of the first iteration of the backward elimination procedure is to identify a subset of n 1 features that can achieve the highest improvement in classification performance. In particular, each individual feature in the candidate feature group, which includes n features, is temporally eliminated and the classification performance obtained by the remaining n 1 features is evaluated. The feature that its elimination leads to the highest improvement in the classification performance is identified and permanently removed to obtain a subset of n 1 features with enhanced classification performance. This process is repetitively applied in the subsequent iterations of the backward elimination procedure to reduce the size of the selected features and, at the same time, improve the classification performance. This iterative process stops when the elimination of any feature leads to a reduction in the classification performance.
The process of classifying the features is performed using a binary SVM classifier [32] that is implemented using the LIBSVM library [41]. The SVM classifier is configured to use the radial basis function (RBF) kernel. Using this configuration, the SVM classifier has two control parameters that require tuning, namely the regularization parameter (C) and the RBF kernel parameter ( σ ). The training and testing of the SVM classifier is performed using a ten-fold cross-validation procedure combined with a grid search approach to tune the parameters C and σ . The ten-fold cross-validation procedure is used to reduce the possibility of overfitting the SVM classifier to the 380 BUS images included in Dataset 1 [42]. To carry out the ten-fold cross-validation procedure, the BUS images in Dataset 1 are divided into ten groups such that each group includes 38 BUS images. In each iteration of the ten-fold cross-validation procedure, nine groups of BUS images are used to train the SVM classifier and the remaining group is used for testing. This train-test process is repeated for ten iterations such that each group of BUS images is used exactly once as a testing group. Moreover, as suggested in [43,44], the tuning of C and σ is carried out using a grid search approach that examines the C and σ values of { 2 5 , 2 4 , 2 3 , , 2 15 } and { 2 15 , 2 14 , 2 13 , , 2 3 } , respectively. The grid search approach is configured to find the values of C and σ that maximize the classification performance that is evaluated using the MCC metric. In fact, the MCC is employed since it provides an effective classification metric for imbalanced data, where such data imbalance can occur due to the unequal numbers of benign and malignant BUS images [40].

2.3. BUS Image Classification by Combining the Deep Features with Handcrafted Features

We investigated the possibility of improving the classification of the BUS images included in Dataset 1 by combining the deep features with handcrafted features that were introduced in previous studies. In particular, the best-performing deep feature set extracted from the pre-trained VGG19 model is combined with handcrafted texture and morphological features that are commonly used for BUS image classification. To extract the handcrafted texture features, the ROI in the BUS image that includes the tumor is analyzed using the Gray-Level Co-occurrence Matrix (GLCM) [45] to quantify the statistics of the pixels’ intensities within the ROI. In the current study, 10 distances ( d = 1 , 2 , 3 , , 10 pixels ) and 4 orientations ( θ = 0 , 45 , 90 , 135 ) are employed to compute the GLCM, as suggested in [14]. This process enabled the creation of 40 GLCMs. Each GLCM is analyzed to extract 20 texture features that are widely employed for BUS image classification [7,14,46]. These texture features are summarized in Table 2. Hence, the total number of texture features that are computed for each BUS image is equal to 800.
In addition, the tumor outline in each BUS image is analyzed to extract 18 handcrafted morphological features, which are summarized in Table 2. In particular, 10 morphological features, namely the tumor area [12], tumor perimeter [12], tumor form factor [10,13], tumor roundness [10,13], tumor aspect ratio [10,13], tumor convexity [10,13], tumor solidity [10,13], tumor extent [10,13], tumor undulation characteristics [47], and tumor compactness [12,48], are extracted directly from the tumor outline. Two morphological features are extracted by computing the the normalized radial length (NRL) of the tumor and calculating its entropy and variance [12,49]. In fact, the NRL of the tumor is the Euclidean distance between the tumor center and the pixels located at the tumor boundary normalized by the maximum distance [49]. The last 6 morphological features are extracted from the best-fit ellipse that is obtained by fitting an ellipse to the tumor outline [12]. These 6 features are the length of the ellipse major axis [12], the length of the ellipse minor axis [12], the ratio between the ellipse major and minor axes [12], the ratio of the ellipse perimeter and the tumor perimeter [12], the angle of the ellipse major axis [12], and the overlap between the ellipse and the tumor [12].
The best-performing deep feature set, the handcrafted texture features, and the handcrafted morphological features are concatenated to form three feature groups that combine deep and handcrafted features. The first group comprises the best-performing deep feature set and the handcrafted texture features. The second group comprises the best-performing deep feature set and the handcrafted morphological features. Finally, the third group is composed of the best-performing deep feature set and the handcrafted texture and morphological features. Each one of these feature groups is normalized using l 2 -normalization and processed using features selection and classification to find the best combination of deep and handcrafted features that optimizes the classification of the BUS images included in Dataset 1.

2.4. Performance Comparison

For the BUS images included in Dataset 1, we compared the classifications obtained using the deep feature sets and the groups of combined deep and handcrafted features with the classifications achieved using three other classification approaches. The first approach aims to investigate the importance of the features selection algorithm, which eliminates the irrelevant and redundant features. In particular, the first approach considers the BUS image classifications that are obtained using the best-performing deep feature set but without applying the features selection algorithm. In other words, the first approach is focused on the BUS image classifications that are achieved by applying all features included in the best-performing deep feature set directly to the SVM classifier. The second approach aims to evaluate the classification performance that can be achieved using the handcrafted texture and morphological features. In particular, the second approach considers the BUS image classifications obtained using the 800 handcrafted texture features, the 18 handcrafted morphological features, and the combined 818 handcrafted texture and morphological features, which are described in Section 2.3, after applying features selection. The third approach considers the BUS image classification performance achieved by fine-tuning the pre-trained VGG19 model using the BUS images included in Dataset 1. The process of fine-tuning the pre-trained VGG19 model is performed using the fine-tuning procedure presented in [22].

2.5. Generalization Performance

The generalization performance of the deep feature sets and the groups of combined deep and handcrafted features was investigated in the current study. In particular, the BUS images in Dataset 1 are used to train two SVM classifiers, where the first classifier is based on the best-performing deep feature set and the second classifier is based on the best-performing group of combined deep and handcrafted features. In fact, the best-performing deep feature set and the best-performing group of combined deep and handcrafted features are identified based on the ten-fold cross-validation analyses performed for Database 1, as described in Section 2.2 and Section 2.3, respectively. The generalization performance of the two SVM classifiers was evaluated by employing each classifier to classify the 163 BUS images included in Dataset 2.

2.6. Performance Metrics

Six performance metrics were used to evaluate the cross-validation classifications performed for the BUS images in Dataset 1 based on the deep feature sets, the groups of combined deep and handcrafted features, and the three classification approaches used in the performance comparison. These metrics are the accuracy, sensitivity, specificity, Matthews correlation coefficient (MCC), positive predictive value (PPV), and negative predictive value (NPV). The mathematical formulas of these metrics are provided below [39,52]:
A c c u r a c y = T P + T N T P + T N + F P + F N , S e n s i t i v i t y = T P T P + F N , S p e c i f i c i t y = T N T N + F P , M C C = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N ) ) , P P V = T P T P + F P , N P V = T N T N + F N ,
where TP is the number of malignant tumors that are classified correctly, TN is the number of benign tumors that are classified correctly, FP is the number of benign tumors that are classified incorrectly, and FN is the number of malignant tumors that are classified incorrectly. In fact, the six performance metrics were computed for each one of the ten folds that carried out during the cross-validation analysis performed for Dataset 1. Moreover, the mean ± standard deviation values of the six metrics are calculated across the ten folds.
In addition to the six performance metrics, the receiver operator characteristic (ROC) curves were computed for the cross-validation classifications performed for Dataset 1 using the best-performing combination of deep features, the best-performing combination of deep and handcrafted features, the best-performing combination of handcrafted features, and the fine-tuned VGG19 model. In fact, the ROC curves aim to study the relation between the classification sensitivity and specificity obtained using each one of these classification approaches. The values of the AUC, which evaluates the overall performance of the BUS image classifier, were also calculated for these four classification approaches. In addition to the ROC curve analysis, these four classification approaches were studied using paired t-tests based on the BUS image classification accuracies. In fact, the aim of the paired t-tests is to investigate if the BUS image classification accuracy obtained using the most powerful approach, out of the four approaches described above, is significantly different than the other three approaches.
The generalization performance of the best-performing deep feature set and the best-performing group of combined deep and handcrafted features, which are achieved by performing the training using Dataset 1 and the testing using Dataset 2, was evaluated using the six metrics described above. In particular, the values these metrics were computed based on the classifications obtained for the 163 BUS images included in Dataset 2.

3. Results and Discussions

3.1. BUS Image Classification Results Obtained Using the Deep Features

The classification results obtained using the ten-fold cross-validation analyses that are performed for the BUS images in Dataset 1 based on the deep feature sets are provided in Table 3. The highest classification performance is achieved using the C O N V feature set that is extracted at CF Extraction Level 3. The C O N V feature set achieved BUS image classification with mean accuracy of 94.2 % , mean sensitivity of 93.3 % , mean specificity of 94.9 % , mean PPV of 93.3 % , mean NPV of 94.9 % , and mean MCC of 88.2 % . In fact, the features selection procedure was able to process the 11,008 features included in the C O N V feature set to select a combination of 25 features that achieved the highest classification performance. These 25 features are extracted from six convolution layers of the pre-trained VGG19 model, which are C O N V 3 _ 3 (1 feature), C O N V 4 _ 2 (3 features), C O N V 4 _ 3 (1 feature), C O N V 4 _ 4 (1 feature), C O N V 5 _ 1 (5 features), and C O N V 5 _ 4 (14 features). Furthermore, Table 3 indicates that the classification results achieved using the C O N V _ F C feature set matches the classification results obtained using the C O N V feature set. This is due to the fact that the feature combination selected for the C O N V _ F C feature set matches the feature combination selected for the C O N V feature set. Hence, the fully connected features included in the C O N V _ F C feature set were unable to improve the classification performance achieved by the convolution features included in both the C O N V _ F C feature set and the C O N V feature set. Hence, the best-performing deep feature set that is extracted from the pre-trained VGG19 model is considered to be the C O N V feature set.
In addition, the results provided in Table 3 show that the classification results achieved using the feature sets extracted at CF Extraction Level 2 are generally higher than the results obtained using the feature sets extracted at CF Extraction Level 1. This can be attributed to the fact that each feature set extracted at CF Extraction Level 2 is composed of the features obtained from a particular convolution block of the VGG19 model, while the feature sets extracted at CF Extraction Level 1 is composed of the features obtained from the individual convolution layers of the VGG19 model. Hence, for the feature sets extracted at CF Extraction Level 2, the features selection procedure can process the features extracted from all layers of each convolution block of the VGG19 model. However, in the case of the feature sets extracted at CF Extraction Level 1, the search space of the features section procedure is limited to the features obtained from the individual convolution layers of the VGG19 model.
For CF Extraction Level 1, the results presented in Table 3 indicate that the feature sets associated with high convolution layers, such as C O N V 5 _ 4 and C O N V 5 _ 3 , generally achieved better classification results compared to the feature sets associated with low convolution layers, such as C O N V 1 _ 1 and C O N V 1 _ 2 . Similarly, for CF Extraction Level 2, the feature sets associated with high convolution blocks, such as C O N V 5 , obtained better classification results compared to the feature sets associated with low convolution blocks, such as C O N V 1 . This can be attributed to the fact that high convolution layers tend to learn abstract features while low convolution layers extract low-level features, such as the edges [53].

3.2. BUS Image Classification Results Obtained by Combining the Deep Features with Handcrafted Features

Table 4 provides a comparison between the classification results achieved by the C O N V feature set and the results obtained by combining the C O N V feature set with the handcrafted texture, morphological, and combined texture and morphological features. In fact, this comparison is performed using the ten-fold cross-validation procedure that is applied on the BUS images in Dataset 1. As shown in Table 4, the highest classification performance is achieved by combining the C O N V feature set (11,008 features) with the handcrafted morphological features (18 features) and processing these combined features using features selection to select a combination that consists of 18 C O N V features and 3 morphological features. In fact, the selected combination of C O N V and morphological features enabled the classification of BUS images with mean accuracy of 96.1 % , mean sensitivity of 95.7 % , mean specificity of 96.3 % , mean PPV of 95.1 % , mean NPV of 96.8 % , and mean MCC of 92.0 % . These results indicate that the morphological features, which quantify the shape and structure of the tumor, include tumor quantification information that complement the tumor quantifications that are achieved by the C O N V feature set. Hence, the combined C O N V feature set and morphological features were able to achieve higher classification performance compared with the C O N V feature set alone. The high classification sensitivity obtained by combining the C O N V feature set with the morphological features indicates that these combined features enable high detection capability of malignant breast tumors. In addition, the high classification specificity achieved by the combined C O N V feature set and morphological features indicate that these combined features reduce the potential of misclassifying benign tumors, which in turn reduces the number of unnecessary biopsies performed for benign tumors.
Furthermore, the results provided in Table 4 indicate that the classification performance achieved by the C O N V feature set matches the classification performance obtained by combining the C O N V feature set with the handcrafted texture features. In addition, the classification performance achieved by combining the C O N V feature set with the handcrafted morphological features matches the classification performance obtained by combining the C O N V feature set with the handcrafted morphological and texture features. This is attributed to the fact that the features selected by the features selection algorithm for the C O N V feature set are the same features that are selected for the combined C O N V feature set and texture features. In addition, the features selected for the combined C O N V feature set and morphological features are the same features that are selected for the combined C O N V feature set, morphological features, and textures features. These findings suggest that the tumor quantifications obtained by the C O N V feature set include the tumor quantifications achieved by the handcrafted texture features.

3.3. Performance Comparison Results

Table 5 provides a comparison between the results obtained using the C O N V feature set with and without features selection, the combined C O N V feature set and handcrafted morphological features with features selection, the handcrafted texture, morphological, and combined texture and morphological features with features selection, and the fine-tuned VGG19 model. In fact, this comparison is performed using the ten-fold cross-validation procedure that is applied on the BUS images in Dataset 1. As shown in the table, the highest classification performance is achieved using the combined C O N V feature set and morphological features after applying the features selection algorithm. Moreover, the second highest classification performance is obtained using the C O N V feature set after applying the features selection algorithm.
Furthermore, Table 5 shows that the classification results obtained using the C O N V feature set degraded when the features selection algorithm is not applied. This is attributed to the fact that the C O N V feature set is composed of 11,008 features that include relevant features, which enable effective BUS image classification, as well as redundant and irrelevant features, which degrade the classification performance. Hence, the performance of the SVM classifier degraded significantly when all 11,800 features are classified directly by the classier. On the other hand, when the C O N V feature set is processed using the features selection algorithm, the most effective features are selected and used to classify the BUS images.
In addition, Table 5 shows that the handcrafted texture, morphological, and combined texture and morphological features, which are processed using the features selection algorithm, achieved low classification performance compared to the C O N V feature set and the combined C O N V feature set and morphological features when the features selection algorithm is applied. In addition, Table 5 indicates that the classification performance obtained using the morphological features is higher than the texture features. Furthermore, the classification performance is improved by combining the texture and morphological features. These findings agree with the BUS image classification results reported in previous studies, such as [7]. Furthermore, Table 5 indicates that the fine-tuned VGG19 model achieved classification results that are lower than the results obtained using the C O N V feature set when the features selection algorithm is applied. On the other hand, the fine-tuned VGG19 model outperformed the C O N V feature set when the features selection algorithm is not applied.
Figure 2 shows the ROC curves obtained using the C O N V feature set and the combined C O N V feature set and morphological features after applying the features selection algorithm. Furthermore, the figure shows the ROC curve achieved using the combined texture and morphological features after applying the features selection algorithm as well as the ROC curve obtained using the fine-tuned VGG19 model. The highest AUC value is obtained using the combined C O N V feature set and morphological features. Moreover, the second highest AUC value is achieved using the C O N V feature set. These results confirm the superior classification performance achieved by combining the C O N V feature set with the morphological features.
The p values obtained using the paired t-tests to compare the classification accuracies of the combined C O N V feature set and morphological features after applying features selection with the C O N V feature set after applying features selection, the combined texture and morphological features after applying features selection, and the fine-tuned VGG19 model are equal to 0.043, 0.003, and 0.001. These p values indicate that the classification accuracy achieved by the combined C O N V feature set and morphological features is significantly different than the three other classification approaches at confidence level of 0.05.

3.4. Generalization Performance Results

For the C O N V feature set, the values of the classification accuracy, sensitivity, specificity, PPV, NPV, and MCC obtained by performing the training using Dataset 1 and the testing using Dataset 2 are equal to 93.3 % , 90.7 % , 94.5 % , 89.1 % , 95.4 % , and 84.8 % , respectively. Furthermore, the values of the classification accuracy, sensitivity, specificity, PPV, NPV, and MCC achieved using the combined C O N V feature set and morphological features are equal to 95.1 % , 94.4 % , 95.4 % , 91.1 % , 97.2 % , and 89.1 % , respectively. These results are close to the corresponding results in Table 4, which are obtained using the ten-fold cross-validation analysis that is applied on Dataset 1. This finding suggests that the classification results reported in the current study can be generalized to other BUS image datasets.

4. Conclusions

The current study contributes to the ongoing efforts to improve BUS image classification by extracting deep features from the pre-trained VGG19 model at six different deep features extraction levels, combining the extracted deep features with handcrafted texture and morphological features, processing the features using a features selection algorithm, and classifying the selected features using a SVM classifier. The results reported in the current study indicate that the highest classification performance that can be achieved using the deep features is enabled using the C O N V feature set, which includes features extracted from the layers of all convolution blocks of the VGG19 model. In addition, the results show that the classification performance of the C O N V feature set can be improved by combining this feature set with handcrafted morphological features. In particular, the combined C O N V feature set and morphological features achieved mean accuracy of 96.1 % , mean sensitivity of 95.7 % , mean specificity of 96.3 % , mean PPV of 95.1 % , mean NPV of 96.8 % , and mean MCC of 92.0 % . On the other hand, combining the C O N V feature set with the handcrafted texture features did not improve the classification performance, which suggests that the tumor quantifications provided by the handcrafted texture features are included in the C O N V feature set. The high sensitivity and specificity values obtained by the combined C O N V feature set and morphological features can enable high detection capability of malignant tumors and reduce the potential of misclassifying benign tumors as malignant tumors. Furthermore, the performance comparison results provided in the current study show that both the C O N V feature set and the combined C O N V feature set and morphological features outperform the handcrafted texture and morphological features and the fine-tuned VGG19 model. The results also show that the performance of the C O N V feature set degrades when the features selection algorithm is not applied. This suggests the importance of processing the deep features using features selection algorithms to enable high classification performance. The generalization performance analysis conducted in the current study indicates that the C O N V feature set and the combined C O N V feature set and morphological features can be used to achieve high classification performance in other BUS image datasets. The future directions of the current study include combining the deep features that are extracted from the BUS image using different pre-trained CNN models with the goal of improving the classification performance.

Author Contributions

Data curation, M.I.D. and M.S.A.-N.; Funding acquisition, M.I.D.; Investigation, M.I.D. and S.A.-R.; Methodology, M.I.D. and S.A.-R.; Project administration, M.I.D. and S.A.-R.; Software, S.A.-R., M.I.D., and T.M.B.; Validation, S.A.-R. and M.I.D.; Visualization, S.A.-R. and M.I.D.; Writing, original draft preparation, M.I.D., S.A.-R., and R.A.; Writing, review and editing, M.I.D., R.A., and F.H.A.-H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Seed Grant Program, German Jordanian University, Amman, Jordan (grant number SEEIT 03/2020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Nothacker, M.; Duda, V.; Hahn, M.; Warm, M.; Degenhardt, F.; Madjar, H.; Weinbrenner, S.; Albert, U. Early detection of breast cancer: Benefits and risks of supplemental breast ultrasound in asymptomatic women with mammographically dense breast tissue. A systematic review. BMC Cancer 2009, 9, 335. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Chotai, N.; Kulkarni, S. Breast Ultrasound. In Breast Imaging Essentials; Springer: Singapore, 2020. [Google Scholar]
  4. Ding, J.; Cheng, H.D.; Huang, J.; Liu, J.; Zhang, Y. Breast ultrasound image classification based on multiple-instance learning. J. Digit. Imaging 2012, 25, 620–627. [Google Scholar] [CrossRef] [PubMed]
  5. Rouhi, R.; Jafari, M.; Kasaei, S.; Keshavarzian, P. Benign and malignant breast tumors classification based on region growing and CNN segmentation. Expert Syst. Appl. 2015, 42, 990–1002. [Google Scholar] [CrossRef]
  6. Prabusankarlal, K.M.; Thirumoorthy, P.; Manavalan, R. Assessment of combined textural and morphological features for diagnosis of breast masses in ultrasound. Human-Centric Comput. Inf. Sci. 2015, 5, 12. [Google Scholar] [CrossRef] [Green Version]
  7. Daoud, M.I.; Bdair, T.M.; Al-Najar, M.; Alazrai, R. A fusion-based approach for breast ultrasound image classification using multiple-ROI texture and morphological analyses. Comput. Math. Methods Med. 2016, 2016, 6740956. [Google Scholar] [CrossRef]
  8. Daoud, M.I.; Saleh, A.; Hababeh, I.; Alazrai, R. Content-based image retrieval for breast ultrasound images using convolutional autoencoders: A Feasibility study. In Proceedings of the 3rd International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris, France, 24–26 April 2019; pp. 1–4. [Google Scholar]
  9. Han, S.; Kang, H.; Jeong, J.; Park, M.; Kim, W.; Bang, W.; Seong, Y. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys. Med. Biol. 2017, 62, 7714–7728. [Google Scholar] [CrossRef] [PubMed]
  10. Wu, W.J.; Lin, S.W.; Moon, W.K. Combining support vector machine with genetic algorithm to classify ultrasound breast tumor images. Comput. Med. Imaging Graph. 2012, 36, 627–633. [Google Scholar] [CrossRef]
  11. Nemat, H.; Fehri, H.; Ahmadinejad, N.; Frangi, A.F.; Gooya, A. Classification of breast lesions in ultrasonography using sparse logistic regression and morphology-based texture features. Med. Phys. 2018, 45, 4112–4124. [Google Scholar] [CrossRef] [PubMed]
  12. Moon, W.K.; Lo, C.M.; Cho, N.; Chang, J.M.; Huang, C.S.; Chend, J.H.; Chang, R.F. Computer-aided diagnosis of breast masses using quantified BI-RADS findings. Comput. Methods Programs Biomed. 2013, 111, 84–92. [Google Scholar] [CrossRef] [PubMed]
  13. Chang, R.F.; Wu, W.J.; Moon, W.K.; Chen, D.R. Automatic ultrasound segmentation and morphology based diagnosis of solid breast tumors. Breast Cancer Res. Treat. 2005, 89, 179–185. [Google Scholar] [CrossRef] [PubMed]
  14. Gomez, W.; Pereira, W.C.A.; Infantosi, A.F.C. Analysis of co-occurrence texture statistics as a function of gray-level quantization for classifying breast ultrasound. IEEE Trans. Med. Imaging 2012, 31, 1889–1899. [Google Scholar] [CrossRef]
  15. Lin, C.; Hou, Y.; Chen, T.; Chen, K. Breast nodules computer-aided diagnostic system design using fuzzy cerebellar model neural networks. IEEE Trans. Fuzzy Syst. 2014, 22, 693–699. [Google Scholar] [CrossRef]
  16. Daoud, M.I.; Abdel-Rahman, S.; Alazrai, R. Breast ultrasound image classification using a pre-trained convolutional neural network. In Proceedings of the 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Sorrento, Italy, 26–29 November 2019; pp. 167–171. [Google Scholar]
  17. Liu, S.; Wang, Y.; Yang, X.; Lei, B.; Liu, L.; Li, S.X.; Ni, D.; Wang, T. Deep learning in medical ultrasound analysis: A review. Engineering 2019, 5, 261–275. [Google Scholar] [CrossRef]
  18. Huang, Q.; Zhang, F.; Li, X. Machine learning in ultrasound computer-aided diagnostic systems: A survey. BioMed Res. Int. 2018, 2018, 5137904. [Google Scholar] [CrossRef] [PubMed]
  19. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Fujioka, T.; Kubota, K.; Mori, M.; Kikuchi, Y.; Katsuta, L.; Kasahara, M.; Oda, G.; Ishiba, T.; Nakagawa, T.; Tateishi, U. Distinction between benign and malignant breast masses at breast ultrasound using deep learning method with convolutional neural network. Jpn. J. Radiol. 2019, 37, 466–472. [Google Scholar] [CrossRef]
  21. Byra, M. Discriminant analysis of neural style representations for breast lesion classification in ultrasound. Biocybern. Biomed. Eng. 2018, 38, 684–690. [Google Scholar] [CrossRef]
  22. Byra, M.; Galperin, M.; Ojeda-Fournier, H.; Olson, L.; O’Boyle, M.; Comstock, C.; Andre, M. Breast mass classification in sonography with transfer learning using a deep convolutional neural network and color conversion. Med. Phys. 2019, 46, 746–755. [Google Scholar] [CrossRef]
  23. Xiao, T.; Liu, L.; Li, K.; Qin, W.; Yu, S.; Li, Z. Comparison of transferred deep neural networks in ultrasonic breast masses discrimination. BioMed Res. Int. 2018, 2018, 4605191. [Google Scholar] [CrossRef]
  24. Tanaka, H.; Chiu, S.W.; Watanabe, T.; Kaoku, S.; Yamaguchi, T. Computer-aided diagnosis system for breast ultrasound images using deep learning. Phys. Med. Biol. 2019, 64, 235013. [Google Scholar] [CrossRef] [PubMed]
  25. Antropova, N.; Huynh, B.; Giger, M. A deep feature fusion methodology for breast cancer diagnosis demonstrated on three imaging modality datasets. Phys. Med. Biol. 2017, 44, 5162–5171. [Google Scholar] [CrossRef] [PubMed]
  26. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2015), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  27. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
  28. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  31. Duda, R.; Hart, P. Pattern Classification and Scene Analysis, 1st ed.; Wiley: New York, NY, USA, 1973. [Google Scholar]
  32. Vapnik, V.N. The Nature of Statistical Learning Theory, 2nd ed.; Springer: New York, NY, USA, 2000. [Google Scholar]
  33. Yap, M.H.; Pons, G.; Marti, J.; Ganau, S.; Sentis, M.; Zwiggelaar, R.; Davison, A.K.; Marti, R. Automated breast ultrasound lesions detection using convolutional neural networks. IEEE J. Biomed. Health Inform. 2018, 22, 1218–1226. [Google Scholar] [CrossRef] [Green Version]
  34. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. Vol. 2015, 115, 159–252. [Google Scholar] [CrossRef] [Green Version]
  35. Zheng, L.; Zhao, Y.; Wang, S.; Wang, J.; Tian, Q. Good practice in CNN feature transfer. arXiv 2016, arXiv:1604.00133. [Google Scholar]
  36. Arandjelović, R.; Zisserman, A. Three things everyone should know to improve object retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2012), Providence, RI, USA, 16–21 June 2012; pp. 2911–2918. [Google Scholar]
  37. Urbanowicz, R.J.; Meeker, M.; La Cava, W.; Olson, R.S.; Moore, J.H. Relief-based feature selection: Introduction and review. J. Biomed. Inform. 2018, 85, 189–203. [Google Scholar] [CrossRef]
  38. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  39. Matthews, B.W. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Biophys. Acta (BBA) Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  40. Chicco, D. Ten quick tips for machine learning in computational biology. BioData Min. 2017, 10, 1–17. [Google Scholar] [CrossRef]
  41. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  42. Hawkins, D.M. The problem of overfitting. J. Chem. Inf. Comput. Sci. 2004, 44, 1–12. [Google Scholar] [CrossRef] [PubMed]
  43. Hsu, C.W.; Chang, C.C.; Lin, C.J. A Practical Guide to Support Vector Classification; Technical Report; Department of Computer Science and Information Engineering, National Taiwan University: Taipei, Taiwan, 2008. [Google Scholar]
  44. Daoud, M.I.; Atallah, A.A.; Awwad, F.; Al-Najjar, M.; Alazrai, R. Automatic superpixel-based segmentation method for breast ultrasound images. Expert Syst. Appl. 2019, 121, 78–96. [Google Scholar] [CrossRef]
  45. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  46. Moon, W.K.; Huang, Y.; Lo, C.; Huang, C.; Bae, M.S.; Kim, W.H.; Chen, J.; Chang, R. Computer-aided diagnosis for distinguishing between triple-negative breast cancer and fibroadenomas based on ultrasound texture features. Med. Phys. 2015, 42, 3024–3035. [Google Scholar] [CrossRef]
  47. Shen, W.C.; Chang, R.F.; Moon, W.K.; Chou, Y.H.; Huang, C.S. Breast ultrasound computer-aided diagnosis using BI-RADS features. Acad. Radiol. 2007, 14, 928–939. [Google Scholar] [CrossRef]
  48. Rangayyan, R.M.; Mudigonda, N.R.; Desautels, J.E.L. Boundary modelling and shape analysis methods for classification of mammographic masses. Med. Biol. Eng. Comput. 2000, 38, 487–496. [Google Scholar] [CrossRef]
  49. Nie, K.; Chen, J.H.; Yu, H.J.; Chu, Y.; Nalcioglu, O.; Su, M.Y. Quantitative analysis of lesion morphology and texture features for diagnostic prediction in breast MRI. Acad. Radiol. 2008, 15, 1513–1525. [Google Scholar] [CrossRef] [Green Version]
  50. Soh, L.K.; Tsatsoulis, C. Texture analysis of SAR sea ice imagery using gray level co-occurrence matrices. IEEE Trans. Geosci. Remote Sens. 1999, 37, 780–795. [Google Scholar] [CrossRef] [Green Version]
  51. Clausi, D.A. An analysis of co-occurrence texture statistics as a function of grey level quantization. Can. J. Remote Sens. 2002, 28, 45–62. [Google Scholar] [CrossRef]
  52. Cheng, H.D.; Shan, J.; Ju, W.; Guo, Y.; Zhang, L. Automated breast cancer detection and classification using ultrasound images: A survey. Pattern Recognit. 2010, 43, 299–317. [Google Scholar] [CrossRef] [Green Version]
  53. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Graphical illustration of the process of extracting deep features from the ROI in the BUS image using the pre-trained VGG19 model.
Figure 1. Graphical illustration of the process of extracting deep features from the ROI in the BUS image using the pre-trained VGG19 model.
Sensors 20 06838 g001
Figure 2. The ROC curves obtained using the C O N V feature set (with features selection), the combined C O N V feature set and morphological features (with features selection), the combined handcrafted texture and morphological features (with features selection), and the fine-tuned VGG19 CNN model.
Figure 2. The ROC curves obtained using the C O N V feature set (with features selection), the combined C O N V feature set and morphological features (with features selection), the combined handcrafted texture and morphological features (with features selection), and the fine-tuned VGG19 CNN model.
Sensors 20 06838 g002
Table 1. The six levels that are employed to extract the deep feature sets from the BUS image.
Table 1. The six levels that are employed to extract the deep feature sets from the BUS image.
Deep Features Extraction LevelFeature SetsDescription
CF Extraction Level 1 C O N V 1 _ 1 , C O N V 1 _ 2 , C O N V 2 _ 1 , C O N V 2 _ 2 , C O N V 3 _ 1 , C O N V 3 _ 2 , C O N V 3 _ 3 , C O N V 3 _ 4 , C O N V 4 _ 1 , C O N V 4 _ 2 , C O N V 4 _ 3 , C O N V 4 _ 4 , C O N V 5 _ 1 , C O N V 5 _ 2 , C O N V 5 _ 3 , C O N V 5 _ 4 A total of 11,008 convolution features organized into 16 feature sets are extracted from the ROI that includes the tumor, where each feature set corresponds to one of the convolution layers of the VGG19 model. To generate a given feature set, the convolution features C L a v g and C L a v g extracted from the convolution layer that corresponds to the feature set are concatenated and normalized.
CF Extraction Level 2 C O N V 1 , C O N V 2 , C O N V 3 , C O N V 4 , C O N V 5 A total of 11,008 convolution features organized into 5 feature sets are extracted from the ROI that includes the tumor, where each feature set corresponds to one of the convolution blocks of the VGG19 model. To generate a given feature set, the feature sets extracted from the layers of the convolution block that corresponds to the feature set are concatenated and normalized.
CF Extraction Level 3 C O N V A total of 11,008 convolution features organized into 1 feature set are extracted from the ROI that includes the tumor. To generate the feature set, the feature sets extracted from all convolution blocks of the VGG19 model are concatenated and normalized.
FCF Extraction Level 1 F C 6 and F C 7 Two feature sets, where each set includes 4096 fully connected features, are extracted from the ROI that includes the tumor. The computation of the two feature sets is achieved by extracting and normalizing the activations of first and second fully connected layers of the VGG19 model.
FCF Extraction Level 2 F C A feature set that includes 8192 fully connected features is extracted from the ROI that includes the tumor. The computation of the feature set is achieved by concatenating and normalizing the two feature sets extracted from the first and second fully connected layers of the VGG19 model.
Combined CF FCF Extraction C O N V _ F C A feature set that includes 19,200 convolution and fully connected features is extracted from the ROI that includes the tumor. The computation of the feature set is performed by extracting deep features from all convolution blocks and all fully connected layers of the VGG19 model and then concatenating and normalizing the extracted features.
Table 2. The handcrafted texture and morphological features that are extracted from the BUS image.
Table 2. The handcrafted texture and morphological features that are extracted from the BUS image.
TypeFeaturesDescription
Texture featuresAutocorrelation [50], contrast [14], correlation [50], cluster prominence [50], cluster shade [50], dissimilarity [50], energy [50], entropy [50], homogeneity [50], maximum probability [50], sum of squares [45], sum average [45], sum entropy [45], sum variance [45], difference variance [45], difference entropy [45], information measure of correlation I [45], information measure of correlation II [45], inverse difference normalized [51], inverse difference moment normalized [51]A total of 800 texture features are extracted from the ROI that includes the tumor. In particular, 40 GLCMs are generated using 10 distances ( d = 1 , 2 , 3 , , 10 pixels ) and 4 orientations ( θ = 0 , 45 , 90 , 135 ) . Moreover, each GLCM is analyzed to extract 20 texture features.
Morphological featuresTumor area [12], tumor perimeter [12], tumor form factor [10,13], tumor roundness [10,13], tumor aspect ratio [10,13], tumor convexity [10,13], tumor solidity [10,13], tumor extent [10,13], tumor undulation characteristics [47], tumor compactness [12,48], NRL entropy [12,49], NRL variance [12,49], length of the ellipse major axis [12], length of the ellipse minor axis [12], ratio between the ellipse major and minor axes [12], ratio of the ellipse perimeter and the tumor perimeter [12], angle of the ellipse major axis [12], overlap between the ellipse and the tumor [12]A total of 18 morphological features are extracted from the tumor outline. In particular, 10 morphological features are computed directly based on the tumor outline. Moreover, 2 morphological features are computed based on the NRL of the tumor. In addition, 6 morphological features are computed by fitting an ellipse to the tumor outline.
Table 3. The BUS image classification results obtained using the feature sets extracted from the pre-trained VGG19 model at six deep features extraction levels. For the six performance metrics, the mean ± standard deviation values are computed across the ten folds of the cross-validation procedure performed using Dataset 1.
Table 3. The BUS image classification results obtained using the feature sets extracted from the pre-trained VGG19 model at six deep features extraction levels. For the six performance metrics, the mean ± standard deviation values are computed across the ten folds of the cross-validation procedure performed using Dataset 1.
Deep Features Extraction LevelFeature SetTotal No. of FeaturesNo. of Selected FeaturesSelected FeaturesAccuracy ( % ) Sensitivity ( % ) Specificity ( % ) PPV ( % ) NPV ( % ) MCC ( % )
CF Extraction Level 1 C O N V 1 _ 1 12834 C O N V 1 _ 1 (34) 78.2 ± 6.9 76.7 ± 7.4 79.3 ± 11.8 73.5 ± 12.0 81.9 ± 6.8 55.7 ± 12.7
C O N V 1 _ 2 12821 C O N V 1 _ 2 (21) 81.8 ± 5.2 81.6 ± 8.9 82.0 ± 9.7 77.3 ± 11.0 85.6 ± 5.5 63.3 ± 10.1
C O N V 2 _ 1 25625 C O N V 2 _ 1 (25) 82.1 ± 5.2 77.3 ± 9.3 85.7 ± 8.9 80.3 ± 8.9 83.4 ± 6.2 63.3 ± 9.8
C O N V 2 _ 2 25638 C O N V 2 _ 2 (38) 84.2 ± 3.7 80.4 ± 9.4 87.1 ± 7.7 82.4 ± 7.4 85.5 ± 7.6 67.7 ± 6.2
C O N V 3 _ 1 51233 C O N V 3 _ 1 (33) 84.7 ± 4.8 80.4 ± 11.9 88.0 ± 4.4 83.4 ± 2.1 85.7 ± 8.1 68.7 ± 9.2
C O N V 3 _ 2 51247 C O N V 3 _ 2 (47) 86.1 ± 4.5 82.2 ± 8.1 88.9 ± 5.5 84.8 ± 7.4 86.9 ± 5.7 71.4 ± 9.0
C O N V 3 _ 3 51229 C O N V 3 _ 3 (29) 85.3 ± 4.5 79.8 ± 10.5 89.4 ± 5.7 85.0 ± 6.2 85.5 ± 7.0 69.8 ± 8.1
C O N V 3 _ 4 51286 C O N V 3 _ 4 (86) 88.9 ± 4.4 85.3 ± 10.0 91.7 ± 3.1 88.5 ± 4.1 89.2 ± 7.1 77.4 ± 8.6
C O N V 4 _ 1 102420 C O N V 4 _ 1 (20) 88.4 ± 5.0 84.0 ± 7.6 91.7 ± 5.6 88.4 ± 7.0 88.4 ± 5.8 76.3 ± 10.2
C O N V 4 _ 2 102461 C O N V 4 _ 2 (61) 87.1 ± 4.0 82.8 ± 8.3 90.3 ± 2.6 86.5 ± 4.4 87.5 ± 6.0 73.6 ± 7.8
C O N V 4 _ 3 102431 C O N V 4 _ 3 (31) 86.8 ± 5.5 82.8 ± 9.3 89.9 ± 5.4 86.0 ± 5.7 87.4 ± 7.0 73.1 ± 10.8
C O N V 4 _ 4 102432 C O N V 4 _ 4 (32) 90.0 ± 4.8 89.0 ± 10.3 90.8 ± 4.9 87.9 ± 9.3 91.6 ± 6.0 79.6 ± 10.5
C O N V 5 _ 1 102458 C O N V 5 _ 1 (58) 89.5 ± 5.0 87.7 ± 11.2 90.8 ± 6.0 87.7 ± 8.1 90.8 ± 6.1 78.5 ± 11.0
C O N V 5 _ 2 102441 C O N V 5 _ 2 (41) 89.7 ± 3.2 88.3 ± 9.1 90.8 ± 5.2 87.8 ± 8.5 91.2 ± 5.5 79.1 ± 6.8
C O N V 5 _ 3 102423 C O N V 5 _ 3 (23) 89.7 ± 4.4 87.1 ± 8.9 91.7 ± 6.5 88.8 ± 10.7 90.5 ± 6.5 79.0 ± 9.1
C O N V 5 _ 4 102431 C O N V 5 _ 4 (31) 91.3 ± 4.1 91.4 ± 5.4 91.2 ± 6.7 88.7 ± 10.1 93.4 ± 3.2 82.4 ± 9.2
CF Extraction Level 2 C O N V 1 25634 C O N V 1 _ 1 (17), C O N V 1 _ 2 (17) 82.6 ± 5.1 82.8 ± 7.3 82.5 ± 7.5 78.0 ± 9.2 86.5 ± 5.5 64.9 ± 10.0
C O N V 2 51223 C O N V 2 _ 1 (5), C O N V 2 _ 2 (18) 84.7 ± 3.7 80.4 ± 9.4 88.0 ± 7.0 83.4 ± 6.9 85.7 ± 7.6 68.7 ± 6.1
C O N V 3 204815 C O N V 3 _ 1 (7), C O N V 3 _ 4 (8) 88.7 ± 4.3 84.0 ± 9.0 92.2 ± 5.8 89.0 ± 7.7 88.5 ± 6.3 76.8 ± 8.3
C O N V 4 409634 C O N V 4 _ 1 (8), C O N V 4 _ 2 (4), CONV4_3 (8), CONV4_4 (14) 90.5 ± 4.8 88.3 ± 10.1 92.2 ± 4.4 89.4 ± 9.1 91.3 ± 5.9 80.6 ± 10.6
C O N V 5 409627 C O N V 5 _ 2 (7), C O N V 5 _ 3 (8), C O N V 5 _ 4 (12) 91.8 ± 4.5 91.4 ± 5.4 92.2 ± 6.5 89.8 ± 10.5 93.5 ± 3.2 83.4 ± 10.0
CF Extraction Level 3 C O N V 11,00825 C O N V 3 _ 3 (1), C O N V 4 _ 2 (3), C O N V 4 _ 3 (1), C O N V 4 _ 4 (1), C O N V 5 _ 1 (5), C O N V 5 _ 4 (14)94.2 ± 2.793.3 ± 5.194.9 ± 4.193.3 ± 5.694.9 ± 4.488.2 ± 5.5
FCF Extraction Level 1 F C 6 409636 F C 6 (36) 90.5 ± 2.8 88.3 ± 7.0 92.2 ± 4.3 89.4 ± 6.2 91.3 ± 5.2 80.6 ± 5.8
F C 7 409698 F C 7 (98) 90.0 ± 4.4 87.7 ± 10.2 91.7 ± 5.1 88.8 ± 7.4 90.9 ± 6.2 79.6 ± 9.4
FCF Extraction Level 2 F C 819236 F C 6 (36) 90.5 ± 2.8 88.3 ± 7.0 92.2 ± 4.3 89.4 ± 6.2 91.3 ± 5.2 80.6 ± 5.8
Combined CF FCF Extraction C O N V _ F C 19,20025 C O N V 3 _ 3 (1), C O N V 4 _ 2 (3), C O N V 4 _ 3 (1), C O N V 4 _ 4 (1), C O N V 5 _ 1 (5), C O N V 5 _ 4 (14)94.2 ± 2.793.3 ± 5.194.9 ± 4.193.3 ± 5.694.9 ± 4.488.2 ± 5.5
Table 4. The BUS image classification results obtained using the C O N V feature set compared with the classification results achieved by combining the C O N V feature set with handcrafted texture and morphological features. For the six performance metrics, the mean ± standard deviation values are computed across the ten folds of the cross-validation procedure performed using Dataset 1.
Table 4. The BUS image classification results obtained using the C O N V feature set compared with the classification results achieved by combining the C O N V feature set with handcrafted texture and morphological features. For the six performance metrics, the mean ± standard deviation values are computed across the ten folds of the cross-validation procedure performed using Dataset 1.
FeaturesTotal No. of FeaturesNo. of Selected FeaturesSelected FeaturesAccuracy ( % ) Sensitivity ( % ) Specificity ( % ) PPV ( % ) NPV ( % ) MCC ( % )
C O N V feature set11,00825 C O N V 3 _ 3 (1), C O N V 4 _ 2 (3), C O N V 4 _ 3 (1), C O N V 4 _ 4 (1), C O N V 5 _ 1 (5), C O N V 5 _ 4 (14)94.2 ± 2.793.3 ± 5.194.9 ± 4.193.3 ± 5.694.9 ± 4.488.2 ± 5.5
C O N V feature set and texture features11,80825 C O N V 3 _ 3 (1), C O N V 4 _ 2 (3), C O N V 4 _ 3 (1), C O N V 4 _ 4 (1), C O N V 5 _ 1 (5), C O N V 5 _ 4 (14)94.2 ± 2.793.3 ± 5.194.9 ± 4.193.3 ± 5.694.9 ± 4.488.2 ± 5.5
C O N V feature set and morphological features11,02621 C O N V 4 _ 4 (4), C O N V 5 _ 4 (14), morphological (3)96.1 ± 2.295.7 ± 4.296.3 ± 3.695.1 ± 5.496.8 ± 3.192.0 ± 5.0
C O N V feature set, texture features, and morphological features11,82621 C O N V 4 _ 4 (4), C O N V 5 _ 4 (14), morphological (3)96.1 ± 2.295.7 ± 4.296.3 ± 3.695.1 ± 5.496.8 ± 3.192.0 ± 5.0
Table 5. Performance comparison between the classification results obtained using the C O N V feature set with and without features selection, the combined C O N V feature set and morphological features with features selection, the handcrafted texture, morphological, and combined texture and morphological features with features selection, and the fine-tuned VGG19 model. For the six performance metrics, the mean ± standard deviation values are computed across the ten folds of the cross-validation procedure performed using Dataset 1.
Table 5. Performance comparison between the classification results obtained using the C O N V feature set with and without features selection, the combined C O N V feature set and morphological features with features selection, the handcrafted texture, morphological, and combined texture and morphological features with features selection, and the fine-tuned VGG19 model. For the six performance metrics, the mean ± standard deviation values are computed across the ten folds of the cross-validation procedure performed using Dataset 1.
FeaturesTotal No. of FeaturesNo. of Selected FeaturesAccuracy ( % ) Sensitivity ( % ) Specificity ( % ) PPV ( % ) NPV ( % ) MCC ( % )
C O N V feature set (with features selection)11,0082594.2± 2.793.3 ± 5.194.9± 4.193.3 ± 5.694.9 ± 4.488.2 ± 5.5
C O N V feature set and morphological features (with features selection)11,0262196.1 ± 2.295.7 ± 4.296.3 ± 3.695.1 ± 5.496.8 ± 3.192.0 ± 5.0
C O N V feature set (without features selection)11,008-80.5 ± 4.582.2 ± 9.079.3 ± 8.074.9 ± 7.185.6 ± 7.460.9 ± 8.6
Texture features (with features selection)8003884.2 ± 4.381.0 ± 9.686.6 ± 6.682.0 ± 10.685.8 ± 5.567.7 ± 9.2
Morphological features (with features selection)18887.1 ± 4.782.2 ± 8.990.8 ± 7.987.0 ± 11.987.2 ± 6.273.6 ± 9.7
Texture and morphological features (with features selection)8182987.9 ± 6.182.8 ± 12.391.7 ± 3.988.2 ± 6.187.7 ± 7.775.2 ± 12.7
Fine-tuning--88.2 ± 4.583.4 ± 8.091.7 ± 5.588.3 ± 8.188.1 ± 6.975.8 ± 9.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Daoud, M.I.; Abdel-Rahman, S.; Bdair, T.M.; Al-Najar, M.S.; Al-Hawari, F.H.; Alazrai, R. Breast Tumor Classification in Ultrasound Images Using Combined Deep and Handcrafted Features. Sensors 2020, 20, 6838. https://0-doi-org.brum.beds.ac.uk/10.3390/s20236838

AMA Style

Daoud MI, Abdel-Rahman S, Bdair TM, Al-Najar MS, Al-Hawari FH, Alazrai R. Breast Tumor Classification in Ultrasound Images Using Combined Deep and Handcrafted Features. Sensors. 2020; 20(23):6838. https://0-doi-org.brum.beds.ac.uk/10.3390/s20236838

Chicago/Turabian Style

Daoud, Mohammad I., Samir Abdel-Rahman, Tariq M. Bdair, Mahasen S. Al-Najar, Feras H. Al-Hawari, and Rami Alazrai. 2020. "Breast Tumor Classification in Ultrasound Images Using Combined Deep and Handcrafted Features" Sensors 20, no. 23: 6838. https://0-doi-org.brum.beds.ac.uk/10.3390/s20236838

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop