Next Article in Journal
Association of In-Hospital Mortality and Trauma Team Activation: A 10-Year Study
Next Article in Special Issue
Proposal to Improve the Image Quality of Short-Acquisition Time-Dedicated Breast Positron Emission Tomography Using the Pix2pix Generative Adversarial Network
Previous Article in Journal
Cardiac Imaging and Management of Cardiac Disease in Asymptomatic Renal Transplant Candidates: A Current Update
Previous Article in Special Issue
Medical Radiation Exposure Reduction in PET via Super-Resolution Deep Learning Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Assessment for Mining Important Medical Image Features of Various Modalities

by
Ioannis D. Apostolopoulos
1,*,
Nikolaos D. Papathanasiou
2,
Nikolaos I. Papandrianos
3,
Elpiniki I. Papageorgiou
3 and
George S. Panayiotakis
1
1
Department of Medical Physics, School of Medicine, University of Patras, 26504 Rio, Greece
2
Laboratory of Nuclear Medicine, University Hospital of Patras, 26504 Rio, Greece
3
Department of Energy Systems, Gaiopolis Campus, University of Thessaly, 41500 Larisa, Greece
*
Author to whom correspondence should be addressed.
Submission received: 25 August 2022 / Revised: 13 September 2022 / Accepted: 22 September 2022 / Published: 27 September 2022
(This article belongs to the Special Issue Intelligent Imaging in Nuclear Medicine)

Abstract

:
Deep learning (DL) is a well-established pipeline for feature extraction in medical and nonmedical imaging tasks, such as object detection, segmentation, and classification. However, DL faces the issue of explainability, which prohibits reliable utilisation in everyday clinical practice. This study evaluates DL methods for their efficiency in revealing and suggesting potential image biomarkers. Eleven biomedical image datasets of various modalities are utilised, including SPECT, CT, photographs, microscopy, and X-ray. Seven state-of-the-art CNNs are employed and tuned to perform image classification in tasks. The main conclusion of the research is that DL reveals potential biomarkers in several cases, especially when the models are trained from scratch in domains where low-level features such as shapes and edges are not enough to make decisions. Furthermore, in some cases, device acquisition variations slightly affect the performance of DL models.

1. Introduction

There is extensive discussion about the quality and usefulness of the image features extracted by deep learning (DL) methods, especially in medical imaging. DL holds first place in terms of the number of features mined and quantified throughout the processing layers of a deep network; however, few studies have performed a deep analysis of the clinical, prognostic, and diagnostic value of these features. This is due to several facts, such as the massive number of features to be analysed, the general behaviour of deep networks as black boxes, and the lack of statistical and mathematical tools to distinguish significant attributes from sets containing thousands or millions of features [1].
The application of DL in processing medical images, especially in developing predictive and diagnostic models, is hindered by the fact that it is neither explainable nor transparent. Moreover, the model’s supervision of the decision-making process is unknown to engineers, let alone to medical staff. On the other hand, the discovery of vital image biomarkers has recently catalysed the emergence of many techniques based on machine learning (ML) and others that export features in a nonautomatic way [2,3].
Despite the significant results, the latter procedure (i.e., manual feature extraction) bears limitations. Manual feature extraction or supervised automatic predefined feature extraction works assuming no other feature exists except those under investigation. For example, research focused on revealing the prognostic or predictive value of a predefined feature or a well-known biomarker, regardless of the results, overlooks the possible presence of other features, which may be more essential than the one studied. This is time-consuming and may overlook potential correlations between discovered and undiscovered features or biomarkers. Frequently, the candidate biomarker is not a new feature discovered in the image but a prominent feature, the extraction of which is possible with conventional tools. The above issues are partially valid in radiomics [4] analysis.

1.1. Related Work

Most related research in the field is focused on predicting clinical outcomes by analysing predefined potential biomarkers extracted from medical and clinical image data [5,6,7,8,9,10]. In addition, there are studies aiming to discover new biomarkers from gene expression data or other clinical features. For example, Akbari et al. [11] utilised DL features and radiomic features (subtle yet spatially complex quantitative imaging phenomic (QIP) features) from Magnetic resonance imaging scans to achieve a noninvasive classification of true progression (TP) versus non-TP and TP versus pseudo-progression (PsP). For the in-depth features, the authors utilised a pretrained network. Courtiol et al. [12] designed a network to predict the survival times of patients with malignant mesothelioma (MM) based on histological criteria. They utilised histopathological images from two publicly available image data repositories. Their methodology yielded better accuracy in predicting patient survival compared to pathology practices. In addition, the proposed network suggested regions of interest contributing to the outcome. Those regions refer to the stroma and contain histological features associated with inflammation, cellular diversity, and vacuolisation.
Discovering novel biomarkers from medical images is under-represented in medical research. Zhuang et al. [13] proposed a novel methodology for classifying autism spectrum disorder (ASD) and identifying biomarkers for ASD, utilising the connectivity matrix calculated from fMRI as the input. They employed invertible networks and determined the decision boundary. The difference between the data point and projection onto the decision boundary was utilised, on the basis of which potential biomarkers were defined. Regression was used for the validation of the suggested biomarkers. The experiments demonstrated that the proposed network is effective and interpretable, thus allowing for new biomarker discoveries.
Lei et al. [14] proposed a survival analysis system utilising DL features combined with feature selection methods and survival analysis strategies, such as the Cox proportional hazards model. Their experiments were performed utilising open lung cancer datasets. The DL framework suggested specific biomarkers during their experiments, which were analysed. Furthermore, the DL methodology suggested regions where discriminating features were discovered.
Waldstein et al. [15] analysed approximately 50,000 retinal optical coherence tomography (OCT) volume scans of roughly 1000 patients with age-related macular degeneration employing DL to generate 20 local and global candidate biomarkers. Those candidate features correlated with traditionally used biomarkers in clinical practice. In addition, they discovered hitherto unknown biomarkers. Their research demonstrated that the discovered image features are closely connected with the traditional biomarkers with the specific DL approach. A correlation analysis was performed to achieve the above.
The studies mentioned above indicate that DL methods could be useful in revealing important image features that correlate well with existing biomarkers or suggest new ones. However, the results of the conducted studies are not easily validated due to dataset size limitations. This happens because the overall importance of the selected potential biomarker cannot be verified outside of the data sample utilised for a particular study. However, a correlation between a certain disease or therapy response and a predefined feature or characteristic may exist in a specific data selection. Therefore, thousands of image data must be analysed to confirm the discovery of a new biomarker, let alone the fact that these data should be representative of the actual population.
Moreover, the inherent ability of DL to reveal such biomarkers is not formally validated. Especially for clinical image analysis, the automatic procedures must be straightforward, transparent, explainable, and reproducible [16]. Every scientific conclusion derived from an artificial intelligence (AI) method, which may affect and redefine the medical pipeline in many domains, must rely on established techniques, analytical experiments, and clinical evaluation. In medical imaging, the definition of quantifiable and reproducible significant biomarkers evolves into a particular field of research. It utilises the full range of tools developed, such as ML, DL, and radiomics.

1.2. The Aim of the Study

The present study intends to assess DL methods from a biomedical engineering perspective rather than a statistical and mathematical evaluation based solely on predefined accuracy metrics. The study evaluates the potential importance of medical imaging features extracted by several DL models by examining the following:
a.
The effectiveness of transfer learning based on the type of the extracted features;
b.
The models’ robustness to acquisition device variation;
c.
The validity of the extracted feature maps across the layers of the best-performing DL models;
d.
The post hoc interpretation of Grad-CAM visualisations presenting the suggested areas of interest.
As discussed in the remainder or the paper, these four methodological steps can offer important information about the extracted imaging features’ type and importance, regardless of the classification metrics achieved by each DL model.
The experiments of the study are performed using several medical imaging datasets, including solitary pulmonary nodule (SPN) computed tomography (CT) scans, COVID-19 CT and X-ray scans, coronary artery disease (CAD) SPECT scans, skin cancer photographs, histopathological images of lung, colon, lymphoma, and prostate cancer, and cervical cancer images. In addition, seven state-of-the-art DL models are employed and compared with a modified version of VGG19 [17] proposed by the author team in recent studies [18,19,20,21,22] that demonstrated promising results.
The contributions of this study can be summarised as follows:
(i).
Deep learning CNNs are benchmarked to inspect their reasoning based on the extracted features. This analysis yields strong evidence that features of medical interest are extracted when training the proposed networks from scratch.
(ii).
The effectiveness of the Grad-CAM algorithm in improving the explainability of deep learning models is evaluated by analysing the suggested areas of interest.
(iii).
An innovative modification of the VGG19 network is extensively evaluated on various image classification tasks.
(iv).
It is concluded that FF-VGG19 and Xception networks are the most suitable CNNs for classifying medical images from various sources, such as CT, SPECT, and X-ray.

2. Materials and Methods

2.1. DL Feature Extraction Methods

The types of features mined by a DL model depend on network structure and the combination of the consisted parameters. For example, different features are extracted by a convolution filter of 3 × 3 pixel size compared to a convolution filter of 5 × 5 pixel size. Thus, an almost infinite number of features can be extracted. It is possible that a large variety of different features contribute equally to the desired results and could theoretically be excluded by employing a feature selection method. The model learns the significance of each feature throughout the loss backpropagation method, which updates the weights assigned to each filter. In this way, features irrelevant to the desired outcome suffer weight penalties.
While many researchers utilise handcrafted CNNs to perform medical image classification tasks, it is accepted that state-of-the-art networks are preferable since they are already validated in similar tasks. Well-established networks commonly yield better results [22,23,24,25,26]. Moreover, the manually designed networks are mostly randomly picked and adjusted following various experiments. With no doubt about their effectiveness, it is rare for the proposed networks to be dedicated to a specific project and for their parameters to be documented on the basis of a specific goal. Networks that follow this process go beyond the narrow boundaries of the domain for which they were designed and are applied to other domains [27].
In DL, employing networks trained on one image task (e.g., classification) to perform the same task on a set of different but related images is called transfer learning. This is the case in most studies employing CNNs. Transfer learning refers to the process wherein knowledge is transferred from one domain to another.

2.1.1. Feature Extraction via Transfer Learning or Learning with Off-the-Shelf Features

This method is often called feature extraction via transfer learning. It involves employing a pretrained network, retaining its architecture and learned filters, and training only some extra dense layers with the domain images under analysis. As CNN maintains knowledge from the first domain, this knowledge transfer limits the learning capacity of CNN. The network attempts to reveal, extract, and rate new features from the target domain task, but it will extract information according to its initial mastery. Usually, the extracted features are validated and classified by an effective classifier, such as a neural network or a support vector machine. In various examples, the retained knowledge (i.e., weights) is adequate to estimate and correctly predict the class of an image of another (but related) domain. The images of the source and the target images share some common decisive characteristics.

2.1.2. Training from Scratch

Conversely, transfer learning with training from scratch borrows only the architecture of the CNN, not its obtained knowledge. The complete network is trained on images of the domain of interest, adjusting the network to expertise on the specific task and the filters according to the desired outcome. Training from scratch is avoided when we face the data scarcity issue, or when the CNN is so deep that it introduces millions of trainable parameters, which are impossible to learn. Training from scratch may also be avoided when the source and target images belong to the same domain and share many characteristics. In such cases, feature extraction via transfer learning may be preferable. Lastly, this learning method severely increases the computational cost, leading to long-term training on a scale of hours, days, or even months [28].

2.1.3. Fine-Tuning

A hybrid learning method, which consists of the abovementioned strategies, is called fine-tuning. This method involves adjusting the trainability of specific filters of the CNN while retaining its original architecture. In this way, useful knowledge from the initial training is transferred to the new task. Some filters are randomly initiated and allowed to be adjusted during the backpropagation procedure of the training and validation process.

2.2. Datasets

The present study utilises 11 medical imaging datasets of various modalities and domains. Table 1 summarises the characteristics of each dataset.

Image Pre-Processing and Data Augmentation

Data augmentation and image processing methods are applied to normalise the images, reduce dimensionality, and prevent overfitting.
Data augmentation methods intend to expand the training data and help the networks ignore spatial variations and irrelevant image characteristics. For example, the random rotations aim to present different representations of the same finding and instruct the network to go beyond the specific location of the finding, i.e., to detect the important finding and focus on it without considering its exact position. Random Gaussian noise helps the networks generalise because they are forced to depend their decisions on holistic features rather than specific ones. Data augmentation in medical imaging must be applied cautiously. Heavy image distortions that do not correspond to realistic representations can be avoided because they can confuse the models.
Moreover, it is often the relevant location of a finding that contributes to the final result. For example, the location of a lung nodule (e.g., near the pleura) is crucial to the diagnosis. Applying augmentations that disrupt this image characteristic may result in a faulty diagnosis.
Data augmentation is not applied before training but simultaneously during each training fold over a 10-fold cross-validation procedure, ensuring that the generated images are not included in the test sets. For each dataset, the image pre-processing parameters are described as follows:
  • For the PET/CT and LIDC-IDRI datasets, the extracted nodule images are rescaled to 80 × 80 pixels, close to the minimum acceptable size for the CNNs. Image augmentations include 10° rotations, horizontal and vertical shifts, and random Gaussian noise additions with a mean of zero and a sigma of 0.010.5.
  • All images are placed and fit to a 266 × 200 black background for the COV-X dataset. Slight augmentations (rotation by 20°) are applied during the training. For the COV-CT dataset, augmentation refers to width and height shifts (at a maximum pixel range of 10%) and the addition of Gaussian noise with a mean of zero and a sigma of 0.010.5.
  • For the Cells and Cells2 dataset, the images are rescaled to 100 × 100 pixels, and no augmentation methods are applied.
  • As far as the MPIP dataset is concerned, the images are rescaled to 200 × 200 pixels, and slight rotations (maximum 10°) are applied to each polar map before concatenating them into a single four-polar-map image.
  • The images of the Leukaemia dataset are rescaled to 80 × 80 pixels, and no data augmentation is applied during training.
  • The images of NV-LES, MNIST, and SKIN datasets are rescaled to 200 × 150 pixels, and no data augmentation is applied.

2.3. DL Networks

The study employs seven state-of-the-art networks that stood out in the ImageNet challenge [39]. In addition, an innovative modification of VGG19 introduced by the authors in a recent study [19] is also employed. Table 2 presents an overview of the employed CNNs. Figure 1 presents the structure of the FF-VGG19 network.

2.4. Research Methodology

An overview of the methodology is presented in Figure 2.

2.4.1. Training from Scratch and Feature Extraction via Transfer Learning

Each network is trained using the aforementioned transfer learning methods in this experiment. The main aim of this experiment is to test the hypothesis that training from scratch is superior to feature extraction via transfer learning. Secondly, the best-performing networks are compared to define a superior network suitable for all the datasets. Table 3 summarises the involved sub-experiments.
The reader can notice that the training dataset is the same as the test dataset during this experiment, and this was achieved following a 10-fold cross-validation procedure.

2.4.2. Investigating the Robustness of DL Methods to Image Acquisition Device Variation

Results of the previous experiment revealed that training from scratch yields better results and may contribute to potential biomarker detection from the input images. To further investigate this, a second experiment was conducted. The test datasets differed from the training datasets to investigate the robustness to variation in image acquisition devices. For example, to investigate the importance of the features extracted by the training of the CNNs on the LIDC-IDRI dataset, the CNN was asked to distinguish benign and malignant nodule images from the PET/CT dataset. This procedure was applied to three major medical classification tasks:
  • Distinguish between benign and malignant SPNs;
  • Distinguish between benign and cancerous cells;
  • Distinguish between lymphocyte and monocyte blood cells.
The source and the target domain classification tasks are the same. However, since there is a variation in the data distributions of images between datasets of the same kind (e.g., SPN datasets), a sub-optimal feature extractor could be misled if the extracted features are dataset-specific. In this scenario, employing this classifier to classify images from an alternative dataset (of the same domain and task) would not be effective. On the other hand, if the trained classifier ignores the irrelevant artefacts and characteristics of the hidden dataset and achieves a high classification score, that would be a prime indicator that the classification is based on realistic and precise features (potential image biomarkers). The experiments of the second phase are explained in Table 4.

2.4.3. Feature and Output Visualisation Methodologies

The previous experiments included performance assessment in terms of quantifiable metrics. In the present experiment, the output feature maps and the Grad-CAM algorithm visualisation results were used as the criteria for rating the performance of DL models. Those two methods are explained below.
  • Grad-CAM algorithm [46]
The Grad-CAM algorithm intends to identify the areas of the input image having a critical effect on the classification decision of the classifier placed at the top of the CNN. Hence, its functionalities are fully exploited in object detection tasks, where a specific image area contains the desired object.
b.
Extracted Feature Map visualisation method
This methodology tracks the output feature maps of every CNN layer and reconstructs them into an image. In this way, the user can inspect what features are learned during the process and decide their importance.
Visualising the areas of interest or the extracted feature maps does not solve the explainability issue. However, it is useful for reaching some conclusions. More precisely, the visualisation of the feature maps can provide relevant information as to how the DL models handle the feature extraction process, how the extracted features are produced during the convolution stages, and whether those features represent expected patterns also visible by the human eye.

3. Results

In this section, the results of the study are analytically presented. Section 3.1 presents the results of transfer learning with “off-the-shelf” features. Next, in Section 3.2, the training from scratch results are discussed. In Section 3.3, the robustness of device variation is presented. In Section 3.4 and Section 3.5, the evaluations of the extracted features maps and the Grad-CAM maps are presented, respectively.

3.1. Transfer Learning with “Off-the-Shelf” Features

The performance of the various CNNs on the 11 medical image datasets is presented in Table 5. Each experiment was conducted three times to inspect the reproducibility of the results. A variation of ±2% was observed for each training–testing run. Hence, we considered the classification accuracies equivalent when their difference was below 2%.
The suboptimal performance observed in Table 5 underlines that the transfer learning strategy utilising off-the-shelf features may not be effective for discovering and revealing significant, deep, and high-level medical-specific features from the inputs. However, slightly better performance was expected and verified in tasks where low-level features, such as the shape of the depicted SPN, are essential predictors of the desired outcome. For example, it was observed that, for the LIDC-IDRI dataset, VGG19 achieved 80.97% classification accuracy, and its predictions were probably based on the shapes and sizes of the SPNs. Low-level features were adequate to successfully characterise 80.97% of the SPN images incorporated into LIDC-IDRI. However, this did not apply to the PET/CT dataset containing SPN images. Good classification accuracies were observed in Cells, NV-LES, and MNIST datasets, achieved by either FF-VGG19 or VGG19. This may reveal that the architecture of those networks aids in extracting deep and useful features, even if the initial training was performed using nonmedical images. This assumption could be validated in future research. In Table 6, more evaluation metrics are given for the best-performing network in each dataset.
FF-VGG19 obtained the best accuracy in most of the experiments, specifically, 72.09% on PET dataset, 84.86% on LIDC dataset, 84.96% on COV_X dataset, 79.35% on COV_CT dataset, 90.74% on Cells dataset, 61.80% on Cells2 dataset, 77.92% on Leukaemia dataset, 63.24% on MPIP dataset, 82.36% on NVLES dataset, 77.74% on SKIN dataset, and 75.55 on MNIST dataset. The classification accuracies were also accompanied by acceptable AUC and F1-scores, as seen in Table 6.

3.2. Feature Extraction via Training from Scratch

The results underline the effectiveness of training from scratch. Specifically, almost every CNN improved its accuracy in every experimental dataset, as demonstrated in Table 7.
FF-VGG19 and Xception stood out in this experiment, achieving high classification accuracies in every dataset. Training the CNNs from scratch yielded better results in most cases. Xception achieved the best accuracy on PET/CT (69.77%), and FF-VGG19 achieved the best accuracy on LIDC-IDRI (82.52%). FF-VGG19 attained the best result on the COV_X dataset, yielding 84.64% accuracy. It also achieved the best scores on COV_CT and Cells datasets, yielding 78.57% and 91.61% accuracy scores. Xception exhibited better results than any other CNN on Cells2 and Leukaemia datasets (59.66% and 78.85% accuracy, respectively). FF-VGG19 achieved optimal results on the MPIP and NVLES datasets (66.08% and 83.66% accuracy). Lastly, the best accuracy on the MNIST dataset was obtained by FF-VGG19 (77.93%).
It can be observed that training from scratch was beneficial for most networks. Furthermore, it was revealed that training from scratch allowed the models to adjust their weights to specific tasks and not base their predictions on pre-learned features.
While VGG19 appeared to yield better results when training with off-the-shelf features, the other networks showed significant improvements. For example, Xception improved its accuracy in classifying the Cells dataset from ~34% to ~91%. A visualisation of those comparisons is illustrated in Figure 3.
As illustrated in Figure 3, it can be assumed that Xception, ResNet, InceptionV3, and DenseNet121 were best for training from scratch. his is not irrelevant because those networks consist of dozens of layers. In contrast with those networks, VGG19 yielded better results when utilising a transfer learning approach. FF-VGG19 was more stable to learning strategy alterations, maintaining its performance at a top level.
FF-VGG19 was the optimal CNN for achieving maximum accuracy for most datasets, while training from scratch and transfer learning had no important impact on its performance. FF-VGG19 yielded good AUC scores for every dataset, as shown in Table 8. In Figure 4, an overview of the CNNs’ performance for both learning strategies is illustrated.
Some concluding remarks are reported as follows:
  • CNNs failed to perform optimally in small datasets. This behaviour was expected, as CNNs require large-scale training sets to learn. The features extracted by every CNN from Cells2, MPIP, and PET/CT dataset were expected to be low-level image features, such as edges and shapes. In these cases, the features did not constitute novel biomarkers;
  • Training from scratch or transfer learning is a suitable learning strategy in cases wherein a feature of vital medical importance can be the shape, the greyness level, or the size of a pathological finding. Preferable CNNs for small datasets include VGG19 and MobileNet;
  • FF-VGG19 was a robust CNN for most of the medical image datasets. Therefore, it could serve as a general model for similar medical image classification tasks, either without modifying its components and parameters or performing a slight modification to improve the accuracy further.

3.3. Robustness to Medical Image Acquisition Device Variation

The best-performing networks (FF-VGG19 and Xception) were utilised in this experiment. Under this setup, the two CNNs were trained using a dataset of one domain (e.g., SPN malignancy identification) but were evaluated using an unseen dataset of the same domain. The classification accuracies obtained are illustrated in Table 9, while the corresponding sensitivities, specificities, AUC scores, and F1-scores are given in Table 10 and Table 11.
Both networks successfully classified the PET/CT dataset when trained using samples from the LIDC-IDRI dataset. However, the PET/CT dataset was more effectively classified when the networks used LIDC-IDRI as their training set. Specifically, training and testing FF-VGG19 on the PET/CT dataset yielded 69.77% accuracy, while training on LIDC-IDRI and predicting PET/CT yielded 74.41% accuracy. This observation shows that both networks learned to extract important features that may define an SPN in general and performed actual reasoning.
The same phenomenon was observed when the target dataset was MNIST. For example, training FF-VGG19 on MNIST and testing on the same dataset returned an accuracy of 77.96%. At the same time, training on the concatenation of SKIN and NVLES datasets yielded a prediction accuracy of 84.01% on the unseen testing dataset (MNIST). Therefore, it is fair to assume that SKIN and NVLES datasets were effective training samples which contained enough images to help FF-VGG19 acquire and learn the right underlying features.
A decrease of ~10% in accuracy was observed for the remaining experimental cases. While this performance raises concerns about the validity of the learned features, other issues may also have affected the results. For example, the soft data augmentation may have oversimplified the training sets and led the network to learn too specific information, making them unable to generalise and handle new test sets, even in the same domain. This issue needs further investigation.

3.4. Feature Map Visualisation

The activation maps from specific layers of the FF-VGG19 CNN were visualised to inspect the learned weight filters. For every dataset, a random image was selected and classified from the trained FF-VGG19. The training was performed on various datasets, as explained in Table 12. Figure 5 and Figure 6 illustrate the feature maps.
Figure 5 and Figure 6 include the activation maps of randomly selected images from the datasets that visualise the outputs of specific convolution layers from the FF-VGG19 network. Five sets of activation maps, including nine random feature maps each, are reported for each sample original image (labelled a, b, c, etc.). Those groups are numbered from one to five.
The outputs of the early layers, i.e., the layers which performed their analysis first (numbers two to four), were less abstract and detected basic edges and patterns of the image. The black colours depict areas with less information, and the brighter parts highlight important characteristics. Deep convolution layers seek to extract more abstract and underlying features to discover new information related to the task. Those feature maps are easily understood because they seem meaningless to the human eye and brain (maps numbered five and six). However, the investigation of this research was directed toward those feature maps, which may have hidden potential biomarkers.
The feature maps extracted from the skin cancer dataset (MNIST, SKIN, NVLES) tended to characterise the skin texture as a special feature in the image, as seen in Figure 5. Although the hardness of the skin, its roughness, and the presence of hair do not constitute predictors for the severity of skin cancer, the early feature maps detected and analysed their patterns captured in the image. Of course, the classifier’s responsibility is to properly adjust the weights to reject those features for the specific task. This ability is not irrelevant to the available data, the network’s learning capacity, and the parameter tuning. However, late convolutional layers tend to ignore such features in an attempt to go deeper.
In Figure 5(a5), the surroundings of the SPN were separately captured. This confirms the assumption that CNNs captured both point and spatial information while performing an automatic segmentation of potential areas of interest. Furthermore, from the feature maps of the SPNs, especially in a2, a3, and a4, one can notice that the outline of the SPN, which was fainter than their main part (the centre), was also captured separately. Following this visualisation, it was assumed that the CNN attempted to separate the detected shapes into complementary shapes.
In Figure 5(b6), it can be noticed that the content behind the imaged bones, i.e., the histological information, acquired its representation in the thousands of the extracted features. The feature maps b2, b3, b4, and b5 did not capture any important information because they focused on geometries and shapes concerning the skeleton, not what was behind, around, or inside. The same phenomenon was confirmed in the case of the CT images (Figure 5c).
In cases where early detected features related to shapes or edges were less relevant for the particular classification task (e.g., myocardial perfusion imaging polar maps—MPIP dataset), the feature maps were not informative (Figure 6(a2–a4)). As seen in a5 and a6, the convolution process captured specific areas inside the polar maps. Those areas were indeed decisive for the diagnostic test outcome, despite the test’s suboptimal sensitivity and precision. Areas in green were assumed to be indicators (although weak) of the risk of coronary artery disease. However, in many cases, the test yielded false positive results, making things more complex for the diagnosis.
In Figure 6(b2), the reader can notice that the outline of the cancerous cell structure was captured, and more vital information about its texture was captured in later feature maps (b5, b6). Again, the skin texture was separately captured in b3 and b4.
The feature maps’ visualisation method is helpful in determining whether the deep convolutional process is learning important features or is at least seeking patterns in the positive direction. From the illustrated feature maps, is the following could be confirmed in most cases:
  • Both relevant and irrelevant features were captured;
  • Most important and decisive features were discovered in deeper convolutions;
  • At least half of the extracted features were not of medical importance, while the significance of the remainder is yet to be investigated (excluding low-level features, which are known to be significant in specific examples).
Since it is implausible to inspect every produced feature map, especially in deep networks, it is not easy to confirm that all potential biomarkers were extracted from FF-VGG19. However, looking beyond a specific network and focusing solely on the potential of the process of the CNNs yields the conclusion that DL, in general, has what it takes to discover significant features.

3.5. Grad-CAM Results

Selected images were processed from the trained FF-VGG19 network to inspect the decisive features’ location using the Grad-CAM algorithm. The training datasets and the selected images are presented in Table 13.
The Grad-CAM algorithm traces the convolutional layers and neurons participating strongly in formulating the final classification outcome. In this way, FF-VGG19 revealed the important areas of the original images, wherein the decisive features were discovered. Hence, the Grad-CAM algorithm helps interpret and evaluate the feature extraction process, which in many cases, has a valuable contribution to improving the model’s explainability.
For the SPNs depicted in Figure 7a, it can be noticed that the benign SPNs are highlighted in red, and their surroundings are not highlighted at all, in contrast with the malignant SPNs, where the red colourmap covers the entire image. It is demonstrated that the model sought information in the correct areas, although the discovered information could not be visualised, at least with the Grad-CAM algorithm.
It can also be noticed that the COVID-19 X-ray images were classified as such on the basis of discovered patterns in the high respiratory system. In contrast, normal cases were classified on the basis of features in other locations. This demonstrates that CNNs were looking for information in the right direction, as it has been proven that the novel coronavirus affects specific areas of the respiratory system. However, it can also be observed that some irrelevant features were extracted and misjudged as decisive ones.
The detection of COVID-19 in CT images has been questioned in the scientific community. For this reason, the highlighted areas of the CT images (Figure 7c) are not reliable. However, the CNNs also highlighted irrelevant areas (e.g., c1 outer left picture, c2 outer left picture). Therefore, it was assumed that the model was confused by the incomplete and irrelevant-to-COVID-19 information discovered in CT cases. Since the CT_COV dataset included few images, no further assumptions can be made at this stage.
For the myocardial perfusion polar maps, the CNNs were based on green areas of the initial image, which is the correct method. However, the initial image itself was not informative in a significant proportion of the subjects and led to false positives.
The highlighted areas in images depicting skin abnormalities (Figure 8) do not provide useful information regarding the important features. It was only confirmed that FF-VGG19 detected the area of interest correctly, as it was not pointing toward irrelevant locations. In combination with the activation maps of Figure 4, it is demonstrated that FF-VGG19 managed to exclude confusing patterns in the images, such as the surrounding hair.

4. Discussion

Training from scratch was superior to transfer learning in most of the utilised datasets and for the majority of the CNNs. Moreover, it was observed that FF-VGG19 effectively trained from scratch and transfer learning, providing a stable and robust framework for developing a general medical image classification model. The results indicate that the special features learned during the training from scratch are potential biomarkers, especially in cases where the amount of the training data allows for deep CNNs.
Next, the covariance shift methodology was followed to perform a deeper analysis of the extracted features. Specifically, it was investigated whether the learned features of a specific dataset (e.g., skin cancer) can be retained for predictions on a separate dataset of the same nature and kind. The results indicate that the DL model could extract important features from the source datasets and achieve top performance on alternative and unseen datasets. This yields strong evidence that CNNs can identify general and decisive image features of medical importance, which are met on a specific training dataset and refer to global features existing in images of the same domain. Activation maps were also visualised. The Grad-CAM algorithm was employed to shed light on those features. It was observed that the DL model extracted relevant and significant feature maps among the numerous extracted maps. Furthermore, the Grad-CAM algorithm revealed that the DL model correctly pointed to the valid areas of interest where human expertise focuses its attention when examining similar medical images.
The findings of the study agree with the findings of recent works. For example, Yang et al. [47] presented an attention-based explainability framework to detect important areas of interest in breast cancer histopathology images. Their framework, although not suggesting novel biomarkers, was able to identify the expected regions correctly. In PET imaging, Choi et al. [48] developed a deep learning-based cognitive signature of FDG brain PET adaptable for Parkinson’s disease (PD) and Alzheimer’s disease (AD). The two-dimensional projection mapping visualised the degree of cognitive dysfunction in agreement with the expected degree. In [49], the authors designed an explainable framework for automated skin lesion segmentation and classification. The visualised class activation maps verified that the model seeks important features in the correct locations of the image. Again, the visualisation could not suggest specific biomarkers but large regions. The review study of van der Velden et al. [50] discussed several recent studies that used explainable methods for various medical imaging modalities, including PET, MRI, and CT.
The contribution of this research is threefold. Firstly, a deep and extensive analysis of the extracted features of CNNs was conducted. This analysis yielded strong evidence that medical-specific features (i.e., image biomarkers) are extracted when training the proposed networks from scratch and observing large-scale datasets, especially in pathological findings. The outcome is derived from obvious features (e.g., colour, size, and shape). Moreover, the study performed a preliminary analysis of the feature maps and employed the Grad-CAM algorithm to inspect the suggested regions.
Secondly, the study identified the most suitable medical image data classification networks. More precisely, FF-VGG19 and Xception stood out in most classification tasks. On the other hand, DenseNet is ideal for skin malignancy detection.
Thirdly, this research proposed a novel modification of the traditional VGG19 network, which exploited the benefits of feature fusion via concatenation and outperformed every competitive network.
Lastly, we suggested a modification for further evaluation in larger-scale datasets of various medical applications.

Limitations

One may highlight several limitations which fall beyond the scope of this study and raise general issues for the scientific community. Firstly, most of the available datasets are not large enough to successfully train deep and complex CNNs. Therefore, it is imperative that institutions and researchers collaborate to release well-labelled and large-scale datasets of various modalities. Secondly, while various datasets were utilised, the actual range of medical image modalities and diseases was not covered. Further research could be conducted to explore a wider range. Thirdly, since more state-of-the-art networks have emerged lately, a limitation of this study is that they were excluded.
Moreover, this study focused solely on automatic feature extraction methods. At the same time, several hybrid methodologies were recently proposed to both extract hand-crafted features of proven importance and let the CNNs automatically extract more features. Future studies could involve such methods to inspect the usefulness of the suggested features and how much they correlate with known biomarkers or other predefined image characteristics. For example, employing explainable ML algorithms, such as random forest and fuzzy cognitive maps, in integrated DL-ML frameworks could provide more details about the importance of the extracted features.
The field of explainable artificial intelligence was not entirely explored. For example, recent algorithms such as LIME and Grad-CAM++ were not evaluated. It is the authors’ intention to benchmark these methods in future studies.

5. Conclusions

In the present research study, DL models were assessed to inspect their capabilities in extracting and presenting potential image biomarkers. Firstly, it was observed that transfer learning via training from scratch is preferable over feature extraction via transfer learning when large-scale and good-quality data are present. Training from scratch makes it possible to detect special features which can be identified as image biomarkers. Secondly, the inspection of the extracted features utilising feature activation maps and algorithms revealing the location of those features verified that those methods are suitable, can improve the explainability of the models, and can provide an analysis of the extracted characteristics. Their results revealed that DL seeks potential biomarkers in the right direction. The deep layers of the networks indeed revealed those important features. Thirdly, the novel FF-VGG19 outperformed every state-of-the-art CNN when training from scratch, while it remained stable when utilised for transfer learning. Therefore, it could serve as a baseline model for many related medical image classification tasks, including PET scans, CT scans, X-ray scans, dermoscopy, and histopathological images.

Author Contributions

Conceptualisation, I.D.A. and N.I.P.; methodology, I.D.A., N.D.P. and G.S.P.; software, I.D.A. and E.I.P.; validation, N.D.P., E.I.P. and G.S.P.; resources, N.D.P.; writing—original draft preparation, I.D.A. and N.I.P.; writing—review and editing, E.I.P. and G.S.P.; supervision, G.S.P. All authors have read and agreed to the published version of the manuscript.

Funding

The research project was supported by the Hellenic Foundation for Research and Innovation (H.F.R.I.) under the “Second Call for H.F.R.I. Research Projects to support Faculty Members & Researchers” (Project Number: 3656).

Institutional Review Board Statement

Ethical review and approval were waived for this retrospective study.

Informed Consent Statement

Patient consent was waived for this retrospective study.

Data Availability Statement

The PET/CT and MPIP datasets are not publicly available due to ethical reasons. The LIDC-IDRI, COV_X, COV_CT, Cells, Cells2, Leukaemia, NV-LES, MNIST, and SKIN datasets are publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Beykikhoshk, A.; Quinn, T.P.; Lee, S.C.; Tran, T.; Venkatesh, S. Deep TRIAGE: Interpretable and individualised biomarker scores using attention mechanism for the classification of breast cancer sub-types. BMC Med. Genom. 2020, 13, 20. [Google Scholar] [CrossRef] [PubMed]
  2. Jain, D.; Singh, V. Feature selection and classification systems for chronic disease prediction: A review. Egypt. Inform. J. 2018, 19, 179–189. [Google Scholar] [CrossRef]
  3. Traverso, A.; Wee, L.; Dekker, A.; Gillies, R. Repeatability and Reproducibility of Radiomic Features: A Systematic Review. Int. J. Radiat. Oncol. Biol. Phys. 2018, 102, 1143–1158. [Google Scholar] [CrossRef] [PubMed]
  4. Lambin, P.; Leijenaar, R.T.H.; Deist, T.M.; Peerlings, J.; de Jong, E.E.C.; van Timmeren, J.; Sanduleanu, S.; Larue, R.T.H.M.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef] [PubMed]
  5. Cano-Espinosa, C.; Gonzalez, G.; Washko, G.R.; Cazorla, M.; Estepar, R.S.J. Biomarker Localization From Deep Learning Regression Networks. IEEE Trans. Med. Imaging 2020, 39, 2121–2132. [Google Scholar] [CrossRef] [PubMed]
  6. Cole, J.H.; Poudel, R.P.K.; Tsagkrasoulis, D.; Caan, M.W.A.; Steves, C.; Spector, T.D.; Montana, G. Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker. NeuroImage 2017, 163, 115–124. [Google Scholar] [CrossRef] [PubMed]
  7. Eaton-Rosen, Z.; Bragman, F.; Bisdas, S.; Ourselin, S.; Cardoso, M.J. Towards Safe Deep Learning: Accurately Quantifying Biomarker Uncertainty in Neural Network Predictions. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018; Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2018; Volume 11070, pp. 691–699. ISBN 978-3-030-00927-4. [Google Scholar]
  8. Okuda, K.; Watanabe, N.; Hashimoto, M.; Doai, M.; Kawai, Y.; Takahashi, T.; Arikawa, T.; Ooiso, K.; Sunatani, Y.; Iwabuchi, K.; et al. Preliminary Quantitative Evaluation of Radiation-Induced DNA Damage in Peripheral Blood Lymphocytes after Cardiac Dual-Isotope Imaging. Appl. Radiat. Isot. 2019, 154, 108890. [Google Scholar] [CrossRef]
  9. Putin, E.; Mamoshina, P.; Aliper, A.; Korzinkin, M.; Moskalev, A.; Kolosov, A.; Ostrovskiy, A.; Cantor, C.; Vijg, J.; Zhavoronkov, A. Deep biomarkers of human aging: Application of deep neural networks to biomarker development. Aging 2016, 8, 1021–1033. [Google Scholar] [CrossRef]
  10. Wang, S.; Liu, Z.; Rong, Y.; Zhou, B.; Bai, Y.; Wei, W.; Wei, W.; Wang, M.; Guo, Y.; Tian, J. Deep learning provides a new computed tomography-based prognostic biomarker for recurrence prediction in high-grade serous ovarian cancer. Radiother. Oncol. 2019, 132, 171–177. [Google Scholar] [CrossRef]
  11. Akbari, H.; Rathore, S.; Bakas, S.; Nasrallah, M.P.; Shukla, G.; Mamourian, E.; Rozycki, M.; Bagley, S.J.; Rudie, J.D.; Flanders, A.E.; et al. Histopathology-validated machine learning radiographic biomarker for noninvasive discrimination between true progression and pseudo-progression in glioblastoma. Cancer 2020, 126, 2625–2636. [Google Scholar] [CrossRef]
  12. Courtiol, P.; Maussion, C.; Moarii, M.; Pronier, E.; Pilcer, S.; Sefta, M.; Manceron, P.; Toldo, S.; Zaslavskiy, M.; Le Stang, N.; et al. Deep learning-based classification of mesothelioma improves prediction of patient outcome. Nat. Med. 2019, 25, 1519–1525. [Google Scholar] [CrossRef] [PubMed]
  13. Zhuang, J.; Dvornek, N.C.; Li, X.; Ventola, P.; Duncan, J.S. Invertible Network for Classification and Biomarker Selection for ASD. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019; Shen, D., Liu, T., Peters, T.M., Staib, L.H., Essert, C., Zhou, S., Yap, P.-T., Khan, A., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2019; Volume 11766, pp. 700–708. ISBN 978-3-030-32247-2. [Google Scholar]
  14. Lei, C.; Li, H.; Hui, W.; Chen, S.; Yang, L.; Kang, Y.; Bo, Q.; Feng, J. A deep learning-based framework for lung cancer survival analysis with biomarker interpretation. BMC Bioinform. 2020, 21, 112. [Google Scholar] [CrossRef]
  15. Waldstein, S.M.; Seeböck, P.; Donner, R.; Sadeghipour, A.; Bogunović, H.; Osborne, A.; Schmidt-Erfurth, U. Unbiased identification of novel subclinical imaging biomarkers using unsupervised deep learning. Sci. Rep. 2020, 10, 12954. [Google Scholar] [CrossRef] [PubMed]
  16. Razzak, M.I.; Naz, S.; Zaib, A. Deep Learning for Medical Image Processing: Overview, Challenges and the Future. In Classification in BioApps; Dey, N., Ashour, A.S., Borra, S., Eds.; Lecture Notes in Computational Vision and Biomechanics; Springer International Publishing: Cham, Switzerland, 2018; Volume 26, pp. 323–350. ISBN 978-3-319-65980-0. [Google Scholar]
  17. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar] [CrossRef]
  18. Apostolopoulos, I.D.; Mpesiana, T.A. COVID-19: Automatic detection from x-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 1, 635–640. [Google Scholar] [CrossRef]
  19. Apostolopoulos, I.D.; Apostolopoulos, D.I.; Spyridonidis, T.I.; Papathanasiou, N.D.; Panayiotakis, G.S. Multi-input deep learning approach for Cardiovascular Disease diagnosis using Myocardial Perfusion Imaging and clinical data. Phys. Med. 2021, 84, 168–177. [Google Scholar] [CrossRef]
  20. Apostolopoulos, I.D.; Apostolopoulos, D.J.; Papathanasiou, N.D. Deep Learning Methods to Reveal Important X-Ray Features in COVID-19 Detection: Investigation of Explainability and Feature Reproducibility. Reports 2022, 5, 20. [Google Scholar] [CrossRef]
  21. Apostolopoulos, I.D.; Pintelas, E.G.; Livieris, I.E.; Apostolopoulos, D.J.; Papathanasiou, N.D.; Pintelas, P.E.; Panayiotakis, G.S. Automatic classification of solitary pulmonary nodules in PET/CT imaging employing transfer learning techniques. Med. Biol. Eng. Comput. 2021, 59, 1299–1310. [Google Scholar] [CrossRef]
  22. Apostolopoulos, I.D.; Papathanasiou, N.D.; Panayiotakis, G.S. Classification of Lung Nodule Malignancy in Computed Tomography Imaging Utilising Generative Adversarial Networks and Semi-Supervised Transfer Learning. Biocybern. Biomed. Eng. 2021, 41, 1243–1257. [Google Scholar] [CrossRef]
  23. Causey, J.L.; Zhang, J.; Ma, S.; Jiang, B.; Qualls, J.A.; Politte, D.G.; Prior, F.; Zhang, S.; Huang, X. Highly accurate model for prediction of lung nodule malignancy with CT scans. Sci. Rep. 2018, 8, 9286. [Google Scholar] [CrossRef] [Green Version]
  24. Ricciardi, R.; Mettivier, G.; Staffa, M.; Sarno, A.; Acampora, G.; Minelli, S.; Santoro, A.; Antignani, E.; Orientale, A.; Pilotti, I.A.M.; et al. A deep learning classifier for digital breast tomosynthesis. Phys. Med. 2021, 83, 184–193. [Google Scholar] [CrossRef]
  25. Sheinfeld, M.; Levinson, S.; Orion, I. Highly Accurate Prediction of Specific Activity Using Deep Learning. Appl. Radiat. Isot. 2017, 130, 115–120. [Google Scholar] [CrossRef]
  26. Zhao, X.; Liu, L.; Qi, S.; Teng, Y.; Li, J.; Qian, W. Agile convolutional neural network for pulmonary nodule classification using CT images. Int. J. CARS 2018, 13, 585–595. [Google Scholar] [CrossRef] [PubMed]
  27. Ali, F.; Khan, S.; Waseem Abbas, A.; Shah, B.; Hussain, T.; Song, D.; EI-Sappagh, S.; Singh, J. A Two-Tier Framework Based on GoogLeNet and YOLOv3 Models for Tumor Detection in MRI. Comput. Mater. Contin. 2022, 72, 73–92. [Google Scholar] [CrossRef]
  28. Seide, F.; Agarwal, A. CNTK: Microsoft’s Open-Source Deep-Learning Toolkit. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, San Francisco, CA, USA, 13 August 2016; p. 2135. [Google Scholar]
  29. Armato, S.G.; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The Lung Image Database Consortium (LIDC) and Image Database Resource Initiative (IDRI): A Completed Reference Database of Lung Nodules on CT Scans: The LIDC/IDRI thoracic CT database of lung nodules. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef] [PubMed]
  30. Cohen, J.P. COVID-19 Image Data Collection. 2020. Available online: https://github.com/ieee8023/covid-chestxray-dataset (accessed on 23 March 2020).
  31. Zhao, J.; Zhang, Y.; He, X.; Xie, P. COVID-CT-Dataset: A CT Scan Dataset about COVID-19. arXiv 2020, arXiv:2003.13865. [Google Scholar] [CrossRef]
  32. Matek, C.; Schwarz, S.; Marr, C.; Spiekermann, K. A Single-Cell Morphological Dataset of Leukocytes from AML Patients and Non-Malignant Controls. 2019. Available online: https://www.leukaemiamedtechresearch.org.uk/open-information/datasets/single-cell-morphological-dataset-of-leukocytes-from-aml-patients (accessed on 23 March 2020).
  33. Shenggan BCCD_Dataset. 2019. Available online: https://github.com/Shenggan/BCCD_Dataset (accessed on 23 March 2020).
  34. Mourya, S.; Kant, S.; Kumar, P.; Gupta, A.; Gupta, R. ALL Challenge Dataset of ISBI. 2019. Available online: https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&ved=2ahUKEwiak_f18LP6AhXJqVYBHaa6AHUQFnoECAkQAQ&url=https%3A%2F%2Fwww.mdpi.com%2F2073-431X%2F11%2F8%2F125%2Fpdf%3Fversion%3D1661330918&usg=AOvVaw3oGFUWx-4H-amzP_sAvDg2 (accessed on 23 March 2020).
  35. Apostolopoulos, I.D.; Papathanasiou, N.D.; Spyridonidis, T.; Apostolopoulos, D.J. Automatic characterization of myocardial perfusion imaging polar maps employing deep learning and data augmentation. Hell. J. Nucl. Med. 2020, 23, 125–132. [Google Scholar]
  36. Grover, V. Skin_Cancer_Large_Dataset. 2021. Available online: https://www.kaggle.com/datasets/vankapgrover/skin-cancer-large-dataset (accessed on 23 March 2020).
  37. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef]
  38. Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Helba, B.; Kalloo, A.; Liopyris, K.; Marchetti, M.; et al. Skin Lesion Analysis toward Melanoma Detection 2018: A Challenge Hosted by the International Skin Imaging Collaboration (ISIC). arXiv 2019, arXiv:1902.03368. [Google Scholar] [CrossRef]
  39. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar] [CrossRef]
  41. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. arXiv 2015, arXiv:1512.00567. [Google Scholar] [CrossRef]
  42. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  43. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  44. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning Transferable Architectures for Scalable Image Recognition. arXiv 2018, arXiv:1707.07012. [Google Scholar] [CrossRef]
  45. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  46. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef]
  47. Yang, H.; Kim, J.-Y.; Kim, H.; Adhikari, S.P. Guided Soft Attention Network for Classification of Breast Cancer Histopathology Images. IEEE Trans. Med. Imaging 2020, 39, 1306–1315. [Google Scholar] [CrossRef]
  48. Choi, H.; Kim, Y.K.; Yoon, E.J.; Lee, J.-Y.; Lee, D.S. Cognitive Signature of Brain FDG PET Based on Deep Learning: Domain Transfer from Alzheimer’s Disease to Parkinson’s Disease. Eur. J. Nucl. Med. Mol. Imaging 2020, 47, 403–412. [Google Scholar] [CrossRef]
  49. Xie, Y.; Zhang, J.; Xia, Y.; Shen, C. A Mutual Bootstrapping Model for Automated Skin Lesion Segmentation and Classification. IEEE Trans. Med. Imaging 2020, 39, 2482–2493. [Google Scholar] [CrossRef]
  50. van der Velden, B.H.M.; Kuijf, H.J.; Gilhuijs, K.G.A.; Viergever, M.A. Explainable Artificial Intelligence (XAI) in Deep Learning-Based Medical Image Analysis. Med. Image Anal. 2022, 79, 102470. [Google Scholar] [CrossRef]
Figure 1. Structure of the FF-VGG19 network.
Figure 1. Structure of the FF-VGG19 network.
Diagnostics 12 02333 g001
Figure 2. Methodology of the current research work.
Figure 2. Methodology of the current research work.
Diagnostics 12 02333 g002
Figure 3. Accuracy improvement of the CNNs when training from scratch on the study’s datasets. It can be observed that, in most of the datasets, training from scratch was beneficial for the networks and improved their diagnostic capabilities.
Figure 3. Accuracy improvement of the CNNs when training from scratch on the study’s datasets. It can be observed that, in most of the datasets, training from scratch was beneficial for the networks and improved their diagnostic capabilities.
Diagnostics 12 02333 g003
Figure 4. Best-performing CNNs and learning strategies for each of the experimental medical imaging datasets.
Figure 4. Best-performing CNNs and learning strategies for each of the experimental medical imaging datasets.
Diagnostics 12 02333 g004
Figure 5. FF-VGG19 extracted feature maps of a random input image, where (a1a6) refer to feature maps from the LIDC-IDRI + PET dataset, (b1b6) refer to features maps from the COV_X dataset, and (c1c6) refer to the COV_CT dataset.
Figure 5. FF-VGG19 extracted feature maps of a random input image, where (a1a6) refer to feature maps from the LIDC-IDRI + PET dataset, (b1b6) refer to features maps from the COV_X dataset, and (c1c6) refer to the COV_CT dataset.
Diagnostics 12 02333 g005
Figure 6. FF-VGG19 extracted feature maps of a random input image, where (a1a6) refer to feature maps from the MPIP dataset, and (b1b6) refer to features maps from the MNIST + SKIN + NVLES dataset.
Figure 6. FF-VGG19 extracted feature maps of a random input image, where (a1a6) refer to feature maps from the MPIP dataset, and (b1b6) refer to features maps from the MNIST + SKIN + NVLES dataset.
Diagnostics 12 02333 g006
Figure 7. Grad-CAM results for the following test dataset samples: (a) LIDC, (b) COV_X, (c) COV_CT, (d) MPIP, (e) Cells, and (f) MNIST.
Figure 7. Grad-CAM results for the following test dataset samples: (a) LIDC, (b) COV_X, (c) COV_CT, (d) MPIP, (e) Cells, and (f) MNIST.
Diagnostics 12 02333 g007
Figure 8. Grad-CAM results for the following test dataset samples: (a) NVLES, (b) MNIST, (c) SKIN, and (d) NVLES.
Figure 8. Grad-CAM results for the following test dataset samples: (a) NVLES, (b) MNIST, (c) SKIN, and (d) NVLES.
Diagnostics 12 02333 g008
Table 1. Description of the study’s datasets.
Table 1. Description of the study’s datasets.
Dataset NameModality and DomainImage Type (Source)ObjectiveSystemClasses (Number of Images)
PET/CT
[21]
PET/CT scans, including SPNsDicomDistinguish between benign and malignant solitary pulmonary nodulesGeneral Electric Healthcare: Discovery iQ3 sl16Benign (61)
Malignant (111)
LIDC-IDRI
[29]
CT scans, including SPNsDicomDistinguish between benign and malignant solitary pulmonary nodulesVarious CT systemsBenign (620)
Malignant (616)
COV_X
[30]
X-ray scans of respiratory diseasesjpegDistinguish between COVID-19 and other diseasesVarious X-ray systemsPulmonary oedema (293)
Pleural effusion (311)
Obstructive pulmonary disease (315)
Pulmonary fibrosis (280)
COVID-19 (455)
Bacterial and viral pneumonia (910)
Normal cases (1341)
COV_CT
[31]
CT scans of respiratory diseasesDicomDistinguish between COVID-19 and non-COVID-19 casesVarious CT systemsCOVID-19 (349)
Normal (397)
Cells
[32]
Microscopic Images of cancerous cellstiffDistinguish between typesPrecipoint GmbH, Freising, GermanyEosinophil (424)
Typical lymphocyte (3937)
Monocyte (1789)
Myeloblast (3268)
Neutrophil (4419)
Cells2
[33]
Microscopic Images of cancerous cellsbmpDistinguish between typesNot availableMonocyte (20)
Eosinophil (88)
Lymphocyte (34)
Leukaemia
[34]
Microscopic images of cancerous cellsjpegHaematological malignancy detectionNot availableCancerous (8491)
Normal (4037)
MPIP
[35]
SPECT MPI polar maps of CADDicomDistinguish between healthy and coronary artery disease subjectsInfinia, Hawkey-4, GE HealthcareCAD (136)
Normal (80)
NV-LES
[36]
Skin images for skin cancer recognitionjpegDistinguish between benign nevus and cancerous melanomaVarious Benign (3000)
Cancerous (3000)
MNIST
[37]
Skin images for skin cancer recognitionjpegDistinguish among 5 skin abnormal classesVariousActinic keratoses (327)
Basal cell carcinoma (514)
Benign keratosis (1099)
Melanoma (1113)
Melanocytic nevi (6705)
SKIN
[38]
Skin images for skin cancer recognitionjpegDistinguish among 3 skin abnormal classesVariousBenign (1843)
Melanoma (753)
Table 2. DL networks employed in the present study.
Table 2. DL networks employed in the present study.
Network NameVersionYearReference
Virtual Geometry Group (VGG)VGG162015[17]
Feature-Fusion VGG19Baseline2021[35]
Residual Networkv.1512015[40]
Inceptionv.32015[41]
XceptionBaseline2017[42]
Densely Connected Networkv.1212018[43]
NasNetMobile version2018[44]
Mobile Networkv.22017[45]
Table 3. Experiment setups for evaluating the training from scratch against feature extraction via transfer learning methodologies.
Table 3. Experiment setups for evaluating the training from scratch against feature extraction via transfer learning methodologies.
Sub-ExperimentTrain DatasetTest DatasetNetworksTransfer Learning
1PET/CTPET/CTVGG19, FF-VGG19 Xception, ResNet50, Inception V3, DenseNet 121, MobileNet v2, NasNet mobileBy training from scratch or by using “off-the-shelf” features
2LIDC-IDRILIDC-IDRI
3COV_XCOV_X
4COV_CTCOV_CT
5CellsCells
6Cells2Cells2
7LeukaemiaLeukaemia
8MPIPMPIP
9NV-LESNV-LES
10MNISTMNIST
11SKINSKIN
Table 4. Experiment setups for evaluating the models’ robustness to image device variation.
Table 4. Experiment setups for evaluating the models’ robustness to image device variation.
ExperimentTrain DatasetTest DatasetLearning
1LIDC-IDRIPET/CTTraining from scratch
3MNISTNV-LESTraining from scratch
4NV-LESSKINTraining from scratch
5SKINMNISTTraining from scratch
Table 5. Classification accuracy (%) of employed CNNS, utilising off-the-shelf features. The table columns refer to the various networks employed for the classification. XCEPT refers to Xception, RES refers to ResNet, INCE refers to InceptionV3, DENSE refers to DenseNet, NAS refers to NasNet, and MOB refers to MobileNet. The evaluation is based on a 10-fold cross-validation.
Table 5. Classification accuracy (%) of employed CNNS, utilising off-the-shelf features. The table columns refer to the various networks employed for the classification. XCEPT refers to Xception, RES refers to ResNet, INCE refers to InceptionV3, DENSE refers to DenseNet, NAS refers to NasNet, and MOB refers to MobileNet. The evaluation is based on a 10-fold cross-validation.
CaseTrain DatasetTest DatasetVGG19
(%)
FF-VGG19
(%)
XCEPT
(%)
RES
(%)
INCE
(%)
DENSE
(%)
NAS
(%)
MOB
(%)
1PET/CTPET/CT72.0955.1948.7247.1255.2862.7155.2851.79
2LIDC-IDRILIDC-IDRI80.9784.8678.5750.1660.1071.9150.7264.97
3COV_XCOV_X79.9284.9644.6351.5758.5660.1052.9056.03
4COV_CTCOV_CT78.1679.3553.8952.0148.7939.9444.1040.34
5CellsCells86.7090.7434.5630.8824.5339.4928.5930.38
6Cells2Cells261.8041.1446.8040.5238.5724.9041.4725.28
7LeukemiaLeukemia72.4777.9267.7767.7364.4846.0361.1763.35
8MPIPMPIP63.2461.0151.8549.5345.3745.8346.7547.68
9NV-LESNV-LES78.4682.3656.9849.2556.4861.0349.0053.46
10MNISTMNIST74.5977.7465.7168.7162.9068.6554.8968.72
11SKINSKIN71.3875.5765.8965.5564.7563.1161.0162.94
Table 6. Classification metrics of the best CNN for each dataset when training with off-the-shelf features. Some metrics are not computed due to multiclass classification.
Table 6. Classification metrics of the best CNN for each dataset when training with off-the-shelf features. Some metrics are not computed due to multiclass classification.
CaseBest-PerformerAccuracy (%)Sensitivity (%)Specificity (%)AUCF1-Score
1VGG1972.0983.5260.9176.6474.61
2FF-VGG1984.8684.8782.9591.4784.25
3FF-VGG1984.96--96.70-
4FF-VGG1979.3585.6773.8087.6477.13
5FF-VGG1990.74--94.85-
6VGG1961.80--63.58-
7FF-VGG1977.9278.2560.2981.2377.27
8VGG1963.2486.7617.5083.3322.00
9FF-VGG1982.3685.1679.5693.1183.27
10FF-VGG1977.74--84.53-
11FF-VGG1975.5751.9787.9882.2882.48
Table 7. Classification accuracy of the CNNs when training from scratch. The table columns refer to the various networks employed for the classification. XCEPT refers to Xception, RES refers to ResNet, INCE refers to InceptionV3, DENSE refers to DenseNet, NAS refers to NasNet, and MOB refers to MobileNet. The evaluation is based on 10-fold cross-validation.
Table 7. Classification accuracy of the CNNs when training from scratch. The table columns refer to the various networks employed for the classification. XCEPT refers to Xception, RES refers to ResNet, INCE refers to InceptionV3, DENSE refers to DenseNet, NAS refers to NasNet, and MOB refers to MobileNet. The evaluation is based on 10-fold cross-validation.
CaseTrain DatasetTest DatasetVGG19
(%)
FF-VGG19
(%)
XCEPT
(%)
RES
(%)
INCE
(%)
DENSE
(%)
NAS
(%)
MOB
(%)
1PET/CTPET/CT51.7650.7269.7750.0046.7759.5853.9257.45
2LIDC-IDRILIDC-IDRI57.2182.5281.9678.4574.1077.5852.5074.35
3COV_XCOV_X52.1184.6460.6163.6377.3675.7261.8652.95
4COV_CTCOV_CT50.9378.5774.1670.2470.9167.2963.2771.98
5CellsCells47.6891.6189.8571.2072.7473.0520.2466.68
6Cells2Cells231.1947.2359.6629.8049.2846.9026.9548.00
7LeukemiaLeukemia46.5778.5878.8568.8662.1568.950.3750.91
8MPIPMPIP52.2766.0859.7256.9450.4658.3359.2554.62
9NV-LESNV-LES64.9683.6677.7570.2464.3876.0555.6657.21
10MNISTMNIST53.5277.9374.2846.2063.9173.3467.4070.38
11SKINSKIN52.5276.7161.9275.5671.9573.9965.7772.52
Table 8. Classification metrics of the best CNN for each dataset when training from scratch.
Table 8. Classification metrics of the best CNN for each dataset when training from scratch.
CaseBest-PerformerAccuracy (%)Sensitivity (%)Specificity (%)AUCF1-Score
1Xception69.7769.7665.7676.6460.93
2FF-VGG1982.5276.2988.7090.6780.80
3FF-VGG1984.94--95.25-
4FF-VGG1978.5787.9670.2783.4879.23
5FF-VGG1991.61--94.46-
6Xception59.66--69.46-
7Xception78.8587.1961.2583.860.5
8FF-VGG1966.0885.2933.7590.7039.26
9FF-VGG1983.6684.7082.6387.7684.24
10FF-VGG1977.93--84.58-
11FF-VGG1976.7147.3692.1383.9883.77
Table 9. FF-VGG19 and Xception network performance under medical image acquisition device variation.
Table 9. FF-VGG19 and Xception network performance under medical image acquisition device variation.
Train DatasetTest DatasetFF-GG19 (%)Xception (%)
LIDC-IDRIPET/CT74.4170.93
MNISTNVLES69.9363.92
SKINMNIST56.2585.25
NVLESSKIN61.6466.86
Table 10. Analytical metrics for the Xception network when evaluating on alternate acquisition device.
Table 10. Analytical metrics for the Xception network when evaluating on alternate acquisition device.
Train DatasetTest DatasetACCSENSPEF1AUC
LIDC-IDRIPET/CT70.930.840.5973.9583.02
MNISTNVLES63.920.850.5967.2879.39
SKINMNIST85.250.210.8692.0259.27
NVLESSKIN66.860.780.6779.6958.85
Table 11. Analytical metrics for the FF-VGG19 network when evaluating on alternate acquisition device.
Table 11. Analytical metrics for the FF-VGG19 network when evaluating on alternate acquisition device.
Train DatasetTest DatasetACCSENSPEF1AUC
LIDC-IDRIPET/CT74.410.860.6376.8484.54
MNISTNVLES69.930.980.4276.4988.61
SKINMNIST56.250.630.5568.3663.98
NVLESSKIN61.640.760.5464.7968.66
Table 12. The training dataset and the testing image of each visualised feature map group.
Table 12. The training dataset and the testing image of each visualised feature map group.
Training DatasetTesting Images
LIDC-IDRI + PETFrom LIDC-IDRI
COV_XFrom COV_X
COV_CTFrom COV_CT
CellsFrom Cells
MPIPFrom MPIP
MNISTFrom MNIST
NV-LESFrom NV-LES
MNIST + SKIN + NVLESFrom NVLES
MNIST + SKIN + NVLESFrom SKIN
MNIST + SKIN + NVLESFrom MNIST
LIDC-IDRI + PETFrom LIDC-IDRI
COV_XFrom COV_X
Table 13. The training dataset and the testing image of each visualised Grad-CAM.
Table 13. The training dataset and the testing image of each visualised Grad-CAM.
Training DatasetTesting Images
LIDC-IDRI + PETFrom LIDC-IDRI
COV_XFrom COV_X
COV_CTFrom COV_CT
CellsFrom Cells
MPIPFrom MPIP
MNISTFrom MNIST
NV-LESFrom NV-LES
SKINFrom SKIN
MNIST + SKIN + NVLESFrom NVLES
MNIST + SKIN + NVLESFrom SKIN
MNIST + SKIN + NVLESFrom MNIST
LIDC-IDRI + PETFrom LIDC-IDRI
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Apostolopoulos, I.D.; Papathanasiou, N.D.; Papandrianos, N.I.; Papageorgiou, E.I.; Panayiotakis, G.S. Deep Learning Assessment for Mining Important Medical Image Features of Various Modalities. Diagnostics 2022, 12, 2333. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12102333

AMA Style

Apostolopoulos ID, Papathanasiou ND, Papandrianos NI, Papageorgiou EI, Panayiotakis GS. Deep Learning Assessment for Mining Important Medical Image Features of Various Modalities. Diagnostics. 2022; 12(10):2333. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12102333

Chicago/Turabian Style

Apostolopoulos, Ioannis D., Nikolaos D. Papathanasiou, Nikolaos I. Papandrianos, Elpiniki I. Papageorgiou, and George S. Panayiotakis. 2022. "Deep Learning Assessment for Mining Important Medical Image Features of Various Modalities" Diagnostics 12, no. 10: 2333. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12102333

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop