Next Article in Journal
24 h Holter ECG Monitoring of Patients with Rheumatoid Arthritis—A Potential Role for a Precise Evaluation of QT Interval Duration and Associated Arrhythmic Complications
Next Article in Special Issue
Influence of the COVID-19 Pandemic on the Subjective Life Satisfaction of South Korean Adults: Bayesian Nomogram Approach
Previous Article in Journal
Applications of Optical Coherence Tomography in the Diagnosis of Enamel Defects
Previous Article in Special Issue
Automated Diagnosis of Cervical Intraepithelial Neoplasia in Histology Images via Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Differentiation of Atypical Parkinsonian Syndromes Using Brain Iron Patterns in Susceptibility Weighted Imaging

1
Department of Information Convergence Engineering, Pusan National University, Busan 46241, Korea
2
Department of Neurology, Pusan National University Yangsan Hospital, Pusan National University School of Medicine, Yangsan 50612, Korea
3
School of Computer Science and Engineering, Pusan National University, Busan 46241, Korea
*
Author to whom correspondence should be addressed.
Submission received: 1 February 2022 / Revised: 23 February 2022 / Accepted: 2 March 2022 / Published: 5 March 2022
(This article belongs to the Special Issue Artificial Intelligence Approaches for Medical Diagnostics in Korea)

Abstract

:
In recent studies, iron overload has been reported in atypical parkinsonian syndromes. The topographic patterns of iron distribution in deep brain nuclei vary by each subtype of parkinsonian syndrome, which is affected by underlying disease pathologies. In this study, we developed a novel framework that automatically analyzes the disease-specific patterns of iron accumulation using susceptibility weighted imaging (SWI). We constructed various machine learning models that can classify diseases using radiomic features extracted from SWI, representing distinctive iron distribution patterns for each disorder. Since radiomic features are sensitive to the region of interest, we used a combination of T1-weighted MRI and SWI to improve the segmentation of deep brain nuclei. Radiomics was applied to SWI from 34 patients with a parkinsonian variant of multiple system atrophy, 21 patients with cerebellar variant multiple system atrophy, 17 patients with progressive supranuclear palsy, and 56 patients with Parkinson’s disease. The machine learning classifiers that learn the radiomic features extracted from iron-reflected segmentation results produced an average area under receiver operating characteristic curve (AUC) of 0.8607 on the training data and 0.8489 on the testing data, which is superior to the conventional classifier with segmentation using only T1-weighted images. Our radiomic model based on the hybrid images is a promising tool for automatically differentiating atypical parkinsonian syndromes.

1. Introduction

In neurodegenerative disease, abnormal neuronal cells die rapidly in parts of the nervous system or the entire brain, resulting in loss of brain function, including cognitive and motor abilities. Parkinson’s disease (PD) is the second most common neurodegenerative disorder after Alzheimer’s and is accompanied by motor symptoms such as bradykinesia, tremor, and gait disturbance, making it difficult to conduct daily activities and many nonmotor symptoms such as cognitive impairment, depression, autonomic dysfunction, and sleep disturbance. Atypical parkinsonian syndromes (APSs), comprising of progressive supranuclear palsy (PSP) and a parkinsonian variant of multiple system atrophy (MSA-P), are degenerative diseases that share similar Parkinsonism symptoms and signs with PD [1] but show additional symptoms and different rates of functional deterioration and prognosis [2]. Therefore, the development of methods for distinguishing between PD and APS has clinical significance.
One of the main pathogenesis of PD is iron accumulation in the substantia nigra area of the brain associated with the degeneration of dopaminergic neurons and accumulation of misfolded proteins [3]. According to recent pathological studies, each parkinsonian syndrome has unique topographic patterns of iron distribution in deep brain nuclei, which are influenced by underlying disease pathologies [4,5].
There have been many studies using advanced magnetic resonance images (MRI) to detect the physiological mechanisms underlying PD and to distinguish APS from PD, such as using resting-state functional MRI (fMRI) [6] or diffusion MRI [7], but these approaches are not easy to apply in general clinical practice because they are time consuming and do not guarantee consistent results [8]. In addition, various studies using other modalities, including PET and SPECT, can achieve significant diagnostic relevance with respect to imaging of PD and APS [9,10]. However, there are also some disadvantages of these modalities, such as radiation exposure of CT [11], and obstacles to the clinical application of PET by limited access and high examination costs [12]. These common advanced neuroimaging techniques are summarized in Table 1.
Susceptibility weighted imaging (SWI), a type of iron-sensitive MRI, is frequently used to detect disease-specific patterns of uneven and localized iron concentration in brain regions [13]. Figure 1 shows the sample SWI axial slices of the MSA-P, MSA-C, PSP, and PD. Increases in iron-related signals in the anterior and medial aspects of the globus pallidus of SWI are highly specific markers of PSP. For MSA-P, a significant accumulation of iron is present in the lateral aspect of the globus pallidus adjacent to the putamen. In addition, the posterolateral putaminal hypointensity and lateral-to-medial gradient appear consistently in MSA-P SWI [14]. However, assessing the putaminal hypointensity by focusing only on the signal intensity without accounting the distributional pattern fails to differentiate between MSA-P from PD [15]. A generic and age-related sign of physiological mineralization is slit-like hypointensity along the lateral margin of the putamen or evenly distributed hypointensity throughout the putamen [16]. Therefore, finding a distinctive pattern that distinguishes parkinsonian syndromes besides nonspecific and age-related signs is challenging.
To analyze the regional iron heterogeneity in deep brain nuclei without an expert radiologist, radiomic features provide advanced quantification and classification methodologies based on machine learning algorithms. Radiomics can extract textural features that express the relationship with neighboring voxels, allowing us to analyze the regional iron deposition in the subcortical structures. It is suitable for SWI, where the signal itself cannot be used because the SWI intensities of non-paramagnetic materials, such as white matter (WM) and cerebrospinal fluid (CSF), are modified through the filtered phase mask to emphasize the susceptibility in the image. There is considerable interest in the potential of radiomics for non-invasive biomarkers in different organs and pathologies, including neurodegenerative diseases [17].
Since radiomics is sensitive to changes in image intensities, accurate and robust segmentation of deep gray matter (DGM) nuclei is required. Although manually viewing the image and judging the lesion or progression is highly accurate when performed by an expert radiologist, it has the disadvantages of high time consumption and monetary costs to diagnose large numbers of patients. To overcome these problems, several automated segmentation tools based on T1-weighted (T1w) images have been developed including FreeSurfer [18,19], FMRIB software library (FSL) integrated registration and segmentation tool (FIRST) [20], and others [21]. These techniques have been applied in multiple brain imaging studies for examining volume and shape changes in subcortical brain regions that may be linked to normal aging or neurodegenerative disorders. DGM segmentation is scan–rescan reliable on the same scanning platform and between separate scanning platforms, indicating that these tools may be used in large-scale longitudinal and multisite studies [22,23]. However, if only T1w images are used as atlas-based tools, the segmentation results tend to be inaccurate [24] and do not represent the patient’s hallmarks, because the spatial correspondence of subcortical structures between an abnormal brain and standard atlas is poor and the contrast of DGM in T1w images is insufficient [25]. Therefore, it is necessary to develop a segmentation method that better reflects the distinctive features of each disease using a modality other than T1w.
In this paper, we propose a novel framework that uses the SWI to automatically analyzed the disease-specific patterns of iron accumulation. Our contributions to this study are listed below:
  • We proposed a fully automatic framework for the analysis of iron deposition patterns in SWI.
  • We developed segmentation that reflects more the contrast of iron accumulation than conventional methods using a hybrid contrast image, which is created by image processing and combining T1w and SWI.
  • We designed machine learning classifiers trained using texture-representing features extracted by our segmentation method.
  • We demonstrated the improved performance of the machine learning classifier for differentiating APS using our segmentation framework.
The remainder of the paper is organized as follows. In Section 2, we propose an automated framework for SWI segmentation and the radiomic learning model, including hybrid image generation, DGM segmentation, radiomic feature extraction and selection, and machine learning classifier validation. Experimental results are presented in Section 3, wherein the proposed algorithm is validated using the patient datasets. Finally, the main conclusions and discussions are made in Section 4.

2. Materials and Methods

In this section, we describe the details of our framework that automatically differentiate APS using brain iron patterns in SWI. Figure 2 presents the overall framework of the proposed method. First, a DGM mask using the advantages of both T1w and SWI was obtained by optimally combining preprocessed and registered images. The radiomic features were retrieved from the brain regions of interest (ROI) by adjusting the distance between the neighboring voxels. Thereafter, a machine learning feature selection algorithm was applied to select meaningful features that distinguish the diseases. Finally, various machine learning classifiers were trained and tested using the selected features.

2.1. Patients

A total of 34 MSA-P, 21 MSA-C, 17 PSP, and 56 PD patients were enrolled from the Pusan University Yangsan Hospital. The following clinical diagnostic criteria were fulfilled by the patients: PSP diagnosed according to the Litvan criteria [26], MSA according to clinical consensus criteria [27], and PD according to the UK Brain Bank criteria [28]. Movement Disorder Society (MDS) PSP criteria were retrospectively applied to all consecutive patients with PSP. Twelve patients were classified as probable PSP Richardson’s syndrome (PSP-RS) and five were classified as probable PSP with predominant Parkinsonism (PSP-P). Subjects with microvascular lesions discovered from brain MRI were excluded. The Hoehn and Yahr (H&Y) stage and motor examination part of the Unified Parkinson’s Disease Rating Scale (UPDRS III) were used to measure disease severity and motor symptoms. Written and informed consent was obtained from all subjects participating in the study, which was approved by the Pusan National University Institutional Review Board, in accordance with the guidelines of the Helsinki Declaration.

2.2. Imaging Acquisition

We obtained the 3D magnetization prepared rapid gradient echo (MPRAGE) axial or sagittal T1w and SWI MRI volumes of 34 MSA-P, 21 cerebellar variants of MSA (MSA-C), 17 PSP, and 56 PD patients from the Pusan National University Yangsan Hospital using protocols approved by the institutional review board. The MRI scans were conducted using a 3.0T MRI scanner (Verio, Siemens, Erlangen, Germany). The T1w data were acquired under the following sequence parameters: echo time (TE) = 2.2 ms, repetition time (TR) = 1900 ms, inversion time (TI) = 900 ms, flip angle (FA) = 9, dimensions = 280 × 320 × 176, and voxel size = 0.75 mm × 0.75 mm × 1 mm. The SWI data were acquired under the following sequence parameters: TE = 20 ms, TR = 28 ms, FA = 15, dimensions = 260 × 320 × 64, and voxel size = 0.6875 mm × 0.6875 mm × 2 mm.

2.3. Data Preprocessing and SWI Registration

We performed SWI postprocessing as the first step. Magnitude, high-pass filtered phase images, and the processed SWI data were reconstructed automatically on a workstation (Syngo, Siemens Medical Solution) as a DICOM file format for analysis. Then, we created an initial segmentation mask for T1w to use when creating HC through FreeSurfer reconstruction. Non-parametric non-uniform and intensity normalization (N4ITK) bias-field correction [29] and intensity normalization were applied. We applied intensity normalization to scale the T1w signal intensity to a predefined mean value of 110 in the white matter (WM).
Subsequently, the SWI images were registered to the T1w images using affine transform. Since the T1w and the SWI images of the same subject have identical anatomy and head motion between scans, the two images were successfully aligned using an affine registration. These data preprocessing are the steps before calculating weights, combining the steps shown in Figure 3.

2.4. SWI Segmentation Using Hybrid Contrast Image

To obtain segmentation results reflecting iron-related signals, we used both the T1w and SWI images simultaneously and merged them into a single hybrid contrast (HC) image [30]. Since SWI provides superior contrast for iron-rich structures, while the T1w images have greater contrast in the curvature of complicated gyrus and sulcus principally used for registration, using the HC results in the DGM segmentation that reflects more iron contents than using T1w alone, which better reflects the disease’s hallmarks such as nuclei atrophy [31] caused by the iron deposition.
The HC image is defined by linearly combining T1w and SWI images:
HC = w 1 · T 1 w + w 2 · SWI ,
where w 1 and w 2 are weighting coefficients for T1w and SWI, respectively.
We adjusted the weighting coefficients w 1 and w 2 to make HC as close as possible to the reference, Montreal Neurological Institute (MNI) template. We employed the MNI template’s contrast as the target for the coefficient optimization because it has a typical T1w contrast with outstanding DGM structural delineation. The optimized values of the weighting coefficients w 1 * , w 2 * can be obtained by minimizing the squared difference of the mean signal intensities in the target brain regions between the HC and MNI template:
( w 1 * , w 2 * ) = arg min w 1 , w 2 I put T 1 w I put SWI I pall T 1 w I pall SWI w 1 w 2 I put MNI I pall MNI 2 2 ,
where I put T 1 w , I put SWI , and I put MNI are the mean values of the T1w, SWI, and MNI template images in the putamen region, respectively, and I pall T 1 w , I pall SWI , and I pall MNI are the mean values of the T1w, SWI, and MNI template images in the globus pallidus region, respectively. We chose the putamen and globus pallidus for the target regions because of high-contrast signals in the broad areas.
Then, we used advanced normalization tools (ANTs) to register the HC to the MNI template by computing an initial affine registration and non-linear registration employing a non-rigid diffeomorphic registration scheme [32]. The ANTs produced the most consistent and reliable registration results among 14 different registration methods [33]. The segmentation results from the MNI space were inversely warped to the individual T1w image space. The overall procedure of SWI segmentation is shown in Figure 3.

2.5. Feature Extraction and Selection

Radiomic features were extracted from the segmented DGM region of the SWI images automatically computed in Section 2.4. The radiomic features included 19 first-order statistical features, 10 2D shape-based features, 16 3D shape-based features and the following texture-based features: 72 gray-level co-occurrence matrix (GLCM) features, 16 gray-level run length matrix (GLRLM) features, 16 gray-level size zone matrix (GLSZM) features, 15 neighboring gray-tone difference matrix (NGTDM) features, and 14 gray-level dependence matrix (GLDM) features [34], as shown in Figure 2. These matrices represent the relationship with the surrounding voxels according to the kernel for each voxel. For example, the ( i , j ) th element of the GLCM represents the number of times the combination of levels i and j occur in two voxels in the image, which are separated by a distance of δ pixels along the angle θ .
We added GLCM and NGTDM features while changing the distance to neighboring voxels for which the relationship was calculated as four and seven voxels to the default python radiomic package [35]. Since SWI does not provide quantitative measurements of susceptibility, we excluded the signal-based features and focused on the texture-based features. We subtracted the signal-based features such as the minimum, maximum, mean, median, 10th percentile, 90th percentile of intensity, gray level range, and others. We use only selected optimal features (see below).
Among the sub-cortical structures in the DGM, we chose the putamen to extract radiomic features because comparing them is easy owing to the putamen’s large size and high contrast. It shows a large difference between the mask segmented from the T1w-only and the proposed methods. In addition, the radiomic results extracted from the putamen showed the best performance in disease classification using machine learning [8].
Next, to avoid overfitting the learning model, feature selection was performed before applying machine learning algorithms [36]. We employed the Fisher score algorithm to rank the radiomic features and a filter-based method for supervised feature selection. It chooses each feature independently according to its scores based on the Fisher criterion. We selected the top-10 ranked features based on Fisher score. We finally applied these selected features to classify the data using machine learning.

2.6. Machine Learning Classifier Training and Testing

To distinguish between subtypes of parkinsonian syndromes, we used the 10 most popular machine learning classifiers [37] such as k-nearest neighbors (kNN) [38], linear support vector machine classifier (linSVC) [39], support vector machine with radial basis function (RBF) kernel classifier (RBFSVC) [40], Gaussian process classifier (GP) [41], random forest classifier (RF) [42], decision tree classifier (DT) [43], multi-layer perceptron classifier (MLP) [44], AdaBoost classifier (ADA) [45], Gaussian Naïve Bayes classifier (GNB) [46], and quadratic discriminant analysis classifier (QDA) [47]. These classifiers have the potential for radiomics to aid in the development of non-invasive biomarkers [48].
The total datasets were divided into training and testing sets at a 7:3 ratio. In the training sets, features were selected, and 10 classifiers were constructed with 3-fold cross-validation. To evaluate the performance of the classifiers for differentiation of APS, the area under receiver operating characteristic curve (AUC), balanced accuracy (bAcc), sensitivity (Sen), specificity (Spe), and accuracy (Acc) were measured as defined by:
Sen = TP TP + FN
Spe = TN TN + FP
bAcc = Sen + Spe 2
Acc = TP + TN TP + TN + FP + FN
where TP denotes the number of the actual positives that are correctly classified as positives, FN denotes the number of the actual positives that are wrongly classified as negatives, TN denotes the number of the actual negatives that are correctly classified as negatives, and FP denotes the number of the actual negatives that are wrongly classified as positives. The AUC metric is defined as the area under the receiver operating characteristic (ROC) curve plotted by true positive rate (TPR, equivalent to sensitivity) against false positive rate (FPR, equivalent to 1 specificity ) with varying thresholds. For statistical evaluation, the performance metrics were obtained by randomly changing the training and testing sets 100 times and averaged. The source code is available in GitHub: https://github.com/KimYunSoo/classify_radiomic (accessed on 22 January 2022).

3. Results

3.1. Demographic Characteristics

The demographic and clinical characteristics of the subject groups are listed in Table 2. There were no significant differences between subject groups in terms of gender distribution. Age was higher in the PSP group than other groups. There was no discernible difference in disease duration between MSA-P, MSA-C, PSP, and PD. The disease severity measured using the UPDRS and H&Y scores was greater in the PSP and MSA groups than the PD group, and MMSE was lower in the PSP and MSA groups than the PD group (p < 0.001).

3.2. SWI Segmentation Results

Figure 4 shows an example of axial slices around the DGM area in the T1w, SWI, and HC images. The DGM contrast is weak and the cortex contrast is clear in the T1w, while the trend is opposite for SWI. Whereas, the HC shows high contrast clearly for both the DGM and cortex. Figure 5 shows that the proposed approach produces segmentation results that better represent hypointensity indicating iron concentration in putamen SWI images. HC segmentation masks that use both T1w and SWI simultaneously reflect more hallmarks of parkinsonian disorders, such as iron accumulation and the resulting putamen atrophy, than T1w-only masks.

3.3. Feature Extraction and Selection Results

Table 3 shows the 10 most significant features selected from SWI for the differentiation of MSA-P and PD when using HC and T1w-only (by FreeSurfer, FS) segmentation masks and their mean values. Autocorrelation7, SumAverage4, JointAverage4, SumAverage7, JointAverage7, and Imc24 in GLCM and HighGrayLevelEmphasis in GLDM were commonly selected both in HC and T1w-only segmentation. Imc24 is the correlation between the probability distribution of intensity and occurrence number, quantifying the complexity of the texture, by neighboring voxel distances of 4. JointAverage7 and SumAverage7 (JointAverage4 and SumAverage4) measure the relationship between occurrences of pairs by neighboring voxel distances of 7 (4, respectively) with lower or higher intensity values. These indicate that the number of pairs of lower or higher intensities helps to differentiate between diseases. HighGrayLevelEmphasis in GLDM measures the distribution of the higher gray-level values with a higher value indicating a greater concentration of high gray-level values in the volume. In addition, in the case of comparison with other disease groups as shown in Table A1, Table A2, Table A3, Table A4 and Table A5, ClusterShade4 and MCC4 were also found to be common in HC and T1w-only. ClusterShade4 is a metric of the skewness and uniformity of the GLCM by neighboring voxel distances of 4 [49]. MCC4 is the maximal correlation coefficient for nearby voxel distances of 4, which also assesses the complexity of the texture. These features represent how dependent and uniform the distributions are.
The significant features only selected using HC mask include Autocorrelation4 in GLCM and HighGrayLevelRunEmphasis in GLRLM. Autocorrelation4 quantifies the magnitude of texture coarseness by neighboring voxel distances of 4; therefore, it operates more effectively in the HC segmentation mask as clusters of similar intensities appear better in HC than in T1w-only mask, which includes regions that are not iron-deposited. HighGrayLevelRunEmphasis in GLRLM measures the distribution of the higher gray-level values. RunEntropy and ShortRunHighGrayLevelEmphasis in GLRLM are also common when using HC masks in other disease group comparisons. RunEntropy is a metric that evaluates the uncertainty and randomness in the distribution of run lengths and gray levels. Therefore, heterogeneity in the texture patterns measure by RunEntropy is helpful in classifying each disorder. ShortRunHighGrayLevelEmphasis assesses the distribution of the high gray-level values and their joint distribution with shorter run lengths in GLRLM. The feature indicates how concentrated hyperintensities in SWI are, which is significant for distinguishing each subtype of parkinsonian disorder.
The significant features selected using T1w-only segmentation include GrayLevelNonUniformity in GLRLM and DependenceVariance in GLDM. GrayLevelNonUniformity is a metric that compares the similarity of the SWI image’s gray-level intensity values. The variance in dependence size in the image is measured by DependenceVariance. Moreover, in other disorder comparison cases, LargeDependenceHighGrayLevelEmphasis in the GLDM and Strength in NGTDM were frequently selected features using the T1w-only mask. LargeDependenceHighGrayLevelEmphasis in GLDM is the metric of joint distribution of substantial reliance on it. Strength in NGTDM measures how easily defined and visible the primitives in the image are. These all work mainly in the T1w-only mask, where there are both hypo- and hyper-intensity clusters together, because the T1w-only mask is likely to include the region without iron deposition (see Figure 5).

3.4. SVM Results

Table 4 lists the training and testing area under the receiver operating characteristic curve (AUC) of the RBF SVM classifier employing features from the T1w-only and HC masks. The SVM with RBF kernel that learns the radiomic features extracted from iron-reflected segmentation results produced an average AUC of 0.8607 in training and 0.8489 in testing. T1w-only mask-based radiomic training classifiers had an average AUC of 0.7570 in training and 0.7866 in testing. The classifier model trained with features extracted using the HC mask shows better performance than the T1w-only mask-based SVM classifier.
The RBF SVM classifier receiver operating characteristic (ROC) curves for each disease distinguishing case are shown in Figure A1, Figure A2, Figure A3, Figure A4, Figure A5 and Figure A6. Through other classification algorithms, it was confirmed that the performance of the proposed method is improved compared to the T1w-only method in the same way as the RBF SVM.
The balanced accuracy, sensitivity, and specificity of the RBF SVM classifier using features from T1w-only masks and HC masks are listed in Table 5. The machine learning classifier that learns the SWI-reflected radiomic features produced an average balanced accuracy of 0.7666 for the training cohort and 0.7992 for the testing cohort. The classifier model trained by radiomics extracted from T1w-only segmentation masks achieved 0.6557 in training and 0.7620 in testing.
The classifier that was trained on the radiomic features extracted by the proposed method achieved an average accuracy of 0.8000 in training and 0.8059 in testing, as shown in Table 6. Conventional T1w-only segmentation classifiers had an accuracy of 0.7352 in training and 0.7653 in testing.
The AUC, balanced accuracy, sensitivity, specificity, and accuracy of all other classifiers are listed in Table A6, Table A7, Table A8, Table A9, Table A10, Table A11, Table A12, Table A13, Table A14, Table A15, Table A16, Table A17, Table A18, Table A19, Table A20, Table A21, Table A22, Table A23, Table A24, Table A25, Table A26, Table A27, Table A28, Table A29, Table A30, Table A31 and Table A32. Similar to the RBF SVM, in other classifier models, the AUC, balanced accuracy, and accuracy increased when HC masks reflecting iron-related signal were used.

4. Discussion and Conclusions

In this paper, we proposed a novel framework that automatically analyzes the disease-specific patterns of iron deposition using SWI. Through this proposed framework, by directly inputting raw data, the results of disease classification by automated processing without any human intervention can be applied to diagnosis.
Atypical Parkinsonian syndromes, such as MSA-P, MSA-C, and PSP, can be mistaken for PD, especially in the early stages of the disease. This is because both APS and PD are present with Parkinsonism. Therefore, it is critical to distinguish between PD and APS; nevertheless, conventional MRI still makes it difficult to discriminate between these neurodegenerative disorders.
We demonstrated that in individuals with abnormal brain anatomy, the commonly used T1w-only segmentation pipeline produces erroneous subcortical segmentation. The goal of this study was to overcome this issue by modifying the conventional pipeline that incorporates nonlinear registration and by using a dedicated hybrid image contrast created by combining standard T1w images with SWI. By using the HC, which is a combination of the T1w and SWI, for the DGM segmentation, it is possible to identify iron deposition automatically without manual segmentation by expert radiologists, as was done in the past. We have visually shown that putamen segmentation performance was improved by using both the T1w and SWI.
We conducted a qualitative assessment of the visual delineation of our segmentation framework results. If there is a manual segmentation mask by an expert, it can be used as the gold standard, and objective and quantitative evaluation can be performed through metrics such as the dice coefficient. However, manual segmentation performed by experts is costly and time consuming. Some studies have used visual ratings as metrics [50].
Another goal of the present study was to a create machine learning classifier that can distinguish APS from PD using image texture-based features derived from basal nuclei on SWI. Different iron deposition patterns for each disease were compared by extracting quantified radiomic features. The distinction between each subtype of parkinsonian disorder groups was better exposed by the features retrieved with the SWI-reflected mask. When classifying diseases using various machine learning algorithms, it was confirmed that the performance of the classifier improved by training features extracted from the HC.
We recognize the lack of pathological confirmation for diagnosis and phenotypic categorization, which remain the gold standard for the diagnosis of PSP. However, we selected patients with the typical clinical characteristics of MSA, PSP and PD, and assessed these patients over several years.
We used the texture features of the signal intensity contrast to train the machine learning classifiers. Since SWI does not represent a quantified value of iron content, the quantitative values of iron deposition were not measured. We used only texture features because we intended to classify disorders by analyzing the image patterns of each disease and not to create a reference point or threshold with a quantified number. Although we did not directly compare the quantitative values, we indirectly demonstrated the improvement of segmentation through outperforming the machine learning classifier.
In future work, we will validate the proposed framework more clinically using R2*. In addition, we will aim to apply our hybrid approach of brain tissue segmentation in other PET-MRI modalities.

Author Contributions

Conceptualization, J.K.G. and J.-H.L.; methodology, J.K.G. and Y.S.K.; software, Y.S.K.; validation, Y.S.K.; formal analysis, Y.S.K.; investigation, J.K.G.; resources, J.K.G. and J.-H.L.; data curation, J.-H.L.; writing—original draft preparation, Y.S.K.; writing—review and editing, J.K.G. and J.-H.L.; visualization, Y.S.K.; supervision, J.K.G.; project administration, J.K.G.; funding acquisition, J.K.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Institute of Information communications Technology Planning Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2020-0-01450, Artificial Intelligence Convergence Research Center [Pusan National University]); and in part by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. NRF-2020R1C1C1008362).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of Pusan National University Yangsan Hospital (protocol code 05-2014-089 and 10302014).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A

Figure A1, Figure A2, Figure A3, Figure A4, Figure A5 and Figure A6 show the receiver operating characteristic (ROC) curves for each differentiating disease case. As in the RBF SVM, in other algorithms, the classifiers that learned the features extracted using HC masks performed better overall in terms of the AUC, balanced accuracy, and accuracy compared to the models trained using the conventional T1w-only masks.
Figure A1. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-P vs. MSA-C.
Figure A1. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-P vs. MSA-C.
Diagnostics 12 00637 g0a1
Figure A2. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-P vs. PD.
Figure A2. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-P vs. PD.
Diagnostics 12 00637 g0a2
Figure A3. Receiver-operating characteristic (ROC) curves of the RBF SVM classifier for MSA-P vs. PSP.
Figure A3. Receiver-operating characteristic (ROC) curves of the RBF SVM classifier for MSA-P vs. PSP.
Diagnostics 12 00637 g0a3
Figure A4. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-C vs. PD.
Figure A4. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-C vs. PD.
Diagnostics 12 00637 g0a4
Figure A5. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-C vs. PSP.
Figure A5. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for MSA-C vs. PSP.
Diagnostics 12 00637 g0a5
Figure A6. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for PD vs. PSP.
Figure A6. Receiver operating characteristic (ROC) curves of the RBF SVM classifier for PD vs. PSP.
Diagnostics 12 00637 g0a6

Appendix B

Appendix B.1

Table A1 list the mean values of the features with HC and T1w-only masks when comparing MSA-P and MSA-C.
Table A1. Mean values of the features of each subtype of disease.
Table A1. Mean values of the features of each subtype of disease.
HC FeaturesMSA-PMSA-CT1w-Only FeaturesMSA-PMSA-C
glszm_GrayLevelVariance6.86454.0758glcm_Id40.44730.5428
glszm_HighGrayLevel-
ZoneEmphasis
70.98140.6072glcm_ClusterTendency78.49453.201
ngtdm_Strength0.25150.0898ngtdm_Strength0.24090.085
ngtdm_Strength40.14090.0534gldm_SmallDependenceEmphasis0.06380.0331
ngtdm_Strength70.12230.0495gldm_SmallDependence-
LowGrayLevelEmphasis
0.00210.0016
glrlm_ShortRunHigh-
GrayLevelEmphasis
51.229124.4967glcm_ClusterShade−61.4694−7.7083
glcm_Imc240.54420.3803glcm_SumSquares46.38071.9578
glcm_JointAverage78.13535.728glcm_Idm40.37880.4915
glcm_SumAverage716.270611.4561glcm_ClusterShade4−12.6393−2.0909
glszm_SmallAreaHigh-
GrayLevelEmphasis
34.835518.1394gldm_DependenceVariance23.220527.458

Appendix B.2

Table A2 lists the mean values of the features with HC and T1w-only masks when comparing MSA-P and PSP.
Table A2. Mean values of the features of each subtype of disease.
Table A2. Mean values of the features of each subtype of disease.
HC FeaturesMSA-PPDT1w-Only FeaturesMSA-PPD
glszm_ZoneVariance10,415.416,002.22ngtdm_Busyness46.235513.1834
ngtdm_Busyness3.28686.393glcm_DifferenceAverage72.69581.9655
glcm_JointAverage47.7625.2133gldm_GrayLevelNonUniformity556.3621736.7501
glcm_ClusterShade4−11.6021−0.0316gldm_LargeDependence-
HighGrayLevelEmphasis
7484.0963824.813
gldm_SmallDependence-
LowGrayLevelEmphasis
0.00270.0028ngtdm_Strength0.24090.1118
gldm_GrayLevelNonUniformity438.2295550.6763gldm_HighGrayLevelEmphasis75.700538.6685
gldm_DependenceNonUniformity186.9766175.5947glcm_Idn40.85950.8683
glcm_Imc14−0.074−0.0451glcm_DifferenceVariance44.83922.0783
glcm_MCC40.38230.2921glrlm_GrayLevelNonUniformity344.9109470.024
glcm_Autocorrelation463.634928.8556glszm_HighGrayLevel-
ZoneEmphasis
75.158847.8979

Appendix B.3

Table A3 lists the mean values of the features with HC and T1w-only masks when comparing MSA-C and PD.
Table A3. Mean values of the features of each subtype of disease.
Table A3. Mean values of the features of each subtype of disease.
HC FeaturesMSA-CPDT1w-Only FeaturesMSA-CPD
glcm_ClusterShade−7.7694−1.3529glcm_ClusterShade4−2.09090.5222
glcm_ClusterShade4−2.4839−0.4058glcm_ClusterShade−7.7083−0.0626
glcm_MCC40.27420.2319glcm_MCC40.31670.2744
glcm_Imc14−0.0384−0.0291glcm_JointAverage75.89465.5243
glcm_Imc240.38030.3244glcm_ClusterShade7−1.1194−0.1304
glrlm_RunEntropy3.89553.7793gldm_DependenceVariance27.45827.8954
glcm_ClusterShade7−1.1463−0.3319glcm_Imc240.4220.364
gldm_DependenceEntropy6.4996.3396glcm_Imc1−0.2039−0.1892
glrlm_GrayLevelNon-
UniformityNormalized
0.20820.2397gldm_DependenceEntropy6.64756.518
glcm_SumEntropy3.18022.9418glcm_MCC0.66020.6362

Appendix B.4

Table A4 lists the mean values of the features with HC and T1w-only masks when comparing MSA-C and PSP.
Table A4. Mean values of the features of each subtype of disease.
Table A4. Mean values of the features of each subtype of disease.
HC FeaturesMSA-PPDT1w-Only FeaturesMSA-PPD
glcm_MCC0.61030.6006gldm_DependenceVariance27.45821.3893
glrlm_RunEntropy3.89553.8289gldm_DependenceNon-
UniformityNormalized
0.05550.0642
glcm_JointAverage75.7285.4607ngtdm_Coarseness0.00240.0028
glcm_Imc240.38030.4196glrlm_RunEntropy4.00754.0395
glcm_MCC40.27420.2921glszm_LargeAreaHigh-
GrayLevelEmphasis
1,495,719946,080.3
gldm_LargeDependence-
LowGrayLevelEmphasis
6.0856.0346glszm_LargeAreaLow-
GrayLevelEmphasis
1948.3491398.383
gldm_DependenceNonUniformity195.5685175.5947glszm_LowGrayLevelZoneEmphasis0.06730.0777
glszm_ZoneVariance31,931.7316,002.22gldm_LargeDependenceEmphasis138.6942109.827
glszm_SmallAreaLow-
GrayLevelEmphasis
0.02750.0305gldm_GrayLevelVariance2.14462.9733
glszm_ZoneEntropy5.07885.0619gldm_SmallDependence-
LowGrayLevelEmphasis
0.00160.002

Appendix B.5

Table A5 lists the mean values of the features with HC and T1w-only masks when comparing PD and PSP.
Table A5. Mean values of the features of each subtype of disease.
Table A5. Mean values of the features of each subtype of disease.
HC FeaturesMSA-PPDT1w-Only FeaturesMSA-PPD
glcm_Autocorrelation727.999831.5698glcm_SumEntropy42.79133.1555
glcm_Contrast72.76994.5716glcm_SumAverage711.048711.8343
gldm_LargeDependenceHigh-
GrayLevelEmphasis
3796.0272722.489gldm_HighGrayLevelEmphasis32.952138.6685
glrlm_RunEntropy3.77933.8289gldm_LargeDependence-
HighGrayLevelEmphasis
4652.9883824.813
glcm_DifferenceAverage41.12381.4629gldm_LowGrayLevelEmphasis0.04740.0488
gldm_DependenceVariance27.564719.9317glszm_ZonePercentage0.01920.0266
glcm_ClusterProminence429.3774.0809glcm_JointEnergy40.07020.0402
glcm_JointAverage45.04435.2133glcm_ClusterShade-0.06260.2446
glcm_Imc240.32440.4196glcm_DifferenceEntropy42.0072.306
glrlm_RunPercentage0.62470.6961glszm_SizeZoneNonUniformity17.044819.3692

Appendix C

Appendix C.1

Table A6, Table A7 and Table A8 list the results of the classifier trained with k-nearest neighbor (kNN).
Table A6. kNN classifier training and testing AUC when using HC and T1w-only segmentation masks.
Table A6. kNN classifier training and testing AUC when using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.85890.76770.86210.7910
MSA-P vs. PD0.88390.82030.88650.8340
MSA-P vs. PSP0.83570.83560.85690.8272
MSA-C vs. PD0.68700.66130.68050.6613
MSA-C vs. PSP0.79080.78550.79320.7813
PD vs. PSP0.88950.73230.87610.8369
Table A7. kNN classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A7. kNN classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.79150.85810.72500.68430.78890.57970.79060.85420.72690.71080.80540.6161
MSA-P vs. PD0.87750.86820.88680.80570.77110.84020.85690.82590.88790.79240.72260.8622
MSA-P vs. PSP0.77910.85740.70090.77120.84030.70200.81010.88030.73990.78670.88600.6874
MSA-C vs. PD0.65130.49440.80820.66040.52920.79150.67790.56170.79490.63470.47760.7918
MSA-C vs. PSP0.71960.73600.70320.67410.72370.62460.72420.73800.71030.69730.73050.6642
PD vs. PSP0.83150.90330.75970.61470.81780.41160.80940.89810.72070.73300.87310.5928
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A8. kNN classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A8. kNN classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.79550.69700.80060.7333
MSA-P vs. PD0.87890.81630.86420.8095
MSA-P vs. PSP0.80550.79290.82770.8123
MSA-C vs. PD0.73870.73790.74200.722
MSA-C vs. PSP0.70700.66440.71670.7016
PD vs. PSP0.86670.73240.85690.8116

Appendix C.2

Comparison of the linear support vector machine (linSVM) classifier is given in Table A9, Table A10 and Table A11.
Table A9. linSVM classifier training and testing AUC when using HC and T1w-only segmentation masks.
Table A9. linSVM classifier training and testing AUC when using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.88090.87900.87030.8631
MSA-P vs. PD0.91590.89020.91560.8799
MSA-P vs. PSP0.88400.88820.89280.8821
MSA-C vs. PD0.73140.70970.74080.7261
MSA-C vs. PSP0.96940.93490.93810.9300
PD vs. PSP0.94330.82320.93460.8294
Table A10. linSVM classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A10. linSVM classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.79950.86930.72970.75400.89070.61730.78320.85620.71020.77000.89620.6437
MSA-P vs. PD0.89980.91780.88180.85930.85240.86620.89640.90750.88540.84950.83830.8607
MSA-P vs. PSP0.77200.87050.67350.74590.83760.65420.79040.88240.69830.77330.88340.6633
MSA-C vs. PD0.78300.77370.79230.78990.81110.76870.82120.87530.76700.78540.80260.7682
MSA-C vs. PSP0.90380.89870.90890.82100.81660.82540.85730.85520.85940.85000.83550.8645
PD vs. PSP0.82910.91080.74750.67470.83290.51660.83140.91400.74890.73770.84810.6272
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A11. linSVM classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A11. linSVM classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.80520.78900.79210.7738
MSA-P vs. PD0.89260.86190.88920.8392
MSA-P vs. PSP0.80020.77320.81610.8019
MSA-C vs. PD0.78830.76970.76910.7670
MSA-C vs. PSP0.89070.81250.84190.8340
PD vs. PSP0.86380.77810.87420.8127

Appendix C.3

Table A12, Table A13 and Table A14 list the results of the Gaussian process (GP) based classifier.
Table A12. GP classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A12. GP classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.74200.69350.85290.6632
MSA-P vs. PD0.92430.87530.91410.8790
MSA-P vs. PSP0.88800.82770.89360.8735
MSA-C vs. PD0.70180.68930.71850.6957
MSA-C vs. PSP0.73540.72740.75740.7322
PD vs. PSP0.63050.50290.52210.5000
Table A13. GP classifier training and testing balanced accuracy, sensitivity, and specificity when using HC and T1w-only segmentation masks.
Table A13. GP classifier training and testing balanced accuracy, sensitivity, and specificity when using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.73570.78260.68870.69480.79900.59050.78680.87380.69980.73880.80150.6762
MSA-P vs. PD0.89160.89820.88490.84490.82380.86610.90240.90970.89510.88000.89580.8642
MSA-P vs. PSP0.74120.84230.64010.71710.81590.61820.75900.85960.65840.73930.85320.6253
MSA-C vs. PD0.78580.79220.77940.77810.75550.80060.79930.8180.78060.78240.78360.7812
MSA-C vs. PSP0.66550.71100.62010.60980.55940.66030.71710.71040.72380.60760.56610.6491
PD vs. PSP0.72600.79090.66110.58030.80320.35740.77180.80810.73550.76990.79900.7408
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A14. GP classifier training and testing accuracy when using HC and T1w-only segmentation masks.
Table A14. GP classifier training and testing accuracy when using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.75150.73690.79330.7682
MSA-P vs. PD0.89010.85300.89640.8571
MSA-P vs. PSP0.79180.76900.80320.7884
MSA-C vs. PD0.79290.780.78160.7770
MSA-C vs. PSP0.67200.55940.70960.5661
PD vs. PSP0.78570.75050.80330.8009

Appendix C.4

Table A15, Table A16 and Table A17 list the performances of the classifier that learned radiomic features based on random forest (RF).
Table A15. RF classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A15. RF classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.87250.78940.88400.8504
MSA-P vs. PD0.91350.84000.90780.8553
MSA-P vs. PSP0.84620.84180.86320.8497
MSA-C vs. PD0.71590.69510.70990.6649
MSA-C vs. PSP0.96410.88800.94580.8152
PD vs. PSP0.88960.82600.88690.8648
Table A16. RF classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A16. RF classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.81680.87370.75990.73660.81700.65610.81380.88630.74120.78260.88470.6804
MSA-P vs. PD0.88790.89240.88330.83010.80880.85140.89240.90130.88350.86450.85730.8717
MSA-P vs. PSP0.77340.82990.71690.72620.81440.63800.77810.85300.70320.77530.85290.6976
MSA-C vs. PD0.71960.63270.80650.71010.62780.79240.73590.68210.78980.71920.64530.7932
MSA-C vs. PSP0.89200.89210.89180.79600.81170.78030.86740.86120.87360.81920.82330.8152
PD vs. PSP0.77980.86960.69010.69680.83220.56140.78220.87720.68720.72340.85230.5945
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A17. RF classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A17. RF classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.8260.74750.82120.7902
MSA-P vs. PD0.88550.83630.88570.8571
MSA-P vs. PSP0.79410.76430.79940.7987
MSA-C vs. PD0.77620.76390.76910.7670
MSA-C vs. PSP0.88840.78390.85170.8083
PD vs. PSP0.82950.77620.83360.8051

Appendix C.5

Decision tree (DT) classifier results are listed in Table A18, Table A19 and Table A20.
Table A18. DT classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A18. DT classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.77890.66150.75280.7000
MSA-P vs. PD0.85680.75560.86880.7521
MSA-P vs. PSP0.74260.74950.76310.7577
MSA-C vs. PD0.64270.61920.64650.6208
MSA-C vs. PSP0.87340.78900.84120.8180
PD vs. PSP0.74530.68020.74330.7119
Table A19. DT classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A19. DT classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.76500.82610.70400.64630.75040.54230.74090.82130.66050.69760.73710.6581
MSA-P vs. PD0.85100.83830.86370.77450.71530.83370.84390.81890.86890.77520.68380.8666
MSA-P vs. PSP0.74870.82680.67060.73750.82440.65070.73170.83530.62800.72710.82100.6332
MSA-C vs. PD0.68420.55870.80960.63660.46930.80390.65290.51050.79530.64830.50170.7948
MSA-C vs. PSP0.87250.89110.85390.79110.80820.77410.84380.86180.82570.83160.82950.8336
PD vs. PSP0.77430.87680.67180.67940.85060.50820.73900.88080.59710.69340.85880.5281
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A20. DT classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A20. DT classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.77120.68590.75390.7132
MSA-P vs. PD0.85280.78880.85170.8035
MSA-P vs. PSP0.77140.77460.75740.7561
MSA-C vs. PD0.74970.72470.73290.7141
MSA-C vs. PSP0.87120.77570.83780.8267
PD vs. PSP0.82950.75810.81160.7878

Appendix C.6

The performances of the classifier trained with multi-layer perceptron (MLP), also known as Neural Net (NN), are listed in Table A21, Table A22 and Table A23.
Table A21. MLP classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A21. MLP classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.87500.78180.87680.8620
MSA-P vs. PD0.90290.87140.91570.8770
MSA-P vs. PSP0.89240.87530.90720.9000
MSA-C vs. PD0.80840.77980.79000.7841
MSA-C vs. PSP0.78500.68700.75970.6539
PD vs. PSP0.85640.80970.81450.7541
Table A22. MLP classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A22. MLP classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.61730.76070.47380.56870.64830.48910.76450.83670.69220.71350.81340.6136
MSA-P vs. PD0.80540.83610.77480.67480.70150.64810.89510.90730.88300.88000.89580.8642
MSA-P vs. PSP0.69040.80710.57380.67860.81460.54260.64160.81380.46930.61280.80510.4205
MSA-C vs. PD0.76800.72750.80840.73640.69290.77980.77070.75140.79000.73860.69320.7841
MSA-C vs. PSP0.59380.65320.53430.51340.56570.46120.54760.62540.46980.50560.41020.6010
PD vs. PSP0.65550.79900.51210.64260.78330.50190.69430.79690.59180.68260.80990.5552
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A23. MLP classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A23. MLP classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.70210.64440.78120.7575
MSA-P vs. PD0.81530.70720.88750.8571
MSA-P vs. PSP0.78860.77400.78900.7748
MSA-C vs. PD0.79560.76850.78290.7725
MSA-C vs. PSP0.66430.56430.64930.5172
PD vs. PSP0.77050.75620.75510.7502

Appendix C.7

The results of the classifier trained based on AdaBoost (ADA) are listed in Table A24, Table A25 and Table A26.
Table A24. ADA classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A24. ADA classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.84700.74860.84870.8287
MSA-P vs. PD0.90790.81640.90210.8328
MSA-P vs. PSP0.86970.85340.84260.8417
MSA-C vs. PD0.70000.65360.69990.6743
MSA-C vs. PSP0.95080.88840.92810.8939
PD vs. PSP0.87890.79600.88060.8538
Table A25. ADA classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A25. ADA classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.76190.81500.70880.64970.73680.56260.74860.81550.68160.74310.81580.6705
MSA-P vs. PD0.84670.81630.87710.75620.67930.83300.83900.79690.88110.77700.72540.8286
MSA-P vs. PSP0.77420.83780.71050.74500.82520.66480.74420.83690.65140.72750.83110.6239
MSA-C vs. PD0.64210.48870.79560.63400.46830.79970.66320.53590.79060.62550.45860.7924
MSA-C vs. PSP0.87250.87330.87160.81760.83360.80150.86560.87920.85210.81890.83210.8057
PD vs. PSP0.77100.88790.65420.68610.84320.52900.74810.87430.62180.71430.86480.5638
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A26. ADA classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A26. ADA classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.770.68960.75290.7563
MSA-P vs. PD0.85350.77600.850.7976
MSA-P vs. PSP0.79600.77280.76900.7587
MSA-C vs. PD0.72620.72540.72790.7145
MSA-C vs. PSP0.86960.80620.85880.8094
PD vs. PSP0.83330.77330.81760.8036

Appendix C.8

Results of classifier using Gaussian naïve Bayes (GNB) are listed in Table A27, Table A28 and Table A29.
Table A27. GNB classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A27. GNB classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.89150.83970.88490.8683
MSA-P vs. PD0.92790.87870.92450.8817
MSA-P vs. PSP0.90540.87650.89610.8777
MSA-C vs. PD0.70880.66100.71200.6652
MSA-C vs. PSP0.96580.88400.94930.8721
PD vs. PSP0.91980.86000.91540.8657
Table A28. GNB classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A28. GNB classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.84170.93170.75170.75960.86220.65710.82250.92310.72180.81380.90660.7210
MSA-P vs. PD0.89380.88600.90160.84340.79910.88760.89220.87130.91320.80730.75180.8629
MSA-P vs. PSP0.77220.87890.66560.76700.90230.63170.78670.88890.68450.76260.88380.6413
MSA-C vs. PD0.67670.53590.81750.66000.50990.81000.71790.63200.80390.68500.55690.8132
MSA-C vs. PSP0.88060.90050.86070.75670.78390.72950.83830.87920.79740.74930.78400.7146
PD vs. PSP0.75660.92180.59130.71790.89220.54350.75150.93530.56760.70830.91230.5042
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A29. GNB classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A29. GNB classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.84650.77050.83080.8272
MSA-P vs. PD0.89480.85570.89640.8125
MSA-P vs. PSP0.80300.79880.81350.7890
MSA-C vs. PD0.75790.73220.76910.7562
MSA-C vs. PSP0.87740.75560.83620.7581
PD vs. PSP0.81620.78570.81330.7807

Appendix C.9

Table A30, Table A31 and Table A32 list the results of the quadratic discriminant analysis (QDA) classifier.
Table A30. QDA classifier training and testing AUC using HC and T1w-only segmentation masks.
Table A30. QDA classifier training and testing AUC using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.89530.85500.90290.8749
MSA-P vs. PD0.93170.88980.92000.9000
MSA-P vs. PSP0.89150.89290.91500.9091
MSA-C vs. PD0.69250.69290.74570.6580
MSA-C vs. PSP0.93680.85890.90560.8382
PD vs. PSP0.88700.79440.89510.8638
Table A31. QDA classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Table A31. QDA classifier training and testing balanced accuracy, sensitivity, and specificity using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.79030.80050.78010.68470.79160.57780.81790.85950.77620.79370.85670.7307
MSA-P vs. PD0.88180.85100.91260.84450.80050.88850.88040.85010.91070.82730.76880.8858
MSA-P vs. PSP0.79960.79160.80760.78770.79750.77790.83810.85610.82010.81990.86550.7742
MSA-C vs. PD0.70510.59710.81300.65550.51790.79320.72040.62950.81120.66690.55670.7772
MSA-C vs. PSP0.88850.82810.94890.76300.76460.76140.79480.78710.80250.76550.76070.7704
PD vs. PSP0.78680.86550.70810.72130.84340.59910.78410.89260.67550.74140.88920.5937
bAcc: balanced accuracy, Sen: sensitivity, Spe: specificity.
Table A32. QDA classifier training and testing accuracy using HC and T1w-only segmentation masks.
Table A32. QDA classifier training and testing accuracy using HC and T1w-only segmentation masks.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.79210.74780.82660.8111
MSA-P vs. PD0.88800.85640.88570.8392
MSA-P vs. PSP0.79440.78160.84450.8374
MSA-C vs. PD0.77060.73620.77410.7341
MSA-C vs. PSP0.84790.72870.78120.7478
PD vs. PSP0.83140.79330.84070.8180

References

  1. Barbagallo, G.; Sierra-Peña, M.; Nemmi, F.; Traon, A.P.L.; Meissner, W.G.; Rascol, O.; Péran, P. Multimodal MRI assessment of nigro-striatal pathway in multiple system atrophy and Parkinson disease. Mov. Disord. 2016, 31, 142–149. [Google Scholar] [CrossRef] [PubMed]
  2. Jellinger, K.A. Neuropathological spectrum of synucleinopathies. Mov. Disord. 2003, 18, 2–12. [Google Scholar] [CrossRef] [PubMed]
  3. Ward, R.J.; Zucca, F.A.; Duyn, J.H.; Crichton, R.R.; Zecca, L. The role of iron in brain ageing and neurodegenerative disorders. Lancet Neurol. 2014, 13, 1045–1060. [Google Scholar] [CrossRef] [Green Version]
  4. Kaindlstorfer, C.; Jellinger, K.A.; Eschlböck, S.; Stefanova, N.; Weiss, G.; Wenning, G.K. The relevance of iron in the pathogenesis of multiple system atrophy: A viewpoint. J. Alzheimer’s Dis. 2018, 61, 1253–1273. [Google Scholar] [CrossRef] [Green Version]
  5. Peeraully, T. Multiple system atrophy. Semin. Neurol. 2014, 34, 290–292. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, N.; Edmiston, E.K.; Luo, X.; Yang, H.; Chang, M.; Wang, F.; Fan, G. Comparing abnormalities of amplitude of low-frequency fluctuations in multiple system atrophy and idiopathic Parkinson’s disease measured with resting-state fMRI. Psychiatry Res. Neuroimaging 2017, 269, 73–81. [Google Scholar] [CrossRef]
  7. Hikishima, K.; Ando, K.; Yano, R.; Kawai, K.; Komaki, Y.; Inoue, T.; Itoh, T.; Yamada, M.; Momoshima, S.; Okano, H.J.; et al. Parkinson disease: Diffusion MR imaging to detect nigrostriatal pathway loss in a marmoset model treated with 1-methyl-4-phenyl-1,2,3,6-tetrahydropyridine. Radiology 2015, 275, 430–437. [Google Scholar] [CrossRef]
  8. Pang, H.; Yu, Z.; Li, R.; Yang, H.; Fan, G. MRI-based radiomics of basal nuclei in differentiating idiopathic Parkinson’s disease From parkinsonian variants of multiple system atrophy: A susceptibility-weighted imaging study. Front. Aging Neurosci. 2020, 12, 587250. [Google Scholar] [CrossRef]
  9. Saeed, U.; Lang, A.E.; Masellis, M. Neuroimaging Advances in Parkinson’s Disease and Atypical Parkinsonian Syndromes. Front. Neurol. 2020, 11, 572976. [Google Scholar] [CrossRef]
  10. Aludin, S.; Schmill, L.A. MRI Signs of Parkinson’s Disease and Atypical Parkinsonism. Rofo 2021, 193, 1403–1410. [Google Scholar]
  11. Sodickson, A.; Baeyens, P.F.; Andriole, K.P.; Prevedello, L.M.; Nawfel, R.D.; Hanson, R.; Khorasani, R. Recurrent CT, cumulative radiation exposure, and associated radiation-induced cancer risks from CT of adults. Radiology 2009, 251, 175–184. [Google Scholar] [CrossRef] [PubMed]
  12. Alster, P.; Nieciecki, M.; Migda, B.; Kutyłowski, M.; Madetko, N.; Duszyńska-Wąs, K.; Charzyńska, I.; Koziorowski, D.; Królicki, L.; Friedman, A. The Strengths and Obstacles in the Differential Diagnosis of Progressive Supranuclear Palsy—Parkinsonism Predominant (PSP-P) and Multiple System Atrophy (MSA) Using Magnetic Resonance Imaging (MRI) and Perfusion Single Photon Emission Computed Tomography (SPECT). Diagnostics 2022, 12, 385. [Google Scholar] [PubMed]
  13. Lee, J.H.; Baik, S.K. Putaminal hypointensity in the parkinsonian variant of multiple system atrophy: Simple visual assessment using susceptibility-weighted imaging. J. Mov. Disord. 2011, 4, 60–63. [Google Scholar] [CrossRef] [PubMed]
  14. Lee, J.H.; Lee, M.S. Brain iron accumulation in atypical parkinsonian syndromes: In vivo MRI evidences for distinctive patterns. Front. Neurol. 2019, 10, 74. [Google Scholar] [CrossRef] [Green Version]
  15. Gupta, D.; Saini, J.; Kesavadas, C.; Sarma, P.S.; Kishore, A. Utility of susceptibility-weighted MRI in differentiating Parkinson’s disease and atypical parkinsonism. Neuroradiology 2010, 52, 1087–1094. [Google Scholar] [CrossRef]
  16. Harder, S.L.; Hopp, K.M.; Ward, H.; Neglio, H.; Gitlin, J.; Kido, D. Mineralization of the deep gray matter with age: A retrospective review with susceptibility-weighted MR imaging. Am. J. Neuroradiol. 2008, 29, 176–183. [Google Scholar] [CrossRef] [Green Version]
  17. Cheng, Z.; Zhang, J.; He, N.; Li, Y.; Wen, Y.; Xu, H.; Tang, R.; Jin, Z.; Haacke, E.M.; Yan, F.; et al. Radiomic features of the nigrosome-1 region of the substantia nigra: Using quantitative susceptibility mapping to assist the diagnosis of idiopathic Parkinson’s disease. Front. Aging Neurosci. 2019, 11, 167. [Google Scholar] [CrossRef] [Green Version]
  18. Fischl, B.; Salat, D.H.; Busa, E.; Albert, M.; Dieterich, M.; Haselgrove, C.; van der Kouwe, A.; Killiany, R.; Kennedy, D.; Klaveness, S.; et al. Whole brain segmentation: Automated labeling of neuroanatomical structures in the human brain. Neuron 2002, 33, 341–355. [Google Scholar] [CrossRef] [Green Version]
  19. Fischl, B. FreeSurfer. NeuroImage 2012, 62, 774–781. [Google Scholar] [CrossRef] [Green Version]
  20. Patenaude, B.; Smith, S.M.; Kennedy, D.N.; Jenkinson, M. A Bayesian model of shape and appearance for subcortical brain segmentation. Neuroimage 2011, 56, 907–922. [Google Scholar] [CrossRef] [Green Version]
  21. Wang, J.; Vachet, C.; Rumple, A.; Gouttard, S.; Ouziel, C.; Perrot, E.; Du, G.; Huang, X.; Gerig, G.; Styner, M. Multi-atlas segmentation of subcortical brain structures via the AutoSeg software pipeline. Front. Neuroinform. 2014, 8, 7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Morey, R.A.; Selgrade, E.S.; Wagner, H.R., 2nd; Huettel, S.A.; Wang, L.; McCarthy, G. Scan-rescan reliability of subcortical brain volumes derived from automated segmentation. Hum. Brain Mapp. 2010, 31, 1751–1762. [Google Scholar] [PubMed] [Green Version]
  23. Nugent, A.C.; Luckenbaugh, D.A.; Wood, S.E.; Bogers, W.; Zarate, C.A., Jr.; Drevets, W.C. Automated subcortical segmentation using FIRST: Test-retest reliability, interscanner reliability, and comparison to manual segmentation. Hum. Brain Mapp. 2013, 34, 2313–2329. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. de Sitter, A.; Verhoeven, T.; Burggraaff, J.; Liu, Y.; Simoes, J.; Ruggieri, S.; Palotai, M.; Brouwer, I.; Versteeg, A.; Wottschel, V.; et al. Reduced accuracy of MRI deep grey matter segmentation in multiple sclerosis: An evaluation of four automated methods against manual reference segmentations in a multi-center cohort. J. Neurol. 2020, 267, 3541–3554. [Google Scholar] [CrossRef] [PubMed]
  25. Derakhshan, M.; Caramanos, Z.; Giacomini, P.S.; Narayanan, S.; Maranzano, J.; Francis, S.J.; Arnold, D.L.; Collins, D.L. Evaluation of automated techniques for the quantification of grey matter atrophy in patients with multiple sclerosis. Neuroimage 2010, 52, 1261–1267. [Google Scholar] [CrossRef]
  26. Litvan, I.; Agid, Y.; Calne, D.; Campbell, G.; Dubois, B.; Duvoisin, R.C. Clinical research criteria for the diagnosis of progressive supranuclear palsy (Steel-Richardson-Olszewski syndrome): Report of the NINDSSPSP international workshop. Neurology 1996, 47, 1–9. [Google Scholar] [CrossRef]
  27. Gilman, S.; Wenning, G.K.; Low, P.A.; Brooks, D.J.; Mathias, C.J.; Trojanowski, J.Q.; Wood, N.W.; Colosimo, C.; Dürr, A.; Fowler, C.J.; et al. Second consensus statement on the diagnosis of multiple system atrophy. Neurology 2008, 71, 670–676. [Google Scholar] [CrossRef]
  28. Hughes, A.J.; Daniel, S.E.; Kilford, L.; Lees, A.J. Accuracy of clinical diagnosis of idiopathic Parkinson’s disease: A clinico-pathological study of 100 cases. J. Neurol. Neurosurg. Psychiatry 1992, 55, 181–184. [Google Scholar] [CrossRef] [Green Version]
  29. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef] [Green Version]
  30. Feng, X.; Deistung, A.; Dwyer, M.G.; Hagemeier, J.; Polak, P.; Lebenberg, J.; Frouin, F.; Zivadinov, R.; Reichenbach, J.R.; Schweser, F. An improved FSL-FIRST pipeline for subcortical gray matter segmentation to study abnormal brain anatomy using quantitative susceptibility mapping (QSM). Magn. Reason. Imaging 2017, 39, 110–122. [Google Scholar] [CrossRef]
  31. Jones, B.C.; Nair, G.; Shea, C.D.; Crainiceanu, C.M.; Cortese, I.C.; Reich, D.S. Quantification of multiple-sclerosis-related brain atrophy in two heterogeneous MRI datasets using mixed-effects modeling. Neuroimage Clin. 2013, 3, 171–179. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Avants, B.B.; Tustison, N.J.; Song, G.; Cook, P.A.; Klein, A.; Gee, J.C. A reproducible evaluation of ANTs similarity metric performance in brain image registration. Neuroimage 2011, 54, 2033–2044. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Klein, A.; Andersson, J.; Ardekani, B.A.; Ashburner, J.; Avants, B.; Chiang, M.C.; Christensen, G.E.; Collins, D.L.; Gee, J.; Hellier, P.; et al. Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration. Neuroimage 2009, 46, 786–802. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Van Griethuysen, J.J.M.; Fedorov, A.; Parmar, C.; Hosny, A.; Aucoin, N.; Narayan, V.; Beets-Tan, R.G.H.; Fillion-Robin, J.C.; Pieper, S.; Aerts, H.J.W.L. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 2017, 77, e104–e107. [Google Scholar] [CrossRef] [Green Version]
  35. Available online: https://pyradiomics.readthedocs.io/en/latest/index.html (accessed on 22 January 2022).
  36. Wu, Y.; Jiang, J.H.; Chen, L.; Lu, J.Y.; Ge, J.J.; Liu, F.T.; Yu, J.T.; Lin, W.; Zuo, C.T.; Wang, J. Use of radiomic features and support vector machine to distinguish Parkinson’s disease cases from normal controls. Ann. Transl. Med. 2019, 7, 773. [Google Scholar] [CrossRef]
  37. Salmanpour, M.R.; Shamsaei, M.; Rahmim, A. Feature selection and machine learning methods for optimal identification and prediction of subtypes in Parkinson’s disease. Comput. Methods Programs Biomed. 2021, 206, 106131. [Google Scholar] [CrossRef]
  38. Suguna, N.; Thanushkodi, K. An improved k-nearest neighbor classification using genetic algorithm. Int. J. Comput. Sci. Issues 2010, 7, 18–21. [Google Scholar]
  39. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  40. Kuo, B.C.; Ho, H.H.; Li, C.H.; Hung, C.C. A kernel-based feature selection method for SVM with RBF kernel for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 317–326. [Google Scholar] [CrossRef]
  41. Rasmussen, C.E. Gaussian processes for machine learning. In Summer School on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2003; pp. 63–71. [Google Scholar]
  42. Ali, J.; Khan, R.; Ahmad, N.; Maqsood, I. Random forests and decision trees. Int. J. Comput. Sci. Issues 2012, 9, 272–278. [Google Scholar]
  43. Chourasia, S. Survey paper on improved methods of ID3 decision tree classification. Int. J. Sci. Res. Publ. 2013, 3, 1–4. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  45. Hastie, T.; Rosset, S.; Zhu, J.; Zuo, H. Multi-class adaboost. Stat. Its Interface 2009, 2, 349–360. [Google Scholar] [CrossRef] [Green Version]
  46. Chan, T.F.; Golub, G.H.; LeVeque, R.J. Updating formulae and a pairwise algorithm for computing sample variances. In COMPSTAT 1982 5th Symposium Held at Toulouse 1982; Physica: Heidelberg, Germany, 1982; pp. 30–41. [Google Scholar]
  47. Srivastava, M.R.; Gupta, M.R.; Frigyik, B.A. Bayesian quadratic discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1277–1305. [Google Scholar]
  48. Feng, Q.; Chen, Y.; Liao, Z.; Jiang, H.; Mao, D.; Wang, M. Corpus callosum radiomics-based classification model in alzheimer’s disease: A case-control study. Front. Neurol. 2018, 9, 618. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Zwanenburg, A.; Leger, S.; Vallieres, M.; Lock, S. Image biomarker standardisation initiative. arXiv 2016, arXiv:1612.07003. [Google Scholar]
  50. Lee, M.J.; Kim, T.H.; Kim, S.J.; Kim, B.K.; Mun, C.W.; Lee, J.H. Quantitative Validation of a Visual Rating Scale for Defining High-Iron Putamen in Patients with Multiple System Atrophy. Front. Neurol. 2019, 10, 1014. [Google Scholar] [CrossRef] [Green Version]
Figure 1. SWI axial view of parkinsonian syndrome patients: parkinsonian variant multiple system atrophy (MSA-P), cerebellar variant multiple system atrophy (MSA-C), progressive supranuclear palsy (PSP), and Parkinson’s disease (PD). Increased iron-related signals in the anterior and medial aspects of the globus pallidus (open arrow) of SWI is a highly specific sign of PSP. For MSA-P, significant accumulation of iron in the lateral aspect of the globus pallidus adjacent to putamen, posterolateral putaminal hypointensity, (closed arrow) and lateral-to-medial gradient appear consistently.
Figure 1. SWI axial view of parkinsonian syndrome patients: parkinsonian variant multiple system atrophy (MSA-P), cerebellar variant multiple system atrophy (MSA-C), progressive supranuclear palsy (PSP), and Parkinson’s disease (PD). Increased iron-related signals in the anterior and medial aspects of the globus pallidus (open arrow) of SWI is a highly specific sign of PSP. For MSA-P, significant accumulation of iron in the lateral aspect of the globus pallidus adjacent to putamen, posterolateral putaminal hypointensity, (closed arrow) and lateral-to-medial gradient appear consistently.
Diagnostics 12 00637 g001
Figure 2. Overall flowchart of combining T1w and SWI, SWI segmentation, feature extraction and selection, and disease classification. We create a hybrid image combining T1w and SWI for iron-reflected DGM segmentation, extract texture representative features, and classify parkinsonian disorders with the significant features selected using various machine learning algorithms.
Figure 2. Overall flowchart of combining T1w and SWI, SWI segmentation, feature extraction and selection, and disease classification. We create a hybrid image combining T1w and SWI for iron-reflected DGM segmentation, extract texture representative features, and classify parkinsonian disorders with the significant features selected using various machine learning algorithms.
Diagnostics 12 00637 g002
Figure 3. Flowchart of making a deep gray matter (DGM) mask using the T1w and SWI images. T1w and SWI were preprocessed through normalization, bias correction, and registration. The merging weight coefficients were calculated from initial DGM mask obtained using only T1w segmentation, and a hybrid contrast image (HC) was created as a result. The DGM mask was obtained by registering the HC to the MNI atlas space using non-linear registration. The final mask was obtained by applying inverse warping to the original coordinates.
Figure 3. Flowchart of making a deep gray matter (DGM) mask using the T1w and SWI images. T1w and SWI were preprocessed through normalization, bias correction, and registration. The merging weight coefficients were calculated from initial DGM mask obtained using only T1w segmentation, and a hybrid contrast image (HC) was created as a result. The DGM mask was obtained by registering the HC to the MNI atlas space using non-linear registration. The final mask was obtained by applying inverse warping to the original coordinates.
Diagnostics 12 00637 g003
Figure 4. Deep gray matter (DGM) axial slice in T1w, SWI, and HC images. HC has both a high contrast cortex, which is the advantage of T1w, and a more prominent DGM boundary, which is visible in the SWI.
Figure 4. Deep gray matter (DGM) axial slice in T1w, SWI, and HC images. HC has both a high contrast cortex, which is the advantage of T1w, and a more prominent DGM boundary, which is visible in the SWI.
Diagnostics 12 00637 g004
Figure 5. Putamen mask of segmentation result of using only T1-weighted image (FreeSurfer) and using T1w and SWI (proposed method) with SWI overlaid. The segmentation result using only T1w includes the part without iron accumulation when overlaid with the SWI (yellow). The proposed method reflects more of the iron deposition (red).
Figure 5. Putamen mask of segmentation result of using only T1-weighted image (FreeSurfer) and using T1w and SWI (proposed method) with SWI overlaid. The segmentation result using only T1w includes the part without iron accumulation when overlaid with the SWI (yellow). The proposed method reflects more of the iron deposition (red).
Diagnostics 12 00637 g005
Table 1. An overview of the common neuroimaging modalities (DTI, PET, SPECT, and SWI), role of modality, and potential of differentiating PD and APS.
Table 1. An overview of the common neuroimaging modalities (DTI, PET, SPECT, and SWI), role of modality, and potential of differentiating PD and APS.
Neuroimaging ModalityRole of ModalityPotential of Differentiating PD and APS
Diffusion-tensor image (DTI) [7]Detect characteristics such as fractional anisotropy (FA) and mean diffusion (MD)Decreased FA and/or increased MD in the substantia nigra, the corpus callosum, the frontal lobes, the cingulum, and the temporal cortex
Positron emission tomography (PET) [9]Measure amyloid pathology, tau pathology, a-Synuclein pathology, metabolic activity by measuring changes in the glucose consumptionPD-related spatial covariance pattern may involve increased pallidothalamic and pontine activity associated with decreased metabolism in supplementary motor area, premotor cortex, and parietal association areas
Single photon emission computed tomography (SPECT) [12]Measure dopamine transporter (DAT) density, dopamine D2 receptor, metabolic activity by measuring changes in the cerebral blood flowDecreased striatal presynaptic DAT binding contralateral to parkinsonian symptomatology with greater reduction in posterior putamen than in anterior putamen or caudate nucleus
Susceptibility weighted image (SWI) [13]Visualize iron-related contents sensitivelySubstantia nigra pars compacta, globus pallidus internus, the putamen, and the red nucleus have been described as regions with increased iron concentration
Table 2. Clinical and demographic characteristics of patients.
Table 2. Clinical and demographic characteristics of patients.
MSA-PMSA-CPSPPDSignificance
Gender (M/F)13/219/1211/632/24p = 0.179
χ 2 (3) = 4.894
Age (years)59.05 ± 7.8358.95 ± 6.3065.64 ± 5.5856.85 ± 7.60p < 0.001
F(3,124) = 6.052
UPDRS-III39.73 ± 12.8630.80 ± 9.7335.94 ± 8.1524.33 ± 9.57p < 0.001
F(3,124) = 16.597
H-Y stage3.10 ± 0.763.14 ± 0.613.5 ± 0.752.02 ± 0.51p < 0.001
F(3,124) = 36.885
Duration (months)30.23 ± 15.2530.52 ± 13.6231.11 ± 18.0935.41 ± 22.23p = 0.661
F(3,124) = 0.532
MMSE25.44 ± 2.7324.76 ± 3.2323.82 ± 4.0326.89 ± 2.41p < 0.001
F(3,124) = 6.395
The data are presented as number or mean ± standard deviation. For continuous variables, values are expressed as F statistics, while for categorical variables, values are expressed as χ 2 statistics. MSA-P : parkinsonian variant of multiple system atrophy, MSA-C: cerebellar variant of multiple system atrophy, PSP: progressive supranuclear palsy, PD: Parkinson’s disease, UPDRS III: motor examination part of the Unified Parkinson’s Disease Rating Scale, H-Y: Hoehn & Yahr.
Table 3. Mean values of top 10 features selected from SWI when comparing MSA-P and PD using HC and T1w-only segmentation masks. Common features found both in HC and T1w-only segmentation are indicated in bold.
Table 3. Mean values of top 10 features selected from SWI when comparing MSA-P and PD using HC and T1w-only segmentation masks. Common features found both in HC and T1w-only segmentation are indicated in bold.
Features by HCMSA-PPDFeatures by T1w-OnlyMSA-PPD
glrlm_ShortRunHigh-
GrayLevelEmphasis
51.229120.3952glcm_MCC40.41660.2744
glcm_Autocorrelation769.240727.9998glcm_Imc240.56650.364
glcm_JointAverage78.13535.1994glcm_JointAverage78.40945.5243
glcm_SumAverage716.270610.3988glcm_SumAverage716.818711.0487
gldm_HighGrayLevelEmphasis64.921927.8777gldm_DependenceVariance23.220527.8954
glrlm_HighGrayLevelRunEmphasis64.008628.202glrlm_GrayLevelNonUniformity344.9109652.5307
glcm_Imc240.54420.3244glcm_Autocorrelation772.681131.5304
glcm_Autocorrelation463.634926.3972glcm_SumAverage416.189910.7769
glcm_SumAverage415.523910.0885glcm_JointAverage48.0955.3884
glcm_JointAverage47.7625.0443gldm_HighGrayLevelEmphasis75.700532.9521
Table 4. RBF SVM classifier training and testing AUC when using HC and T1w-only segmentation masks. The classifier model trained with features extracted using HC masks showed 0.1037 higher AUC for training and 0.062 higher AUC for testing compared to the T1w-only mask-based SVM classifier.
Table 4. RBF SVM classifier training and testing AUC when using HC and T1w-only segmentation masks. The classifier model trained with features extracted using HC masks showed 0.1037 higher AUC for training and 0.062 higher AUC for testing compared to the T1w-only mask-based SVM classifier.
Differentiating
Diseases
Train AUCTest AUC
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.88560.82420.86990.8263
MSA-P vs. PD0.89380.85370.90320.8561
MSA-P vs. PSP0.88250.82450.88690.8499
MSA-C vs. PD0.67310.58780.68200.6193
MSA-C vs. PSP0.88830.67960.81800.7578
PD vs. PSP0.94110.77240.93380.8104
AUC: area under the receiver operating characteristic (ROC) curve
Table 5. RBF SVM classifier training and testing balanced accuracy, sensitivity, and specificity when using HC and T1w-only segmentation masks. The classifier model trained with features extracted using HC masks outperformed the SVM classifier based on T1w-only masks by 0.1109 in training and 0.0372 in testing in terms of the balanced accuracy.
Table 5. RBF SVM classifier training and testing balanced accuracy, sensitivity, and specificity when using HC and T1w-only segmentation masks. The classifier model trained with features extracted using HC masks outperformed the SVM classifier based on T1w-only masks by 0.1109 in training and 0.0372 in testing in terms of the balanced accuracy.
Differentiating
Diseases
TrainTest
HCT1w-OnlyHCT1w-Only
bAccSenSpebAccSenSpebAccSenSpebAccSenSpe
MSA-P vs. MSA-C0.79310.84720.73900.70050.80450.59630.79220.86620.71830.73130.82980.6327
MSA-P vs. PD0.91200.88650.89370.84820.83160.86470.89810.90460.89170.88000.89580.8642
MSA-P vs. PSP0.77900.87070.68740.60230.78540.41930.78620.88020.69220.75350.83450.6725
MSA-C vs. PD0.78630.77270.79990.75160.73350.76980.78990.79880.78100.78720.80310.7714
MSA-C vs. PSP0.74700.80450.68950.54910.67140.42690.72620.80200.65050.68280.68380.6818
PD vs. PSP0.58230.79140.37320.48260.77570.18940.80270.81940.78600.73760.78070.0776
bAcc : balanced accuracy, Sen: sensitivity, Spe: specificity.
Table 6. RBF SVM classifier training and testing accuracy when using HC and T1w-only segmentation masks. The HC trained classifier distinguishes disorders better than the T1w-only trained classifier by 0.0648 in training and 0.0406 in testing.
Table 6. RBF SVM classifier training and testing accuracy when using HC and T1w-only segmentation masks. The HC trained classifier distinguishes disorders better than the T1w-only trained classifier by 0.0648 in training and 0.0406 in testing.
Differentiating
Diseases
Train ACCTest ACC
HCT1w-OnlyHCT1w-Only
MSA-P vs. MSA-C0.79720.73360.80180.7552
MSA-P vs. PD0.89440.85440.89280.8571
MSA-P vs. PSP0.80870.69020.81350.7692
MSA-C vs. PD0.79600.76820.78040.7708
MSA-C vs. PSP0.71720.59730.72880.6616
PD vs. PSP0.78670.76760.81840.778
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, Y.S.; Lee, J.-H.; Gahm, J.K. Automated Differentiation of Atypical Parkinsonian Syndromes Using Brain Iron Patterns in Susceptibility Weighted Imaging. Diagnostics 2022, 12, 637. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12030637

AMA Style

Kim YS, Lee J-H, Gahm JK. Automated Differentiation of Atypical Parkinsonian Syndromes Using Brain Iron Patterns in Susceptibility Weighted Imaging. Diagnostics. 2022; 12(3):637. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12030637

Chicago/Turabian Style

Kim, Yun Soo, Jae-Hyeok Lee, and Jin Kyu Gahm. 2022. "Automated Differentiation of Atypical Parkinsonian Syndromes Using Brain Iron Patterns in Susceptibility Weighted Imaging" Diagnostics 12, no. 3: 637. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics12030637

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop