Next Article in Journal
Smartphone Photogrammetric Assessment for Head Measurements
Next Article in Special Issue
A Machine Learning App for Monitoring Physical Therapy at Home
Previous Article in Journal
User Experience Evaluation of Upper Limb Rehabilitation Robots: Implications for Design Optimization: A Pilot Study
Previous Article in Special Issue
A Dynamic Procedure to Detect Maximum Voluntary Contractions in Low Back
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Approach for Automatic and Objective Grading of the Motor Impairment Severity in Parkinson’s Disease for Use in Tele-Assessments

by
Mehar Singh
1,2,*,
Prithvi Prakash
3,
Rachneet Kaur
4,
Richard Sowers
4,
James Robert Brašić
5,6,7 and
Manuel Enrique Hernandez
8,9,10,*
1
Computer Science and Engineering Division, University of Michigan, Ann-Arbor, MI 48109, USA
2
Department of Computer Science, University of Illinois Urbana-Champaign, Urbana, IL 61801, USA
3
School of Information Sciences, University of Illinois Urbana-Champaign, Champaign, IL 61820, USA
4
Department of Industrial and Enterprise Systems Engineering, University of Illinois Urbana-Champaign, Urbana, IL 61801, USA
5
Section of High Resolution Brain Positron Emission Tomography Imaging, Division of Nuclear Medicine and Molecular Imaging, Russell H. Morgan Department of Radiology and Radiological Sciences, The Johns Hopkins University School of Medicine, Baltimore, MD 21205, USA
6
Department of Behavioral Health, New York City Health + Hospitals/Bellevue, 462 First Avenue, New York, NY 10016, USA
7
Department of Psychiatry, New York University Grossman School of Medicine, New York University Langone Health, New York University, 550 First Avenue, New York, NY 10016, USA
8
Neuroscience Program, Beckman Institute, College of Liberal Arts & Sciences, University of Illinois Urbana-Champaign, Urbana, IL 61801, USA
9
Department of Biomedical and Translational Sciences, University of Illinois Urbana-Champaign, Urbana, IL 61801, USA
10
Department of Kinesiology and Community Health, University of Illinois Urbana-Champaign, Urbana, IL 61801, USA
*
Authors to whom correspondence should be addressed.
Submission received: 20 September 2023 / Revised: 18 October 2023 / Accepted: 25 October 2023 / Published: 6 November 2023

Abstract

:
Wearable sensors provide a tool for at-home monitoring of motor impairment progression in neurological conditions such as Parkinson’s disease (PD). This study examined the ability of deep learning approaches to grade the motor impairment severity in a modified version of the Movement Disorders Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS) using low-cost wearable sensors. We hypothesized that expanding training datasets with motion data from healthy older adults (HOAs) and initializing classifiers with weights learned from unsupervised pre-training would lead to an improvement in performance when classifying lower vs. higher motor impairment relative to a baseline deep learning model (XceptionTime). This study evaluated the change in classification performance after using expanded training datasets with HOAs and transferring weights from unsupervised pre-training compared to a baseline deep learning model (XceptionTime) using both upper extremity (finger tapping, hand movements, and pronation–supination movements of the hands) and lower extremity (toe tapping and leg agility) tasks consistent with the MDS-UPDRS. Overall, we found a 12.2% improvement in accuracy after expanding the training dataset and pre-training using max-vote inference on hand movement tasks. Moreover, we found that the classification performance improves for every task except toe tapping after the addition of HOA training data. These findings suggest that learning from HOA motion data can implicitly improve the representations of PD motion data for the purposes of motor impairment classification. Further, our results suggest that unsupervised pre-training can improve the performance of motor impairment classifiers without any additional annotated PD data, which may provide a viable solution for a widely deployable telemedicine solution.

1. Introduction

Parkinson’s disease (PD) is a prevalent neurodegenerative disorder that is characterized by motor symptoms including bradykinesia, rigidity, resting tremor, and postural instability [1,2], while there is no cure for PD, access to multidisciplinary medical care can greatly benefit people with PD [3]. However, clinical evaluations typically require limited-availability in-person appointments with a movement disorders specialist at limited locations, requiring significant travel time and hampering access [4]. Given the expected increase in the prevalence of PD expected in the population [5], there is a significant need for telemedicine solutions to improve access to healthcare for people with PD [6].
Traditional clinical evaluations of PD patients can often be time-consuming and inefficient for both patients and clinicians. While multiple rating scales have been proposed, the gold standard evaluation of motor impairment in people with PD is the Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS) [7]. The MDS-UPDRS motor subsection provides a qualitative assessment of bradykinesia, resting tremor, and postural instability symptoms, but not rigidity. However, there is a lack of automatic and objective evaluations of motor impairment for potential integration in telemedicine applications.
Telemedicine solutions may be well suited to evaluate people with PD, particularly in rural and underserved communities that usually lack access to healthcare [8,9]. Teleconsultations may provide an opportunity for the remote administration of neurological examinations, but subtle features such as bradykinesia may be difficult to capture using video alone [10]. Additionally, rigidity cannot be assessed without a physical examination, with hands-on manipulations of the joints of the patient by the examiner. While the COVID-19 pandemic fast tracked the deployment of telemedicine applications to address the needs of people with PD, there are still challenges related to inconsistent monitoring images, privacy, poor connectivity, and access to technology.
The use of machine learning with behavioral data has shown great promise in differentiating the pathological and physiological motor responses arising from PD [11,12,13] and several other neurological conditions [14,15] from healthy controls. The integration of low-cost wearable technology and deep learning approaches may provide a viable approach towards the development of robust and more widely deployable telemedicine solutions.
However, one of the major challenges with machine learning in telemedicine is the extensive amounts of labeled data required to train supervised learning models [16]. There exists a shortage of available datasets on motion data from PD patients in addition to the costliness and inefficiency of manually annotating data. As such, further development of unsupervised learning and transfer learning techniques is needed in telemedicine. Unsupervised learning is the process of learning from unlabeled data, while transfer learning is the process of using the knowledge acquired from one objective, perhaps with more abundant data, to an adjacent objective. However, it is important to establish specific approaches for a given application when using transfer learning principles [17].
In this study, we explored how transferring unsupervised pre-trained weights onto supervised models can improve the performance of motor impairment classification for PD patients. We experimented with models trained on datasets containing motion sequences from only PD patients as well as both PD patients and healthy older adults (HOAs). Firstly, we hypothesized that the inclusion of HOA motion sequences in training datasets would lead to performance improvements, since such sequences could act as a healthy control for anomaly detection. Secondly, we predicted that using unsupervised pre-trained weights to initialize our supervised classification models would improve the performance relative to random weight initialization (RandInit). We suspected that the unsupervised models from which we transferred weights would implicitly learn meaningful representations of motion sequences for the motor impairment classification objective.
Overall, we sought to determine if unsupervised learning, transfer learning, and the inclusion of HOA motion sequences in training datasets can improve motor impairment classification for PD patients without using any additional data from PD patients.

2. Materials and Methods

2.1. Protocol

We collected continuous motion data from the extremities of both PD and HOA participants using a custom-built, low-cost quantitative measurement system. It consisted of a 3-axis accelerometer (ADXL335) and an evaluation board (EVAL-ADXL335Z) with USB connectors [18]. We received ground-truth evaluation of movement impairment from an examiner certified in the administration of the Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS) [7,18].
For upper extremity movements (finger tapping (FT), hand movements (HMs), and pronation–supination movement of the hands (PS)), repetitive movements were recorded at 80 Hz using two accelerometers placed at:
  • The dorsal surface of the second (middle) phalanx of each index finger;
  • The dorsum of each arm, midway between the radius and the ulna and two inches proximal to each wrist joint.
For lower extremity movements (e.g., toe tapping (TT) and leg agility (LA)), repetitive movements were recorded at 80 Hz using two accelerometers placed at:
  • The anterior surface of each tibia, two inches proximal to the medial malleolus;
  • The dorsal surface of the proximal phalanx of each big toe.
Participants returned after a week or more for repeat testing. Figure 1 illustrates the sensor placement setup used for data collection.

2.2. Participants

We used an open-source dataset [19] that recorded 28 participants composed of 20 individuals with PD (14 males, mean age = 67 ± 10 years) and 8 HOAs (five males, mean age = 64 ± 7 years). Additionally, 19 of these participants had a retest session. This study utilized 16 recorded sessions from HOAs and 32 recorded sessions from individuals with PD. The data were collected in accordance with the Code of Ethics of the World Medical Association (Declaration of Helsinki) and the protocol was approved by the Johns Hopkins Institutional Review Board. Written informed consent was obtained from all participants utilizing a protocol (Protocol Number: IRB00110 166 and Initial Approval Date: 22 September 2016) approved by the Johns Hopkins Medical Institutions Institutional Review Board, Baltimore, MD, USA.

2.3. Deep Learning Hardware and Software Tools

All of our experiments were performed on an Apple M2 Max machine Ventura 13.4 using a 12-core CPU. We trained our models using tsai v0.3.6 [20], an open-source deep learning tool built on top of fastai [21] and PyTorch [22], in Python v3.9.17.

2.4. Data Preprocessing

For each participant’s session in which a task (FT, HM, PS, TT, LA) was performed, the resultant acceleration of the sensor i at timestep j was calculated as:
A i j = a x i j 2 + a y i j 2 + a z i j 2
where a x , a y , a z represent acceleration in the x-axis, y-axis, and z-axis. We then defined the acceleration sequence as:
S = A 1 , 1 A 2 , 1 A 1 , 2 A 2 , 2 A 1 , N A 2 , N
where the length N of the recording was arbitrary, and two accelerometers were used. The derivative of the resultant of each accelerometer, or jerk, was calculated. Further, the root mean square was taken over the time series sequences using a 1 s window. Since the length N of each time series sequence was arbitrary, overlapping snippets of five seconds in duration (i.e., 400 data points) from the recordings were segmented out. Each segmented sequence was standardized at an individual sample level (mean and standard deviation were calculated for each segmented sequence). Additionally, each of these segmented sequences shared a motor impairment severity label associated with the movement of the participant for the given task. These labels, originally scaled in the range of 0 to 4, were aggregated to either low (0/1) or high (3/4) levels of motor impairment. We ignored all sequences labelled as severity level 2.

2.5. Training, Validation, and Test Sets

The first step was to divide our sequences and labels into training, validation, and test sets, as Figure 2 displays. For each task, we selected the corresponding segmented sequences. All segmented sequences that originally belonged to the same original sequence were placed in the same group. In doing so, we prevented segmented sequences from the same test session from crossing over from training or validation sets to the test sets, which could artificially boost the model performance. We then filtered these groups into a set of groups for HOA participants and a set of groups for PD participants. The HOA groups were divided into a training set T r n H O A and a validation set V a l H O A with an 80-20% random split.
The set of PD groups went through a stratified K-fold, which divided the set of groups into K equally sized folds while maintaining the associated label distribution in each fold. We used K = 5 in our implementation. We iterated from i = 1 to K and set the test set T s t P D i to the i th fold. The training set T r n P D i and the validation set V a l P D i were set by a 70-30% random split of the remaining K 1 folds. The stratified K-fold guarantees that every segmented PD sequence is used at least once to evaluate the model’s performance via the test set. As such, it mitigates some of the bias surrounding model performance estimations that is caused by the arbitrary selection of training, validation, and test datasets.

2.6. XceptionTime Architecture Selection

XceptionTime [23] is a convolutional neural network architecture for time-series data, as shown in Figure 3. We used XceptionTime for in deep learning model in this paper since it has empirically been shown to perform well on motor impairment classification for individuals with PD [24]. Our experiment can be reproduced with other neural network architectures. The architecture used for the unsupervised models must be the same as the architectures used for the supervised models since our method involves transferring weights from the unsupervised models to the supervised models, which is only possible if their architectures are identical.

2.7. Unsupervised Learning: Procedure

Recent work [25] has demonstrated the ability of unsupervised neural networks to implicitly learn deep representations of multivariate time-series inputs when trained to predict masked portions. The pre-trained weights from such unsupervised models have been shown to offer performance benefits when transferred to supervised models for classification, even when the training datasets are small. We employed this method by independently masking a proportion r of the input with zeros. The masking was applied such that:
  • The length of each masked segment was drawn from a geometric distribution with a mean of l m .
  • Each succeeding unmasked segment had a mean length l u = l m 1 r r .
Our procedure used the hyper-parameters r = 0.15 and l m = 3 . We trained unsupervised XceptionTime models for 200 epochs with a maximum learning rate of 10 3 . They were optimized using the mean squared error loss function:
L M S E = 1 N i = 1 N ( y i y ^ i ) 2
where y i is the ith masked input, y ^ i is the predicted ith masked input, and N is the total number of masked inputs. Figure 4 displays an example of the masked predictions made by one of our unsupervised XceptionTime models.

2.8. Unsupervised Learning: Pre-Trained Weight Generation

Using the specifications detailed in Section 2.7, we trained unsupervised XceptionTime models for each task and saved their weights. A workflow of our unsupervised learning procedure can be found in Figure 5. For each task, we retrieved T r n H O A and V a l H O A . Then, we iterated from i = 1 to K and retrieved T r n P D i and V a l P D i . We used T r n P D i and V a l P D i to train and validate an unsupervised learning model, whose learned weights we saved as W P D i . Then, we trained another unsupervised learning model using the sequences from both T r n P D i and T r n H O A . This model was validated from both V a l P D i and V a l H O A . The learned weights of this model were saved as W P D , H O A i .

2.9. Supervised Learning Procedure

For each task, we iterated from i = 1 to K and trained 10 different supervised XceptionTime models. Of the ten models, five used T r n P D i as the training dataset, V a l P D i as the validation dataset, and T s t P D i as the test dataset to evaluate motor impairment classification performance. The other five used T r n P D i T r n H O A as the training dataset, V a l P D i V a l H O A as the validation dataset, and T s t P D i as the test dataset. Note that every test set neglected HOA data to only record the model performance on classifying motor impairment for PD patients.
In each group of five supervised XceptionTime models, one model used random initialization for its weights. It underwent training for 50 epochs with a maximum learning rate of 10 3 . Among the remaining four models, two were initialized with the pre-trained weights W P D i , while the other two were initialized with the pre-trained weights W P D , H O A i . For each pair of models using the same pre-trained weights for initialization, the first model had its head classification layers fine-tuned for 10 epochs. Subsequently, every layer was fine-tuned for an additional 50 epochs with a maximum learning rate of 10 3 . This technique is referred to as “fine-tune last, then all” (FTL). In contrast, for the second model in each pair, every layer was fine-tuned for 50 epochs with a maximum learning rate of 10 3 . This technique is referred to as “fine-tune all” (FTA).
All of the supervised XceptionTime models were optimized using the binary cross entropy loss function:
L B C E = 1 N i = 1 N ( y i log ( p i ) + ( 1 y i ) log ( 1 p i ) )
where N is the number of training sequences, y i is the true label (motor impairment level), and p i is the predicted probability of high motor impairment.
We considered the supervised XceptionTime model trained with T r n P D and initialized with random weights in the baseline model, since it does not use pre-trained weights or HOA data. A summary of all the models created can be found in Table 1.

2.10. Model Inference

Recall that the original sequences were segmented into 400 time steps. We evaluated two forms of inference from our supervised XceptionTime models on test datasets containing PD sequences. The first was segmented inference, where a motor impairment prediction was made from a single segmented sequence. The other was max-vote inference, where the motor impairment prediction for an original sequence was based on the mode of the predictions from its segmented sequences. For each form of inference, we saved and averaged metrics including the accuracy and F1 score.

2.11. Aggregate Models

In addition to classifying motor impairment in individual motor tasks (PD, HM, FT, TT, LA), we also trained aggregate unsupervised and supervised XceptionTime models. To do this, we followed the same steps as described in the previous Section 2.5, Section 2.6, Section 2.7, Section 2.8 and Section 2.9 without filtering the segmented sequences for a particular task.

2.12. Evaluating Robustness

In real-world applications, factors including administrator inexperience, positioning, and calibration issues can make accelerometer sensor measurements unexpected and noisy. To analyze how our models’ robustness to perturbations is correlated with their training set ( T r n P D , T r n P D T r n H O A ) and weight initialization (RandInit, W P D , W P D , H O A ), we added Gaussian noise to each segmented sequence in our test sets and re-evaluated the accuracies. The Gaussian noise was generated with mean μ = 0 and standard deviation σ = 0.05 .

2.13. Quantifying Representational Similarities

To understand how the internal behaviors of our supervised XceptionTime models for each task changed depending on their training set and initialization, we used centered kernel alignment (CKA) [26] to measure the feature similarity of the representations from convolutional layers between models with different training sets ( T r n P D , T r n P D T r n H O A ) and with different weight initializations (RandInit, W P D , W P D , H O A ). CKA is a recently introduced similarity index that measures the similarity of deep neural network representations. Specifically, we evaluated the similarity of the representations from both the first and last XceptionTime modules (Figure 3) within the XceptionTime block. We omitted all models that used the FTL fine-tuning scheme in order to obtain the same amount of models initialized with RandInit, W P D , and W P D , H O A . Figure 6 displays how the representations were aggregated by model attributes to compute CKA similarities.

3. Results and Discussion

3.1. Best Performing Models

Table 2 and Table 3 show the best performing models in each task using segmented and max-vote inference, respectively. For segmented inference, we show that expanding the training dataset to include T r n H O A , using pre-trained weights during initialization, or both yielded improvements in accuracy over the baseline for every task, as well as for aggregates. The same is reflected in max-vote inference, except for the TT task, in which the baseline model was tied for the highest accuracy.
Under max-vote inference, every model that showed at least 10% improvement over the baseline used T r n P D T r n H O A as the training set and W P D for weight initialization. Overall, the best performing models used T r n P D T r n H O A for four of five tasks (as well as for aggregates), and were initialized with W P D for four of five tasks. This provides some indication that HOA training data and pre-training can both boost accuracy for motor impairment classifiers. The model with the highest max-vote accuracy was trained on the HM task with T r n P D as its training set and was intialized with W P D , H O A . It achieved an average accuracy of 92%, which was a 12.2% improvement over the baseline model for the HM task.

3.2. Benefits of HOA Training Data

Table 4 shows the average accuracy for our supervised XceptionTime models, grouped by their training set attribute ( T r n P D , T r n P D T r n H O A ) and their weight initialization attribute (RandInit, W P D , W P D , H O A ). The models for three of the five tasks, as well as for aggregates, exhibited max-vote accuracy improvements when using T r n P D T r n H O A compared to just T r n P D . The models for PS improved the most with the inclusion of HOA training data, by 5.9% on average. Our results show that, generally, HOA data should be considered as an addition to training datasets for motor impairment classification. For every task except TT (and including the aggregate), adding HOA data to the training set produced comparable or improved average accuracies.
We suspect that the addition of HOA training data improves the performance by making models more generalizable and robust. Recent work [27] has demonstrated that for machine learning models on medical tasks, adding more subtypes to training data increases the variability of the dataset, reduces overfitting, and improves model robustness. Our experiments reflect this as well. As Table 4 displays, the average accuracies of our supervised XceptionTime models reduced considerably with Gaussian noise applied to each segmented sequence. Despite this, when trained on T r n P D T r n H O A compared to just T r n P D , the models for four of five tasks (with TT being the exception), as well as for the aggregate, exhibited both an improvement in the average max-vote accuracy and a reduction in the variance of model accuracy across all K trials. The models for FT improved the most, by 5.83% on average. The improvements that training with T r n P D T r n H O A provide for both the original and noisy motion sequences are equal [28], which highlights a correlation between better performing neural networks and neural networks that are more robust to perturbations.

3.3. Benefits of Pre-Training

As displayed by Table 4, the models for all five tasks exhibited max-vote accuracy improvements when initialized with pre-trained weights compared to random weights. While random initialization performed the best for the aggregate case, using either set of pre-trained weights ( W P D or W P D , H O A ) led to comparable results that were only 0.8% less accurate on average. In contrast, the models for HM and LA improved the most with pre-trained weight initialization. Specifically, the models for HM saw an average improvement of 5.6% when initialized with W P D , H O A compared to random weights. The models for LA saw an average improvement of 6.1% when initialized with W P D compared to random weights. Weight initialization with W P D , H O A performed best for LA, while initialization with W P D performed best for PS, HM, FT, and TT.
Overall, we have demonstrated that the use of pre-trained weights, learned from masked unsupervised learning (Section 2.7), increases the motor impairment classification performance. The recent literature [29,30] attributes this fact, in part, to the increased generalization that pre-trained weights provide, as they have already learned the prevalent features from the source domain. Our findings are consistent with prior work, which shows that for small datasets, transferring and fine-tuning pre-trained weights for PD classification specifically can lead to a better performance compared to models initialized with random weights [31].
Pre-trained weights have also been shown to improve the robustness of machine learning models [32,33,34] for downstream tasks, including PD classification [35]. However, as Table 4 shows, we observed no significant performance improvements resulting from pre-trained weights compared to randomly initialized weights when Gaussian noise was applied to the motion sequences. A potential reason for this disconnect may be that most of the literature on the effect of pre-training on robustness uses models trained on large datasets, such as ImageNet [36], to generate pre-trained weights. Our method, in contrast, re-uses the same data for both pre-training and classification (Section 2.8 and Section 2.9). Since the extensive variation in a large training set often leads to models that are more resistant to perturbations in input data [27], it might be the case that models pre-trained on ImageNet’s 14 million images are more robust simply due to the nature of the dataset size.
Despite the small dataset size used to generate our pre-trained weights, the max-vote accuracy improvements that we presented in all tasks in the non-perturbed case support the notion that large-scale datasets are not necessary for pre-training; even unsupervised pre-training on small-scale datasets can provide a boost in performance [25,37].

3.4. Similarities between Learned Representations

Table 5 and Table 6, respectively, show the CKA similarity for representations from the first and last XceptionTime modules between models with different training sets ( T r n P D , T r n P D T r n H O A ) and with different weight initializations (RandInit, W P D , W P D , H O A ). Overall, we can observe a dissimilarity between the representations from models with different attributes, supporting the idea that the choice of training data and weight initialization for a model has a significant impact on its feature space [26,38]. We do note some exceptions to this, namely for the aggregate case. The similarities between representations from the last XceptionTime module seem to be consistently higher than those from other individual tasks. One possible explanation is that since filtering motion sequences for specific tasks is not executed in the aggregate case, the models are trained on more data, which can lead to similar feature spaces. However, more research on this is required to give a more conclusive explanation.
Another important finding is that the feature similarities between models initialized with W P D and W P D , H O A are higher than the similarities between the same models and those using RandInit for the majority of the cases. This corroborates the findings in [38], highlighting the role of feature re-use when using pre-trained weights, a phenomenon not seen with random initialization.

3.5. Performance Differences among Tasks

Consistent with prior work [24], we found hand movement tasks to yield the best overall classification performance, considering both the segmented and max-vote accuracy (see Table 2 and Table 3). Using segmented and max-vote data, the best overall classification performance was achieved using the hand movement task, consistent with the identification of PD-related motor changes during virtual reach-to-grasp movements [39]. However, the greatest improvement from the baseline segmented performance was in finger tapping tasks, which is consistent with the use of kinematic data from finger tapping in prior efforts to classify PD-related motor changes [40].
Table 4 shows that, among all tasks, the max-vote accuracy increased the most when adding HOA training data for the pronation–supination hand task (about 6% improvement). This finding may be due to the fact that patients who claim to be healthy often actually have limited pronation in the forearm, since pronation can be recompensed by shoulder abduction and elbow flexion [41]. As a result, motion data from HOA may be characteristically more similar to motion data from PD patients for pronation–supination compared to other tasks. Thus, future studies with healthy young adults as participants would help in providing a benchmark for both aging and PD-related changes in upper extremity functions.
Furthermore, while most tasks demonstrated a considerable improvement over baseline models, toe tapping tasks did not demonstrate an improvement in performance. These findings may be partly due to the dissimilarity of kinematics in healthy controls relative to PD participants while toe tapping, but necessitate further investigation. The current findings are consistent with the prior findings of a good reliability of bradykinesia evaluations using individual upper extremity movements [42] and bradykinesia evaluations using hand grasping tasks [43].

4. Conclusions

In this study, we find that expanding training datasets with HOA motion data using unsupervised learning and transferring pre-trained weights for initialization can improve the motor impairment classification performance of a continuous motion data monitor for PD patients in telemedicine applications. Our method shows particular promise in low-data domains as it does not use additional PD or annotated data compared to the baseline classification model. Thus, the incorporation of data from healthy individuals could be instrumental in refining deep learning classification models tailored for neurological disorders, serving as both a baseline and a point of reference. Such an approach could pave the way for the development of extensive datasets, since acquiring data from healthy individuals is more straightforward and may provide an easier way to enhance models in future work.
In particular, we find that the addition of HOA training data improves both the classification accuracy and the robustness for most tasks in a clinical motor impairment evaluation. Furthermore, we show that unsupervised pre-training, even on small scale datasets, can provide a boost in motor impairment classification performance. Our observation that CKA feature similarities are generally higher between pre-trained classifiers than between a pre-trained and randomly initialized classifier suggests that pre-trained models exhibit feature re-use.
While these result are promising, future work should focus on evaluating the current approach in a larger cohort by either collecting more PD and HOA training data, collecting healthy young adult data to establish baseline age-related changes, or by generating synthetic data with generative machine learning models. Additionally, a natural expansion of this work is to consider methods to expand the feature selection of the input data, for instance, by using data from additional IMU sensors placed on different parts of the upper or lower extremities, using spectral components of IMU motion data (with fast Fourier transforms or spectrograms), or using multiple orders of differentiation of IMU motion data. A multi-modal approach should also be considered, which uses data from different domains such as muscle activations (retrieved from electromyogram sensors).

Author Contributions

Conceptualization, R.S., J.R.B. and M.E.H.; methodology, P.P. and M.S.; formal analysis, P.P., M.S. and R.K.; data curation, J.R.B., P.P., M.S. and R.K.; writing—original draft preparation, P.P., M.S. and M.E.H.; visualization, M.S.; supervision, R.S., J.R.B. and M.E.H.; project administration, J.R.B. and M.E.H.; funding acquisition, J.R.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. Written informed consent was obtained from all participants utilizing a protocol (Protocol Number: IRB00110 166 and Initial Approval Date: 22 September 2016) approved by Johns Hopkins Medical Institutions Institutional Review Board, Baltimore, MD, USA.

Informed Consent Statement

Written informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data used for analysis are available in an open-source dataset [19].

Acknowledgments

The International Parkinson and Movement Disorder Society kindly gave permission to use the Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS) [7]. We would like to thank Timothy Harrigan for instrumentation and data collection, Joshua Levy for assisting with date pre-processing, as well as all the volunteers who participated in this study.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CKAcentered kernel alignment
FTAfine-tune all
FTLfine-tune last, then all
FTfinger tapping
HMhand movement
HOAhealthy older adult
LAleg agility
MVmax-vote
PDParkinson’s disease
PSpronation–supination
RandInitrandom initialization
TTtoe tapping
Wweights

References

  1. de Lau, L.M.; Breteler, M.M. Epidemiology of Parkinson’s disease. Lancet Neurol. 2006, 5, 525–535. [Google Scholar] [CrossRef] [PubMed]
  2. Postuma, R.B.; Berg, D.; Stern, M.; Poewe, W.; Olanow, C.W.; Oertel, W.; Obeso, J.; Marek, K.; Litvan, I.; Lang, A.E.; et al. MDS clinical diagnostic criteria for Parkinson’s disease. Mov. Disord. 2015, 30, 1591–1601. [Google Scholar] [CrossRef]
  3. van der Eijk, M.; Faber, M.J.; Al Shamma, S.; Munneke, M.; Bloem, B.R. Moving towards patient-centered healthcare for patients with Parkinson’s disease. Park. Relat. Disord. 2011, 17, 360–364. [Google Scholar] [CrossRef] [PubMed]
  4. Pretzer-Aboff, I.; Prettyman, A. Implementation of an integrative holistic healthcare model for people living with Parkinson’s disease. Gerontologist 2015, 55, S146–S153. [Google Scholar] [CrossRef] [PubMed]
  5. Butler, R.N. Population aging and health. BMJ 1997, 315, 1082–1084. [Google Scholar] [CrossRef] [PubMed]
  6. Ben-Pazi, H.; Browne, P.; Chan, P.; Cubo, E.; Guttman, M.; Hassan, A.; Hatcher-Martin, J.; Mari, Z.; Moukheiber, E.; Okubadejo, N.; et al. The promise of telemedicine for movement disorders: An interdisciplinary approach. Curr. Neurol. Neurosci. Rep. 2018, 18, 26. [Google Scholar] [CrossRef]
  7. Goetz, C.G.; Tilley, B.C.; Shaftman, S.R.; Stebbins, G.T.; Fahn, S.; Martinez-Martin, P.; Poewe, W.; Sampaio, C.; Stern, M.B.; Dodel, R.; et al. Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS): Scale presentation and clinimetric testing results. Mov. Disord. Off. J. Mov. Disord. Soc. 2008, 23, 2129–2170. [Google Scholar] [CrossRef]
  8. Shalash, A.; Spindler, M.; Cubo, E. Global perspective on telemedicine for Parkinson’s disease. J. Park. Dis. 2021, 11, S11–S18. [Google Scholar] [CrossRef]
  9. Achey, M.; Aldred, J.L.; Aljehani, N.; Bloem, B.R.; Biglan, K.M.; Chan, P.; Cubo, E.; Ray Dorsey, E.; Goetz, C.G.; Guttman, M.; et al. The past, present, and future of telemedicine for Parkinson’s disease. Mov. Disord. 2014, 29, 871–883. [Google Scholar] [CrossRef] [PubMed]
  10. van den Bergh, R.; Bloem, B.R.; Meinders, M.J.; Evers, L.J. The state of telemedicine for persons with Parkinson’s disease. Curr. Opin. Neurol. 2021, 34, 589. [Google Scholar] [CrossRef] [PubMed]
  11. Vásquez-Correa, J.C.; Arias-Vergara, T.; Orozco-Arroyave, J.R.; Eskofier, B.; Klucken, J.; Nöth, E. Multimodal Assessment of Parkinson’s Disease: A Deep Learning Approach. IEEE J. Biomed. Health Inform. 2019, 23, 1618–1630. [Google Scholar] [CrossRef]
  12. Kotsavasiloglou, C.; Kostikis, N.; Hristu-Varsakelis, D.; Arnaoutoglou, M. Machine learning-based classification of simple drawing movements in Parkinson’s disease. Biomed. Signal Process. Control 2017, 31, 174–180. [Google Scholar] [CrossRef]
  13. Kaur, R.; Motl, R.W.; Sowers, R.; Hernandez, M.E. A Vision-Based Framework for Predicting Multiple Sclerosis and Parkinson’s Disease Gait Dysfunctions-A Deep Learning Approach. IEEE J. Biomed. Health Inform. 2022, 27, 190–201. [Google Scholar] [CrossRef]
  14. Hu, Y.; Bishnoi, A.; Kaur, R.; Sowers, R.; Hernandez, M.E. Exploration of machine learning to identify community dwelling older adults with balance dysfunction using short duration accelerometer data. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 812–815. [Google Scholar] [CrossRef]
  15. Kaur, R.; Chen, Z.; Motl, R.; Hernandez, M.E.; Sowers, R. Predicting multiple sclerosis from gait dynamics using an instrumented treadmill: A machine learning approach. IEEE Trans. Biomed. Eng. 2020, 68, 2666–2677. [Google Scholar] [CrossRef]
  16. Willemink, M.J.; Koszek, W.A.; Hardell, C.; Wu, J.; Fleischmann, D.; Harvey, H.; Folio, L.R.; Summers, R.M.; Rubin, D.L.; Lungren, M.P. Preparing Medical Imaging Data for Machine Learning. Radiology 2020, 295, 4–15. [Google Scholar] [CrossRef] [PubMed]
  17. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A comprehensive survey on transfer learning. Proc. IEEE 2020, 109, 43–76. [Google Scholar] [CrossRef]
  18. McKay, G.N.; Harrigan, T.P.; Brašić, J.R. A low-cost quantitative continuous measurement of movements in the extremities of people with Parkinson’s disease. MethodsX 2019, 6, 169–189. [Google Scholar] [CrossRef]
  19. Harrigan, T.P.; Hwang, B.J.; Mathur, A.K.; Mills, K.A.; Pantelyat, A.Y.; Bang, J.A.; Syed, A.B.; Vyas, P.; Martin, S.D.; Jamal, A.; et al. Dataset of quantitative structured office measurements of movements in the extremities. Data Brief 2020, 31, 105876. [Google Scholar] [CrossRef]
  20. Oguiza, I. Tsai—A State-of-the-Art Deep Learning Library for Time Series and Sequential Data; Github: San Francisco, CA, USA, 2022. [Google Scholar]
  21. Howard, J.; Gugger, S.; Mueller, Z.; Warner, B.; Bird, K.; Moe, M.; Abraham, T.; Capelle, T.; Malleswaram, M.; Husian, H.; et al. Fastai. 2018. Available online: https://github.com/fastai/fastai (accessed on 6 January 2023).
  22. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
  23. Rahimian, E.; Zabihi, S.; Atashzar, S.F.; Asif, A.; Mohammadi, A. XceptionTime: A Novel Deep Architecture based on Depthwise Separable Convolutions for Hand Gesture Classification. arXiv 2019, arXiv:1911.03803. [Google Scholar]
  24. Prakash, P.; Kaur, R.; Levy, J.; Sowers, R.; Brašić, J.R.; Hernandez, M.E. A Deep Learning Approach for Grading of Motor Impairment Severity in Parkinson’s Disease. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, Australia, 24–27 July 2023. in press. [Google Scholar]
  25. Zerveas, G.; Jayaraman, S.; Patel, D.; Bhamidipaty, A.; Eickhoff, C. A Transformer-based Framework for Multivariate Time Series Representation Learning. arXiv 2020, arXiv:2010.02803. [Google Scholar]
  26. Kornblith, S.; Norouzi, M.; Lee, H.; Hinton, G.E. Similarity of Neural Network Representations Revisited. arXiv 2019, arXiv:1905.00414. [Google Scholar]
  27. Tang, H.; Sun, N.; Shen, S. Improving Generalization of Deep Learning Models for Diagnostic Pathology by Increasing Variability in Training Data: Experiments on Osteosarcoma Subtypes. J. Pathol. Inform. 2021, 12, 30. [Google Scholar] [CrossRef]
  28. Rozsa, A.; Günther, M.; Boult, T.E. Are Accuracy and Robustness Correlated. In Proceedings of the 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA), Anaheim, CA, USA, 18–20 December 2016; pp. 227–232. [Google Scholar] [CrossRef]
  29. Peng, A.Y.; Sing Koh, Y.; Riddle, P.; Pfahringer, B. Using Supervised Pretraining to Improve Generalization of Neural Networks on Binary Classification Problems. In Proceedings of the Machine Learning and Knowledge Discovery in Databases, Dublin, Ireland, 10–14 September 2019; Berlingerio, M., Bonchi, F., Gärtner, T., Hurley, N., Ifrim, G., Eds.; Springer: Cham, Switzerland, 2019; pp. 410–425. [Google Scholar]
  30. Chen, S.; Ma, K.; Zheng, Y. Med3D: Transfer Learning for 3D Medical Image Analysis. arXiv 2019, arXiv:1904.00625. [Google Scholar]
  31. Arasteh, E.; Mahdizadeh, A.; Mirian, M.S.; Lee, S.; McKeown, M.J. Deep Transfer Learning for Parkinson’s Disease Monitoring by Image-Based Representation of Resting-State EEG Using Directional Connectivity. Algorithms 2022, 15, 5. [Google Scholar] [CrossRef]
  32. Hendrycks, D.; Lee, K.; Mazeika, M. Using Pre-Training Can Improve Model Robustness and Uncertainty. arXiv 2019, arXiv:1901.09960. [Google Scholar]
  33. Hendrycks, D.; Mazeika, M.; Kadavath, S.; Song, D. Using Self-Supervised Learning Can Improve Model Robustness and Uncertainty. arXiv 2019, arXiv:1906.12340. [Google Scholar]
  34. Djolonga, J.; Yung, J.; Tschannen, M.; Romijnders, R.; Beyer, L.; Kolesnikov, A.; Puigcerver, J.; Minderer, M.; D’Amour, A.; Moldovan, D.; et al. On Robustness and Transferability of Convolutional Neural Networks. arXiv 2021, arXiv:/2007.08558. [Google Scholar]
  35. Rezaee, K.; Savarkar, S.; Yu, X.; Zhang, J. A hybrid deep transfer learning-based approach for Parkinson’s disease classification in surface electromyography signals. Biomed. Signal Process. Control 2022, 71, 103161. [Google Scholar] [CrossRef]
  36. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  37. El-Nouby, A.; Izacard, G.; Touvron, H.; Laptev, I.; Jegou, H.; Grave, E. Are Large-scale Datasets Necessary for Self-Supervised Pre-training? arXiv 2021, arXiv:2112.10740. [Google Scholar]
  38. Neyshabur, B.; Sedghi, H.; Zhang, C. What is being transferred in transfer learning? arXiv 2020, arXiv:2008.11687. [Google Scholar]
  39. Weyhenmeyer, J.; Hernandez, M.E.; Lainscsek, C.; Poizner, H.; Sejnowski, T.J. Multimodal classification of Parkinson’s disease using delay differential analysis. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Republic of Korea, 16–19 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 2868–2875. [Google Scholar] [CrossRef]
  40. Lainscsek, C.; Rowat, P.; Schettino, L.; Lee, D.; Song, D.; Letellier, C.; Poizner, H. Finger tapping movements of Parkinson’s disease patients automatically rated using nonlinear delay differential equations. Chaos Interdiscip. J. Nonlinear Sci. 2012, 22, 013119. [Google Scholar] [CrossRef] [PubMed]
  41. Rickert, M.; Bürger, A.; Günther, C.M.; Schulz, C.U. Forearm rotation in healthy adults of all ages and both sexes. J. Shoulder Elb. Surg. 2008, 17, 271–275. [Google Scholar] [CrossRef]
  42. Heldman, D.A.; Giuffrida, J.P.; Chen, R.; Payne, M.; Mazzella, F.; Duker, A.P.; Sahay, A.; Kim, S.J.; Revilla, F.J.; Espay, A.J. The modified bradykinesia rating scale for Parkinson’s disease: Reliability and comparison with kinematic measures. Mov. Disord. 2011, 26, 1859–1863. [Google Scholar] [CrossRef] [PubMed]
  43. Lin, Z.; Xiong, Y.; Cai, G.; Dai, H.; Xia, X.; Tan, Y.; Lueth, T.C. Quantification of parkinsonian bradykinesia based on axis-angle representation and SVM multiclass classification method. IEEE Access 2018, 6, 26895–26903. [Google Scholar] [CrossRef]
Figure 1. The setup for sensor placement for upper extremity and lower extremity data collection. Images of the hand and the foot are reproduced with permission [18].
Figure 1. The setup for sensor placement for upper extremity and lower extremity data collection. Images of the hand and the foot are reproduced with permission [18].
Sensors 23 09004 g001
Figure 2. The divisions of PD and HOA sequences used to create training, validation, and test groups. Abbreviations: HOA = healthy older adult; PD = Parkinson’s disease.
Figure 2. The divisions of PD and HOA sequences used to create training, validation, and test groups. Abbreviations: HOA = healthy older adult; PD = Parkinson’s disease.
Sensors 23 09004 g002
Figure 3. XceptionTime (XT) block architecture. Abbreviation: B = batch size; BN = batch normalization; f = the number of output channels 4 .
Figure 3. XceptionTime (XT) block architecture. Abbreviation: B = batch size; BN = batch normalization; f = the number of output channels 4 .
Sensors 23 09004 g003
Figure 4. Example predictions, made by one of our unsupervised XceptionTime models, of masked portions of an input motion sequence from an HOA participant. The two colors correspond to the two sensors that make up the multivariate time-series sequence. Abbreviations: HOA = healthy older adult.
Figure 4. Example predictions, made by one of our unsupervised XceptionTime models, of masked portions of an input motion sequence from an HOA participant. The two colors correspond to the two sensors that make up the multivariate time-series sequence. Abbreviations: HOA = healthy older adult.
Sensors 23 09004 g004
Figure 5. The combination of training and validation sets used to train unsupervised learning models and generate pre-trained weights. Abbreviations: HOA = healthy older adult; PD = Parkinson’s disease; W = weights.
Figure 5. The combination of training and validation sets used to train unsupervised learning models and generate pre-trained weights. Abbreviations: HOA = healthy older adult; PD = Parkinson’s disease; W = weights.
Sensors 23 09004 g005
Figure 6. A workflow depicting the process of computing CKA similarity indexes for the representations among models with different initializations and training sets. Abbreviations: CKA = centered kernel alignment; FTA = fine-tune all; FTL = fine-tune last, then all; HOA = healthy older adult; PD = Parkinson’s disease; RandInit = random initialization; W = weights.
Figure 6. A workflow depicting the process of computing CKA similarity indexes for the representations among models with different initializations and training sets. Abbreviations: CKA = centered kernel alignment; FTA = fine-tune all; FTL = fine-tune last, then all; HOA = healthy older adult; PD = Parkinson’s disease; RandInit = random initialization; W = weights.
Sensors 23 09004 g006
Table 1. A summary of the attributes of each of the 10 supervised XceptionTime models we trained. Abbreviations: FTA = fine-tune all; FTL = fine-tune last, then all; HOA = healthy older adult; PD = Parkinson’s disease; RandInit = random initialization; W = weight.
Table 1. A summary of the attributes of each of the 10 supervised XceptionTime models we trained. Abbreviations: FTA = fine-tune all; FTL = fine-tune last, then all; HOA = healthy older adult; PD = Parkinson’s disease; RandInit = random initialization; W = weight.
Training DatasetWeight InitializationFine-Tuning Scheme
T r n P D RandInit-
T r n P D W P D FTL
T r n P D W P D FTA
T r n P D W P D , H O A FTL
T r n P D W P D , H O A FTA
T r n P D T r n H O A RandInit-
T r n P D T r n H O A W P D FTL
T r n P D T r n H O A W P D FTA
T r n P D T r n H O A W P D , H O A FTL
T r n P D T r n H O A W P D , H O A FTA
Table 2. The models with the best segmented accuracy (and F1 score as tie-breaker) in each task. The best accuracy and best accuracy increases from baseline have been highlighted in bold. Abbreviations: FTA = fine-tune all; FTL = fine-tune last, then all; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; PD = Parkinson’s disease; PS = pronation–supination; TT = toe tapping; W = weights.
Table 2. The models with the best segmented accuracy (and F1 score as tie-breaker) in each task. The best accuracy and best accuracy increases from baseline have been highlighted in bold. Abbreviations: FTA = fine-tune all; FTL = fine-tune last, then all; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; PD = Parkinson’s disease; PS = pronation–supination; TT = toe tapping; W = weights.
TaskBaseline Model Best Model
Avg. Seg. Acc.Train SetWeight Init.Fine-TuningAvg. Seg. Acc.Δ from Baseline
PS0.612 ± 0.087 T r n P D T r n H O A W P D FTA0.701 ± 0.0660.089
HM0.744 ± 0.059 T r n P D W P D , H O A FTA0.789 ± 0.0720.045
FT0.586 ± 0.096 T r n P D T r n H O A W P D FTL0.680 ± 0.0900.094
TT0.693 ± 0.115 T r n P D W P D FTA0.728 ± 0.0980.035
LA0.695 ± 0.077 T r n P D W P D , H O A FTL0.773 ± 0.1340.078
Agg.0.745 ± 0.069 T r n P D W P D FTA0.760 ± 0.0160.015
Table 3. The models with the best max-vote accuracy (and F1 score as tie-breaker) on each task. The best accuracy and best accuracy increases from baseline have been highlighted in bold. Abbreviations: FTA = fine-tune all; FTL = fine-tune last, then all; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; MV = max-vote; PD = Parkinson’s disease; PS = pronation–supination; RandInit = random initialization; TT = toe tapping; W = weights.
Table 3. The models with the best max-vote accuracy (and F1 score as tie-breaker) on each task. The best accuracy and best accuracy increases from baseline have been highlighted in bold. Abbreviations: FTA = fine-tune all; FTL = fine-tune last, then all; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; MV = max-vote; PD = Parkinson’s disease; PS = pronation–supination; RandInit = random initialization; TT = toe tapping; W = weights.
TaskBaseline Model Best Model
Avg. MV Acc.Train SetWeight Init.Fine-TuningAvg. MV Acc.Δ from Baseline
PS0.594 ± 0.197 T r n P D T r n H O A W P D FTA0.694 ± 0.1210.100
HM0.798 ± 0.087 T r n P D T r n H O A W P D FTL0.920 ± 0.0750.122
FT0.647 ± 0.098 T r n P D T r n H O A RandInit-0.713 ± 0.1420.066
TT0.775 ± 0.122 T r n P D W P D FTA0.775 ± 0.1220.000
LA0.756 ± 0.046 T r n P D T r n H O A W P D FTL0.860 ± 0.0800.104
Agg.0.777 ± 0.047 T r n P D T r n H O A RandInit-0.812 ± 0.0360.035
Table 4. The average accuracy of models, grouped by attribute, on test sets for each task. Performance metrics on both test sets with and without Gaussian noise added are recorded. Bold indicates the best max-vote accuracy among the different choices for each model attribute. Abbreviations: FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; MV = max-vote; PD = Parkinson’s disease; PS = pronation–supination; RandInit = random initialization; TT = toe tapping; W = weights.
Table 4. The average accuracy of models, grouped by attribute, on test sets for each task. Performance metrics on both test sets with and without Gaussian noise added are recorded. Bold indicates the best max-vote accuracy among the different choices for each model attribute. Abbreviations: FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; MV = max-vote; PD = Parkinson’s disease; PS = pronation–supination; RandInit = random initialization; TT = toe tapping; W = weights.
TaskModel AttributesWithout Gaussian NoiseWith Gaussian Noise
Avg. Seg. Acc.Avg. MV Acc.Avg. Seg. Acc.Avg. MV Acc.
PSTrain Set T r n P D 0.602 ± 0.0970.603 ± 0.1480.596 ± 0.0780.518 ± 0.093
T r n P D T r n H O A 0.649 ± 0.0970.662 ± 0.1130.606 ± 0.0900.556 ± 0.087
Weight InitRandInit0.631 ± 0.0840.632 ± 0.1620.619 ± 0.0630.544 ± 0.069
W P D 0.628 ± 0.0880.642 ± 0.1270.599 ± 0.0850.544 ± 0.069
W P D , H O A 0.620 ± 0.1160.624 ± 0.1260.594 ± 0.0910.526 ± 0.117
HMTrain Set T r n P D 0.753 ± 0.0620.837 ± 0.0770.631 ± 0.0880.655 ± 0.133
T r n P D T r n H O A 0.748 ± 0.0690.832 ± 0.0900.649 ± 0.0860.702 ± 0.097
Weight InitRandInit0.739 ± 0.0660.796 ± 0.0920.624 ± 0.1060.636 ± 0.176
W P D 0.756 ± 0.0540.852 ± 0.0780.642 ± 0.0810.683 ± 0.109
W P D , H O A 0.751 ± 0.0750.837 ± 0.0770.645 ± 0.0820.695 ± 0.08
FTTrain Set T r n P D 0.612 ± 0.1020.668 ± 0.1110.538 ± 0.1040.587 ± 0.145
T r n P D T r n H O A 0.633 ± 0.0860.690 ± 0.0930.565 ± 0.0860.640 ± 0.099
Weight InitRandInit0.598 ± 0.0930.680 ± 0.1260.527 ± 0.1070.573 ± 0.175
W P D 0.636 ± 0.0950.698 ± 0.0880.539 ± 0.0910.614 ± 0.109
W P D , H O A 0.622 ± 0.0930.658 ± 0.1010.576 ± 0.0910.632 ± 0.109
TTTrain Set T r n P D 0.693 ± 0.1120.775 ± 0.1060.554 ± 0.1090.450 ± 0.071
T r n P D T r n H O A 0.619 ± 0.1250.715 ± 0.1820.523 ± 0.1300.415 ± 0.126
Weight InitRandInit0.673 ± 0.1120.738 ± 0.1310.565 ± 0.0970.450 ± 0.061
W P D 0.652 ± 0.1320.756 ± 0.1650.523 ± 0.1300.415 ± 0.126
W P D , H O A 0.652 ± 0.1230.738 ± 0.1470.541 ± 0.1230.438 ± 0.093
LATrain Set T r n P D 0.732 ± 0.1040.810 ± 0.08190.726 ± 0.1190.798 ± 0.066
T r n P D T r n H O A 0.746 ± 0.0930.830 ± 0.0670.750 ± 0.1010.806 ± 0.052
Weight InitRandInit0.715 ± 0.0890.777 ± 0.0570.739 ± 0.1160.798 ± 0.06
W P D 0.735 ± 0.1040.823 ± 0.0750.731 ± 0.1090.798 ± 0.060
W P D , H O A 0.755 ± 0.0950.838 ± 0.0740.744 ± 0.1090.808 ± 0.059
Agg.Train Set T r n P D 0.752 ± 0.0350.783 ± 0.0390.588 ± 0.0520.532 ± 0.097
T r n P D T r n H O A 0.753 ± 0.0470.793 ± 0.0620.603 ± 0.0460.563 ± 0.089
Weight InitRandInit0.748 ± 0.0580.794 ± 0.0450.613 ± 0.0580.585 ± 0.106
W P D 0.753 ± 0.0300.786 ± 0.050.609 ± 0.0370.564 ± 0.083
W P D , H O A 0.753 ± 0.0420.786 ± 0.0580.573 ± 0.0470.513 ± 0.086
Table 5. Similarities between the representations from the first and last XceptionTime module for each supervised model, grouped by training set. Abbreviations: CKA = centered kernel alignment; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; PD = Parkinson’s disease; PS = pronation–supination; TT = toe tapping.
Table 5. Similarities between the representations from the first and last XceptionTime module for each supervised model, grouped by training set. Abbreviations: CKA = centered kernel alignment; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; PD = Parkinson’s disease; PS = pronation–supination; TT = toe tapping.
ModuleTrain Sets to CompareCKA Similarity
PSHMFTTTLAAgg.
First T r n P D / T r n P D T r n H O A 0.4060.4510.3440.4810.5540.338
Last T r n P D / T r n P D T r n H O A 0.4560.1580.4600.6500.4310.823
Table 6. Similarities between the representations from the first and last XceptionTime module for each supervised model, grouped by weight initialization. Abbreviations: CKA = centered kernel alignment; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; PD = Parkinson’s disease; PS = pronation–supination; RandInit = random initialization; TT = toe tapping; W = weights.
Table 6. Similarities between the representations from the first and last XceptionTime module for each supervised model, grouped by weight initialization. Abbreviations: CKA = centered kernel alignment; FT = finger tapping; HM = hand movements; HOA = healthy older adult; LA = leg agility; PD = Parkinson’s disease; PS = pronation–supination; RandInit = random initialization; TT = toe tapping; W = weights.
ModuleInits. to CompareCKA Similarity
PSHMFTTTLAAgg.
FirstRandInit/ W P D 0.1340.1180.1250.2210.0900.143
RandInit/ W P D , H O A 0.0990.1770.2910.0630.0900.097
W P D / W P D , H O A 0.5580.2090.4820.2330.5840.225
LastRandInit/ W P D 0.1230.4370.3710.3050.3440.855
RandInit/ W P D , H O A 0.1510.1020.4840.3070.4750.629
W P D / W P D , H O A 0.2450.1680.4530.3270.8080.842
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Singh, M.; Prakash, P.; Kaur, R.; Sowers, R.; Brašić, J.R.; Hernandez, M.E. A Deep Learning Approach for Automatic and Objective Grading of the Motor Impairment Severity in Parkinson’s Disease for Use in Tele-Assessments. Sensors 2023, 23, 9004. https://0-doi-org.brum.beds.ac.uk/10.3390/s23219004

AMA Style

Singh M, Prakash P, Kaur R, Sowers R, Brašić JR, Hernandez ME. A Deep Learning Approach for Automatic and Objective Grading of the Motor Impairment Severity in Parkinson’s Disease for Use in Tele-Assessments. Sensors. 2023; 23(21):9004. https://0-doi-org.brum.beds.ac.uk/10.3390/s23219004

Chicago/Turabian Style

Singh, Mehar, Prithvi Prakash, Rachneet Kaur, Richard Sowers, James Robert Brašić, and Manuel Enrique Hernandez. 2023. "A Deep Learning Approach for Automatic and Objective Grading of the Motor Impairment Severity in Parkinson’s Disease for Use in Tele-Assessments" Sensors 23, no. 21: 9004. https://0-doi-org.brum.beds.ac.uk/10.3390/s23219004

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop