Next Article in Journal
Aggressive Variant of Hepatic Epstein–Barr Virus-Associated Inflammatory Pseudotumor-like Follicular Dendritic Cell Sarcoma with PD-L1 and SSTR2a Expression
Next Article in Special Issue
Artificial Intelligence Image Recognition System for Preventing Wrong-Site Upper Limb Surgery
Previous Article in Journal
Automatic Heart Rate Detection during Sleep Using Tracheal Audio Recordings from Wireless Acoustic Sensor
Previous Article in Special Issue
Predicting CTS Diagnosis and Prognosis Based on Machine Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Machine Learning Algorithms for Prognostic Assessment in Rotator Cuff Pathologies: A Clinical Data-Based Approach

1
Orthopaedic and Trauma Surgery, Fondazione Policlinico Universitario Campus Bio-Medico, Via Alvaro del Portillo 200, 00128 Rome, Italy
2
Research Unit of Orthopaedic and Trauma Surgery, Department of Medicine and Surgery, Università Campus Bio-Medico di Roma, Via Alvaro del Portillo 21, 00128 Rome, Italy
3
Institute for Biomedical Research and Innovation (IRIB), National Research Council of Italy (CNR), 98164 Messina, Italy
4
Department of Engineering, Università Campus Bio-Medico di Roma, Via Alvaro del Portillo 21, 00128 Rome, Italy
5
Hospital for Special Surgery, New York, NY 10021, USA
6
S’Anna Institute, 88900 Crotone, Italy
7
Pharmacotechnology Documentation and Transfer Unit, Preclinical and Translational Pharmacology, Department of Pharmacy, Health Science and Nutrition, University of Calabria, 87036 Arcavacata di Rende, Italy
*
Authors to whom correspondence should be addressed.
Submission received: 1 August 2023 / Revised: 30 August 2023 / Accepted: 4 September 2023 / Published: 11 September 2023
(This article belongs to the Special Issue Artificial Intelligence in Orthopedic Surgery and Sport Medicine)

Abstract

:
Aim: The overall aim of this proposal is to ameliorate the care of rotator cuff (RC) tear patients by applying an innovative machine learning approach for outcome prediction after arthroscopic repair. Materials and Methods: We applied state-of-the-art machine learning algorithms to evaluate the best predictors of the outcome, and 100 RC patients were evaluated at baseline (T0), after 1 month (T1), 3 months (T2), 6 months (T3), and 1 year (T4) from surgical intervention. The outcome measure was the Costant–Murley Shoulder Score, whereas age, sex, BMI, the 36-Item Short-Form Survey, the Simple Shoulder Test, the Hospital Anxiety and Depression Scale, the American Shoulder and Elbow Surgeons Score, the Oxford Shoulder Score, and the Shoulder Pain and Disability Index were considered as predictive factors. Support vector machine (SVM), k-nearest neighbors (k-NN), naïve Bayes (NB), and random forest (RF) algorithms were employed. Results: Across all sessions, the classifiers demonstrated suboptimal performance when using both the complete and shrunken sets of features. Specifically, the logistic regression (LR) classifier achieved a mean accuracy of 46.5% ± 6%, while the random forest (RF) classifier achieved 51.25% ± 4%. For the shrunken set of features, LR obtained a mean accuracy of 48.5% ± 6%, and RF achieved 45.5% ± 4.5%. No statistical differences were found when comparing the performance metrics of ML algorithms. Conclusions: This study underlines the importance of extending the application of AI methods to new predictors, such as neuroimaging and kinematic data, in order to better record significant shifts in RC patients’ prognosis. Limitations: The data quality within the cohort could represent a limitation, since certain variables, such as smoking, diabetes, and work injury, are known to have an impact on the outcome.

1. Introduction

The most common etiology of shoulder discomfort is rotator cuff (RC) disease, which may account for up to 70% of all consultations involving the shoulder [1,2,3]. It has been demonstrated that RC tears result in substantial discomfort and impairment, as well as reduced performance in daily living tasks [3,4]. Even though surgery is widely accepted around the world [5,6], the possibility of re-tearing is still a significant postoperative obstacle [5,6,7,8,9]. A re-tear is considered a significant risk according to the age, the extent of the first lesion, and the fatty degeneration of the patient [10,11].
About 25% of patients with RCTs experience anxiety or depression, and psychological health may be a key indicator of how well a patient will recover from arthroscopic rotator cuff surgery [12]. Nonetheless, in medical care, the functional assessment frequently focuses on the objective aspects of the illness, such as measuring range of motion (ROM) and strength [12,13]. In light of this, orthopedic research has progressed, and the creation of well-established, patient-oriented metrics has offered clinical outcome evaluation a new perspective, in addition to objective measurements [14]. Patient-Reported Outcome Measures (PROMs) are frequently employed to assess the patients’ wellness [15]. Bypassing the clinician’s evaluation, PROMs are subjective patient-reported scales designed to offer clinical status outcomes [16,17]. Insufficient pain alleviation for patients with shoulder discomfort may lead to the development of mental health issues such as depression and anxiety [18,19]. These psychological disorders represent strong predictors of worse postoperative functional outcomes [20,21]. An example of PROM used as a screening tool to evaluate individuals with musculoskeletal disorders’ psychological health is the Hospital Anxiety and Depression Scale (HADS) [22,23]. The HADS is designed to detect depression and anxiety among non-psychiatric ward patients [24]. Physical symptoms of emotional distress including headaches, weight loss, and insomnia are not included in this questionnaire since they may be the outcome of a medical illness rather than emotional distress itself [25]. The Shoulder Pain and Disability Index (SPADI), on the other hand, was one of the first PROMs to be established specifically for patients with shoulder diseases, to assess pain and impairment [26,27]. Due to its proven test–re-test reliability [28] and change sensitivity, it is commonly used in orthopedic clinical practice [29]. Accordingly, the Simple Shoulder Test (SST) appears to share the same assets [30]. The SST was created to determine the functional limitations of the diseased shoulder in relation to the patient’s day-to-day activities [31]. Again, the 36-Item Short-Form Health Survey (SF-36) is a 36-item questionnaire that is widely included in orthopedics research and represents a valid approach for assessing health-related quality of life [32,33]. It examines the social, emotional, and physical functioning of the patient [34]. The SF-36 is frequently correlated with the Oxford Shoulder Score (OSS) [35]. The latter is a 12-item, subjective questionnaire, based precisely on the parameters of pain and function of the injured shoulder [36]. The OSS consistency, repeatability, and validity have all been established [37]. Finally, the American Shoulder and Elbow Surgeons (ASES) score was designed to evaluate shoulder discomfort and functional limitations in adults with musculoskeletal disorders [38]. The American Academy of Orthopedic Surgeons and the ASES Value Committee have recognized it as an outcome instrument that should be utilized for all shoulder pathology patients due to its applicability [39,40,41].
According to the studies carried out to date, the ASES, DASH, SPADI, OSS, and SST are the most employed questionnaires for the evaluation of the general health of orthopedic patients in both the preoperative and postoperative settings [42]. However, in the current literature, there is no investigation of the reliability of these measures as predictors of outcome in RC patients. The aim of this study was to apply state-of-the-art machine learning algorithms to evaluate the validity of these scales to predict the outcome after arthroscopic repair in RC patients. Longitudinal evaluations at 1, 3, 6, and 12 months from surgical intervention were calculated.

2. Materials and Methods

2.1. Population

We conducted a prospective cohort study enrolling patients admitted to the outpatient department of Orthopedics from October 2021 to April 2023. Here, 100 patients who underwent arthroscopic rotator cuff repair for rotator cuff tears of any grade were consecutively enrolled. For each patient, demographic information and comorbidities were collected at study entry, as reported in Table 1. Moreover, patients underwent clinical assessment on admission (T0) and after one month (T1), three months (T2), and six months (T3) at discharge, whereas the outcome variable was assessed after twelve months at discharge (T4).
Informed consent was obtained for all the enrolled patients. The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of Campus Bio-Medico University of Rome (COSMO study, protocol number: 78/18 OSS ComEt CBM, 16/10/18). The study was developed following the Good Clinical Practice (GCP) guidelines.

2.2. Proposed Approach

As an outcome measure, we used the Costant–Murley Shoulder Score, which is a 100-point scale recommended by the European Society for Shoulder and Elbow Surgery (ESSES) to assess shoulder function. It evaluates four subscales related to shoulder pathology: pain, daily living activities, strength, and range of motion [43]. The higher the score, the better the quality of the function. Then, to make prediction easier for the machine learning algorithms, the outcome variable was divided into four classes (poor, fair, good, and very good), as explained in Table 2.
By converting the continuous Constant–Murley Score into categorical values, only three classes were obtained, that is poor, fair, and good. No very good class was found (Figure 1).

2.2.1. Predictors’ Selection

As previously mentioned, several clinical scales were tested as predictors of the outcome variable at admission (T0) and during follow-up (T1–T3):
  • 36-Item Short-Form Survey (SF-36): A short-form, patient-reported assessment that provides an eight-scale profile of scores for both physical and mental health [44].
  • Simple Shoulder Test (SST): A shoulder-specific scale that assesses the affected shoulder’s functional limits in people with shoulder disorders [45]. It includes 12 questions in a dichotomous (yes/no) style, with scores ranging from 0 (worst) to 12 (best).
  • Hospital Anxiety and Depression Scale (HADS): A self-reported questionnaire for detecting states of anxiety and depression in a non-psychiatric setting. The anxiety and depression subscales are evaluated using seven questions each, on a four-point (0–3) response scale [46] (Table 3).
  • American Shoulder and Elbow Surgeons Score (ASES): A 100-point scale that estimates two dimensions of shoulder function—pain and performance—in activities of daily living. The ASES score allocates 50 points for measuring function and 50 points for pain [47].
  • Oxford Shoulder Score (OSS): A 12-item, patient-reported scale created especially for evaluating shoulder surgery results. Each of the 12 questions can be scored from 1 (best) to 5 (worst), so the final score ranges from 12 to 60 [48].
  • Shoulder Pain And Disability Index (SPADI): A patient-reported score aimed at quantifying pain and disability in patients with a shoulder injury. It includes 13 items, and there are 2 subscales: pain and disability. The pain subscale has five items, whereas the disability scale consists of eight items. The total score is computed by averaging the 2 subscale scores, thus ranging from 0 to 100 [49].
  • Additional demographic and clinical variables, such as age, sex, weight, height, and the Goutallier grade, were also considered potential predictors (Table 1). The Goutallier classification was used to quantify the amount of fatty degeneration of the rotator cuff musculature. It has five stages, starting from Stage 0, meaning normal muscle, to Stage 4, where fat is more present than muscle [50].
The features were chosen based on their proven use in orthopedic patient assessment and their potential to reflect various aspects of patient well-being, pain, and functional ability. We aimed to incorporate a balanced combination of established clinical metrics and additional variables that could contribute to the prediction of the outcome. The aforementioned metrics capture different facets of patient well-being and functional limitations, making them plausible candidates for predicting the outcome.

2.2.2. Statistical Analysis

In this research, we used a ML approach to predict clinical outcomes based on data collected at different times (T0–T3).
Since the variables were not normally distributed, they were compared at admission and during follow-up using the Friedman test, and the Durbin–Conover test for post-hoc pairwise comparison. The Friedman test is suitable for repeated measure data, examining the null hypothesis that there are no differences in the distributions of the variables between the timepoints. The statistical analysis was conducted using the Jamovi statistical software (version 2.3.19.0). We set the significance level (α) at α = 0.05 to determine statistical significance. Table 4 explains the results of the preliminary statistical analysis and the average ± standard deviation of each variable throughout the sessions. The results were highly significant for all variables, with p-values consistently lower than 0.001.

2.2.3. Machine Learning Approach

At first, considering the limited number of features provided, all the features were selected as input of various canonical machine learning classifiers, including support vector machine (SVM), k-nearest neighbors (k-NN), naïve Bayes (NB), and random forest (RF), in addiction to logistic regression (LR). All the classifiers were tested separately among each session (T0–T3) to assess their ability to predict the output at one year from discharge.
Additionally, after extensive testing, we discovered that the LR and RF classifiers performed better than the others, offering the best predicting performances as well as additional benefits for our specific objective. In particular, LR is a linear classifier that employs a logistic function to model the relationship between the input features and output classes. Its clarity and interpretability fit the need to comprehend the importance of various factors in making predictions about the course of rotator cuff disease. RF, on the other hand, is an ensemble method that builds a number of decision trees and combines their predictions to reach a decision. The RF classifier was preferred to a simple decision tree (DT) classifier because it mitigates overfitting and captures complex relationships in the data. Furthermore, the ensemble nature of RF further improves its robustness and predictive accuracy, making it suitable for our objective of precise outcome prediction.
Later, correlations between features were computed, keeping only those variables that were highly correlated with the output and poorly correlated with other features. Moreover, a statistical-based feature selection method (Chi-squared test) was performed to both reduce the computational cost of modeling and to assess whether the performances improved with the shrunken set of features. This step aimed at eliminating redundant or irrelevant features, focusing on those with stronger predictive power. All these steps were performed using Python 3 on the Jupyter Notebook platform.
The data preprocessing was firstly performed to treat missing data and scale variables, ensuring the effectiveness of the predictive models. Missing values were replaced with the median value of the predictor. All the variables were then standardized using the RobustScaler module provided in the scikit-learn library, which removes the median value and scales the data according to the quantile range [51], mitigating the effect of outliers.
For each session and for each machine learning model, we employed the nested cross-validation (Nested CV), a procedure for model hyperparameters’ optimization and for preventing overfitting problems. After defining the hyperparameter space, the procedure involves two loops:
  • The outer loop uses K-fold cross-validation (K-fold CV) to divide the dataset into training set and test sets, assessing the quality of the model trained on the inner loop.
  • The inner loop performs hyperparameter tuning by using K-fold CV on the training set.
The strength of this procedure is that it does not use the same data to model parameters and estimate the performance of the model, reducing the risk of overfitting associated, instead, with traditional cross-validation techniques. The two k values for the inner and outer loops were set to balance the computational cost of the procedure and to provide an unbiased and reliable estimate of the performance. In this direction, we selected a value equal to 10 for the outer loop and 3 for the inner loop [52].
Accuracy was provided as the score of the estimator for each run of the cross-validation. Classification performances were then assessed as mean ± standard deviation across the folds.
For each run of the cross-validation, the estimator’s scores were provided as accuracy, precision, recall, F1-score, and AUC. By calculating recall metrics for each severity category (poor, fair, good), we extended our analysis. This refinement shed light on the model’s proficiency in identifying patients at various severity levels.
Classification performances were then assessed as mean ± standard deviation across the folds (Table 5 and Table 6). This approach provided a thorough understanding of the models’ predictive abilities across various sessions and spotted any potential changes in performance over time.

2.2.4. Feature Selection

A statistical feature selection method was also applied to each session to reduce data dimensionality and explore whether models’ performances improved. Specifically, the ANOVA F-value was computed to evaluate the relationship between each input variable and the output, obtaining a score. The higher the score, the more output-dependent the variable is.
The optimal subset of features was obtained by defining a breakpoint as the highest difference between consecutive scores and selecting only the predictors with a score greater than it. Figure 2 displays the scores obtained for each session and the features selected by implementing the previous procedure.
Before performing feature selection, a correlation matrix was extracted to quantify the association between pairs of variables. OOS, SPADI, and SST features were highly correlated (rho > 0.8) during each session, which is not surprising since they all provide insight about the level of shoulder pain perceived by the patient. In order to prevent multicollinearity problems, we decided to keep only the variable that correlated the most with the output, which was the SST scale.

3. Results

No good performances were obtained for any of the classifiers during any of the sessions by using all the features, as detailed in Table 5. An accuracy greater than 50% was not achieved by any classifier across the sessions (Figure 3), resulting in a bad predictive capability of the ML models.
Following experimentation with various machine learning models it was observed that their performance remained comparable to that of the simplest logistic regression model. Consequently, we decided to present results solely for the logistic regression and random forest classifiers, the latter of which achieved superior overall accuracy.
We found the same results after performing feature selection, obtaining an accuracy of less than 55% for all sessions (Table 6). Notably, no increasing performances were obtained by using the shrunken set of features, as displayed in Figure 3.
Considering the obtained results, we conducted a Mann–Whitney U test to directly compare the accuracy of the two classifiers, namely, logistic regression and random forest. No statistically significant differences in the performance were found between the two models, as shown in Figure 3. Despite extensive exploration of various classifiers and feature sets, the models struggled to surpass a certain accuracy threshold, highlighting the challenges faced in achieving satisfactory predictive capabilities.

4. Discussion

Artificial intelligence (AI) has the ability to completely transform the practice of medicine by increasing precision and effectiveness, cutting costs, and delivering individualized treatment regimens [53]. In the orthopedic field, AI is being used to develop clinical prediction models for orthopedic patients that can help doctors make more informed decisions about patient care. Algorithms have been used to predict the outcomes of total joint replacement surgery [54], to predict fracture risk in orthopedic patients [55], and to diagnose orthopedic conditions such as osteoarthritis and spinal stenosis. To reach this performance level, AI algorithms need to analyze multidimensional patient data, such as age, medical history, motion, pain levels, medication therapy, bone density, and medical images such as X-rays and MRI scans, in order to recognize patients at higher risk for complications or poor outcomes after surgery.
In this study, we demonstrated that clinical data alone are not useful to reach a robust outcome prediction. Indeed, by using demographic information, comorbidities, and clinical metric scores, we were able to train a classifier with 55% accuracy at most in the third session.
According to the literature, another important function of ML approaches in the clinical setting appears to be the diagnostic one. This is confirmed by Ho et al. [56], who evaluated 194 ultrasound images of RCTs by fine-tuning pre-trained deep models in order to analyze and classify the RCTs. Among them, DenseNet121 demonstrated the best classification performance, with 88.2% accuracy, 93.8% sensitivity, 83.6% specificity, and an AUC score of 0.832. Kim et al. [57] developed a deep learning algorithm to rule out significant RCT based on conventional shoulder radiographs in 1095 shoulders suspected of RCT. The sensitivity of this approach was 97.3%, and it could rule out significant RCT with a negative likelihood ratio of 0.06 and a negative predictive value of 96.6%.
Concerning the reliability of clinical data as outcome measures of orthopedic surgical interventions, recently, Martin et al. [58] combined datasets from the Norwegian and Danish knee ligament registers in order to create an algorithm that can predict outcomes after primary anterior cruciate ligament reconstruction (ACLR). Using clinical data from 63,000 patients, ML analysis enabled prediction of the ACLR risk with moderate accuracy (68%). However, other publications found in the current literature positively identified significant clinical predictors. In fact, the research conducted by Li et al. [59] reported that ML models, in particular the XGBoost model, successfully recognized important clinical variables for the prediction of outcomes of 417 patients with RCTs. The most statistically significant values for prediction were the Jobe test, Bear hug test, and the age of the patients, with mean Shapley additive explanation (SHAP) values of 1.458, 0.950, and 0.790, respectively. Similarly, Dong et al. [60] studied a cohort of 1967 patients through a human–computer interactive Electronic Medical System (EMS) and demonstrated the presence of predictors of RC calcific tendinitis stratified according to the patients’ sex: women diagnosed with diabetes mellitus and men diagnosed with hyperlipidemia, diabetes mellitus, and hypothyroidism showed a higher risk of developing RC calcific tendinitis. Clinical factors such as age and sex were investigated in both articles, similarly to this paper. However, in the present paper, these features did not represent significant outcome predictors.
On the other hand, the most common functional outcomes reviewed as clinical outcome predictors by other articles in the current literatures were ASES, Visual Analog Scale (VAS), University of California Los Angeles (UCLA) Shoulder Score, CMS, and ROM. This is illustrated by Potty et al. [61] and Kumar et al. [62]. The former [61] applied the XGBoost algorithm to 631 patients undergoing arthroscopic RC repair to produce an expected post-operative ASES score. The anticipated improvement in the ASES score was within the Minimal Clinical Important Difference (MCID) value, indicating that ASES, even though the predictions did not exactly correspond to the actual results, along with BMI, age, and tendon quality, is a key predictor of clinical outcomes [9]. The latter [62], aiming to quantify the accuracy of prediction of postoperative ASES, UCLA, CMS, VAS, and ROM on 6210 patients, employed 3 different ML approaches: linear regression, XGBoost, and Wide and Deep. The Wide and Deep technique was associated with the smallest mean absolute error and predicted the postoperative ASES score to ±10.1 to 11.3 points, the UCLA score to ±2.5 to 3.4, the CMS to ±7.3 to 7.9, and the VAS pain score to ±1.2 to 1.4, demonstrating that preoperatively recognizing which patient characteristics may be predictive of a worse clinical outcome and issues of major clinical benefit improvement allows ML approaches to accurately risk-stratify patients. Moreover, the study by Vassalou et al. [63] demonstrated the validity of XGboost, which achieved an AUC of 69.2% (95%CI from 54.5% to 83.8%) for the prediction of complete resolution of pain at 1 year for 100 patients undergoing ultrasound-guided percutaneous irrigation of calcific tendinopathy of RC. Age, VAS, and the size of the calcification were found to be the three most important variables for the classification performance.
The present study evaluated various clinical scales, such as the SST, HADS, SF-36, OSS, and SPADI, which were not found in other publications on the same topic. However, it also assessed the ASES score, which was common to Potty et al. [61] and Kumar et al.’s publications [62]. Nonetheless, differently from the other articles, this paper did not find important predictive values of the ASES score, nor of the other clinical scales and clinical factors proposed.
The potential application of a predictive model based on subjective clinical scales such as the SST, HADS, SF-36, OSS, SPADI, and ASES would allow a deeper understanding of the patients’ current own health perception and allow the clinicians to better organize the treatment plan.
The lack of a significant application of AI algorithms to clinical prediction of RC patients could be dependent on several factors. The main limitation concerning the present paper is the small number of the cohort. However, it is important to note that research with similar inclusion and exclusion criteria is scarce within the existing literature. On the other hand, the data quality within the cohort poses additional limitations. Certain variables, such as smoking, diabetes, and work injury, are known to have an impact on the outcome. However, due to the nature of data collection, it was not feasible to include all relevant variables in the database. Moreover, some variables within the dataset have inherent uncertainties. For instance, the Goutallier grade, used to assess a specific aspect of the data, exhibits poor interrater reliability. The lack of consensus among raters regarding the Goutallier grade introduces ambiguity and inconsistency, which can adversely affect the performance of algorithms employed for predictions. In addition, we did not record other relevant clinical indicators such as diagnosis, implant type, ROM, and radiographic findings [64]. Another limitation is due to the fact that the three classes were not well balanced (Figure 1), which can cause the estimator to incorrectly discern the classes. Finally, the type of predictors, such as continuous and categorized (operator-dependent) variables, and the lack of objective biological high-dimensional data (i.e., neuroimaging, genetics), might have also limited the performance of our ML approach with respect to other studies.

5. Conclusions

We found that ML algorithms are not able to predict outcomes of patients with RC with sufficient accuracy using only clinical data. In fact, utilizing demographic data, comorbidities, and well-known clinical metrics, we were able to train a classifier session with a maximum accuracy of 55%. Despite that AI application in orthopedic surgery is in its relative infancy, our negative findings highlight the need to consider metrics capturing dynamic changes in prognosis, extending the current models with new objective predictors, such as kinematic and neuroimaging data. Computer-aided diagnostics may improve doctors’ ability to correctly identify musculoskeletal disorders and enhance the patients’ overall experience.
Finally, this study lays solid foundations for future studies to be carried out with greater cohorts and additional clinical scales.

Author Contributions

Conceptualization, A.C., G.T. and U.G.L.; methodology, R.B. and C.C.; software, G.T. and S.C.; validation, C.D.N. and R.B.; formal analysis, R.B., S.C. and G.T.; investigation, C.D.N. and C.C.; resources, G.P. and A.P.; data curation, A.C.; writing—original draft preparation, A.C., G.T. and B.B.; writing—review and editing, A.C., G.T. and B.B.; supervision, G.P. and A.P.; project administration, U.G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was approved by the Institutional Review Board of Campus Bio-Medico University of Rome (COSMO study, protocol number: 78/18 OSS ComEt CBM, 16/10/18) according to the Helsinki Declaration.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request. Access to the database is available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Brindisino, F.; Salomon, M.; Giagio, S.; Pastore, C.; Innocenti, T. Rotator cuff repair vs. nonoperative treatment: A systematic review with meta-analysis. J. Shoulder Elbow Surg. 2021, 30, 2648–2659. [Google Scholar] [CrossRef] [PubMed]
  2. Longo, U.G.; Salvatore, G.; Rizzello, G.; Berton, A.; Ciuffreda, M.; Candela, V.; Denaro, V. The burden of rotator cuff surgery in Italy: A nationwide registry study. Arch. Orthop. Trauma Surg. 2017, 137, 217–224. [Google Scholar] [CrossRef] [PubMed]
  3. Maffulli, N.; Longo, U.G.; Berton, A.; Loppini, M.; Denaro, V. Biological factors in the pathogenesis of rotator cuff tears. Sports Med. Arthrosc. Rev. 2011, 19, 194–201. [Google Scholar] [CrossRef] [PubMed]
  4. Longo, U.G.; Berton, A.; Papapietro, N.; Maffulli, N.; Denaro, V. Epidemiology, genetics and biological factors of rotator cuff tears. Med. Sport Sci. 2012, 57, 1–9. [Google Scholar] [CrossRef] [PubMed]
  5. Franceschi, F.; Longo, U.G.; Ruzzini, L.; Rizzello, G.; Maffulli, N.; Denaro, V. The Roman Bridge: A “double pulley—Suture bridges” technique for rotator cuff repair. BMC Musculoskelet. Disord. 2007, 8, 123. [Google Scholar] [CrossRef]
  6. Franceschi, F.; Longo, U.G.; Ruzzini, L.; Rizzello, G.; Denaro, V. Arthroscopic management of calcific tendinitis of the subscapularis tendon. Knee Surg. Sports Traumatol. Arthrosc. 2007, 15, 1482–1485. [Google Scholar] [CrossRef]
  7. Longo, U.G.; Carnevale, A.; Piergentili, I.; Berton, A.; Candela, V.; Schena, E.; Denaro, V. Retear rates after rotator cuff surgery: A systematic review and meta-analysis. BMC Musculoskelet. Disord. 2021, 22, 749. [Google Scholar] [CrossRef]
  8. Kim, D.M.; Jeon, I.H.; Yang, H.S.; Shin, M.J.; Park, J.H.; Kholinne, E.; Kim, H.; Park, D.; Koh, K.H. Poor Prognostic Factors in Patients With Rotator Cuff Retear. Orthop. J. Sports Med. 2021, 9, 2325967121992154. [Google Scholar] [CrossRef]
  9. Ruzzini, L.; Abbruzzese, F.; Rainer, A.; Longo, U.G.; Trombetta, M.; Maffulli, N.; Denaro, V. Characterization of age-related changes of tendon stem cells from adult human tendons. Knee Surg. Sports Traumatol. Arthrosc. 2014, 22, 2856–2866. [Google Scholar] [CrossRef]
  10. Aguado, G.; Obando, D.V.; Herrera, G.A.; Ramirez, A.; Llinás, P.J. Retears of the Rotator Cuff: An Ultrasonographic Assessment During the First Postoperative Year. Orthop. J. Sports Med. 2019, 7, 2325967119889049. [Google Scholar] [CrossRef]
  11. Franceschi, F.; Longo, U.G.; Ruzzini, L.; Rizzello, G.; Maffulli, N.; Denaro, V. Soft tissue tenodesis of the long head of the biceps tendon associated to the Roman Bridge repair. BMC Musculoskelet. Disord. 2008, 9, 78. [Google Scholar] [CrossRef] [PubMed]
  12. Beaton, D.E.; Richards, R.R. Measuring function of the shoulder. A cross-sectional comparison of five questionnaires. J. Bone Jt. Surg. Am. 1996, 78, 882–890. [Google Scholar] [CrossRef] [PubMed]
  13. Schmidt, S.; Ferrer, M.; González, M.; González, N.; Valderas, J.M.; Alonso, J.; Escobar, A.; Vrotsou, K.; Group, E. Evaluation of shoulder-specific patient-reported outcome measures: A systematic and standardized comparison of available evidence. J. Shoulder Elbow Surg. 2014, 23, 434–444. [Google Scholar] [CrossRef] [PubMed]
  14. Padua, R.; de Girolamo, L.; Grassi, A.; Cucchi, D. Choosing patient-reported outcome measures for shoulder pathology. EFORT Open Rev. 2021, 6, 779–787. [Google Scholar] [CrossRef]
  15. Frizziero, A.; Trainito, S.; Oliva, F.; Nicoli Aldini, N.; Masiero, S.; Maffulli, N. The role of eccentric exercise in sport injuries rehabilitation. Br. Med. Bull. 2014, 110, 47–75. [Google Scholar] [CrossRef] [PubMed]
  16. Carrozzino, D.; Patierno, C.; Guidi, J.; Berrocal Montiel, C.; Cao, J.; Charlson, M.E.; Christensen, K.S.; Concato, J.; De Las Cuevas, C.; de Leon, J.; et al. Clinimetric Criteria for Patient-Reported Outcome Measures. Psychother. Psychosom. 2021, 90, 222–232. [Google Scholar] [CrossRef]
  17. Churruca, K.; Pomare, C.; Ellis, L.A.; Long, J.C.; Henderson, S.B.; Murphy, L.E.D.; Leahy, C.J.; Braithwaite, J. Patient-reported outcome measures (PROMs): A review of generic and condition-specific measures and a discussion of trends and issues. Health Expect. 2021, 24, 1015–1024. [Google Scholar] [CrossRef]
  18. Badcock, L.J.; Lewis, M.; Hay, E.M.; McCarney, R.; Croft, P.R. Chronic shoulder pain in the community: A syndrome of disability or distress? Ann. Rheum. Dis. 2002, 61, 128–131. [Google Scholar] [CrossRef]
  19. Longo, U.G.; Facchinetti, G.; Marchetti, A.; Candela, V.; Risi Ambrogioni, L.; Faldetta, A.; De Marinis, M.G.; Denaro, V. Sleep Disturbance and Rotator Cuff Tears: A Systematic Review. Medicina 2019, 55, 453. [Google Scholar] [CrossRef]
  20. Baker, M.; Albelo, F.; Zhang, T.; Schneider, M.B.; Foster, M.J.; Aneizi, A.; Hasan, S.A.; Gilotra, M.N.; Henn, R.F. PROMIS Depression and Anxiety in shoulder surgery patients. Bone Jt. J. 2022, 104-B, 479–485. [Google Scholar] [CrossRef]
  21. Berton, A.; Longo, U.G.; Candela, V.; Fioravanti, S.; Giannone, L.; Arcangeli, V.; Alciati, V.; Berton, C.; Facchinetti, G.; Marchetti, A.; et al. Virtual Reality, Augmented Reality, Gamification, and Telerehabilitation: Psychological Impact on Orthopedic Patients’ Rehabilitation. J. Clin. Med. 2020, 9, 2567. [Google Scholar] [CrossRef]
  22. Pallant, J.F.; Bailey, C.M. Assessment of the structure of the Hospital Anxiety and Depression Scale in musculoskeletal patients. Health Qual. Life Outcomes 2005, 3, 82. [Google Scholar] [CrossRef] [PubMed]
  23. Härter, M.; Reuter, K.; Gross-Hardt, K.; Bengel, J. Screening for anxiety, depressive and somatoform disorders in rehabilitation--validity of HADS and GHQ-12 in patients with musculoskeletal disease. Disabil. Rehabil. 2001, 23, 737–744. [Google Scholar] [CrossRef] [PubMed]
  24. Bjelland, I.; Dahl, A.A.; Haug, T.T.; Neckelmann, D. The validity of the Hospital Anxiety and Depression Scale. An updated literature review. J. Psychosom Res. 2002, 52, 69–77. [Google Scholar] [CrossRef] [PubMed]
  25. Annunziata, M.A.; Muzzatti, B.; Bidoli, E.; Flaiban, C.; Bomben, F.; Piccinin, M.; Gipponi, K.M.; Mariutti, G.; Busato, S.; Mella, S. Hospital Anxiety and Depression Scale (HADS) accuracy in cancer patients. Support. Care Cancer 2020, 28, 3921–3926. [Google Scholar] [CrossRef]
  26. Ebrahimzadeh, M.H.; Vahedi, E.; Baradaran, A.; Birjandinejad, A.; Seyyed-Hoseinian, S.H.; Bagheri, F.; Kachooei, A.R. Psychometric Properties of the Persian Version of the Simple Shoulder Test (SST) Questionnaire. Arch. Bone Jt. Surg. 2016, 4, 387–392. [Google Scholar]
  27. Breckenridge, J.D.; McAuley, J.H. Shoulder Pain and Disability Index (SPADI). J. Physiother. 2011, 57, 197. [Google Scholar] [CrossRef]
  28. Schmitt, J.S.; Di Fabio, R.P. Reliable change and minimum important difference (MID) proportions facilitated group responsiveness comparisons using individual threshold criteria. J. Clin. Epidemiol. 2004, 57, 1008–1018. [Google Scholar] [CrossRef]
  29. MacDermid, J.C.; Solomon, P.; Prkachin, K. The Shoulder Pain and Disability Index demonstrates factor, construct and longitudinal validity. BMC Musculoskelet. Disord. 2006, 7, 12. [Google Scholar] [CrossRef]
  30. Angst, F.; Schwyzer, H.K.; Aeschlimann, A.; Simmen, B.R.; Goldhahn, J. Measures of adult shoulder function: Disabilities of the Arm, Shoulder, and Hand Questionnaire (DASH) and its short version (QuickDASH), Shoulder Pain and Disability Index (SPADI), American Shoulder and Elbow Surgeons (ASES) Society standardized shoulder assessment form, Constant (Murley) Score (CS), Simple Shoulder Test (SST), Oxford Shoulder Score (OSS), Shoulder Disability Questionnaire (SDQ), and Western Ontario Shoulder Instability Index (WOSI). Arthritis Care Res. 2011, 63 (Suppl. S11), S174–S188. [Google Scholar] [CrossRef]
  31. Godfrey, J.; Hamman, R.; Lowenstein, S.; Briggs, K.; Kocher, M. Reliability, validity, and responsiveness of the simple shoulder test: Psychometric properties by age and injury type. J. Shoulder Elbow Surg. 2007, 16, 260–267. [Google Scholar] [CrossRef] [PubMed]
  32. Jenkinson, C.; Coulter, A.; Wright, L. Short form 36 (SF36) health survey questionnaire: Normative data for adults of working age. BMJ 1993, 306, 1437–1440. [Google Scholar] [CrossRef] [PubMed]
  33. Lins, L.; Carvalho, F.M. SF-36 total score as a single measure of health-related quality of life: Scoping review. SAGE Open Med. 2016, 4, 2050312116671725. [Google Scholar] [CrossRef]
  34. Yalvaç, B.; Mesci, N.; Geler Külcü, D.; Yurdakul, O.V. Comparison of ultrasound and extracorporeal shock wave therapy in lateral epicondylosis. Acta Orthop. Traumatol. Turc. 2018, 52, 357–362. [Google Scholar] [CrossRef]
  35. Clement, N.D.; Court-Brown, C.M. Oxford shoulder score in a normal population. Int. J. Shoulder Surg. 2014, 8, 10–14. [Google Scholar] [CrossRef] [PubMed]
  36. Russell, S.; Jariwala, A.; Conlon, R.; Selfe, J.; Richards, J.; Walton, M. A blinded, randomized, controlled trial assessing conservative management strategies for frozen shoulder. J. Shoulder Elbow Surg. 2014, 23, 500–507. [Google Scholar] [CrossRef] [PubMed]
  37. Dawson, J.; Fitzpatrick, R.; Carr, A. Questionnaire on the perceptions of patients about shoulder surgery. J. Bone Jt. Surg. Br. 1996, 78, 593–600. [Google Scholar] [CrossRef]
  38. Wylie, J.D.; Beckmann, J.T.; Granger, E.; Tashjian, R.Z. Functional outcomes assessment in shoulder surgery. World J. Orthop. 2014, 5, 623–633. [Google Scholar] [CrossRef]
  39. Razmjou, H.; Bean, A.; van Osnabrugge, V.; MacDermid, J.C.; Holtby, R. Cross-sectional and longitudinal construct validity of two rotator cuff disease-specific outcome measures. BMC Musculoskelet. Disord. 2006, 7, 26. [Google Scholar] [CrossRef]
  40. Baumgarten, K.M. The American Shoulder and Elbow Surgeons score has excellent correlation with the Western Ontario Osteoarthritis score. J. Shoulder Elbow Surg. 2021, 30, 2604–2610. [Google Scholar] [CrossRef]
  41. Hawkins, R.J.; Thigpen, C.A. Selection, implementation, and interpretation of patient-centered shoulder and elbow outcomes. J. Shoulder Elbow Surg. 2018, 27, 357–362. [Google Scholar] [CrossRef]
  42. Roy, J.S.; MacDermid, J.C.; Woodhouse, L.J. Measuring shoulder function: A systematic review of four questionnaires. Arthritis Rheum. 2009, 61, 623–632. [Google Scholar] [CrossRef]
  43. Vrotsou, K.; Ávila, M.; Machón, M.; Mateo-Abad, M.; Pardo, Y.; Garin, O.; Zaror, C.; González, N.; Escobar, A.; Cuéllar, R. Constant-Murley Score: Systematic review and standardized evaluation in different shoulder pathologies. Qual. Life Res. 2018, 27, 2217–2226. [Google Scholar] [CrossRef] [PubMed]
  44. Ware, J.E. SF-36 health survey update. Spine 2000, 25, 3130–3139. [Google Scholar] [CrossRef]
  45. Roy, J.S.; Macdermid, J.C.; Faber, K.J.; Drosdowech, D.S.; Athwal, G.S. The simple shoulder test is responsive in assessing change following shoulder arthroplasty. J. Orthop. Sports Phys. Ther. 2010, 40, 413–421. [Google Scholar] [CrossRef] [PubMed]
  46. Zigmond, A.S.; Snaith, R.P. The hospital anxiety and depression scale. Acta Psychiatr. Scand. 1983, 67, 361–370. [Google Scholar] [CrossRef] [PubMed]
  47. Sallay, P.I.; Reed, L. The measurement of normative American Shoulder and Elbow Surgeons scores. J. Shoulder Elbow Surg. 2003, 12, 622–627. [Google Scholar] [CrossRef]
  48. Dawson, J.; Rogers, K.; Fitzpatrick, R.; Carr, A. The Oxford shoulder score revisited. Arch. Orthop. Trauma Surg. 2009, 129, 119–123. [Google Scholar] [CrossRef]
  49. Roach, K.E.; Budiman-Mak, E.; Songsiridej, N.; Lertratanakul, Y. Development of a shoulder pain and disability index. Arthritis Care Res. 1991, 4, 143–149. [Google Scholar] [CrossRef] [PubMed]
  50. Somerson, J.S.; Hsu, J.E.; Gorbaty, J.D.; Gee, A.O. Classifications in Brief: Goutallier Classification of Fatty Infiltration of the Rotator Cuff Musculature. Clin. Orthop. Relat. Res. 2016, 474, 1328–1332. [Google Scholar] [CrossRef]
  51. Abraham, A.; Pedregosa, F.; Eickenberg, M.; Gervais, P.; Mueller, A.; Kossaifi, J.; Gramfort, A.; Thirion, B.; Varoquaux, G. Machine learning for neuroimaging with scikit-learn. Front. Neuroinform. 2014, 8, 14. [Google Scholar] [CrossRef] [PubMed]
  52. Cawley, G.a.T.N. On Over-fitting in Model Selection and Subsequent Selection Bias in Performance Evaluation. J. Mach. Learn. Res. 2010, 11, 2079–2107. [Google Scholar]
  53. Gheisari, M. Deep learning: Applications, architectures, models, tools, and frameworks: A comprehensive survey. CAAI Trans. Intell. Technol. 2023, 1–26. [Google Scholar] [CrossRef]
  54. Klemt, C.; Uzosike, A.C.; Esposito, J.G.; Harvey, M.J.; Yeo, I.; Subih, M.; Kwon, Y.M. The utility of machine learning algorithms for the prediction of patient-reported outcome measures following primary hip and knee total joint arthroplasty. Arch. Orthop. Trauma Surg. 2023, 143, 2235–2245. [Google Scholar] [CrossRef] [PubMed]
  55. Kunze, K.N.; Polce, E.M.; Patel, A.; Courtney, P.M.; Sporer, S.M.; Levine, B.R. Machine learning algorithms predict within one size of the final implant ultimately used in total knee arthroplasty with good-to-excellent accuracy. Knee Surg. Sports Traumatol. Arthrosc. 2022, 30, 2565–2572. [Google Scholar] [CrossRef] [PubMed]
  56. Ho, T.T.; Kim, G.T.; Kim, T.; Choi, S.; Park, E.K. Classification of rotator cuff tears in ultrasound images using deep learning models. Med. Biol. Eng. Comput. 2022, 60, 1269–1278. [Google Scholar] [CrossRef]
  57. Kim, Y.; Choi, D.; Lee, K.J.; Kang, Y.; Ahn, J.M.; Lee, E.; Lee, J.W.; Kang, H.S. Ruling out rotator cuff tear in shoulder radiograph series using deep learning: Redefining the role of conventional radiograph. Eur. Radiol. 2020, 30, 2843–2852. [Google Scholar] [CrossRef]
  58. Martin, R.K.; Wastvedt, S.; Pareek, A.; Persson, A.; Visnes, H.; Fenstad, A.M.; Moatshe, G.; Wolfson, J.; Lind, M.; Engebretsen, L. Ceiling Effect of the Combined Norwegian and Danish Knee Ligament Registers Limits Anterior Cruciate Ligament Reconstruction Outcome Prediction. Am. J. Sports Med. 2023, 51, 2324–2332. [Google Scholar] [CrossRef]
  59. Li, C.; Alike, Y.; Hou, J.; Long, Y.; Zheng, Z.; Meng, K.; Yang, R. Machine learning model successfully identifies important clinical features for predicting outpatients with rotator cuff tears. Knee Surg. Sports Traumatol. Arthrosc. 2023, 31, 2615–2623. [Google Scholar] [CrossRef]
  60. Dong, S.; Li, J.; Zhao, H.; Zheng, Y.; Chen, Y.; Shen, J.; Yang, H.; Zhu, J. Risk Factor Analysis for Predicting the Onset of Rotator Cuff Calcific Tendinitis Based on Artificial Intelligence. Comput. Intell. Neurosci. 2022, 2022, 8978878. [Google Scholar] [CrossRef]
  61. Potty, A.G.; Potty, A.S.R.; Maffulli, N.; Blumenschein, L.A.; Ganta, D.; Mistovich, R.J.; Fuentes, M.; Denard, P.J.; Sethi, P.M.; Shah, A.A.; et al. Approaching Artificial Intelligence in Orthopaedics: Predictive Analytics and Machine Learning to Prognosticate Arthroscopic Rotator Cuff Surgical Outcomes. J. Clin. Med. 2023, 12, 2369. [Google Scholar] [CrossRef] [PubMed]
  62. Kumar, V.; Roche, C.; Overman, S.; Simovitch, R.; Flurin, P.H.; Wright, T.; Zuckerman, J.; Routman, H.; Teredesai, A. What Is the Accuracy of Three Different Machine Learning Techniques to Predict Clinical Outcomes After Shoulder Arthroplasty? Clin. Orthop. Relat. Res. 2020, 478, 2351–2363. [Google Scholar] [CrossRef]
  63. Vassalou, E.E.; Klontzas, M.E.; Marias, K.; Karantanas, A.H. Predicting long-term outcomes of ultrasound-guided percutaneous irrigation of calcific tendinopathy with the use of machine learning. Skeletal. Radiol. 2022, 51, 417–422. [Google Scholar] [CrossRef] [PubMed]
  64. Kumar, V.; Roche, C.; Overman, S.; Simovitch, R.; Flurin, P.H.; Wright, T.; Zuckerman, J.; Routman, H.; Teredesai, A. Using machine learning to predict clinical outcomes after shoulder arthroplasty with a minimal feature set. J. Shoulder Elbow Surg. 2021, 30, e225–e236. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Outcome variable distribution.
Figure 1. Outcome variable distribution.
Diagnostics 13 02915 g001
Figure 2. Features selected by performing statistical FS. (a) ASES, gender, and smoking were selected in T0, (b) SST and SF-36 were selected in T1, (c) HADS was selected in T2, and (d) HADS was selected in T3.
Figure 2. Features selected by performing statistical FS. (a) ASES, gender, and smoking were selected in T0, (b) SST and SF-36 were selected in T1, (c) HADS was selected in T2, and (d) HADS was selected in T3.
Diagnostics 13 02915 g002
Figure 3. Visual comparison of accuracy scores achieved across sessions for both the original and shrunken datasets (µ ± SD).
Figure 3. Visual comparison of accuracy scores achieved across sessions for both the original and shrunken datasets (µ ± SD).
Diagnostics 13 02915 g003
Table 1. Demographic information and clinical characteristics of the patients.
Table 1. Demographic information and clinical characteristics of the patients.
PredictorsRC (n = 102)
Age (years)60.21 ± 7.77
Sex40 F/60 M
Height (cm)169.61 ± 8.09
Weight (kg)76.38 ± 10.93
Smoking20 Yes/80 No
Arterial Hypertension52 Yes/48 No
Goutallier Grade of Rotator Cuff Muscle Degeneration (%)
Stage 00%
Stage 172%
Stage 227%
Stage 30%
Stage 41%
Table 2. Conversion between Constant-Murley Score and corresponding qualitative classes.
Table 2. Conversion between Constant-Murley Score and corresponding qualitative classes.
Constant–Murley Score Conversion
RangeClass
86–100Very Good
71–85Good
56–70Fair
<56Poor
Table 3. Hospital Anxiety and Depression Scale (HADS) Scoring and Classification.
Table 3. Hospital Anxiety and Depression Scale (HADS) Scoring and Classification.
ScoreAnxious and Depressive State
8–10Mild
11–14Moderate
15–21Severe
Table 4. Clinical scales of the patients at admission and during follow-up sessions.
Table 4. Clinical scales of the patients at admission and during follow-up sessions.
Variableµ(SD) T0µ(SD) T1µ(SD) T2µ(SD) T3Friedman Test
(p-Value)
SF-36100.34 ± 6.9297.87 ± 7.59103.24 ± 6.61103.61 ± 6.12<0.001
SST3.80 ± 2.741.95 ± 2.226.82 ± 2.598.18 ± 2.86<0.001
HADS10.06 ± 7.477.76 ± 6.327.61 ± 7.688.59 ± 8.34<0.001
ASES36.34 ± 17.1437.72 ± 13.6264.33 ± 16.3471.95 ± 10.26<0.001
OSS33.85 ± 10.3939.87 ± 9.4423.10 ± 8.0618.40 ± 5.75<0.001
SPADI78.15 ± 26.4188.75 ± 26.4741.80 ± 23.3931.88 ± 18.57<0.01
Table 5. Overview of the performance metrics obtained from the application of the logistic regression (LR) and random forest (RF) classifiers using Nested CV across multiple sessions (T0–T3). Results are reported as percentage mean, along with the corresponding standard deviation, of the classifiers’ performance.
Table 5. Overview of the performance metrics obtained from the application of the logistic regression (LR) and random forest (RF) classifiers using Nested CV across multiple sessions (T0–T3). Results are reported as percentage mean, along with the corresponding standard deviation, of the classifiers’ performance.
Full Dataset Classification Model
SessionMetric (%)SeverityLRRF
T0AccuracyOverall39 ± 1246 ± 7
PrecisionOverall46 ± 2359 ± 13
RecallOverall39 ± 1246 ± 7
Poor5 ± 150 ± 0
Fair61 ± 2877 ± 16
Good37 ± 1538 ± 19
F1-ScoreOverall34 ± 1139 ± 6
AUCOverall53 ± 1657 ± 13
T1AccuracyOverall49 ± 1351 ± 10
PrecisionOverall58 ± 1661 ± 19
RecallOverall49 ± 1351 ± 10
Poor5 ± 150 ± 0
Fair62 ± 2275 ± 23
Good62 ± 2357 ± 25
F1-ScoreOverall44 ± 1244 ± 9
AUCOverall62 ± 1067 ± 8
T2AccuracyOverall45 ± 1554 ± 14
PrecisionOverall47 ± 1859 ± 16
RecallOverall45 ± 1554 ± 14
Poor38 ± 2148 ± 24
Fair52 ± 2755 ± 30
Good43 ± 2258 ± 26
F1-ScoreOverall44 ± 1552 ± 15
AUCOverall60 ± 1669 ± 7
T3AccuracyOverall53 ± 1354 ± 14
PrecisionOverall60 ± 1557 ± 17
RecallOverall53 ± 1354 ± 14
Poor53 ± 2845 ± 36
Fair57 ± 2452 ± 27
Good51 ± 1965 ± 23
F1-ScoreOverall52 ± 1251 ± 17
AUCOverall69 ± 1072 ± 12
Table 6. Overview of the performance metrics obtained from the application of the logistic regression (LR) and random forest (RF) classifiers using Nested CV across multiple sessions (T0–T3) after feature selection. Results are reported as percentage accuracy, along with the corresponding standard deviation, of the classifiers’ performance.
Table 6. Overview of the performance metrics obtained from the application of the logistic regression (LR) and random forest (RF) classifiers using Nested CV across multiple sessions (T0–T3) after feature selection. Results are reported as percentage accuracy, along with the corresponding standard deviation, of the classifiers’ performance.
Shrunken Dataset Classification Model
SessionMetric (%)SeverityLRRF
T0AccuracyOverall48 ± 745 ± 7
PrecisionOverall62 ± 1556 ± 13
RecallOverall48 ± 745 ± 7
Poor5 ± 155 ± 15
Fair82 ± 1571 ± 27
Good36 ± 1241 ± 12
F1-ScoreOverall42 ± 739 ± 5
AUCOverall60 ± 1361 ± 13
T1AccuracyOverall51 ± 846 ± 12
PrecisionOverall66 ± 1355 ± 19
RecallOverall51 ± 846 ± 12
Poor0 ± 05 ± 15
Fair71 ± 1856 ± 24
Good59 ± 2962 ± 25
F1-ScoreOverall43 ± 940 ± 11
AUCOverall63 ± 1359 ± 13
T2AccuracyOverall40 ± 1040 ± 14
PrecisionOverall56 ± 1443 ± 19
RecallOverall40 ± 1040 ± 14
Poor17 ± 2148 ± 28
Fair62 ± 2939 ± 32
Good32 ± 3339 ± 24
F1-ScoreOverall33 ± 1337 ± 16
AUCOverall62 ± 1559 ± 11
T3AccuracyOverall55 ± 1151 ± 14
PrecisionOverall61 ± 1157 ± 13
RecallOverall55 ± 1151 ± 14
Poor33 ± 3238 ± 37
Fair58 ± 1846 ± 25
Good71 ± 2267 ± 25
F1-ScoreOverall53 ± 1357 ± 17
AUCOverall76 ± 974 ± 10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Longo, U.G.; Di Naro, C.; Campisi, S.; Casciaro, C.; Bandini, B.; Pareek, A.; Bruschetta, R.; Pioggia, G.; Cerasa, A.; Tartarisco, G. Application of Machine Learning Algorithms for Prognostic Assessment in Rotator Cuff Pathologies: A Clinical Data-Based Approach. Diagnostics 2023, 13, 2915. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics13182915

AMA Style

Longo UG, Di Naro C, Campisi S, Casciaro C, Bandini B, Pareek A, Bruschetta R, Pioggia G, Cerasa A, Tartarisco G. Application of Machine Learning Algorithms for Prognostic Assessment in Rotator Cuff Pathologies: A Clinical Data-Based Approach. Diagnostics. 2023; 13(18):2915. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics13182915

Chicago/Turabian Style

Longo, Umile Giuseppe, Calogero Di Naro, Simona Campisi, Carlo Casciaro, Benedetta Bandini, Ayoosh Pareek, Roberta Bruschetta, Giovanni Pioggia, Antonio Cerasa, and Gennaro Tartarisco. 2023. "Application of Machine Learning Algorithms for Prognostic Assessment in Rotator Cuff Pathologies: A Clinical Data-Based Approach" Diagnostics 13, no. 18: 2915. https://0-doi-org.brum.beds.ac.uk/10.3390/diagnostics13182915

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop