Next Article in Journal
Heritability Estimation of Multiple Sclerosis Related Plasma Protein Levels in Sardinian Families with Immunochip Genotyping Data
Next Article in Special Issue
Predicting Health Risks of Adult Asthmatics Susceptible to Indoor Air Quality Using Improved Logistic and Quantile Regression Models
Previous Article in Journal
Prediction and Classification of COVID-19 Admissions to Intensive Care Units (ICU) Using Weighted Radial Kernel SVM Coupled with Recursive Feature Elimination (RFE)
Previous Article in Special Issue
MegaD: Deep Learning for Rapid and Accurate Disease Status Prediction of Metagenomic Samples
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Intelligent Learning System for Unbiased Prediction of Dementia Based on Autoencoder and Adaboost Ensemble Learning

by
Ashir Javeed
1,2,†,
Ana Luiza Dallora
2,†,
Johan Sanmartin Berglund
2 and
Peter Anderberg
2,3,*
1
Aging Research Center, Karolinska Institutet, 171 65 Stockholm, Sweden
2
Department of Health, Blekinge Institute of Technology, 371 79 Karlskrona, Sweden
3
School of Health Sciences, University of Skövde, 541 28 Skövde, Sweden
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Submission received: 26 May 2022 / Revised: 22 June 2022 / Accepted: 28 June 2022 / Published: 21 July 2022

Abstract

:
Dementia is a neurological condition that primarily affects older adults and there is still no cure or therapy available to cure it. The symptoms of dementia can appear as early as 10 years before the beginning of actual diagnosed dementia. Hence, machine learning (ML) researchers have presented several methods for early detection of dementia based on symptoms. However, these techniques suffer from two major flaws. The first issue is the bias of ML models caused by imbalanced classes in the dataset. Past research did not address this issue well and did not take preventative precautions. Different ML models were developed to illustrate this bias. To alleviate the problem of bias, we deployed a synthetic minority oversampling technique (SMOTE) to balance the training process of the proposed ML model. The second issue is the poor classification accuracy of ML models, which leads to a limited clinical significance. To improve dementia prediction accuracy, we proposed an intelligent learning system that is a hybrid of an autoencoder and adaptive boost model. The autoencoder is used to extract relevant features from the feature space and the Adaboost model is deployed for the classification of dementia by using an extracted subset of features. The hyperparameters of the Adaboost model are fine-tuned using a grid search algorithm. Experimental findings reveal that the suggested learning system outperforms eleven similar systems which were proposed in the literature. Furthermore, it was also observed that the proposed learning system improves the strength of the conventional Adaboost model by 9.8% and reduces its time complexity. Lastly, the proposed learning system achieved classification accuracy of 90.23%, sensitivity of 98.00% and specificity of 96.65%.

1. Introduction

Dementia is a mental condition defined by a steady decline in cognitive processes that interfere with everyday living tasks such as memory, problem solving, visual perception, and capacity to focus on a specific task [1]. Usually, older adults are more prone to dementia and many people believe that it is an unavoidable result of aging, which is perhaps incorrect perception. Dementia is not a natural part of the aging process; instead, it should be regarded as a substantial cognitive deterioration that interferes with everyday life. The fundamental cause of dementia development is a variety of disorders and traumas to the human brain [2]. The number of dementia patients is rapidly increasing worldwide and statistical projections suggest that 135 million people might be affected with dementia by 2050 [3]. Furthermore, dementia is ranked on the seventh place in the leading causes of deaths in the world [4] and it is the major cause of disability and dependency among older adults globally [4].
The conventional diagnostic assessment of dementia involves medical history, clinical examinations (e.g., neurological, mental state, and cognitive examinations) and an interview with a relative other than the informant [5]. Current early-stage dementia diagnosis is based on pathological characteristics or cognitive diagnostic tests. Neuroimaging can detect pathology characteristics. Magnetic resonance imaging (MRI) is used to examine the change in neuron-structure [6,7]. The electroencephalography (EEG) is used to evaluate event-related potentials to diagnose early stages of dementia in patients [8,9]. Patel et al., combined EEG and MRI imaging to improve the detection of the early stage of dementia [10]. However, such tools are insufficient for identifying dementia since the cost of testing is prohibitively high, and the testing method is too lengthy and intrusive. Furthermore, recent research suggests computed tomography (CT) or MRI of the brain to rule out structural causes for the clinical phenotype [1,11]. It has been estimated that primary-care clinicians fail to diagnose anywhere from 29% to 76% of patients with dementia or probable dementia [11].
Along with a reliable diagnostic process, appropriate handling must be simple for dementia patients. There are benefits for employing cognitive tests to determine the early stage of dementia since they are quick and easy to do; nevertheless, it is difficult for paramedics to contact patients and promote the testing because elderly individuals often dread attending hospitals. The only way to do tests is through unskilled relatives, who do not completely comprehend the scales. As a result, test findings are often inaccurate. ML algorithms provide a novel answer to this challenge. Paramedics now have an improved access to patients’ lives because of information technology, and they can detect poor cognitive function at an early stage. Additionally, ML algorithms can provide expert medical knowledge. An automated diagnostic systems based on ML techniques can give a high accuracy and user-friendly method of detecting the early stages of dementia. Based on ML approaches, scientists have developed several automated diagnostics systems for various diseases e.g., heart failure [12,13,14,15,16], Parkinson [17], hepatitis [18] and clinical decision support systems [19].
Ana W. Capuano et al. presented an assessment of dementia risk for older adults based on derivation and validation [20]. In their study, the RADaR (Rapid Risk Assessment of Dementia) discrimination was good for the derivation and external-validation cohorts (AUC of the 3-year prediction = 0.82–0.86), compared to age alone (AUC of the 3-year prediction = 0.73), which is a key risk factor for dementia. The inclusion of genetic information did not improve the discrimination. F. V. Cederwald et al., investigated how the continuing trajectory of cardiovascular risk impacts the likelihood of dementia and memory impairment in the future. For this purpose, they used a Bayesian additive regression tree as a multistate survival analysis method [21]. J. R. Cejudo et al. used the cumulative incidence function and inverse probability weighted Cox proportional hazards regression models with adjustments for demographic and clinical covariates, to investigate whether platelet function is associated with dementia risk in the Framingham Heart Study [22]. Statistical models are useful for determining relationships between variables, but they perform poorly when it comes to predicting outcomes, such as disease prediction. ML models often perform better at predicting results, therefore, researchers are employing ML for disease detection.

1.1. Machine Learning for Dementia

Several automated diagnostic systems were proposed in the literature for the early detection of dementia using ML approaches. Dallora et al. [23] investigated predictive factors for the 10-year prediction of dementia based on decision trees (DT) using the Swedish National study on Aging and Care (SNAC) database. In their proposed method, they deployed a recursive feature elimination (RFE) feature selection method in order to select the most important variables from dataset for the classification of dementia. Their proposed method based on RFE and DT had achieved the highest area under the curve (AUC) of 74.50%. D. Stamate et al., developed a framework for the prediction of mild cognitive impairment (MCI) and dementia. Their proposed framework was based on the Relief approach paired with statistical permutation tests for feature selection, model training, tweaking and testing using ML algorithms such as random forest (RF), support vector machine (SVM), gaussian processes, stochastic gradient boosting and extreme gradient boosting. The stability of model performances were studied using computationally expensive monte carlo simulations. Their results for the dementia detection were: an accuracy of 88.00%, sensitivity of 93.00%, and the specificity of 94.00%, whereas the results for the moderate cognitive impairment detection showed a sensitivity of 86.00% and specificity of 90% [24]. Visser et al., developed a system for detecting subtypes of dementia from blood samples while utilizing deep learning (DL) and other supervised ML approaches such as RF and extreme gradient boosting. The AUC for proposed DL method was 85% (0.80–0.89), for xgboost it was 88% (0.86–0.89), and for RF it was 85% (0.83–0.87). In comparison, cerebrospinal fluid (CSF) measurements of amyloid, p-tau, and t-tau (together with age and gender) gave AUC values of 78%, 83%, and 87%, respectively by using xgboost [25]. P. Gurevich et al., used SVM and neuropsychologic factors and achieved 89.00% accuracy through their proposed method [26]. M. Karaglani et al., proposed an automated diagnosis system for Alzheimer’s disease (AD) by using blood-based biosignatures. In their proposed method, they employed mRNA-based statistically equivalent signatures for feature ranking and a RF model for classification. Their proposed automated diagnostics system reported an accuracy of 84.60% [27]. E. Ryzhikova et al., analyzed CSF using ML algorithms for the diagnosis of AD. For the classification purpose, artificial neural networks (ANN) and SVM discriminant analysis (SVM-DA) methods were applied for distinguishing AD and hippocampal (HC) participants with 84.00% sensitivity and specificity. The proposed classification models had a high discriminative power, implying that the technique had a lot of potential for AD diagnosis [28]. P.C Cho & W.H Chen designed a double layer dementia diagnosis system based on ML where fuzzy cognitive maps (FCMs) and probability neural networks (PNNs) were used to provide the initial diagnosis at the base layer and Bayesian networks (BNs) were used to provide final diagnosis at the top layer. The highest accuracy reported by their proposed system was 83.00% [29]. Multimodal medical signals fusion for smart healthcare systems also important for designing and development of automated diagnostic systems for the prediction of diseases [30].

1.2. State-of-the-Art Work

F. A. Salem et al., presented a regression-based ML model for the prediction of dementia. In their proposed method, they investigated ML approaches for imbalanced classes in the dataset. They started with intentionally oversampling the minority class and undersampling the majority class, in order to reduce the biasness of ML model. Furthermore, they deployed cost-sensitive strategies to penalize the ML models when an instance is misclassified in the minority class. According to their findings, the balanced RF was the most resilient probabilistic model (using only 20 features/variables) with an F1-score of 0.82, G-Mean of 0.88, and AUC of 0.88 using ROC [31]. F.G. Gutierrez et al. had designed an automated diagnostic system for the detection of AD and frontotemporal dementia (FTD) by using feature engineering and genetic algorithms. Their proposed system had obtained an accuracy of 84% [32]. G. Mirzaei & H. Adeli analyzed state-of-the-art ML techniques for the detection and classification of AD [33]. H. Hsiu et al. studied ML algorithms for early identification of cognitive impairment. Their proposed model obtained an accuracy of 70.32% by threefold cross-validation scheme [5]. Several classification models were constructed using different ML models and feature selection methodologies to automate MCI detection based on gait biomarkers. The ML model by A. Shahzad et al. [34] used for mild cognitive impairment (MCI) pre-screening based on inertial sensor-derived gait biomarkers achieved an accuracy of 71.67 % and sensitivity of 83.33 %.
ML algorithms work best when the samples are roughly evenly split in the dataset. However, dementia has a rather uncommon occurrence, thus balancing sampling the sample must occur in order to build datasets.

1.3. Aim of Study

In this article, we have addressed two challenges of dementia prediction using the SNAC dataset, such as bias in the developed ML models and lower accuracy of dementia detection. To show the problem of bias in ML models, we have constructed and trained six distinct ML models i.e., Logistic Regression (LR), K Nearest Neighbors (KNN), Gaussian Naive Bayes (GNB), Support Vector Machine (SVM), Decision Tree (DT), and Random Forest (RF). We used the synthetic minority oversampling technique (SMOTE) method to overcome this problem. The second issue is a poor rate of accuracy for dementia prediction while using SNAC dataset. We develop an intelligent learning system that is a hybrid with autoencoder and adaptive boosting (Adaboost) learning models to address the issue of low accuracy of dementia detection. The autoencoder is utilized for feature extraction, whereas Adaboost is employed for the classification of dementia patients versus healthy subjects. The experimental findings clearly reveal that the offered solutions assist in the alleviation of both problems to some extent.
It is important to note that dementia has numerous subtypes with the most prevalent being Alzheimer’s disease, Vascular dementia, dementia with Lewy Bodies, and Frontotemporal dementia. However, mixed pathologies are not uncommon, particularly Alzheimer’s disease often coexists with Vascular or Lewy Bodies dementia. In addition, unusual subtypes are sometimes mistakenly diagnosed for Alzheimer’s disease [35]. The research described here makes no difference between subtypes, and the word “dementia” refers to all types of dementia.

2. Materials and Methods

2.1. Dataset Description

The data utilized in this study is a subset of the Swedish National Study on Aging and Care (SNAC). The SNAC is a longitudinal cohort that has been collecting multifactorial data from the Swedish older adult population with the goal of “creating trustworthy, comparable, longitudinal datasets” that will represent an effective infrastructure for aging research and care provision to the elderly [36]. The SNAC (https://www.snac-k.se/) was created as a multipurpose project to study the health and social care of the aging population, and it includes a database consisting of records about physical examination, psychological assessment, social factors, lifestyle factors, medical history etc.
The SNAC data is gathered from four different locations, which represent two Swedish counties i.e., borough and municipality: Skåne, Blekinge, Kungsholmen, and Nordanstig. The SNAC-Blekinge baseline assessment is selected in this study, with data collected from 2000 to 2003. Although there is evidence in the literature that environmental variables may have a role in the incidence of dementia [37,38], this study is based on generic criteria and no distinctions are made between urban and rural locations. Subjects are excluded from this study based on the following criteria: (i) subjects who already had dementia at baseline; (ii) subjects who have missing values at the outcome variable (dementia diagnosis at the 10-year mark); (iii) subjects who have more than 10% missing values in the input variables; (iv) subjects who died before the 10-year study mark; and (v) subjects who were diagnosed with dementia before the 10-year mark, as they could already have advanced dementia.
The SNAC Blekinge baseline included 1402 people. Following the application of aforementioned exclusion criteria, the research sample consisted of 726 people (313 males and 413 females), of which 91 (12.5 %) had dementia at the 10-year point and 635 (87.5 %) did not. Table 1 shows the demographics of research sample in the selected dataset. The variables selection from the SNAC-Blekinge database was based on information from the literature that indicate the impact of selected variables on the dementia disorder [39,40]. It is noteworthy during the selection of variables from SNAC-Blekinge database that there were no differences established between dementia subtypes since mixed pathologies are widespread and rare subtypes are frequently misdiagnosed as Alzheimer’s disease [35].
It is also worth mentioning that all of the variables used for the SNAC project were chosen based on evidence of importance in the aging process (health/disease, social and support network, lifestyle factors, material conditions, and personal resources), as well as statistics on care service utilization [36]. At the study’s baseline (2000–2003) 75 variables were chosen from the following categories: demographic, social, lifestyle, medical history, blood test, physical examination, psychological, and the assessment of numerous health instruments related to dementia evaluation. The list of selected variables can be depicted from Table 2.
The target variable that is used to predict the dementia by the proposed model is given by medical doctors at the mark of 10 years following the SNAC baseline. The International Statistical Classification of Diseases and Related Health Problems-10th Revision (ICD-10) and the Diagnostic and Statistical Manual of Mental Disorders (DSM-IV) were used to make the dementia diagnosis (DSM-5).

2.2. Data Preparation

To deal with missing data, the K-Nearest Neighbors (KNN) multiple imputation method was used [51]. This strategy works by locating the K data entries which are most similar (near) to a missing data item. The KNN imputation fills in the missing values with the mean (for numeric variables) or the most common value (for categorical variables) of the K, the most similar neighbors [51]. In this study, the KNN imputation was used independently on items from the majority (no dementia at 10 years) and minority classes (dementia at 10 years mark) because of the significant class imbalance (12.5% on the minority class against 87.5% on the majority class). This way the danger of contaminating the minority class with data from the majority class was reduced. This is consistent with the literature on missing values on binary answer decision trees, which has demonstrated that when imputation is done independently, classification performance improves [52]. After dealing with missing values in the dataset, we have performed a normalization and standardization operation on the selected dataset to improve the quality of data [53]. The performance of ML algorithms improves after data standardization.
Since the dementia dataset has only numeric values with different scales, we have applied a standard-scaler function on the data. The standard-scalar function work as rescaling the distribution of the data values so that the mean of observed value is 0 and standard deviation is 1 [54].

2.3. Proposed Model

In this paper, we developed an intelligent learning system for dementia detection using electronic health record (EHR) data. The proposed learning system is cascaded by three algorithms i.e., autoencoder with synthetic minority oversampling technique (SMOTE) and Adaboost ensemble learning model in order to improve the performance of the proposed model. The autoencoder is used for features extraction so that the proposed model does not learn noisy or irrelevant information from the feature space which causes overfitting in the ML model. SMOTE is deployed to deal with the problem of imbalance classes in the dataset and Adaboost model is used as a predictive model to detect presence or absence of dementia in the population of older adults. To understand the working of proposed learning system, Figure 1 presents an overview of the newly developed model.
Hereby, the operation of proposed learning system components, namely autoencoder, SMOTE, and Adaboost, are given as follows:
The architecture of the autoencoder consists of two parts, one is the encoder while other is the decoder. The encoder part is used to extract features from the reduced dataset. The feedforward neural network is the simplest form of autoencoder with a single layer perceptron that participates in a multilayer perceptron (MLP) with one or more hidden layers between the input and output layer of the neural network. The number of neurons in input and output layer are equal. The purpose behind the equal number of neurons is to reduce the information loss from the original dataset. Thus, autoencoder uses unsupervised learning. The mathematical formation of autoencoder and decoder is given as:
σ : λ ω , : ω λ : σ , = a r g σ , m i n λ ( σ ) λ 2
where σ and are the transition of autoencoder and λ is a given input to target value ω . The hidden layer of neural network take input at the encoding stage is given as:
X ε t = λ a n d m a p t o h ε n = ω
h = Υ ( β X + u )
where h is the data and referred as code, latent variables σ is an activation function, such as sigmoid function. β is a weight matrix and u is a bias vector. Weight and biases values are iteratively updated by backpropagation during training phase.
The decoder stage of the autoencoder maps h to the features extraction X of the same shape of X.
h = Υ ( β X + u )
where Υ , β and u might be irrelevant to corresponding Υ , β and u for the encoder.
For minimize the reconstruction errors of the autoencoders during training phase is referred as information loss and given as:
Ψ ( X , X ) = X X 2 = x Υ ( β ( Υ ( β X + u ) ) + u ) 2
where, X is the average on the training set and autoencoder training is performed through backpropagation of the error, such as other feedforward neural networks.
Following the features extraction from autoencoder, data partitioning occurs for training and testing of the proposed predictive model. The SMOTE was employed to prevent biasness in proposed learning model for an unbiased prediction of dementia [55]. SMOTE generates synthetic samples of the minority class that results in balanced classes in the dataset. Thus, proposed learning system is trained on balance data by avoiding the biasness factors of ML model due to imbalance classes in the dataset. It is important to mention that the SMOTE is applied on training data following data partitioning. If the SMOTE algorithm is used for balancing the classes on the whole dataset (i.e., prior to data partitioning) then, it would result in biased performance of ML model since samples from the testing dataset would also be included in the training dataset [56]. SMOTE, in contrast to other oversampling approaches, acts in the feature space rather than the data space [55]. It synthesizes (i.e., oversamples) minority class samples by taking a sample from the minority class and creating new samples along the line that links any or all of the k-minority class nearest neighbors. Figure 2 presents the minority and majority class distribution in the dataset before and after the implementation of SMOTE. In this study, we use “imbalanced learn”, a Python-based library to employ the SMOTE technique [57].
Boosting is an ensemble learning strategy that combines the learning ability of weak learners to construct a strong learning model. Freund and Schapire introduced the first practical boosting ensemble model, adaptive boosting or Adaboost [58]. In other words, the Adaboost model transforms a collection of weak classifiers or estimators into a powerful one. It combines the result of various learning algorithms (weak learners or estimators) by assessing their weighted total, which represents the boosting ensemble model’s ultimate output. The final equation of the Adaboost model for classification is as follows:
U ( x ) = s i g n t = 1 T λ t u t ( x )
where μ t represents the tth weak classifier and λ t is its associated weight. Equation (6) shows that the Adaboost model is a weighted mixture of T weak learners or estimators. Details on the operation and formulation of the Adaboost model can be found in [59,60]. In this study, we briefly explore the Adaboost model formulation as follows:
For a given dataset with n occurrences and binary labels (i.e., taking the case of binary classification as studied in this research), the feature vector v and class label c may be represented as v i ε R h , c i ε 1 , + 1 where −1 represents the negative class (absence of dementia) and +1 represents the positive class(presence of dementia). Weights for each data point are initialized in the first phase as follows:
ϖ v i , c i = 1 n i = 1 , 2 , 3 , . . . n
Then, we iterate from t = 1 to T, applying weak classifiers to the dataset and selecting the one with the lowest weighted classification error.
k t = K ϖ t 1 c u ( x )
The weight for t t h weak classifier or estimator is then determined as follows:
γ t = 1 2 l n 1 k t k t
Any classifier (weak estimator) with an accuracy greater than 50% will have a positive weight. Furthermore, larger weights will be assigned to more accurate classifiers. Classifiers with less than 50% accuracy, on the other hand, will have negative weights. Adaboost combines such classifier predictions by flipping their sign. As a result of the sign flipping of its prediction, a classifier with 30% accuracy can be changed into one with 70% accuracy. Only classifiers with an exact 50% accuracy have no contribution to the final prediction.
ϖ t + 1 v i , c i = ϖ t v i , c i e x p γ t c i u t ( x i ) Z t
where Z t is a normalizing factor used to make the sum of all instance weights equal to one. Additionally, it is evident from Equation (10), that the “exp” term will always be greater than 1 when the misclassified example is from a positive weighted classifier (i.e., γ t is always positive and c × u is always negative). After each cycle, the misclassified instances will be updated with higher weights. The same concept is used for negative weighted classifiers, with the exception that the initial accurate classifications become misclassifications once the sign is flipped. Finally, after T iterations, the Adaboost model will acquire a final prediction by averaging each classifier’s weighted prediction (i.e., weak estimator).
In this research work, we implemented the Adaboost ensemble model in Python software package using scikit-learn module [61]. The Adaboost model’s hyperparameter, i.e., the number of estimators used to generate the final ensemble model, will be indicated by N e s t . Furthermore, the decision tree classifier is employed as the basis estimator. To improve classification performance, we use an exhaustive search technique to find the ideal value of hyperparameter of Adaboost model (i.e., N e s t , learning rate: l r ) that results as the optimal Adaboost model which helped to yield best performance.

2.4. Validation & Evaluation

To test the efficacy of the proposed learning system, we employ holdout validation scheme and cross-validation. For hold-out validation, we split the dataset into 70% and 30% ratio for training and testing purposes, respectively. To establish the efficacy of the proposed learning system, we tested the proposed model against a range of evaluation metrics. i.e., accuracy, sensitivity, specificity, F-score or F-measure, and Mathew’s correlation coefficient (MCC). To test the efficacy of the proposed learning system using receiver operator curve (ROC) and area under the curve (AUC), we employ a stratified k-fold validation strategy with k = 6. Traditional accuracy metrics fail to reflect a model’s genuine behavior, as illustrated in experiment 1 of Section 3 of this study. Thus, we used the balanced accuracy metric, which more accurately reflects the real behavior of the built models [62,63,64]. Pereira et al. [65] utilized a similar accuracy metric (global accuracy [65]) that was proposed by Papa et al. [66]. This accuracy metric is also a suitable choice for reflecting a model’s genuine behavior when trained on imbalanced data. In the following formulation, ACC stands for the commonly used accuracy metric, while A C C b a l stands for the balanced accuracy metric. The mathematical formulation of the used assessment metrics is given as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
where TP stands for the number of true positives, FP stands for the number of false positives, TN stands for the number of true negatives, and FN stands for the number of false negatives.
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P
A C C b a l = S e n s i t i v i t y + S p e c i f i c i t y 2
M C C = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
F = 2 T P T P + F N + F P
In a statistical analysis of binary classification problem, F signifies F-score, also known as F-measure or F1 score. F-score yields a value range between 0 and 1, where 1 represents perfect forecasts and 0 represents the poorest. MCC is used to assess the correctness of a test. MCC can have a value range between −1 and 1, where 1 represents the perfect forecast and −1 represents the poorest forecasts. Consider an example of 100 individuals among them 90 people sufferers from dementia and 10 individuals are healthy subjects, to highlight the benefits of employing a balanced accuracy metric. If we build a model that always predicts a subject to be a dementia patient, then it would have 100% sensitivity but 0% specificity and traditional accuracy of 90%. However, the balance accuracy will be 50%. It is evident that the true behavior of the constructed model is reflected by balanced accuracy, because model can detect only one class but completely failed to detect the second class. However, the traditional accuracy failed to describe the genuine behavior of the constructed model.

3. Results

Three different types of experiments were carried out to rigorously assess the performance and efficacy of the newly proposed system for dementia prediction. In the first experiment, we have demonstrated the impact of imbalance classes in the dataset using six conventional ML models. While in the second experiment, the traditional Adaboost algorithm is fine-tuned using a grid search algorithm and tested on the balance dementia dataset. The second experiment is extended and in the second phase of the experiment, the dataset is preprocessed through data standardization and normalization. Following that, the newly proposed method based on autoencoder and Adaboost is tested on the processed data with balanced classes in dataset using the SMOTE method. We have also compared the results of the newly proposed model against the traditional Adaboost model on the balanced dataset. In the third experiment, other conventional ML methods are fine-tuned and tested with features extracted from the autoencoder on the same balanced dementia dataset for performance comparison. All experiments are carried out on a system powered by an Intel (R) Core (TM) i5-8250U CPU running at 1.60GHz and running Windows 10 Home 64bits(Blekinge Institute of Technology, Karlskrona, Sweden) as the operating system. All of the experiments make use of the Python software package as a software tool.

3.1. Experiment 1: Impact of Imbalance Classes in the Dataset

In this section, we have employed several ML models (NB, LR, kNN, SVM, RF, DT) to demonstrate the impact of imbalanced data for the prediction of dementia. From the Table 3, it can be depicted that ML models are sensitive to the imbalanced data. The employed ML models are clearly biased in favor of majority class. For instance, it can be observed from the Table 3 that we obtained high a rate of specificity and low rate of sensitivity (see Table 3) when ML models are trained on imbalanced data.
To avoid this biasness problem, we take a step to balance the training data and for this purpose we deployed the SMOTE method [55] to balance the size of each class in the training data. After balancing the data, it is evident from Table 4 that the performance of ML models is improved i.e., the ML models do not suffer from the biased performance as it can be seen from the values of specificity and sensitivity. Furthermore, we have also studied the performance of ML models based on the ROC evaluation metric while using imbalance data. Figure 3 presents the performance evaluation of ML models based on AUC by using ROC curve.

3.2. Experiment 2: Comparative Study with Conventional Adaboost for Dementia Prediction

This experiment has two phases: in the first phase, we deployed a conventional Adaboost model with hyperparameters fine-tuned using a grid search algorithm on the balanced dataset. The performance of this model is assessed using the ROC. Figure 4b, presents the results of this experiment, in which the conventional Adaboost obtained an average accuracy of 82.00% using all 75 features of the balanced dataset based on the k-fold evaluation metric. While in second phase of this experiment, we evaluated the performance of the newly proposed autoencoder-SMOTE-Adaboost model for dementia patient classification. We employed the autoencoder to extract features from the dataset, which not only helped to increase the Adaboost accuracy but also greatly decreased the time complexity of the proposed model by reducing the data dimensionality. The extracted features by the autoencoder are given as input to the Adaboost model and the hyperparameters of the Adaboost was fine-tuned using a grid search algorithm that assisted in determining the optimal number of estimators (N e s t ) and the learning rate of the Adaboost model. The obtained accuracy along with other performance evaluation metrics are given in Table 5. It can be depicted from Table 5 that the newly proposed autoencoder-SMOTE-Adaboost model has achieved the best accuracy on testing data of 90.23% and an accuracy of 92.10% on training data. This was achieved with the best number of estimators (N e s t ) of 10 and learning rate (l r ) of 0.05.
Furthermore, the result of this experiment can be observed from the Figure 4a where the proposed model achieved an average AUC of 90.00% based on the k-fold evaluation metric. For both phases of the experiments, we have taken the same value of K = 6, so that fair comparison is done. The overall performance comparison based on AUC between conventional Adaboost model and the proposed model on balanced dataset is shown in the Figure 4.

3.3. Experiment 3: Performance Comparison of the Proposed Model with Other ML Models

We have constructed various comparable prediction systems such as hybridizing autoencoder with Naive Bayes (NB), Logistic Regression (LR), Random Forest (RF), Decision Tree (DT), K Nearest Neighbors (KNN) and Support vector machine (SVM) to test the efficiency of the newly proposed learning system. Table 6 presents the outcomes of each constructed hybrid model. It is noteworthy that all of these constructed hybrid models are tested on balanced data using SMOTE technique for data balancing.
From Table 6 it can be observed that the newly proposed model has achieved the accuracy of 90.23% while using only a small subset of extracted features (06) in comparison to the rest of the ML models. Hence, the proposed model also reduces the complexity of the Adaboost predictive model as training on a smaller number of features will result in reducing training time of the ML model.

4. Discussion

In this study, an intelligent learning system is presented for the prediction of dementia using the SNAC dataset. We used 75 features from the SNAC dataset related to demographic, social, lifestyle, medical history, biochemical tests, physical examination, psychological assessment and diverse health instruments relevant to the dementia disorder.
To improve the accuracy of the proposed model along with lower time complexity, we have deployed an autoencoder to reduce the data dimensionality. Based on an artificial neural network, the autoencoder helped to extract useful features from the feature space. After extracting features from the dataset, it was observed that the classes in the dataset were highly imbalanced. To balance the class distribution in the training set of the classifier, we have used SMOTE and for the classification of dementia patients the Adaboost ensemble model was employed. The hyperparameters of the Adaboost model were fine tunned using a grid search algorithm. Thus, the proposed learning model consist of two modules which are hybrid as a single system.
From the results, it can be observed that the proposed model dealt effectively with both problems, imbalance classes in dataset and lower accuracy of ML models for dementia prediction. Experiment 1, addressed the impact of imbalanced classes in the dataset for the prediction of dementia. From Table 3, it can be observed that performance of ML models tends to bias toward the majority class with the ML models achieving higher results of specificity and lower results for sensitivity. This means that the ML models tend to bias the majority class in the dataset. The proposed model has not only resolved the issue of bias results but also improved the accuracy of dementia prediction. Table 5 presents the performance of the proposed model along with hyperparameter values of the Adaboost model. It can be depicted from the Table 5 that the newly proposed model achieved the highest accuracy of 90.23% on testing data, training accuracy of 92.10%, sensitivity of 98.00% and specificity of 96.65% while using only 6 features, which are extracted by the autoencoder. The learning rate of the Adaboost was (l r ) 0.05 and the number of estimators were (N e s t ) 10.
Furthermore, we have also compared the results of the newly proposed model with other state-of-the-art ML models which were proposed in the literature for dementia prediction. It can be observed from the Table 7 that the proposed model has achieved significantly improved results when compared to other ML models.

5. Conclusions & Future Work

In this paper, we have identified the problems of lower accuracy and bias in the ML models due to imbalanced classes in the dataset for dementia prediction. From experiments, it is demonstrated that when ML models are trained on imbalanced data, their performance is skewed towards the data’s majority class. As a result, for the dementia detection problem, we found a high rate of specificity, but a poor rate of sensitivity since the dementia patient’s class was in the minority and healthy subject class was in the majority. To deal with the bias problem, we presented a novel diagnostic system for the detection of dementia. In our proposed model, the SMOTE technique is employed to eliminate the problem of imbalanced classes in the dataset. The proposed model has two main components which are hybridized and work as a single learning system. The first component work to extract useful features from the dataset for reducing data dimensionality, which helps to lower the computational complexity of the proposed model and improve the accuracy of dementia prediction. For this purpose, we have employed an autoencoder which has reduced the number of features from 75 to 6. The second component of the newly proposed model works as classifier and for this task, we utilized the Adaboost classifier. The hyperparameters of Adaboost model were fine-tuned using a grid search algorithm. From the experimental results, it is observed that the newly proposed model has outperformed the traditional Adaboost model along with other state-of-the-art ML models that also used extracted features from the autoencoder. It was also observed that the proposed learning system improved the performance of a conventional Adaboost model by 9.8%. Moreover, the proposed model has also demonstrated a lower time complexity as compared to the traditional Adaboost model because the proposed model uses a fewer number of features than the traditional Adaboost model.
Therefore, the problem of bias in the developed ML models was avoided in this work, and an unbiased learning model was designed that enhanced dementia diagnosis accuracy while also lowering the complexity of ML models by reducing the number of features. However, the achieved accuracy still needs significant improvement. This is a shortcoming of this study. Future research should focus on developing more robust models that can enhance dementia diagnosis accuracy while keeping the unbiased behavior of the developed models. This could be possible by combining feature extraction approaches with deep learning models. Furthermore, when the number of samples in the dataset is large, the performance of ML techniques improves. The dataset employed in this study has only 721 samples, which is rather small in terms of sample size. As a result, researchers must develop datasets with large sample sizes, particularly for dementia.

Supplementary Materials

The following supporting information can be downloaded at: https://0-www-mdpi-com.brum.beds.ac.uk/article/10.3390/life12071097/s1. Table S1: Description of all input variables. Refs. [41,43,44,45,46,47,48,50,67,68] are cited in Supplementary materials.

Author Contributions

Conceptualization and validated by P.A., Data curation and formal analysis by A.L.D., Methodology and write up by A.J., Supervised by J.S.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was carried out in accordance with the Declaration of Helsinki and was approved by the Research Ethics Committee at Blekinge Institute of Technology (BTH).

Informed Consent Statement

Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

Data can be available in Supplementary Materials.

Acknowledgments

The first author’s learning process was supported by the National E-Infrastructure for Aging Research (NEAR), Sweden. NEAR is working on improving the health condition of older adults in Sweden.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PNNsProbability neural networks
SVMSupport vector machine
MLPMultilayer perceptron
MLMachine learning
RFRandom forest
GAGenetic algorithm
DTDecision tree
LRLogistic regression
NBNaive bayes
kNNk-nearest neighbors
SMOTESynthetic minority oversampling technique
ROCReceiver operator curve
AUCArea under the curve

References

  1. Creavin, S.T.; Wisniewski, S.; Noel-Storr, A.H.; Trevelyan, C.M.; Hampton, T.; Rayment, D.; Thom, V.M.; Nash, K.J.; Elhamoui, H.; Milligan, R.; et al. Mini-Mental State Examination (MMSE) for the detection of dementia in clinically unevaluated people aged 65 and over in community and primary care populations. Cochrane Database Syst. Rev. 2016. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Lo, R.Y. The borderland between normal aging and dementia. Tzu-Chi Med. J. 2017, 29, 65. [Google Scholar]
  3. Vrijsen, J.; Matulessij, T.; Joxhorst, T.; de Rooij, S.E.; Smidt, N. Knowledge, health beliefs and attitudes towards dementia and dementia risk reduction among the Dutch general population: A cross-sectional study. BMC Public Health 2021, 21, 857. [Google Scholar] [CrossRef]
  4. WHO. Dementia. Available online: https://www.who.int/news-room/fact-sheets/detail/dementia (accessed on 25 May 2022).
  5. Hsiu, H.; Lin, S.K.; Weng, W.L.; Hung, C.M.; Chang, C.K.; Lee, C.C.; Chen, C.T. Discrimination of the Cognitive Function of Community Subjects Using the Arterial Pulse Spectrum and Machine-Learning Analysis. Sensors 2022, 22, 806. [Google Scholar] [CrossRef] [PubMed]
  6. Studholme, C.; Cardenas, V.; Song, E.; Ezekiel, F.; Maudsley, A.; Weiner, M. Accurate template-based correction of brain MRI intensity distortion with application to dementia and aging. IEEE Trans. Med. Imaging 2004, 23, 99–110. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Duchesne, S.; Caroli, A.; Geroldi, C.; Barillot, C.; Frisoni, G.B.; Collins, D.L. MRI-based automated computer classification of probable AD versus normal controls. IEEE Trans. Med. Imaging 2008, 27, 509–520. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Ahiskali, M.; Polikar, R.; Kounios, J.; Green, D.; Clark, C.M. Combining multichannel ERP data for the early diagnosis of Alzheimer’s Disease. In Proceedings of the 2009 4th International IEEE/EMBS Conference on Neural Engineering, Antalya, Turkey, 29 April–2 May 2009; pp. 522–525. [Google Scholar]
  9. Lai, C.L.; Lin, R.T.; Liou, L.M.; Liu, C.K. The role of event-related potentials in cognitive decline in Alzheimer’s disease. Clin. Neurophysiol. 2010, 121, 194–199. [Google Scholar] [CrossRef] [PubMed]
  10. Patel, T.; Polikar, R.; Davatzikos, C.; Clark, C.M. EEG and MRI data fusion for the early diagnosis of Alzheimer’s disease. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–25 August 2008; pp. 1757–1760. [Google Scholar]
  11. Patnode, C.D.; Perdue, L.A.; Rossom, R.C.; Rushkin, M.C.; Redmond, N.; Thomas, R.G.; Lin, J.S. Screening for cognitive impairment in older adults: Updated evidence report and systematic review for the US Preventive Services Task Force. JAMA 2020, 323, 764–785. [Google Scholar] [CrossRef]
  12. Javeed, A.; Rizvi, S.S.; Zhou, S.; Riaz, R.; Khan, S.U.; Kwon, S.J. Heart risk failure prediction using a novel feature selection method for feature refinement and neural network for classification. Mob. Inf. Syst. 2020, 2020. [Google Scholar] [CrossRef]
  13. Javeed, A.; Zhou, S.; Yongjian, L.; Qasim, I.; Noor, A.; Nour, R. An intelligent learning system based on random search algorithm and optimized random forest model for improved heart disease detection. IEEE Access 2019, 7, 180235–180243. [Google Scholar] [CrossRef]
  14. Ali, L.; Rahman, A.; Khan, A.; Zhou, M.; Javeed, A.; Khan, J.A. An automated diagnostic system for heart disease prediction based on χ2 statistical model and optimally configured deep neural network. IEEE Access 2019, 7, 34938–34945. [Google Scholar] [CrossRef]
  15. Javeed, A.; Khan, S.U.; Ali, L.; Ali, S.; Imrana, Y.; Rahman, A. Machine learning-based automated diagnostic systems developed for heart failure prediction using different types of data modalities: A systematic review and future directions. Comput. Math. Methods Med. 2022, 2022. [Google Scholar] [CrossRef] [PubMed]
  16. Ali, L.; Bukhari, S. An approach based on mutually informed neural networks to optimize the generalization capabilities of decision support systems developed for heart failure prediction. IRBM 2021, 42, 345–352. [Google Scholar] [CrossRef]
  17. Ali, L.; Zhu, C.; Golilarz, N.A.; Javeed, A.; Zhou, M.; Liu, Y. Reliable Parkinson’s disease detection by analyzing handwritten drawings: Construction of an unbiased cascaded learning system based on feature selection and adaptive boosting model. IEEE Access 2019, 7, 116480–116489. [Google Scholar] [CrossRef]
  18. Akbar, W.; Wu, W.p.; Saleem, S.; Farhan, M.; Saleem, M.A.; Javeed, A.; Ali, L. Development of Hepatitis Disease Detection System by Exploiting Sparsity in Linear Support Vector Machine to Improve Strength of AdaBoost Ensemble Model. Mob. Inf. Syst. 2020, 2020, 8870240. [Google Scholar] [CrossRef]
  19. Javeed, A.; Ali, L.; Mohammed Seid, A.; Ali, A.; Khan, D.; Imrana, Y. A Clinical Decision Support System (CDSS) for Unbiased Prediction of Caesarean Section Based on Features Extraction and Optimized Classification. Comput. Intell. Neurosci. 2022, 2022, 1901735. [Google Scholar] [CrossRef]
  20. Capuano, A.W.; Shah, R.C.; Blanche, P.; Wilson, R.S.; Barnes, L.L.; Bennett, D.A.; Arvanitakis, Z. Derivation and validation of the Rapid Assessment of Dementia Risk (RADaR) for older adults. PLoS ONE 2022, 17, e0265379. [Google Scholar] [CrossRef]
  21. Farnsworth von Cederwald, B.; Josefsson, M.; Wåhlin, A.; Nyberg, L.; Karalija, N. Association of Cardiovascular Risk Trajectory With Cognitive Decline and Incident Dementia. Neurology 2022, 98, e2013–e2022. [Google Scholar] [CrossRef]
  22. Ramos-Cejudo, J.; Johnson, A.D.; Beiser, A.; Seshadri, S.; Salinas, J.; Berger, J.S.; Fillmore, N.R.; Do, N.; Zheng, C.; Kovbasyuk, Z.; et al. Platelet Function Is Associated with Dementia Risk in the Framingham Heart Study. J. Am. Heart Assoc. 2022, 11, e023918. [Google Scholar] [CrossRef]
  23. Dallora, A.L.; Minku, L.; Mendes, E.; Rennemark, M.; Anderberg, P.; Sanmartin Berglund, J. Multifactorial 10-year prior diagnosis prediction model of dementia. Int. J. Environ. Res. Public Health 2020, 17, 6674. [Google Scholar] [CrossRef]
  24. Stamate, D.; Alghamdi, W.; Ogg, J.; Hoile, R.; Murtagh, F. A machine learning framework for predicting dementia and mild cognitive impairment. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 671–678. [Google Scholar]
  25. Visser, P.J.; Lovestone, S.; Legido-Quigley, C. A metabolite-based machine learning approach to diagnose Alzheimer-type dementia in blood: Results from the European Medical Information Framework for Alzheimer disease biomarker discovery cohort. Alzheimer’s Dement. Transl. Res. Clin. Interv. 2019, 5, 933–938. [Google Scholar]
  26. Gurevich, P.; Stuke, H.; Kastrup, A.; Stuke, H.; Hildebrandt, H. Neuropsychological testing and machine learning distinguish Alzheimer’s disease from other causes for cognitive impairment. Front. Aging Neurosci. 2017, 9, 114. [Google Scholar] [CrossRef] [PubMed]
  27. Karaglani, M.; Gourlia, K.; Tsamardinos, I.; Chatzaki, E. Accurate blood-based diagnostic biosignatures for Alzheimer’s disease via automated machine learning. J. Clin. Med. 2020, 9, 3016. [Google Scholar] [CrossRef]
  28. Ryzhikova, E.; Ralbovsky, N.M.; Sikirzhytski, V.; Kazakov, O.; Halamkova, L.; Quinn, J.; Zimmerman, E.A.; Lednev, I.K. Raman spectroscopy and machine learning for biomedical applications: Alzheimer’s disease diagnosis based on the analysis of cerebrospinal fluid. Spectrochim. Acta Part Mol. Biomol. Spectrosc. 2021, 248, 119188. [Google Scholar] [CrossRef] [PubMed]
  29. Cho, P.C.; Chen, W.H. A double layer dementia diagnosis system using machine learning techniques. In International Conference on Engineering Applications of Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 402–412. [Google Scholar]
  30. Muhammad, G.; Alshehri, F.; Karray, F.; El Saddik, A.; Alsulaiman, M.; Falk, T.H. A comprehensive survey on multimodal medical signals fusion for smart healthcare systems. Inf. Fusion 2021, 76, 355–375. [Google Scholar] [CrossRef]
  31. Salem, F.A.; Chaaya, M.; Ghannam, H.; Al Feel, R.E.; El Asmar, K. Regression based machine learning model for dementia diagnosis in a community setting. Alzheimer’s Dement. 2021, 17, e053839. [Google Scholar] [CrossRef]
  32. Garcia-Gutierrez, F.; Delgado-Alvarez, A.; Delgado-Alonso, C.; Díaz-Álvarez, J.; Pytel, V.; Valles-Salgado, M.; Gil, M.J.; Hernández-Lorenzo, L.; Matías-Guiu, J.; Ayala, J.L.; et al. Diagnosis of Alzheimer’s disease and behavioural variant frontotemporal dementia with machine learning-aided neuropsychological assessment using feature engineering and genetic algorithms. Int. J. Geriatr. Psychiatry 2022, 37. [Google Scholar] [CrossRef]
  33. Mirzaei, G.; Adeli, H. Machine learning techniques for diagnosis of alzheimer disease, mild cognitive disorder, and other types of dementia. Biomed. Signal Process. Control 2022, 72, 103293. [Google Scholar] [CrossRef]
  34. Shahzad, A.; Dadlani, A.; Lee, H.; Kim, K. Automated Prescreening of Mild Cognitive Impairment Using Shank-Mounted Inertial Sensors Based Gait Biomarkers. IEEE Access 2022, 10, 15835–15844. [Google Scholar] [CrossRef]
  35. World Health Organization. Dementia: A Public Health Priority; World Health Organization: Geneva, Switzerland, 2012. [Google Scholar]
  36. Lagergren, M.; Fratiglioni, L.; Hallberg, I.R.; Berglund, J.; Elmståhl, S.; Hagberg, B.; Holst, G.; Rennemark, M.; Sjolund, B.M.; Thorslund, M.; et al. A longitudinal study integrating population, care and social services data. The Swedish National study on Aging and Care (SNAC). Aging Clin. Exp. Res. 2004, 16, 158–168. [Google Scholar] [CrossRef]
  37. Nunes, B.; Silva, R.D.; Cruz, V.T.; Roriz, J.M.; Pais, J.; Silva, M.C. Prevalence and pattern of cognitive impairment in rural and urban populations from Northern Portugal. BMC Neurol. 2010, 10, 42. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Killin, L.O.; Starr, J.M.; Shiue, I.J.; Russ, T.C. Environmental risk factors for dementia: A systematic review. BMC Geriatr. 2016, 16, 175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Yu, J.T.; Xu, W.; Tan, C.C.; Andrieu, S.; Suckling, J.; Evangelou, E.; Pan, A.; Zhang, C.; Jia, J.; Feng, L.; et al. Evidence-based prevention of Alzheimer’s disease: Systematic review and meta-analysis of 243 observational prospective studies and 153 randomised controlled trials. J. Neurol. Neurosurg. Psychiatry 2020, 91, 1201–1209. [Google Scholar] [CrossRef] [PubMed]
  40. Arvanitakis, Z.; Shah, R.C.; Bennett, D.A. Diagnosis and management of dementia. JAMA 2019, 322, 1589–1599. [Google Scholar] [CrossRef]
  41. Antonovsky, A. The structure and properties of the sense of coherence scale. Soc. Sci. Med. 1993, 36, 725–733. [Google Scholar] [CrossRef]
  42. Saklofske, D.H.; Schoenberg, M.R. Wechsler adult intelligence scale (all versions). J. Encycl. Clin. Neuropsychol. 2017, 1–11.
  43. Livingston, G.; Blizard, B.; Mann, A. Does sleep disturbance predict depression in elderly people? A study in inner London. Br. J. Gen. Pract. 1993, 43, 445–448. [Google Scholar]
  44. Brooks, R.; Group, E. EuroQol: The current state of play. Health Policy 1996, 37, 53–72. [Google Scholar] [CrossRef]
  45. Katz, S. Assessing self-maintenance: Activities of daily living, mobility, and instrumental activities of daily living. J. Am. Geriatr. Soc. 1983, 31, 721–727. [Google Scholar] [CrossRef]
  46. Lawton, M.P.; Brody, E.M. Assessment of older people: Self-maintaining and instrumental activities of daily living. Gerontologist 1969, 9, 179–186. [Google Scholar] [CrossRef]
  47. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. “Mini-mental state”: A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef]
  48. Agrell, B.; Dehlin, O. The clock-drawing test. Age Ageing 1998, 27, 399–404. [Google Scholar] [CrossRef]
  49. Sriram, V.; Jenkinson, C.; Peters, M. Using rapid cycle tests of change to develop the Carers Assistive Technology Experience Questionnaire: A cognitive interview study in the UK. BMJ Open 2021, 11, e042361. [Google Scholar] [CrossRef] [PubMed]
  50. Montgomery, S.A.; Åsberg, M. A new depression scale designed to be sensitive to change. Br. J. Psychiatry 1979, 134, 382–389. [Google Scholar] [CrossRef]
  51. Zhang, S. Nearest neighbor selection for iteratively kNN imputation. J. Syst. Softw. 2012, 85, 2541–2552. [Google Scholar] [CrossRef]
  52. Ding, Y.; Simonoff, J.S. An investigation of missing data methods for classification trees applied to binary response data. J. Mach. Learn. Res. 2010, 11, 131–170. [Google Scholar]
  53. Pathak, J.; Bailey, K.R.; Beebe, C.E.; Bethard, S.; Carrell, D.S.; Chen, P.J.; Dligach, D.; Endle, C.M.; Hart, L.A.; Haug, P.J.; et al. Normalization and standardization of electronic health records for high-throughput phenotyping: The SHARPn consortium. J. Am. Med. Inform. Assoc. 2013, 20, e341–e348. [Google Scholar] [CrossRef] [Green Version]
  54. Frazier, L.; Clifton, C., Jr.; Stolterfoht, B. Scale structure: Processing minimum standard and maximum standard scalar adjectives. Cognition 2008, 106, 299–324. [Google Scholar] [CrossRef] [Green Version]
  55. Lemaître, G.; Nogueira, F.; Aridas, C.K. Imbalanced-learn: A python toolbox to tackle the curse of imbalanced datasets in machine learning. J. Mach. Learn. Res. 2017, 18, 559–563. [Google Scholar]
  56. Kline, D.M.; Berardi, V.L. Revisiting squared-error and cross-entropy functions for training neural network classifiers. Neural Comput. Appl. 2005, 14, 310–318. [Google Scholar] [CrossRef]
  57. Nasr, G.E.; Badr, E.; Joun, C. Cross entropy error function in neural networks: Forecasting gasoline demand. In Proceedings of the FLAIRS Conference, Pensacola Beach, FL, USA, 14–16 May 2002; pp. 381–384. [Google Scholar]
  58. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  59. Ou, Z.; Tang, X.; Su, T.; Zhao, P. Cascade AdaBoost classifiers with stage optimization for face detection. In International Conference on Biometrics; Springer: Berlin/Heidelberg, Germany, 2006; pp. 121–128. [Google Scholar]
  60. Prabhakar, S.K.; Rajaguru, H. Adaboost Classifier with dimensionality reduction techniques for Epilepsy Classification from EEG. In International Conference on Biomedical and Health Informatics; Springer: Berlin/Heidelberg, Germany, 2017; pp. 185–189. [Google Scholar]
  61. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  62. Brodersen, K.H.; Ong, C.S.; Stephan, K.E.; Buhmann, J.M. The balanced accuracy and its posterior distribution. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 3121–3124. [Google Scholar]
  63. Velez, D.R.; White, B.C.; Motsinger, A.A.; Bush, W.S.; Ritchie, M.D.; Williams, S.M.; Moore, J.H. A balanced accuracy function for epistasis modeling in imbalanced datasets using multifactor dimensionality reduction. Genet. Epidemiol. Off. Publ. Int. Genet. Epidemiol. Soc. 2007, 31, 306–315. [Google Scholar] [CrossRef]
  64. Tong, T.; Ledig, C.; Guerrero, R.; Schuh, A.; Koikkalainen, J.; Tolonen, A.; Rhodius, H.; Barkhof, F.; Tijms, B.; Lemstra, A.W.; et al. Five-class differential diagnostics of neurodegenerative diseases using random undersampling boosting. Neuroimage Clin. 2017, 15, 613–624. [Google Scholar] [CrossRef] [PubMed]
  65. Pereira, C.R.; Pereira, D.R.; Silva, F.A.; Masieiro, J.P.; Weber, S.A.; Hook, C.; Papa, J.P. A new computer vision-based approach to aid the diagnosis of Parkinson’s disease. Comput. Methods Programs Biomed. 2016, 136, 79–88. [Google Scholar] [CrossRef]
  66. Papa, J.P.; Falcao, A.X.; Suzuki, C.T. Supervised pattern classification based on optimum-path forest. Int. J. Imaging Syst. Technol. 2009, 19, 120–131. [Google Scholar] [CrossRef] [Green Version]
  67. Wechsler, D. The Wechsler Adult Intelligence Scale-III; Psychological Corporation: San Antonio, TX, USA, 1997. [Google Scholar]
  68. Jenkinson, C.; Layte, R. Development and testing of the UK SF-12. J. Health Serv. Res. Policy 1997, 2, 14–18. [Google Scholar] [CrossRef]
Figure 1. Schematic overview of the proposed intelligent learning system.
Figure 1. Schematic overview of the proposed intelligent learning system.
Life 12 01097 g001
Figure 2. Class distribution before and After applying SMOTE.
Figure 2. Class distribution before and After applying SMOTE.
Life 12 01097 g002
Figure 3. ROC of ML models for dementia prediction.
Figure 3. ROC of ML models for dementia prediction.
Life 12 01097 g003
Figure 4. Performance comparison of proposed model with conventional Adaboost model in term of area under the cruve.
Figure 4. Performance comparison of proposed model with conventional Adaboost model in term of area under the cruve.
Life 12 01097 g004
Table 1. Demographic overview of the samples in the dataset.
Table 1. Demographic overview of the samples in the dataset.
Age_GroupMaleFemaleSubj. Sum Diagnosis. Dementia
60828216402
66759517006
72507412410
7841509117
8135468119
8426426822
8704192314
90+00050501
Total31341372691
Table 2. Overview of selected variable.
Table 2. Overview of selected variable.
Variable_CategoryVariables_NamesSum
DemographicAge, Gender02
SocialEducation, Religious Belief, Religious Activities, Voluntary Association, Social Network, Support Network, Loneliness07
LifestyleLight Exercise, Alcohol Consumption, Alcohol Quantity, Work Status, Physical-Workload, Present Smoker, Past Smoker, Number of Cigarettes a Day, Social Activities, Physically Demanding Activities, Leisure Activities11
Medical HistoryNumber of Medications, Family History of Importance, Myocardial Infarction, Arrhythmia, Heart Failure, Stroke, TIA/RIND, Diabetes Type 1, Diabetes Type 2, Thyroid Disease, Cancer, Epilepsy, Atrial Fibrillation, Cardiovascular Ischemia, Parkinson’s Disease, Depression, Other Psychiatric Diseases, Snoring, Sleep Apnea, Hip Fracture, Head Trauma, Developmental Disabilities, High Blood Pressure22
Biochemical TestHemoglobin Analysis, C-Reactive Protein Analysis02
Physical ExaminationBody Mass Index (BMI), Pain in the last 4 weeks, Heart Rate Sitting, Heart Rate Lying, Blood Pressure on the Right Arm, Hand Strength in Right Arm in a 10s Interval, Hand Strength in Left Arm in a 10s Interval, Feeling of Safety from Rising from a Chair, Assessment of Rising from a Chair, Single-Leg Standing with Right Leg, Single Leg Standing with Left Leg, Dental Prosthesis, Number of Teeth13
PsychologicalMemory Loss, Memory Decline, Memory Decline 2, Abstract Thinking, Personality Change, Sense of Identity06
Health InstrumentsSense of Coherence [41], Digit Span Test [42], Backwards Digit Span Test [42], Livingston Index [43], EQ5D Test [44], Activities of Daily Living [45], Instrumental Activities of Daily Living [46], Mini-Mental State Examination [47], Clock Drawing Test [48], Mental Composite Score of the SF-12 Health Survey [49], Physical Composite Score of the SF-12 Health Survey [49], Comprehensive Psychopathological Rating Scale [50]12
Table 3. Performance of conventional ML predictive models on imbalanced dataset, Where Acc T r a i n : Accuracy on training data, Acc T e s t : Accuracy on test data, Sens: Sensitivity, Spec: Specificity, MCC: Matthews correlation coefficient.
Table 3. Performance of conventional ML predictive models on imbalanced dataset, Where Acc T r a i n : Accuracy on training data, Acc T e s t : Accuracy on test data, Sens: Sensitivity, Spec: Specificity, MCC: Matthews correlation coefficient.
ModelAcc Train  (%)Acc Test  (%)Sens. (%)Spec. (%)F1_ScoreMCC
NB82.5774.1022.2291.1074.000.1428
LR85.3271.1523.5390.5571.000.1228
RF89.5576.5015.3689.4077.000.2278
DT71.4566.5025.9391.6267.000.1882
kNN78.5648.4016.6789.6249.000.0335
SVM86.6965.6031.2591.0966.000.1896
Table 4. Performance of conventional ML predictive models on balanced dataset, Where Acc T r a i n : Accuracy on training data, Acc T e s t : Accuracy on test data, Sens: Sensitivity, Spec: Specificity, MCC: Matthews correlation coefficient.
Table 4. Performance of conventional ML predictive models on balanced dataset, Where Acc T r a i n : Accuracy on training data, Acc T e s t : Accuracy on test data, Sens: Sensitivity, Spec: Specificity, MCC: Matthews correlation coefficient.
ModelAcc Train  (%)Acc Test  (%)Sens. (%)Spec. (%)F1_ScoreMCC
NB75.3770.7098.5778.8970.000.2287
LR82.7476.8585.3580.5577.000.4038
RF98.9685.9552.7387.6886.000.4264
DT80.4473.5180.5891.6274.000.3526
kNN78.5667.4975.1655.6267.000.2534
SVM96.2675.8292.5284.2076.000.3596
Table 5. Classification accuracy of the proposed autoencoder-SMOTE-Adaboost model with optimal hyperparameters of Adaboost on balance dataset, where N e s t : number of estimators, l r : learning rate of adaboost, F e : Feature extracted, Acc T r a i n : Accuracy on training data, Acc T e s t : Accuracy on test data, Sens: Sensitivity, Spec: Specificity.
Table 5. Classification accuracy of the proposed autoencoder-SMOTE-Adaboost model with optimal hyperparameters of Adaboost on balance dataset, where N e s t : number of estimators, l r : learning rate of adaboost, F e : Feature extracted, Acc T r a i n : Accuracy on training data, Acc T e s t : Accuracy on test data, Sens: Sensitivity, Spec: Specificity.
N est l r F e Acc Train  (%)Acc Test  (%)Sens. (%)Spec. (%)
4000.050690.4488.2989.8582.66
1000.010288.4889.5482.6391.58
1000.050288.4889.5485.6378.98
1000.011087.1290.0092.1483.56
3000.11289.5490.1686.3291.74
4000.11592.4187.5891.0586.48
3000.10389.3289.5486.0090.55
1000.050588.4890.0087.8295.74
4000.050692.1090.2397.8698.12
2000.050188.7689.5485.0081.41
100.050692.1090.2398.0096.65
500.10489.4890.0078.3688.00
500.050790.1386.3689.0595.48
2000.10694.0886.3698.0090.00
Table 6. Performance of autoencoder-based predictive models on balanced dataset, where Hyp.: hyperparameters value; F e : feature extracted; Acc T r a i n : accuracy on training data; Acc T e s t : accuracy on test data, Sens: sensitivity; Spec: specificity.
Table 6. Performance of autoencoder-based predictive models on balanced dataset, where Hyp.: hyperparameters value; F e : feature extracted; Acc T r a i n : accuracy on training data; Acc T e s t : accuracy on test data, Sens: sensitivity; Spec: specificity.
ModelHyp.F e Acc Train (%)Acc Test (%)Sens. (%)Spec. (%)
AEC * + NBV = 0.821487.2587.2295.5682.37
AEC + LRC = 101584.4087.1585.3590.87
AEC + RF N e = 1001010086.0052.7383.45
AEC + DT N e = 201886.2388.1880.5889.68
AEC + kNNk = 142010083.4879.1695.32
AEC + SVMC = 0.51287.1586.2292.5280.28
AEC + Adaboost  N e = 100692.1090.2398.0096.65
AEC *: Autoencoder.
Table 7. Classification accuracies comparison with previously proposed methods for dementia prediction.
Table 7. Classification accuracies comparison with previously proposed methods for dementia prediction.
Study (Year)MethodAccuracy (%)Balancing
P.C. Cho & W.H. Chen (2012) [29]PNNs83.00No
P.Gurevich et al. (2017) [26]SVM89.00Yes
D.Stamate et al. (2018) [24]Gradient Boosting88.00Yes
Visser et al. (2019) [25]XGBoost+ RF88.00No
Dallora et al. (2020) [23]DT74.50Yes
M.Karaglani et al. (2020) [27]RF84.60No
E. Ryzhikova et al. (2021) [28]ANN + SVM84.00No
F.A salem et al. (2021) [31]RF88.00Yes
F. G. Gutierrez et al. (2022) [32]GA84.00No
G. Mirzaei,& H. Adeli (2022) [5]MLP70.32No
A. Shahzad et al. (2022) [34]SVM71.67No
Proposed Model (2022)Autoencoder + SMOTE + Adaboost90.23Yes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Javeed, A.; Dallora, A.L.; Berglund, J.S.; Anderberg, P. An Intelligent Learning System for Unbiased Prediction of Dementia Based on Autoencoder and Adaboost Ensemble Learning. Life 2022, 12, 1097. https://0-doi-org.brum.beds.ac.uk/10.3390/life12071097

AMA Style

Javeed A, Dallora AL, Berglund JS, Anderberg P. An Intelligent Learning System for Unbiased Prediction of Dementia Based on Autoencoder and Adaboost Ensemble Learning. Life. 2022; 12(7):1097. https://0-doi-org.brum.beds.ac.uk/10.3390/life12071097

Chicago/Turabian Style

Javeed, Ashir, Ana Luiza Dallora, Johan Sanmartin Berglund, and Peter Anderberg. 2022. "An Intelligent Learning System for Unbiased Prediction of Dementia Based on Autoencoder and Adaboost Ensemble Learning" Life 12, no. 7: 1097. https://0-doi-org.brum.beds.ac.uk/10.3390/life12071097

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop