Next Article in Journal
A Source Localization Method Using Complex Variational Mode Decomposition
Next Article in Special Issue
Provoking Artificial Slips and Trips towards Perturbation-Based Balance Training: A Narrative Review
Previous Article in Journal
BepFAMN: A Method for Linear B-Cell Epitope Predictions Based on Fuzzy-ARTMAP Artificial Neural Network
Previous Article in Special Issue
Fall Risk Assessment Using Wearable Sensors: A Narrative Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inertial Data-Based AI Approaches for ADL and Fall Recognition

by
Luís M. Martins
1,2,3,
Nuno Ferrete Ribeiro
1,2,3,4,*,
Filipa Soares
1,2,3 and
Cristina P. Santos
1,2,3
1
Center for MicroElectroMechanical Systems (CMEMS), University of Minho, 4800-058 Guimarães, Portugal
2
LABBELS—Associate Laboratory, 4710-057 Braga, Portugal
3
LABBELS—Associate Laboratory, 4710-058 Guimarães, Portugal
4
MIT Portugal Program, School of Engineering, University of Minho, 4800-058 Guimarães, Portugal
*
Author to whom correspondence should be addressed.
Submission received: 28 March 2022 / Revised: 18 May 2022 / Accepted: 23 May 2022 / Published: 26 May 2022
(This article belongs to the Special Issue Sensors and Technologies for Fall Risk Awareness)

Abstract

:
The recognition of Activities of Daily Living (ADL) has been a widely debated topic, with applications in a vast range of fields. ADL recognition can be accomplished by processing data from wearable sensors, specially located at the lower trunk, which appears to be a suitable option in uncontrolled environments. Several authors have addressed ADL recognition using Artificial Intelligence (AI)-based algorithms, obtaining encouraging results. However, the number of ADL recognized by these algorithms is still limited, rarely focusing on transitional activities, and without addressing falls. Furthermore, the small amount of data used and the lack of information regarding validation processes are other drawbacks found in the literature. To overcome these drawbacks, a total of nine public and private datasets were merged in order to gather a large amount of data to improve the robustness of several ADL recognition algorithms. Furthermore, an AI-based framework was developed in this manuscript to perform a comparative analysis of several ADL Machine Learning (ML)-based classifiers. Feature selection algorithms were used to extract only the relevant features from the dataset’s lower trunk inertial data. For the recognition of 20 different ADL and falls, results have shown that the best performance was obtained with the K-NN classifier with the first 85 features ranked by Relief-F (98.22% accuracy). However, Ensemble Learning classifier with the first 65 features ranked by Principal Component Analysis (PCA) presented 96.53% overall accuracy while maintaining a lower classification time per window (0.039 ms), showing a higher potential for its usage in real-time scenarios in the future. Deep Learning algorithms were also tested. Despite its outcomes not being as good as in the prior procedure, their potential was also demonstrated (overall accuracy of 92.55% for Bidirectional Long Short-Term Memory (LSTM) Neural Network), indicating that they could be a valid option in the future.

1. Introduction

The recognition of Activities of Daily Living (ADL) has been a widely debated topic of study for the past several years with applications in a vast range of fields, from medicine to supervision of a persons’ driving style, and even to sports training analysis, passing through surveillance [1,2]. Several activities are being recognized nowadays. ADL related to human locomotion, such as walking, running, moving up and down stairs, or just sitting or lying down, are identified in several papers [2,3,4,5,6]. Other activities involving finer gestures with the upper limbs, such as driving, talking on the phone, or eating are also addressed [3,7]. Fall detection [8], recognition of sedentary behavior [9] and comfort in smart homes [10] are just a few more examples of automatic ADL recognition applications. The richness and diversity of human activities, as well as the large dimensionality of the data gathered, make the recognition of human activities a difficult but promising endeavor [11]. Furthermore, intraclass variability (the same activity may differ from one subject to the other) and interclass similarity (various activities may exhibit similar forms) make recognizing human activities even more challenging [12].
The identification of human activities can be accomplished by processing signals or image sequences gathered from one or more sensors, such as examining videos and photos from cameras or motion data from inertial sensors [11]. When compared to video and environmental sensor-based systems [13], wearable sensors show to be a suitable option for ADL recognition in uncontrolled environments due to their small size, low cost, and adaptability [5,6]. Several studies on activity recognition used wearable sensing technologies [6], with simpler wearable systems relying on a single accelerometer sensor, and others combining different sets of sensors on different parts of the body to fully capture the subject’s movements [5,14]. Moreover, researchers have been creating various publicly available benchmark datasets to evaluate human activity recognition algorithms [5]. The UCI HAR [15], Sisfall [16], Opportunity [17], and PAMAP2 [18] are some of the most commonly used public datasets to train and evaluate inertial sensor-based ADL recognition algorithms.
Offline activity recognition is the most basic method of activity recognition [14]. Sensors are placed in the surroundings or directly on the subject in most applications. Data are gathered, labeled, and supplied to classification algorithms for activity recognition [5,6,14]. Most offline approaches rely on supervised Machine Learning (ML) models for activity recognition, such as Support Vector Machine (SVM), Decision Trees (DTs) and K-Nearest Neighbors (K-NN) [6,14]. Moufawad et al. [19] used several features in a DTs algorithm to differentiate nine classes of ADL, and an overall accuracy of 97% was obtained across all activities. An accuracy of 80% for a set of 14 ADL classified was attained by Gomaa et al. [4], and the smallest average sensitivity and specificity achieved by the proposed Random Forest (RF) classification algorithm were 81% and 98%, respectively. Two ML algorithms were implemented by Gupta et al. [20] for the classification of six classes of ADL—Naive Bayes (NB) and K-NN. The overall accuracy of the system was about 98% from both the classifiers and showed accuracy of more than 95% for all the activities. Awais et al. [21] used an SVM classifier with RBF kernel to classify the four activity classes with data acquired from several wearable sensors. Among all single-sensor solutions, the best performance was accomplished by the sensor at the lower back (L5), with an F-measure of above 80%, and there was a notable improvement in the performance of 7.3% from single-sensor solution to a two-sensor solution.
Moreover, on several benchmark datasets used by ML-based solutions, conventional (CNN and LSTM) and hybrid (CNN-LSTM) Deep Learning (DL) models have lately shown improved and promising performances, cutting down the time spent on data processing and feature extraction [5,14]. Experimental results attained by Wang et al. [8] show that lightweight neural networks have obtained better results than ML methods previously used in fall detection. The best result was attained with the supervised Convolutional Neural Network (CNN), which achieved an accuracy beyond 99%. Despite the great Fall Detection accuracy results, other activities were not recognised. Altuve et al. [2] used sliding window-based segments and several sequential Long Short-Term Memory (LSTM) neural network to identify six human activities, obtaining 92.91% average accuracy. A Many-to-One LSTM network architecture was used by Chung et al. [22] for the recognition of nine ADL and the results show that activity gathered from inertial data with a sampling rate as low as 10 Hz from four sensors is sufficient to recognize ADL with 93% accuracy. In the work developed by Murad et al. [3], unidirectional, bidirectional, and cascaded architectures based on LSTMs were created, and their effectiveness was evaluated on several benchmark datasets. As it can be observed, overall accuracies over 92% were attained for each dataset, which proved the efficiency of the created models for a broad range of activity recognition tasks. A three-part modular system based on CNNs was proposed by Gil-Martín et al. [23] in order to classify 12 different physical activities. Results show that when applying post-processing techniques, isolated classification errors were reduced, improving the accuracy from 89.83% to 96.62%. Table 1 describes the datasets used by the analyzed works as well as the results obtained by their Deep Learning algorithms.
For the most part, the results achieved by Artificial Intelligence (AI)-based algorithms for ADL recognition are highly satisfactory. However, the number of activities recognized is still generally small, with the analyzed state-of-the-art detecting a maximum of 14 ADL (in this case, with 80% accuracy) [4]. Moreover, the amount of data used, whether through data acquisition protocols or through public datasets use, is small, since the largest public dataset identified in our research contains data from 38 people [16], and the average number of subjects found is less than 22, and with low variability. As a result, there is a greater demand for large-scale data collection campaigns. Studies in the field of merging datasets and pre-processing pipelines are also required in order to properly combine and decrease disparities between data obtained from various sources [27]. Another issue discovered is that the validation processes used in the algorithms are not stated or are inadequately explained in some research. This may cause the developed models to perform well just on the used dataset, but they perform poorly when evaluated on new datasets [28]. Furthermore, despite the fact that some articles implement feature selection techniques, they do not indicate which features and information are most essential for ADL recognition.
This manuscript addresses the aforementioned drawbacks and, in addition, addresses a higher number of ADL than any other of the works analyzed before, recognizing a total of 20 classes, including transitional activities (Sit-to-Stand, Stand-to-Sit, Lying-to-Stand, Stand-to-Lying, pick objects from the ground, bending and turning) and falls (16 ADL and four types of falls). Furthermore, this manuscript aims to answer the following research questions concerning recognition/classification of 16 ADL and four types of fall events from data collected from waist-located inertial sensors: (i) Which is the most suitable classifier and what are the most relevant features? and (ii) Which approach, ML or DL-based models, presents better performance?
With these objectives in mind, an AI-based framework was developed that allows for the benchmarking of several classifiers, including Discriminant Analysis (DA), K-NN, Ensemble Learning, DTs, SVM, CNN, mono and bidirectional LSTM, and CNN-LSTM neural networks. Simultaneously, feature selection algorithms are used to extract only the relevant features from the dataset’s lower trunk inertial data. Furthermore, a dataset fusion and normalization procedure is applied in order to gather a large amount of data to improve the robustness of the suggested algorithms during the several validation steps.
The remainder of this paper is organized as follows. Section 2 explains the inertial data-based dataset fusion and normalization process, feature extraction and selection processes, and finally, the comparative analysis carried out during the models building and evaluation. Then, Section 3 presents multiple results obtained with the proposed methods for validation, evaluation and best model choosing, followed by comparative studies, which try to uncover how changes in window size and model type can influence the final performance of classification models. Finally, Section 5 concludes the document, offering some remarks of the work developed along with directions for future improvements.

2. Methods

2.1. Public Dataset Fusion and Normalization

The proposed algorithm makes use of a large number of publicly available datasets to validate activity recognition models based on inertial data collected from the participants’ waists. Thus, a vast dataset was built from the fusion and normalization of several ADL public datasets. This process, to the best of our knowledge, has not been carried out before in any other works of the kind, despite its importance being highlighted multiple times [27,28]. Datasets that met certain requirements were searched: (i) be publicly available online for download; (ii) contain inertial data collected from the lower trunk (back) of at least accelerometers and gyroscopes; and (iii) contain postural daily activities and/or fall events. The research carried out resulted in the gathering of the public datasets and the three other team-own datasets, which are described in Table 2. The gathered public datasets were as follows.
  • Sisfall [16]: Data acquired with 23 healthy young adults (19–30 years, 149–183 cm, 42–81 kg) and 15 healthy elderly participants (60–75 years, 150–171 cm, 50–102 kg) with a device composed of two types of accelerometer and one gyroscope fixed to the waist of the participants, who performed 19 ADL and 15 fall types.
  • FallAllD [29]: Data acquired from 15 healthy subjects (21–53 years, 158–187 cm, 48–85 kg) who used 3 devices equipped with an accelerometer, a gyroscope, a magnetometer and a barometer. A total of 44 classes of ADL and 35 classes of falls were performed.
  • FARSEEING [30]: Large-scale collaborative database to collect and share sensor signals from real-world falls. Real fall data are acquired from either 2 locations: waist or thigh, and the acquisition devices are equipped with up to 3 sensors, namely accelerometer, gyroscope and magnetometer.
  • UCI HAR [15]: Dataset recorded from 30 healthy subjects (19–48 years) by using a waist-mounted smartphone with an embedded 3-axis accelerometer, gyroscope, and magnetometer. This dataset contains six classes of ADL: walking, ascending stairs, descending stairs, sitting, standing, and laying.
  • Cotechini et al. [31]: Dataset acquired from 8 healthy subjects (22–29 years old, 173–187 cm, 60–94 kg) using a wearable device containing a 3-axis accelerometer and gyroscope, tied to the subject’s waist, that recorded subject’s acceleration and orientation. Subjects simulated 13 typologies of falls and 5 types of ADL.
  • UMAFall [32]: A dataset acquired from a total of 17 healthy subjects (18–55 years, 50–93 kg, 155–195 cm). Accelerometer, gyroscope and magnetometer data were colected from five wearable sensing devices, located on the subject’s chest, waist, wrist, ankle and pocket. The participants performed 8 different ADL and 3 different typologies of falls (except by those older than 50 years, who did not perform falls).
The used team-owned datasets were as follows.
  • +Sense [33]: Dataset with data acquired from 10 healthy subjects (44.02 ± 16.42 years, 67.5 ± 16.06 kg, 172 ± 7.93 cm) and 40 subjects with Parkinson’s disease (64.00 ± 10.60 years, 69.93 ± 11.41 kg, 165.93 ± 8.65 cm). A waist-mounted waistband, equipped with an accelerometer, a gyroscope and a magnetometer recorded subject’s data in walking activity protocols.
  • SafeWalk [34]: Dataset acquired with 12 healthy subjects (25.33 ± 6.33 years old, 66.92 ± 10.07 kg, 1.74 ± 0.11 m). Five IMUs were attached to the lower back, both back thighs, and to both feet of the subjects, who performed walking trials and front fall events.
  • InertialLab [35]: Dataset which includes data from 11 able-bodied subjects (24.53 ± 2.09 years old, 171 ± 10 cm, 65.29 ± 9.02 kg). Gyroscopes and accelerometers were attached to six lower limbs and trunk segments. Walking in varying speed and terrain (flat, ramp, and stairs) and including turns were the activities carried out by the subjects.
The fusion of these various datasets, which include information from both adult and elderly subjects, as well as healthy and unhealthy individuals, served two purposes. First, to create a large and diverse dataset suitable for training ML models. The second goal was to develop AI-based models capable of recognizing ADL regardless of the subject’s age and health condition.
A global dataset containing 6702 files covering a total of 20 ADL and falls conducted by 180 subjects (age = 33.60 ± 16.84 years, weight = 69.98 ± 10.99 kg, height = 168.99 ± 9.42 cm) was generated. This vast amount of data is crucial for validation purposes since it is far larger than any other dataset used in the AI-based ADL recognition research carried out. Furthermore, the global dataset presents a balanced distribution regarding the subject’s gender (M = 54%, F = 46%), also containing data from both young adults and elderly people. However, despite the presence of elderly data, the average and standard deviation values show that the global dataset is still made up mostly of young adults, with the percentage of people over 65 years old being just over 20% of the total dataset (37/180 subjects).
Due to the great variability found between datasets, it was necessary to normalize the data from all datasets, according to the normalization procedure depicted in Figure 1a. First, only data corresponding to the acceleration (accelerometer) and angular velocities (gyroscope) of the sensors located in the subjects’ waist were considered. Then, the sensor reorientation method was applied so that the axis orientation corresponded to the one depicted in Figure 1b. Finally, all datasets underwent a resampling process so that the sampling frequency was normalized to 50 Hz.

2.2. ADL and Falls

The datasets, whether public or private, contain the great majority of ADL used for activity recognition in the literature. Therefore, a total of 20 labels, including periodic activities, static postures, transitions between postures and falls, were used in order to cover all ADL listed in every dataset. It should be noted that some activities whose labels in public datasets were considered different were recognized as the same activity in this work, since their basic body movement is similar, e.g., the cases of Sisfall’s activity of sitting in high chairs or low chairs were included in the “Stand to Sit” class; or even cases of standing in different places, such as in the room and in the elevator, were all included in the “Standing” class. Table 3 lists the ADL that were adressed in this work.
A study carried out on how the ADL were distributed showed that the global dataset is unbalanced, with a greater tendency toward cyclical activities, such as walking or lying (29.73% and 18.52%, respectively), with only a small percentage of transitions between activities and fall events, such as the fall by syncope, which is the activity with the least amount of data in the constructed dataset (0.27%). Figure 2 shows the percentage amounts of each activity present in the created dataset.

2.3. Machine and Deep Learning Classifiers: Comparative Analysis

We performed a comparative analysis, whose strategy is illustrated in Figure 3, to determine the most suitable AI-based classification model and the subset of features that attains the best performance. In the next sections, every step of the proposed strategy will be addressed. In this comparative analysis, the first approach consisted of training several ML classifiers to classify the addressed ADL and fall events with different feature subsets. The subsets of features with the best performance were then used in a second approach, where several Neural Networks architectures were proposed, and their performances were compared with the ML classifiers. Finally, for the best models found between ML and DL approaches, a study to assess the influence of the window size used for feature extraction on the classification model’s performance was carried out. It should be noted that all operations were performed on the global dataset without the use of any noise filters or other sorts of processing, i.e., the raw inertial data were directly implemented in the referred procedures. All the processes used for the development, validation and evaluation of these ADL recognition algorithms were implemented offline using the Matlab 2021b version on a Lenovo Legion Y540: processor—intel core i5, 9th Gen; graphics card—NVIDIA® GeForce® GTX 1650; memory—8 GB DDR4 at 2666 MHz and SSD PCIe of 512 GB.

2.3.1. Feature Extraction

Feature extraction was achieved through the sliding window method, where a signal is segmented into several windows of equal size, on which different features can be calculated. The most used sliding windows’ size corresponded to approximately 1 s for this type of activity classification, and the overlap between consecutive windows can vary from 50% to 87% [1,5,14]. Within the scope of this work, firstly, a one-second window was selected for the comparative analysis, which corresponds to a 50-sample window, with an overlap of 80% (Figure 4). In addition to the initial window size, 4 other different sizes were explored in the window size study: 0.5 s; 1 s; 1.5 s; and 2 s. The overlap was kept at 80% for all tests, despite the literature suggesting that it can also have a high impact on the classification performance and computational cost for real-time applications [1,36]. The segmentation in windows with a size of 1s and an overlap of 80% resulted in a total set of 666,660 windows for the models’ training and evaluation. Similarly, for a window size of 0.5 s, 1.5 s, and 2 s, a total of 1,366,289, 435,111, and 318,516 windows were obtained, respectively.
Thus, for each window, several features, such as the averages, maximums, minimums, and standard deviations of each of the acceleration and angular velocity signals, among other metrics, were extracted, making a total of 199 features calculated. A summary of the extracted features can be seen in Table 4. The window labeling was carried out according to the Mode Labeling Method, where the label of a given window would be the mode of the labels present in that window’s samples [36].

2.3.2. Pre-Processing and Feature Selection

After the aforementioned extraction, min–max scaling [0, 1] was carried out to ensure a low computational cost when building the models [37]. Since each of the datasets collected data with different devices and sensors with various sensitivities, the normalization process was carried out for each public dataset separately in order to reduce the possible bias derived from the different types, ranges, and sensitivities of the sensors used.
In order to understand what kind of improvements this feature selection process can bring, nine Feature Selection Methods (FSM) with diverse types of selection (Filtering, Wrapper and Embedded) were applied to the extracted features [38,39]. Covering the three different types of selection methods mentioned above, the selected methods and their respective type is indicated in Table 5.
Feature selection was also performed by the Principal Component Analysis (PCA), as shown in Figure 5. Principal components (PCs) with a cumulative percent explained of 70% were chosen [47], and a resultant and proportional PC was created and used to rank features. PCA also serves a secondary goal, which is to reduce the computational cost of the comparative analysis. Rather than using all 199 features, the number of features used was decided by multiplying the number of features having a PC value greater than 1 199 by 2 [34,48]. As a result, we chose a higher number of features for comparative analysis than those chosen for PCA and with significant contributions to the variability of the data. The success of this feature selection procedure will be determined by achieving the best performance with a lower number of features than the number of features extracted (199).

2.3.3. Model Building and Evaluation

We compared the following five ML classifiers by using a set of procedures depicted in Figure 3: DA with linear and quadratic approaches; K-NN with squared inverse distances; Ensemble Learning; and DTs. Table 6 presents a description of these classifiers. A progressive comparative analysis was carried out for each of these classifiers, where using the several FSM (Table 5), it was studied with how many of the ranked features (from most to least important), each of the classifiers obtains the best performance results. As previously mentioned, achieving the best classification performance with a lower number of features than the total number of extracted features in the previous step will show a successful application of this feature selection procedure and progressive analysis.
Initially, we performed the hold-out (HO) method to split 70% of the created dataset’s data for training and 30% for testing. The mentioned classifiers were used to obtain the subset of the most relevant features by performing an initial five-fold cross-validation (CV) with one repetition and using only training data. Once we determined the subset of the most relevant features, we performed a five-fold CV with ten repetitions to the four best classifiers from the previous step in order to evaluate the generalization capabilities of each model. The two best classifiers from this step were chosen, and its hyperparameters were optimized through a gridsearch process. The optimized ML classifiers and the neural networks architectures were further trained with all training data and tested with the test data from the HO method and the final performance of each classifier was compared in order to choose the best AI-based classification models.
In order to understand if these can be a viable alternative to ML methods, the classification of ADL and fall events was also carried out using the four Neural Networks architectures presented in Figure 6. As a result, the two best feature subsets acquired from the validation steps were used as input to train and evaluate the DL models. The performance of neural networks with the two best set of features was analyzed, and from the performance results obtained, a choice between the ML-based and DL-based models was made, for the most suitable classification model for ADL recognition. During all of the operations, test specifications such as the loss function used, number of epochs, the optimizer employed, number of hidden layers, batch size and the Learning Rate were kept constant for all the architectures. Table 7 provides a summary of all these characteristics and respective values.
The last study evaluates which of the chosen window sizes provides better performance results to the selected classifiers. Due to the great diversity in the classes to be classified (static postures, cyclical activities, transitions between postures and falls) and their duration, this study became imperative, in order to find the size that best fits the activities to be classified. Furthermore, the time required to perform the training and testing of the K-NN and Ensemble Learning classifiers for each of the window sizes was also computed. This exercise had the objective of investigating the possibility of using one of the algorithms that presented better performance in real-time situations.
The key performance indicators in the developed work were the Mathew’s correlation coefficient (MCC) and the accuracy (ACC). The first evaluation metric exhibits good representational features of unbalanced classes, as occurs in this study [55]. We used it for model performance comparison and reporting. The ACC, as well as the sensitivity (Sens), specificity (Spec), precision (Prec), and F1-score (F1S), were calculated for benchmarking the findings of the literature.

3. Results

3.1. PCA Outcomes

We determined that 11 PCs were necessary for a cumulative percent explained greater than 70%. Furthermore, the resulting PC demonstrated that there were 55 features with PC values greater than 1/199. After performing the PCA, we reduced the number of features to lower the computational cost of the comparative analysis (Figure 5); i.e., instead of using all 199 features, we only used the first 110 features ranked by any feature selection method included in this study. Only training data split from HO method were used.

3.2. ADL and Fall Events Classification

The results attained from the five-fold CV with one repetition disclosed the Ensemble Learning classifier as the one that presented the best performance among the used classifiers (MCC = 85.78%; ACC = 94.59%) when using the first 65 features ranked by the PCA method (Appendix A, Table A1). With the first 85 features ranked by the Relief-F (Appendix A, Table A1), K-NN produced similar but inferior results (MCC = 85.10%; ACC = 93.63%). DTs performed worse with the first 74 features ordered by the same technique (MCC = 70.65%; ACC = 88.22%); and Quadratic and Linear DA had the worst performance results with the first 55 and 66 features ranked by the Relief-F method, respectively. The two best classifiers went through a five-fold CV with ten repetitions, and we realized that increasing the number of repetitions did not change significantly the CV results either for the Ensemble Learning (MCC = 85.79%; ACC = 94.59%) or K-NN classifier (MCC = 85.05%; ACC = 93.62%). From this second CV stage, the Ensemble Learning using the first 65 features ranked by PCA and the K-NN using the first 85 features ranked by Relief-F were chosen for the next phases. Table 8 presents the main results for the two phases of the CV process.
When using test data from the HO data split method, the two best classifiers presented slight improvements in their performance in comparison to the results shown in Table 8. However, the Ensemble Learning model presented lower results (MCC = 88.36%; ACC = 95.44%) than the K-NN classifier (MCC = 93.19%; ACC = 97.27%) when tested with unseen data, contrary to what was verified during the CV process. After the optimization stage, K-NN hyperparameters were: (i) distance—minkowski; (ii) distance weight—squared inverse; (iii) exponent—0.5; and (iv) number of neighbors—1. Ensemble Learning hyperparameters were: (i) Method—Bag; and (ii) number of learning cycles—37. Table 9 depicts the main results obtained for the HO validation process.

3.3. Deep Learning Outcomes

The BiLSTM stood out among the neural networks in both case studies, using the first 85 features ranked by the Relief-F method (MCC = 82.83%; ACC = 92.55%) and the first 65 features ranked by PCA (MCC = 80.52%; ACC = 91.48%). However, in both cases, it presented lower results when compared to Ensemble Learning and K-NN. On the contrary, CNN presented the lowest performance results for both case studies, using the first 85 features ranked by the Relief-F method (MCC = 37.87%; ACC = 57.01%) and the first 65 features ranked by PCA (MCC = 24.90%; ACC = 42.67%). Thus, the ML-based methods, K-NN and Ensemble Learning, were considered the classifiers with better performance for ADL and fall events recognition, among all tested classifiers, being selected for the window size study and classification time analysis. Table 10 contains the main results for the DL-based classification problem.

3.4. Window Size and Classification Time

The results attained in this last analysis, for the optimized K-NN and Ensemble Learning classifiers, are depicted in Table 11. It should be noted that the same labeling method was used during the feature extraction step for each window size selected for this analysis. Both tables show a decreasing trend in the performance of the two classifiers as the window size increases from 0.5 to 2 s.
In addition to the results obtained regarding performance metrics, the time required to perform the training and testing of the K-NN and Ensemble Learning classifiers for each of the window sizes used in this last study was also computed. This exercise has the objective of studying the possibility of using one of the combinations which presented better performance results in real-time situations. Table 12 depicts the results obtained for the training and testing time of each classifier and window size combination.
Through direct observation of Table 12, the K-NN classifier has a training time of around four seconds, regardless of the size of the windows. The Ensemble classifier’s training time shows an increasing trend as the window size decreases. On the other hand, the time required to test only one window (last column of Table 12) is lower than 4.5 × 10 5 s for every window size tested in the case of the Ensemble Learning classification model. The test time per window for the K-NN classifier shows an increasing trend as the window size decreases.

4. Discussion

The ML-based classification was established with a combination of different classifiers and feature selection methods, and it proved to be an accurate strategy. This process allowed the most relevant features to be found for the classification of 20 distinct activities and fall classes (Appendix A), which should aid in a time-effective and low computation strategy for real-time ADL recognition. The PCA-based approach for reducing the computational burden of the comparison analysis may therefore be judged effective, since for most of the tested models and feature selection methods, the CV best performance was achieved with fewer features than those initially extracted.
The success achieved regarding the performance of the classification models through different validation methods, namely hold-out and cross-validation, shows the robustness of the applied processes. The creation of a vast dataset also positively influenced the performance results obtained. According to the results presented in Table 11 and Table 12, the Ensemble Learning classifier with a subset of the 65 first features ranked by the feature selection method PCA, with a window of 0.5 s and an overlap of 80%, was the model which showed the most potential for the classification of the 20 ADL classes in real time. Despite not being the best performer in terms of evaluation metrics, it had a classification time lower than the window advance time, allowing, in theory, its deployment in real-time situations. In addition, for the Ensemble Learning classifier, the comparative analysis of the set of features extracted by the different feature selection models provided a reduction in the number of features by more than 65% from the 199 initial features, to a subset of 65 features, also supporting the possibility of applying this type of models in real-time systems.
According to the literature, windows with a larger size have the tendency to perform worst in the recognition of shorter activities or transitions and better in the classification of cyclical or static activities, which are maintained for longer periods of time. Furthermore, as the window size increases, the model’s capability to recognize ADL in real time decreases [1,36]. As stated in Table 11, for smaller windows, the performance of the classifiers analyzed in this dissertation increases, thus corroborating what is described in the literature.
A direct comparison between the two best results obtained in tests performed with Neural Networks (Table 10) and the process developed for ML classifiers (Table 11) shows that the use of different architectures with the same features ranked by the best feature selection methods as inputs did not produce as good results as when using the K-NN and Ensemble Learning classifiers. The tested LSTM and BiLSTM achieved the best results in terms of performance metrics when compared to the CNN and CNN-LSTM architectures as well as in relation to the computation time of training and testing them. As mentioned in the literature, LSTM has a more suitable architecture for classifying sequential data [3,22]. Moreover, as investigated in the literature, the results achieved by BiLSTM were slightly superior to the basic LSTM [5,56]. The 85 features ranked by Relief-f returned performance metrics results slightly higher than the 65 features ranked by the PCA, as what happened in the studies carried out on ML models. CNNs are usually used in problems involving image inputs, and in this process, the implemented CNN architecture received as input 1D arrays of features extracted from sequential data. This method of use may justify the poor results obtained with the CNNs in terms of performance metrics, given that the convolution processes applied to extract features from the arrays used were not carried out in the most effective way. Despite these poor performance results, other architectures involving this type of networks should continue to be tested, with different type of inputs, such as the raw inertial data, since several studies claim to obtain positive results in the recognition of ADL with the use of this type of neural network [8,14,23].
Thus, this initial study of the behavior of several neural networks showed that despite positive results in some of the cases tested, it did not reach the desired potential when compared to the results obtained with the ML classifiers. This fact raises the need to carry out several studies and some changes in the future in order to obtain performance results higher than those found in the literature. The architectures used must be improved in the search for better results in order to be comparable with the studies currently carried out in the literature. Finally, and as mentioned in most of the works analyzed in the literature, new tests should be carried out in the future with windows segmented from raw inertial data as inputs rather than features, given the ability of neural networks to carry out a process extraction of features [23].

5. Conclusions and Future Work

An algorithm to recognize sixteen ADL and four different types of fall (twenty classes in total) from several AI-based classification models and feature selection models was built in order to find the combination that presents better performance in this type of classification. The performance of the different combinations was evaluated using the following parameters: (i) performance evaluation metrics, (ii) subset of features used, and (iii) classification time per window of the models.
Two different approaches (ML and DL) were investigated and compared: when performing one waist-located inertial sensor-based ADL and fall events recognition. Furthermore, a new procedure of fusion and normalization of public datasets was carried out in order to generate a big enough dataset to validate the activity classification models in order to battle concerns observed in the literature. Our long-term objective is to evaluate these techniques on elderly waistband users to evaluate whether the results of this study translate to identical outcomes in continuous real-life usage.
Taking into account the performance values as well as the classification times found in this work for the machine used, it is concluded that the most effective AI-based classifier was the Ensemble Learning classifier with the first 65 features ranked by the PCA feature selection method. Moreover, the classification time per window in this combination was lower than the window advance time in every window tested, which represents an encouraging result for the application of this algorithm in real time in the future. The DL outcomes were not as good as in the prior procedure; however, their potential was demonstrated, indicating that they could be a good option in the future with the appropriate future work regarding the input data used and its architectures. Although the established methodology allows for a reduction in computing cost, we recommend using a sufficient processor so that all classifiers respond rapidly.
The constructed dataset contains a large amount of successfully normalized data for the validation of activity recognition algorithms. Different steps can still be applied for its continuous improvement, such as the addition of more senior subjects’ data and data-balancing techniques, such as data augmentation, to balance the amount of samples of ADL represented on a smaller scale. However, the choice between using balanced data or preserving the overall distribution of activities should be backed by a critical analysis of the results obtained through a validation with data collected in day-to-day circumstances over long periods of time. Data fusion from other sensors, such as magnetometers or barometers, as well as the addition of other devices for data collection (such as smartwatches, which do not impose any restrictions on the subject’s movements), may aid in the differentiation of activities with similar movement patterns. Improved data-splitting methods must be verified in the future to guarantee that the results attained were not due to the usage of comparable data during the train and test stages of the model. Despite attaining performance results comparable with the literature for bidirectional LSTM, considering the higher number of classes classified, continued development of the tested neural networks is needed. New and improved architectures, as well as ablation studies, should also be conducted to determine the impact of the various settings and stages of each architecture on ADL and fall events recognition. Furthermore, the constructed classifiers should be trained and tested on the various public datasets independently in order to make a more direct comparison between the methods covered in this study and those listed in the literature that use the same datasets for evaluation.
This study contributes to the state-of-the-art with two versatile ADL and fall events detection methods, which are capable of discriminating 20 classes of events. There is also evidence that dataset fusion and normalization is imperative to guarantee a vast and diverse amount of data for ADL recognition algorithms’ validation. Furthermore, it may lead to the incorporation of these tools on instrumented waistbands or other devices for real-time fall risk assessment in the future [57].

Author Contributions

Conceptualization: L.M.M., N.F.R. and C.P.S.; Funding acquisition: C.P.S., N.F.R. and L.M.M.; Investigation: L.M.M., N.F.R., F.S. and C.P.S.; Methodology: L.M.M., N.F.R., F.S. and C.P.S.; Project administration: C.P.S.; Resources: C.P.S.; Supervision: N.F.R. and C.P.S.; Visualization: L.M.M.; Writing—original draft: L.M.M. and N.F.R.; Writing—review and editing: L.M.M., N.F.R. and C.P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by the FCT—Fundação para a Ciência e Tecnologia—national funds, under the scholarship references UMINHO-VC/BII/2021/03 and PD/BD/141515/2018, and the national support to R&D units grant, through the reference project UIDB/04436/2020 and UIDP/04436/2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Joana Figueiredo, Helena Gonçalves and Cristiana Pinheiro (LABBELS, CMEMS, University of Minho) for providing some of the data used in this work.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In this appendix, the complete list of the selected features by the two feature selection methods that presented the best performance results in Section 3 are presented. Table A1 shows the 85 first features ranked by Relief-F method and the 65 main features ranked by the PCA feature selection method. The feature numbers listed in this table correspond to the numbering used in Table 4.
Table A1. Features ranked in descending order per feature selection model.
Table A1. Features ranked in descending order per feature selection model.
FSMNo. of
Features
Ranked Features
Relief-F8566,69,70,68,67,65,110,128,142,143,144,35,31,101,12,9,111,148,15,14,112,114,
151,45,11,20,13,188,10,42,109,102,113,85,153,96,154,17,147,23,41,103,145,146,
116,21,22,18,190,19,16,152,74,83,149,84,24,88,197,64,123,90,89,155,28,46,194,
174,59,48,71,60,29,61,191,62,115,97,32,40,91,80,87
PCA659,97,188,42,102,43,101,128,144,113,148,110,184,31,142,154,116,83,41,103,111,
143,114,183,182,33,176,30,109,86,149,47,151,74,153,112,126,84,44,26,147,69,
180,127,100,145,115,27,155,146,120,4,35,85,36,1,42,7,91,46,45,186,175,192,96

References

  1. Banos, O.; Galvez, J.M.; Damas, M.; Pomares, H.; Rojas, I. Window Size Impact in Human Activity Recognition. Sensors 2014, 14, 6474–6499. [Google Scholar] [CrossRef] [Green Version]
  2. Altuve, M.; Lizarazo, P.; Villamizar, J. Human activity recognition using improved complete ensemble EMD with adaptive noise and long short-term memory neural networks. Biocybern. Biomed. Eng. 2020, 40, 901–909. [Google Scholar] [CrossRef]
  3. Murad, A.; Pyun, J.Y. Deep Recurrent Neural Networks for Human Activity Recognition. Sensors 2017, 17, 2556. [Google Scholar] [CrossRef] [Green Version]
  4. Gomaa, W.; Elbasiony, R.; Ashry, S. ADL Classification Based on Autocorrelation Function of Inertial Signals. In Proceedings of the 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA), Cancun, Mexico, 18–21 December 2017; pp. 833–837. [Google Scholar] [CrossRef]
  5. Elangovan, R.; Perumal, T.; Padmavathi, S. Human Activity Recognition with Smartphone and Wearable Sensors Using Deep Learning Techniques: A Review. IEEE Sens. J. 2021, 21, 13029–13040. [Google Scholar] [CrossRef]
  6. Wang, Q.; Luo, H.; Wang, J.; Sun, L.; Ma, Z.; Zhang, C.; Fu, M.; Zhao, F. Recent Advances in Pedestrian Navigation Activity Recognition: A Review. IEEE Sens. J. 2022, 22, 7499–7518. [Google Scholar] [CrossRef]
  7. Muhammad, F.; Shahid, N.; Omair, S.M.; Haque, M.Z.U. Integrated Wristband using an Inertial Measurement Unit and Electromyography Sensors to Control Robotic Car. J. Inf. Commun. Technol. Robot. Appl. 2019, 10, 36–41. [Google Scholar]
  8. Wang, G.; Li, Q.; Wang, L.; Zhang, Y.; Liu, Z. Elderly Fall Detection with an Accelerometer Using Lightweight Neural Networks. Electronics 2019, 8, 1354. [Google Scholar] [CrossRef] [Green Version]
  9. Kańtoch, E. Recognition of Sedentary Behavior by Machine Learning Analysis of Wearable Sensors during Activities of Daily Living for Telemedical Assessment of Cardiovascular Risk. Sensors 2018, 18, 3219. [Google Scholar] [CrossRef] [Green Version]
  10. Chen, L.; Nugent, C.D.; Wang, H. A Knowledge-Driven Approach to Activity Recognition in Smart Homes. IEEE Trans. Knowl. Data Eng. 2012, 24, 961–974. [Google Scholar] [CrossRef]
  11. Zhang, S.; Wei, Z.; Nie, J.; Huang, L.; Wang, S.; Li, Z. A Review on Human Activity Recognition Using Vision-Based Method. J. Healthc. Eng. 2017, 2017, e3090343. [Google Scholar] [CrossRef]
  12. Bulling, A.; Blanke, U.; Schiele, B. A Tutorial on Human Activity Recognition Using Body-Worn Inertial Sensors. ACM Comput. Surv. 2014, 46, 1–33. [Google Scholar] [CrossRef]
  13. Roy, N.; Misra, A.; Cook, D. Ambient and smartphone sensor assisted ADL recognition in multi-inhabitant smart environments. J. Ambient Intell. Humaniz. Comput. 2016, 7, 1–19. [Google Scholar] [CrossRef] [Green Version]
  14. Baghezza, R.; Bouchard, K.; Bouzouane, A.; Gouin-Vallerand, C. From Offline to Real-Time Distributed Activity Recognition in Wireless Sensor Networks for Healthcare: A Review. Sensors 2021, 21, 2786. [Google Scholar] [CrossRef]
  15. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A Public Domain Dataset for Human Activity Recognition Using Smartphones. In Proceedings of the 21th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
  16. Sucerquia, A.; López, J.D.; Vargas-Bonilla, J.F. SisFall: A Fall and Movement Dataset. Sensors 2017, 17, 198. [Google Scholar] [CrossRef]
  17. Chavarriaga, R.; Sagha, H.; Calatroni, A.; Digumarti, S.T.; Tröster, G.; del R. Millán, J.; Roggen, D. The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition. Pattern Recognit. Lett. 2013, 34, 2033–2042. [Google Scholar] [CrossRef] [Green Version]
  18. Reiss, A.; Stricker, D. Introducing a New Benchmarked Dataset for Activity Monitoring. In Proceedings of the 16th International Symposium on Wearable Computers, Newcastle, UK, 18–22 June 2012. [Google Scholar] [CrossRef]
  19. el Achkar, C.M.; Lenoble-Hoskovec, C.; Paraschiv-Ionescu, A.; Major, K.; Büla, C.; Aminian, K. Instrumented shoes for activity classification in the elderly. Gait Posture 2016, 44, 12–17. [Google Scholar] [CrossRef]
  20. Gupta, P.; Dallas, T. Feature selection and activity recognition system using a single triaxial accelerometer. IEEE Trans. Biomed. Eng. 2014, 61, 1780–1786. [Google Scholar] [CrossRef]
  21. Awais, M.; Chiari, L.; Ihlen, E.A.F.; Helbostad, J.L.; Palmerini, L. Physical Activity Classification for Elderly People in Free-Living Conditions. IEEE J. Biomed. Health Inform. 2019, 23, 197–207. [Google Scholar] [CrossRef]
  22. Chung, S.; Lim, J.; Noh, K.J.; Kim, G.; Jeong, H. Sensor Data Acquisition and Multimodal Sensor Fusion for Human Activity Recognition Using Deep Learning. Sensors 2019, 19, 1716. [Google Scholar] [CrossRef] [Green Version]
  23. Gil-Martín, M.; San-Segundo, R.; Fernández-Martínez, F.; Ferreiros-López, J. Improving physical activity recognition using a new deep learning architecture and post-processing techniques. Eng. Appl. Artif. Intell. 2020, 92, 103679. [Google Scholar] [CrossRef]
  24. Zhang, M.; Sawchuk, A.A. USC-HAD: A Daily Activity Dataset for Ubiquitous Activity Recognition Using Wearable Sensors. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing. Association for Computing Machinery, 2012, UbiComp ’12, Pittsburgh, PA, USA, 5–8 September 2012; pp. 1036–1043. [Google Scholar] [CrossRef]
  25. Bachlin, M.; Plotnik, M.; Roggen, D.; Maidan, I.; Hausdorff, J.M.; Giladi, N.; Troster, G. Wearable Assistant for Parkinson’s Disease Patients With the Freezing of Gait Symptom. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 436–446. [Google Scholar] [CrossRef]
  26. Zappi, P.; Lombriser, C.; Stiefmeier, T.; Farella, E.; Roggen, D.; Benini, L.; Tröster, G. Activity Recognition from On-Body Sensors: Accuracy-Power Trade-Off by Dynamic Sensor Selection. In Proceedings of the Wireless Sensor Networks; Verdone, R., Ed.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 17–33. [Google Scholar]
  27. Ferrari, A.; Micucci, D.; Mobilio, M.; Napoletano, P. Trends in human activity recognition using smartphones. J. Reliab. Intell. Environ. 2021, 7, 189–213. [Google Scholar] [CrossRef]
  28. Shany, T.; Wang, K.; Liu, Y.; Lovell, N.H.; Redmond, S.J. Review: Are we stumbling in our quest to find the best predictor? Over-optimism in sensor-based models for predicting falls in older adults. Healthc. Technol. Lett. 2015, 2, 79–88. [Google Scholar] [CrossRef] [Green Version]
  29. Saleh, M.; Abbas, M.; Le Jeannès, R.B. FallAllD: An Open Dataset of Human Falls and Activities of Daily Living for Classical and Deep Learning Applications. IEEE Sens. J. 2021, 21, 1849–1858. [Google Scholar] [CrossRef]
  30. Klenk, J.; Schwickert, L.; Palmerini, L.; Mellone, S.; Bourke, A.; Ihlen, E.A.F.; Kerse, N.; Hauer, K.; Pijnappels, M.; Synofzik, M.; et al. The FARSEEING real-world fall repository: A large-scale collaborative database to collect and share sensor signals from real-world falls. Eur. Rev. Aging Phys. Act. 2016, 13, 8. [Google Scholar] [CrossRef] [Green Version]
  31. Cotechini, V.; Belli, A.; Palma, L.; Morettini, M.; Burattini, L.; Pierleoni, P. A dataset for the development and optimization of fall detection algorithms based on wearable sensors. Data Brief 2019, 23, 103839. [Google Scholar] [CrossRef]
  32. Casilari, E.; Santoyo-Ramón, J.A.; Cano-García, J.M. UMAFall: A Multisensor Dataset for the Research on Automatic Fall Detection. Procedia Comput. Sci. 2017, 110, 32–39. [Google Scholar] [CrossRef]
  33. Gonçalves, H.; Moreira, R.; Rodrigues, A.; Santos, C. Finding Parameters around the Abdomen for a Vibrotactile System: Healthy and Patients with Parkinson’s Disease. J. Med. Syst. 2018, 42, 232. [Google Scholar] [CrossRef]
  34. Ribeiro, N.F.; André, J.; Costa, L.; Santos, C.P. Development of a Strategy to Predict and Detect Falls Using Wearable Sensors. J. Med. Syst. 2019, 43, 134. [Google Scholar] [CrossRef]
  35. Figueiredo, J.; Carvalho, S.P.; Vilas-Boas, J.P.; Gonçalves, L.M.; Moreno, J.C.; Santos, C.P. Wearable Inertial Sensor System towards Daily Human Kinematic Gait Analysis: Benchmarking Analysis to MVN BIOMECH. Sensors 2020, 20, 2185. [Google Scholar] [CrossRef] [Green Version]
  36. Fida, B.; Bernabucci, I.; Bibbo, D.; Conforto, S.; Schmid, M. Varying behavior of different window sizes on the classification of static and dynamic physical activities from a single accelerometer. Med. Eng. Phys. 2015, 37, 705–711. [Google Scholar] [CrossRef]
  37. Zhang, J.; Lockhart, T.E.; Soangra, R. Classifying Lower Extremity Muscle Fatigue During Walking Using Machine Learning and Inertial Sensors. Ann. Biomed. Eng. 2014, 42, 600–612. [Google Scholar] [CrossRef] [Green Version]
  38. Roffo, G.; Melzi, S.; Castellani, U.; Vinciarelli, A. Infinite Latent Feature Selection: A Probabilistic Latent Graph-Based Ranking Approach. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1407–1415. [Google Scholar] [CrossRef] [Green Version]
  39. Roffo, G. Feature Selection Library (MATLAB Toolbox). arXiv 2016, arXiv:1607.01327. [Google Scholar]
  40. Guo, J.; Guo, Y.; Kong, X.; He, R. Unsupervised feature selection with ordinal locality. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 1213–1218. [Google Scholar]
  41. Du, L.; Shen, Y.D. Unsupervised Feature Selection with Adaptive Structure Learning. arXiv 2015, arXiv:1504.00736. [Google Scholar]
  42. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  43. de Amorim, R.C. Computational Methods of Feature Selection; CRC Press: Boca Raton, FL, USA, 2007. [Google Scholar]
  44. Zaffalon, M.; Hutter, M. Robust Feature Selection by Mutual Information Distributions. In Proceedings of the Eighteenth Conference on Uncertainty in Artificial Intelligence, Edmonton, AB, Canada, 1–4 August 2002; Morgan Kaufmann Publishers Inc.: Burlington, ON, Canada, 2002; pp. 577–584. [Google Scholar]
  45. Bradley, P.; Mangasarian, O.L. Feature Selection via Concave Minimization and Support Vector Machines. In Proceedings of the Fifteenth International Conference (ICML ’98. Morgan Kaufmann, Madison, WI, USA, 24–27 July 1998; pp. 82–90. [Google Scholar]
  46. Roffo, G.; Melzi, S. Ranking to Learn: Feature Ranking and Selection via Eigenvector Centrality. arXiv 2017, arXiv:1704.05409. [Google Scholar]
  47. Jolliffe, I.T. Introduction. In Principal Component Analysis; Jolliffe, I.T., Ed.; Springer Series in Statistics; Springer: New York, NY, USA, 2002; pp. 1–9. [Google Scholar] [CrossRef]
  48. Ribeiro, N.F.; Mouta, P.; Santos, C.P. Two kinematic data-based approaches for cane event detection. J. Ambient. Intell. Humaniz. Comput. 2021. [Google Scholar] [CrossRef]
  49. Silva, A.P.D.; Stam, A. Discriminant analysis. In Reading and Understanding Multivariate Statistics; American Psychological Association: Washington, DC, USA, 1995; pp. 277–318. [Google Scholar]
  50. Abu Alfeilat, H.A.; Hassanat, A.B.; Lasassmeh, O.; Tarawneh, A.S.; Alhasanat, M.B.; Eyal Salman, H.S.; Prasath, V.S. Effects of Distance Measure Choice on K-Nearest Neighbor Classifier Performance: A Review. Big Data 2019, 7, 221–248. [Google Scholar] [CrossRef] [Green Version]
  51. Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
  52. Kingsford, C.; Salzberg, S.L. What are decision trees? Nat. Biotechnol. 2008, 26, 1011–1013. [Google Scholar] [CrossRef]
  53. Frank, A.E.; Kubota, A.; Riek, L.D. Wearable activity recognition for robust human-robot teaming in safety-critical environments via hybrid neural networks. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 449–454. [Google Scholar] [CrossRef]
  54. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; Adaptive Computation and Machine Learning Series; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  55. Jurman, G.; Riccadonna, S.; Furlanello, C. A Comparison of MCC and CEN Error Measures in Multi-Class Prediction. PLoS ONE 2012, 7, e41882. [Google Scholar] [CrossRef] [Green Version]
  56. Ashry, S.; Ogawa, T.; Gomaa, W. CHARM-Deep: Continuous Human Activity Recognition Model Based on Deep Neural Network Using IMU Sensors of Smartwatch. IEEE Sens. J. 2020, 20, 8757–8770. [Google Scholar] [CrossRef]
  57. Danielsen, A.; Olofsen, H.; Bremdal, B.A. Increasing fall risk awareness using wearables: A fall risk awareness protocol. J. Biomed. Inform. 2016, 63, 184–194. [Google Scholar] [CrossRef]
Figure 1. (a) Normalization process steps implemented in order to normalize the public datasets for ADL recognition. (b) Desired sensors orientation. The inertial sensors have to be located on the subjects’ waist. The arrows and letters x, y and z indicate the positive direction of the anteroposterior, mediolateral and longitudinal axes, respectively.
Figure 1. (a) Normalization process steps implemented in order to normalize the public datasets for ADL recognition. (b) Desired sensors orientation. The inertial sensors have to be located on the subjects’ waist. The arrows and letters x, y and z indicate the positive direction of the anteroposterior, mediolateral and longitudinal axes, respectively.
Sensors 22 04028 g001
Figure 2. Percentage quantity of windows created of each activity present in the created dataset. The activities are named according to Table 3.
Figure 2. Percentage quantity of windows created of each activity present in the created dataset. The activities are named according to Table 3.
Sensors 22 04028 g002
Figure 3. Complete process for all the validation, training and evaluation of different Machine Learning and Deep Learning models alongside the best feature set selected by diverse feature selection methods.
Figure 3. Complete process for all the validation, training and evaluation of different Machine Learning and Deep Learning models alongside the best feature set selected by diverse feature selection methods.
Sensors 22 04028 g003
Figure 4. Signal segmentation example, using the sliding window technique example for feature extraction. A sliding window across time is represented by windows 1 to 3.
Figure 4. Signal segmentation example, using the sliding window technique example for feature extraction. A sliding window across time is represented by windows 1 to 3.
Sensors 22 04028 g004
Figure 5. PCA-based procedure to rank and obtain the most crucial features and to limit the computational cost of comparative analysis.
Figure 5. PCA-based procedure to rank and obtain the most crucial features and to limit the computational cost of comparative analysis.
Sensors 22 04028 g005
Figure 6. Neural Networks architectures used in this manuscript. The input shown represents a single feature window. (a) The CNN identifies correlations from the various features provided. (b) The LSTM network detects crucial temporal features. (c) The BiLSTM network has a similar operating mode as LSTM but with bidirectional LSTM layers. (d) The hybrid CNN-LSTM extracts temporal patterns using convolutional features from the CNN convolutional layer. Based on [53].
Figure 6. Neural Networks architectures used in this manuscript. The input shown represents a single feature window. (a) The CNN identifies correlations from the various features provided. (b) The LSTM network detects crucial temporal features. (c) The BiLSTM network has a similar operating mode as LSTM but with bidirectional LSTM layers. (d) The hybrid CNN-LSTM extracts temporal patterns using convolutional features from the CNN convolutional layer. Based on [53].
Sensors 22 04028 g006
Table 1. Datasets used for the evaluation of the Deep Learning algorithms analyzed in the literature and respective description regarding sensing methods, sample frequency, participants, number of activities (classes) recorded and algorithm performance (accuracy). In this table: A = Accelerometer, G = Gyroscope, M = Magnetometer, B = Barometer and ADL = Activities of Daily Living.
Table 1. Datasets used for the evaluation of the Deep Learning algorithms analyzed in the literature and respective description regarding sensing methods, sample frequency, participants, number of activities (classes) recorded and algorithm performance (accuracy). In this table: A = Accelerometer, G = Gyroscope, M = Magnetometer, B = Barometer and ADL = Activities of Daily Living.
DataSetWorkSensorsSample
Frequency
ParticipantsN° of
Classes
Accuracy
Private
Dataset
Chung et al. [22]A, G, M100 Hz5993%
SisFall [16]Wang et al. [8]A, G200 Hz382<99%
PAMAP2 [18]Gil-Martín et al. [23]A, G, M 9 Hz91896.62%
UCI-HAD [15]Altuve et al. [2]
Murad et al. [3]
A, G, M50 Hz306[2]-96.7%
[3]-92.9%
USC-HAD [24]Murad et al. [3]A, G100 Hz141297.8%
Opportunity [17]Murad et al. [3]A, G, M30 Hz418 192.5%
Daphnet FOG [25]Murad et al. [3]A64 Hz10294.1%
Skoda [26]Murad et al. [3]A98 Hz111 192.6%
1 ADL related to hand gestures.
Table 2. Datasets description regarding sensing methods and location, sample frequency, participants and activities recorded, where: A = Accelerometer, G = Gyroscope, M = Magnetometer, B = Barometer and ADL = Activities of Daily Living.
Table 2. Datasets description regarding sensing methods and location, sample frequency, participants and activities recorded, where: A = Accelerometer, G = Gyroscope, M = Magnetometer, B = Barometer and ADL = Activities of Daily Living.
DataSetAvailabilitySensorsLocationSample
Frequency
ParticipantsADL
Falls
SisFall [161PublicA, GWaist200 Hz23 subjects <30 years
15 subjects >60 years
19 ADL
15 Falls
FALLALLD [291PublicA, G
M, B
Chest, Waist,
Wrist
238 Hz15 subjects
21–53 years
44 ADL
35 falls
FARSEEING [301PublicA, G, MWaist, Thigh20 Hz
100 Hz
20 subjects 2Real falls
UCI HAR [15]PublicA, GWaist50 Hz30 subjects
19–48 years
12 ADL
Cotechini [31] 1PublicA, GWaist33, 33 Hz8 subjects5 ADL
13 Falls
UMAFall [32]PublicA, G, MWaist, Chest,
Wrists, Ankle,
Front pocket
20 Hz17 subjects
18–55 years
8 ADL
3 Falls
+Sense [33]PrivateA, G, MWaist100 Hz10 Healthy
40 Pathological
1 ADL
SafeWalk [34]PrivateA, G, MWaist, Thighs,
Feet
30 Hz12 subjects
25.33 ± 6.33 years
1 ADL
Fall
InertialLab [35]PrivateA, G, MWaist, Thighs,
Shank, Feet
200 Hz7 subjects
23–26 years
5 ADL
1 Several activities in these datasets were grouped into one single class of basic activities. 2 Only data from 3 subjects were suitable to use.
Table 3. Static postures and locomotion daily activities, postural transitions and fall events selected to be recognized by the Machine and Deep Learning models.
Table 3. Static postures and locomotion daily activities, postural transitions and fall events selected to be recognized by the Machine and Deep Learning models.
Periodic Activities and Static PosturesTransitionsFall Events
WalkingLying to StandForwards
StandingStand to SitBackwards
SittingSit to StandLateral
LyingStand to Pick to StandSyncope
UpstairsStand to Lying
DownstairsChange Position (Lying)
JumpingTurning
JoggingBending
Table 4. List of all extracted features from each window created. AP, V and ML refer to the anteroposterior, vertical and mediolateral axis, respectively.
Table 4. List of all extracted features from each window created. AP, V and ML refer to the anteroposterior, vertical and mediolateral axis, respectively.
Feature NumberFeature Description
[1–6]Acceleration and Angular velocity (AP, V, ML)
[7–8]SumVM of acceleration and Angular velocity
[9–24]Skewness and kurtosis of acceleration, Angular velocity (AP, V, ML)
and SumVM signals
[25–64]Min, max, mean, variance and Std deviation of acceleration, angular
velocity (AP, V, ML) and SumVM signals
[65–70]Correlation between V-ML, V-AP and ML-AP axis of acceleration and
Angular velocity
[71–77]Slope, Total angular change, Resultant angular acceleration, ASMA, SMA,
Absolute vertical acceleration, Cumulative horizontal displacement
[78–102]Peak-to-Peak, Root Mean Square and Ratio Index
of Acceleration, Angular velocity (AP, V, ML) and SumVM signals
[103–115]Resultant angle change, Flutuation frequency, Resultant
of average acceleration and Resultant of standard deviation (AP, V, ML)
[116–117]Resultant of Delta changes of acceleration and Angular velocity
[118–133]Gravity component, Displacement, Displacement range,
Cumulative sway length and Mean sway velocity (AP, V, ML)
Slope changes, Zero crossings, Waveform length of acceleration,
Angular velocity (AP, V, ML) and SumVM signals
[133–189]Energy, Mean frequency, Peak frequency and magnitude of acceleration,
Angular velocity (AP, V, ML) and SumVM signals
[190–195]SumVM of resultant angular velocity, average acceleration
and Standard deviation, Maximum resultant angular velocity
and Acceleration in the horizontal plane
[196–199]Acceleration exponential moving average, Rotational angle
of acceleration SumVM, Z-Score, Magnitude of angular displacement
Table 5. Feature Selection Methods Tested for ADL Recognition.
Table 5. Feature Selection Methods Tested for ADL Recognition.
Feature Selection Methods (FSM)FSM Type
Infnite Latent Feature Selection (ILFS) [38]Filtering
Unsupervised Feature Selection with Ordinal Locality (UFSOL) [40]Wrapper
Feature Selection with Adaptive Structure Learning (FSASL) [41]Wrapper
Minimum-Redundancy Maximum-Relevancy (MRMR) [42]Filtering
Relief-F [43]Filtering
Mutual Information Feature Selection (MutInfFS) [44]Filtering
Feature Selection Via Concave Minimization (FSV) [45]Embedded
Correlation-Based Feature Selection (CFS) [46]Filtering
Least Absolute Shrinkage and Selection Operator (LASSO) [43]Embedded
Principal Component Analysis (PCA) [47]Filtering
Table 6. Machine Learning models evaluated in this work and respective descriptions.
Table 6. Machine Learning models evaluated in this work and respective descriptions.
ModelReferenceDescription
DA[49]A method that finds combinations of features that separate two or more classes of objects or events, searching for the most variance between classes, and information that maximizes the difference between classes.
K-NN[50]Compares each new instance with all datasets available and the instance closest by distance metrics is used to perform classification. Since every sample of the dataset must be checked for every instance, the time and complexity of the method rises according to the dataset size.
Ensemble
Learning
[51]Creates multiple instances of traditional ML methods and combines them to evolve a single optimal solution to a problem. This approach is capable of producing better predictive models compared to the traditional approach.
DTs[52]A model that predicts the value of a target variable based on numerous input variables. A decision tree is constituted by an internal node, based on which the tree splits into branches. The end of the branch that does not split any longer is the decision.
Table 7. Specifications for the use of the Deep Learning models depicted in Figure 6.
Table 7. Specifications for the use of the Deep Learning models depicted in Figure 6.
SpecificationValue
Epoch Number100
Hidden Layers150
Batch Size64
OptimizerAdam [54]
Learning Rate0.001 (Constant)
Loss FunctionCross-Entropy
Table 8. Comparison of the best classification results (ACC, Sens, Spec, Prec, F1S, MCC), attained after the 5-1 and 5-10 k-fold cross-validation steps for the K-NN and Ensemble Learning classifiers.
Table 8. Comparison of the best classification results (ACC, Sens, Spec, Prec, F1S, MCC), attained after the 5-1 and 5-10 k-fold cross-validation steps for the K-NN and Ensemble Learning classifiers.
ML ModelFSMCV StepN° of
Features
ACC (%)Sens (%)Spec (%)Prec (%)F1S (%)MCC (%)
K-NNRelief-F 8593.6384.1799.6486.8085.4385.10
PCA5 Fold
(1 rep.)
8592.9984.0899.6086.0185.0184.63
FSASL7091.4981.3999.5183.6682.4882.02
Ensemble
Learning
PCA 6594.5982.2299.6890.5485.8085.78
K-NNRelief-F 8593.62 ± 0.01684.12 ± 0.06699.64 ± 0.00186.75 ± 0.05585.38 ± 0.05685.05 ± 0.056
PCA5 Fold
(10 rep.)
8592.95 ± 0.02183.91 ± 0.09499.60 ± 0.00185.88 ± 0.08584.86 ± 0.08584.48 ± 0.086
FSASL7091.48 ± 0.02681.40 ± 0.06399.51 ± 0.00183.59 ± 0.07982.45 ± 0.06681.99 ± 0.067
Ensemble
Learning
PCA 6594.59 ± 0.01582.18 ± 0.06799.68 ± 0.00190.64 ± 0.07385.79 ± 0.06185.79 ± 0.060
Table 9. Hold-out test results for the Ensemble Learning with the first 65 features ranked by the PCA and for the K-NN classifier with the first 85 features ranked by the Relief-f.
Table 9. Hold-out test results for the Ensemble Learning with the first 65 features ranked by the PCA and for the K-NN classifier with the first 85 features ranked by the Relief-f.
ML ModelFSMN° of
Features
ACC (%)Sens (%)Spec (%)Prec (%)F1S (%)MCC (%)
K-NNRelief-F8597.2792.9099.8493.7993.3493.19
Ensemble
Learning
PCA6595.4485.9799.7391.6788.4388.36
Table 10. Results for the test of the 4 Deep Learning architectures with the 85 first features ranked by Relief-f and 65 first features ranked by PCA.
Table 10. Results for the test of the 4 Deep Learning architectures with the 85 first features ranked by Relief-f and 65 first features ranked by PCA.
FSMFeature
Number
ArchitectureACC (%)Sens (%)Spec (%)Prec (%)F1S (%)MCC (%)
Relief-F85CNN57.0137.0697.2254.6735.4737.87
LSTM92.0679.5899.5584.2581.0281.01
CNN-LSTM88.8474.4899.3675.2474.5374.06
BiLSTM92.5581.1499.5785.5683.1482.83
PCA65CNN42.6726.4696.1554.4922.2724.90
LSTM91.4677.8199.5184.3880.6180.38
CNN-LSTM88.5574.3399.3575.0974.3673.88
BiLSTM91.4879.3399.5283.3280.6780.52
Table 11. Window size comparative study results for the K-NN best optimized model with the Relief-F feature selection model.
Table 11. Window size comparative study results for the K-NN best optimized model with the Relief-F feature selection model.
ML Model + FSMWindow Size (s)Window
Overlap (%)
ACC
(%)
Sens
(%)
Spec
(%)
Precn
(%)
F1S
(%)
MCC
(%)
K-NN + Relief-f0.58098.2295.2099.9096.0495.6295.52
197.2792.9099.8493.7993.3493.19
1.596.3091.7399.7991.1591.4191.22
295.3390.5399.7488.5189.4489.22
Ensemble + PCA0.596.5388.9499.7994.0991.2991.21
195.4485.9799.7391.6788.4388.36
1.595.0185.6099.7190.7687.6487.62
294.5185.2199.6889.3786.9286.79
Table 12. Classification time for the training and testing of the two best combinations of Machine Learning (ML) model and Feature Selection Method (FSM), for each of the selected windows for the window size study.
Table 12. Classification time for the training and testing of the two best combinations of Machine Learning (ML) model and Feature Selection Method (FSM), for each of the selected windows for the window size study.
ML Model + FSMWindow
Size (s)
Window
Overlap (%)
Test
Windows
Train
Time (s)
Test
Time (s)
Test Time
per Window (s)
K-NN + Relief-f0.580409,7404.36213,588.880.521
1199,9974.1866,782.580.334
1.5130,4214.7012,633.080.097
295,4824.096752.470.071
Ensemble + PCA0.5409,740829.5515.99 3.90 × 10 5
1199,997279.038.54 4.27 × 10 5
1.5130,421145.215.68 4.35 × 10 5
295,482100.233.94 4.13 × 10 5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Martins, L.M.; Ribeiro, N.F.; Soares, F.; Santos, C.P. Inertial Data-Based AI Approaches for ADL and Fall Recognition. Sensors 2022, 22, 4028. https://0-doi-org.brum.beds.ac.uk/10.3390/s22114028

AMA Style

Martins LM, Ribeiro NF, Soares F, Santos CP. Inertial Data-Based AI Approaches for ADL and Fall Recognition. Sensors. 2022; 22(11):4028. https://0-doi-org.brum.beds.ac.uk/10.3390/s22114028

Chicago/Turabian Style

Martins, Luís M., Nuno Ferrete Ribeiro, Filipa Soares, and Cristina P. Santos. 2022. "Inertial Data-Based AI Approaches for ADL and Fall Recognition" Sensors 22, no. 11: 4028. https://0-doi-org.brum.beds.ac.uk/10.3390/s22114028

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop