Next Article in Journal
Low-Computational-Cost Hybrid FEM-Analytical Induction Machine Model for the Diagnosis of Rotor Eccentricity, Based on Sparse Identification Techniques and Trigonometric Interpolation
Next Article in Special Issue
A Semi-Supervised Transfer Learning with Dynamic Associate Domain Adaptation for Human Activity Recognition Using WiFi Signals
Previous Article in Journal
Interactive Fluid Coupling Effects of Non-Neighbouring Members
Previous Article in Special Issue
A Semi-Supervised Transfer Learning with Grid Segmentation for Outdoor Localization over LoRaWans
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards a Clustering Guided Hierarchical Framework for Sensor-Based Activity Recognition

1
School of Electronic Information Engineering, Foshan University, Foshan 528225, China
2
School of Computer and Information Engineering, Chuzhou University, Chuzhou 239000, China
3
Department of Computer Science and Information Engineering, Tamkang University, New Taipei City 25137, Taiwan
4
National Institute of Technology, Shillong 793003, India
*
Author to whom correspondence should be addressed.
Submission received: 26 September 2021 / Revised: 16 October 2021 / Accepted: 18 October 2021 / Published: 20 October 2021
(This article belongs to the Special Issue Advanced Wireless Sensing Techniques for Communication)

Abstract

:
Human activity recognition plays a prominent role in numerous applications like smart homes, elderly healthcare and ambient intelligence. The complexity of human behavior leads to the difficulty of developing an accurate activity recognizer, especially in situations where different activities have similar sensor readings. Accordingly, how to measure the relationships among activities and construct an activity recognizer for better distinguishing the confusing activities remains critical. To this end, we in this study propose a clustering guided hierarchical framework to discriminate on-going human activities. Specifically, we first introduce a clustering-based activity confusion index and exploit it to automatically and quantitatively measure the confusion between activities in a data-driven way instead of relying on the prior domain knowledge. Afterwards, we design a hierarchical activity recognition framework under the guidance of the confusion relationships to reduce the recognition errors between similar activities. Finally, the simulations on the benchmark datasets are evaluated and results show the superiority of the proposed model over its competitors. In addition, we experimentally evaluate the key components of the framework comprehensively, which indicates its flexibility and stability.

1. Introduction

Benefiting from the rapid development and fusion of sensing technology, pervasive computing, and artificial intelligence, researchers have designed and implemented a wealth of Internet of Things (IoTs) systems and applications, such as behavior analysis, sports and games, the elderly healthcare, chronic disease management, smart buildings, smart homes and ambient intelligence, etc. [1,2]. From the perspective of system architecture, these application services have a typical three-layer architecture of IoTs (sensing layer, network layer, and application layer) [3,4]. With the complex characteristics of human behavior, it is not very easy to build an accurate activity recognizer for practical applications [5]. For example, there exist intra-subject and inter-subject variations of how people perform an activity, which is difficult to the generalization ability of activity recognizers [6]. Besides performing activities sequentially, an individual can conduct concurrent and interleaved activities, and there are activities involved in a specific application having similar sensor signals, which confuses an activity recognizer to a certain extent [7,8]. Therefore, automatically and accurately recognizing human activities is still a challenging issue [9,10].
In the literature, many activity recognition models had been proposed. They were categorized into three types based on the technology, including vision-based, environment interaction-based, and wearable-based methods [11,12,13]. Vision-based mechanisms exploit the computer vision techniques to analyze and infer the human activities in camera captured content. The recent years have witnessed enormous growth in their applications such as video surveillance, games, and security, while their performance is sensitive to background noise and illumination variation [14]. Moreover, their use often raises privacy issues and is further limited to the fixed places [11]. For the second type methods, they analyzed the on-going activities by recording the interaction between the person and the surroundings via the sensing units placed on household objects. Obviously, they are usually limited to indoor scenarios [15]. Different from the above two types of methods, wearable-based methods largely benefit from the communication capabilities of wearable devices. Typically, they first collected the sensor data when one performed activities and trained an activity recognizer, which was used for making predictions [16]. That is to say, a person could wear multiple homogeneous or heterogeneous sensing devices on the body [17], which was more suitable for both scenarios.
In terms of the activity recognition models, there were many models from the generative and discriminative models to unsupervised learning, ensemble learning, and deep learning models [18,19,20]. Although these models advance the study of human activities, most of the studies recognized the predefined activities in a one-step, which we called the flat model for illustration purposes. However, for the activity recognition supported applications, different activities (even if they have totally different semantics) could have quite similar sensor signals, which reduced the ability of activity recognizers in distinguishing confusing activities and inevitably led to degraded accuracy [7,8]. One feasible solution is to develop a hierarchical activity recognition model in order to derive a superior decision between similar activities towards performance improvement, where there are two key factors that determine its practical use and performance. First, determining the relationships among activities is a prerequisite for constructing a hierarchical activity recognizer. Different from the simple and easy case where we can use prior knowledge to group activities, it is a challenging task for new and under-explored cases with few or even no domain expert knowledge about the activities available for use. Accordingly, how to quantitatively measure the confusion among activities largely influences its generalization ability across different scenarios. Second, how to effectively utilize the relationship to build an activity recognizer affected its overall performance in analyzing confusing activities. This requires us to develop a model to better distinguish different activities and to maximize the decision distance between confusing activities. To handle the key issues, we herein propose a clustering guided hierarchical activity recognition framework that consists of two critical components. Specifically, we first introduce a clustering-based activity confusion index and use it to quantitatively measure the confusion among activities in a data-driven way. Afterwards, we develop a hierarchical activity recognition model under the guidance of the relationships to reduce the recognition errors between similar activities. The contributions of the proposed study are listed. (1) Instead of depending on prior knowledge, a data-driven approach, which quantitatively measures the confusion among activities has been proposed. This potentially facilitates the development of the hierarchical activity recognition model and its extension to new and under-explored situations that have limited prior domain knowledge. (2) We formalize the problem and accordingly propose a hierarchical human activity recognition framework guided by the proposed activity confusion index, which has the capability of connecting any two activities and returning a superior decision between activities having similar sensor signals. This is expected to reduce confusion errors. (3) We also experimentally investigate the power of tri-axial accelerometers and gyroscopes and study the parameter sensitivity of the components of the framework, which indicates the power of fusing distinct sensors and the flexibility and stability of the proposed model.
The structure of this paper is shown as follows. Section 2 discusses the related work on activity recognition models. The proposed clustering guided hierarchical framework is given in Section 3. Section 4 and Section 5 analyze the experimental settings and the performance results. The conclusion is given in Section 6.

2. Related Work

Different from image processing and text analysis, the development of activity recognition includes many challenges such as modeling, evaluation, and human behavior, which is characterized by inherent complexity. To get enhanced performance and generalization ability across different scenarios, researchers have proposed many activity recognition models, which can be categorized into two categories: knowledge-driven and data-driven approaches [6,21,22].
Knowledge-driven approaches utilized domain knowledge for activity specification and further used it for sensor modeling, context fusion, activity modeling and recognition [23]. Commonly used modeling tools include, but are not limited to, ontology modeling, evidential theory, and logical modeling [24]. For example, an activity recognition model based on the ontologies and reasoning is used to infer on-going activities [21]. They implemented the proposed approach in a software system and evaluated it as a smart home research laboratory and achieved an overall accuracy of 94.44%. Although the knowledge-driven approaches have advantages of high efficiency, robustness to noise, and easy interpretation, they usually need domain expert knowledge for activity and context modeling, which is not a trivial task, especially for a new and underexplored domain.
Data-driven approaches, also called data-centric approaches, aim to learn an activity recognition model from the collected sensor data [25]. Study [16] used k nearest neighbors and naïve Bayes to train sensor-based activity recognizers. Another study [26] utilized the random forest algorithm to recognize activities such as jump, run, static, walking, go-upstairs, and go-downstairs. There are also researches that use deep learning models to jointly optimize the feature learning and classifier training [27,28,29]. Study [30] proposed a deep convolutional neural network (CNN) based activity recognizer to predict the activities based on the mobile phone data. Ordóñez and Roggen introduced a convolutional long short-term memory (LSTM) model, and they experimentally evaluated its power, demonstrating its superiority over its competitors [17].
Besides, some studies combine knowledge-driven and data-driven approaches. Azkune et al. proposed a model that uses the data-driven approach to evolving knowledge-driven approach aiming to obtain knowledge-driven activity recognition models [31]. Sukor et al. presented a hybrid activity recognition model that first used knowledge-driven reasoning to get an initial activity model and then optimized it using a data-driven approach. They experimentally evaluated the proposed model, which showed significantly higher recognition rates [32].
To better handle the activities that have similar sensor signals, some studies present the hierarchical activity recognition models. Different from the flat models that classify the activities of interest in a single step, hierarchical models usually adopt a coarse-to-fine scheme to infer the activity label via a multi-stage decision, where they transform the original problem into a series of classification problems [33]. Study [25] developed a hierarchical activity recognition model that recognized human gestures and predicted the final activity. Liu et al. presented an activity recognition model that recognized the actions and then utilized the temporal patterns of the actions to recognize activities [34]. From the perspective of the prediction process, we call such a scheme the bottom-up approach. Instead, some studies work in a top-down scheme, where it first predicts an abstract activity and then infers the specific label of the activity. Study [35] proposed a two-stage model that classified activities into dynamic, static, or transition states and then inferred the specific activity within the activity group. Cho and Yoon proposed an activity recognition model based on two-stage 1D CNN to first predict the test sample within an abstract activity group and then make predictions within the group [8]. Wang et al. used prior domain knowledge of the activities to manually organize them into a tree structure and proposed a tree-based activity recognizer [7]. Although top-down methods make no assumption about the temporal relationship and obtain enhanced performance, most of the studies rely on prior knowledge about the activities to group activities and to build a hierarchical activity recognizer [7,8,34,36]. This hinders them when handling new and complex scenarios of which there is little expert knowledge. Moreover, wrong grouping can result in degraded accuracy. In addition, one major limitation of the tree-based models is that they probably induce the accumulation of errors in the hierarchical predictions. Therefore, the above issues motivate us to explore the data-driven way to automatically and quantitatively measure the relationships among activities and then build an activity recognizer for reducing the accumulation of errors. Accordingly, we present a clustering guided hierarchical activity recognition model. Compared with previously work, our proposed data-driven model utilizes the clustering theory to identify confusing activities and quantifies the relationships among activities with the two proposed confusing indices (i.e., cluster confusion index and activity confusion index). Different from the tree-based model, our proposed model allows the flexible connections between any two activities, which helps return a good decision boundary between similar activities. Particularly, the proposed method is a general framework that can take as its building blocks various clustering and classification models and even deep learning models.

3. Clustering Guided Activity Recognition Framework

The characteristics of the complex human behavior lead to the difficulty of designing an activity recognizer, specifically in handling cases where activities with different semantics have similar sensor readings. As a result, this degrades the discriminant capability of an activity recognizer. For example, we experimentally observe that the sensor readings between dynamic activities are more similar than the ones between static activity and dynamic activity. Hence, training a flat activity recognizer to classify the three activities pays much attention to the decision boundary between static activity and dynamic activities and fails to determine a good decision between the two dynamic activities.
One feasible solution is to construct a hierarchical activity recognizer, where we infer the activity labels in a multi-stage way. Accordingly, the key is how to determine the relationships among activities and use them to direct the design of an activity recognition model, especially for the case with little or no prior knowledge. Figure 1 depicts the proposed clustering guided hierarchical activity recognition model, which consists of two-level classification. The training data set contains two parts: the sensing data and the label of activity. Initially, the sensing data will be the input of the designed model, which is the first level of coarse-grained classification. Since the sensing data of some activities, such as sitting and lying, are similar, they might fall in the same cluster. This occurs because their features represented by the sensing data are similar. To further distinguish them in a way of fine-grained feature extraction, a Confusion Analyzer is designed to assign each activity with an activity confusion index, which represents the similarity between activities. The output of the first level coarse-grained classification will further play the role of input data of the second level fine-grained classification model, aiming to further distinguish similar activities. According to the assigned activity confusion index, the output of the first level classification model can be the final decision without feeding into the lower-level classification, if it does not fall in the same group of other activities. Otherwise, the output of the first level coarse-grained classification model will be further processed by applying the fine-grained classification. Herein, the component of Confusion Analyzer aims to assign each activity with an activity confusion index according to the clustering results. The coarse-grained classification and fine-grained classification could use the same machine learning model to design their recognizers. For a specific application, we train a coarse-grained classification model and selectively train a fine-grained classification model for each activity. That is, if there exist confusing activities for one activity, we train a fine-grained model associated with it. Accordingly, the number of classes to be predicted for coarse-grained classification and fine-grained classification is different, and the coarse-grained model usually has higher storage and computation complexity. Although the example given in Figure 1 only depicts the two-level activity recognition. However, the general model of hierarchical classification, as shown in the block of the left bottom of Figure 1, can have multi-level classification models where the number of levels is larger than two.
The framework contains two core components: determining the confusion among activities, which is implemented by the Confusion Analyzer, and constructing a hierarchical model. The idea is to build the hierarchy of the predefined activities under the guidance of the relationships among activities, which is determined by the Confusion Analyzer. Specifically, we first measure the activity relationships without the reliance on prior knowledge by introducing a clustering-guided confusion index. Guided by the relationships, we then design and train a hierarchical activity recognition model. Finally, we apply it to recognize on-going activities associated with the sensor signals. In the next subsections, we detail the former two components. For illustration purposes, D denotes the training set, and L states the predefined activity labels. D = (X, Y) is a labeled dataset, where X is the feature matrix and Y is the corresponding label of each sample of X. tx is a test sample, and |L| indicates the number of predefined activities.

3.1. Measuring the Confusion among Activities

Considering that similar activities tend to generate similar sensor signals from the view of data distributions, we utilize the idea of clustering techniques to partition the data points of the predefined activities into multiple groups and analyze the group purities to quantify the confusion among activities. We herein introduce two definitions before illustrating the procedure.
Definition 1.
(Cluster Confusion Index). Given a cluster C consisting of a subset of samples from D, the class of C is set as the label Li (1 ≤ i ≤ |L|) that has the maximum data points in C. The number of samples with label Lj (1 ≤ j ≤ |L|, i ≠ j) is defined as the cluster confusion index between Lj and Li and is referred to as confc(Lj → Li).
L i = max L k L { x C I ( y x = L k ) } ,
where I (a = b) is an indicator function that returns 1 only if a equals b, x is a sample of C, and yx is the class label of x. The confc(Lj → Li) indicates how many samples of Lj in C are labeled as Li and denotes the confusion between Lj and Li in C. We can then measure the overall confusion index between Lj and Li in D.
Definition 2.
(Activity Confusion Index). Given the k clusters obtained by manual assignment or returned by a clustering algorithm, the activity confusion index conf(Lj → Li) between Lj and Li is defined as the sum of cluster confusion index of the k clusters, as given in Equation (2).
c o n f ( L j L i ) = c = 1 k c o n f c ( L j L i )
According to Definitions 1 and 2, we can apply a data-driven or knowledge-driven approach on D to getting k clusters and further obtaining the confusion index conf(LjLi) between Lj and Li. Due to the advantage of the data-driven approach over the knowledge-driven approach, we exploit the power of clustering techniques in partitioning D.
Next, when considering each pair of the predefined activities, we can organize the activity confusion index into a confusion matrix CM, also called the relationship matrix. Let CM denote a square matrix, and the number of rows equals |L|. The entry CMji of the j-th row and i-th column is the samples from the j-th activity that is categorized as the i-th activity. A large value of CMij leads to more confusion between the activities.
After quantifying the relationships among activities, we then detail how to use them to direct the construction of an activity recognition model in better discriminating confusing activities.

3.2. Hierarchical Activity Recognition Model

The above designed activity confusion matrix indicates that if the predicted label of a test sample tx is Li, its true label probably comes from Lj. We herein use a confusion threshold θ to decide whether Lj is a potential confusing activity of Li. Specifically, for the i-th column, if the ratio η(Lj, Li) of CMji to the sum of the i-th column of CM exceeds θ, Lj is taken as a confusing activity of Li and a mechanism is expected to distinguish Lj and Li.
η ( L j , L i ) = C M j i j = 1 | L | C M j i θ
Particularly, for each activity LAL, we can find a subset of activities S(LA) that are the confusing activities of LA. That is, it is necessary to further distinguish between S(LA) and LA if the predicted activity label is LA in the first step. Accordingly, we propose a hierarchical human activity recognition framework that works in the following scheme. In the training phase, initially, a top-level classification model cls_all are trained to differentiate among all the predefined activities, which takes essentially the same steps of the flat model. Afterwards, for each activity LAL, whose S(LA) is not empty, then a second-level classifier cls_LA is trained to classify LA and S(LA). Note, S(LA) is obtained from the activity confusion matrix in the previous step. In the prediction stage, given a test sample tx, we first classify it using cls_all and get the predicted result LA. If S(LA) is not empty, we then apply cls_LA on tx to return the inferred label; otherwise, report LA.
According to subsections A and B, the proposed framework is shown in Algorithm 1. The lines 1–2 show the steps of quantifying the confusion among activities, lines 3–5 denote the classification model training that mainly describes how to build a hierarchical activity recognizer with the guidance of the activity relationships, and lines 6–8 show the procedure of obtaining the predicted label of a test sample, which involves two-level classifications. Particularly, we present the roles of the two core components in Algorithm 1, where lines 1–2 illustrate the first component and lines 3–5 correspond to the second component, to state how to integrate them into the proposed framework.
Algorithm 1: Clustering Guided Hierarchical Activity Recognition Framework.
Input: Training set D and labels L for activities, Threshold θ, test set tx
Output: The prediction of activity with label LA of tx
// TRAINING STAGE
1. Divide D into groups CLU; // Component #1
2. calculate the activity confusion matrix CM of CLU by applying (1) and (2);
3. Construct cls_all to identify all activities; // Component #2
4. for each activity LA of L do
   (4.1) S(LA) = { }; // initialization of LA
5. for each activity LA of L do
   (5.1) for each activity LB of L do
      obtain η(LB, LA) by applying (3);
      if LA != LB and η(LB, LA) ≥ θ do
        S(LA).add(LB);
   (5.2) if not_empty(S(LA)) do
        construct a cls_LA to distinguish LA and S(LA);
// PREDICTION STAGE
6. LA = cls_all (tx); // infer the label of tx by applying the top-level classifier
7. if not_empty(S(LA)) do
    LA = cls_LA(tx); // infer the label of tx by applying the second-level classifier
8. return LA

4. Experimental Setup

4.1. Experimental Data

This section evaluates the comparative experiments on the UCI-HAR dataset, which consists of six human activities [37]. The smartphone, which has a 3-axis accelerometer and a gyroscope, works with a sampling rate of 50 Hz. For illustration purposes, X, Y, and Z refer to the three axes of the sensing unit, and ‘t’ (‘f’) denotes the time domain information. The features are associated with raw accelerometer signals tAcc-XYZ and gyroscope signals tGyto-XYZ. The frequency-domain features are obtained by applying a Fast Fourier Transform on tAcc-XYZ, and we get fBodyAcc-XYZ and fBodyAccJerk-XYZ. Similarly, for the gyroscope, we extract Jerk signals tBodyGyroJerk-XYZ, tBodyGyroJerkMag, and tBodyGyroMag from tGyto-XYZ. We then extract the frequency-domain features fBodyGyro-XYZ, fBodyGyroMag, and fBodyGyroJerkMag. In summary, 348 and 211 features are extracted from the accelerometer and gyroscope data. Extra two features are correlated with both the accelerometer and the gyroscope, so there are 561 features in total. The dataset consists of 7352 and 2947 training and test set samples, respectively. Furthermore, to study the power of different sensing units in activity recognition, we perform experiments with a single type of sensing unit (i.e., accelerometer or gyroscope) and two types of sensing units (i.e., accelerometer and gyroscope). Table 1 presents the dataset description, where “Acce & Gyro” refers to the use of both accelerometer and gyroscope.

4.2. Model Selection and Performance Metrics

As discussed above, the clustering-guided model takes various partitioning techniques and classification models as its building blocks. As for the data partition, we can do it manually or utilize a clustering algorithm. Since knowledge-driven methods have rather limited generalization to new and under-explored cases, we utilize clustering algorithms to perform the procedure in a data-driven way. Herein, we adopt the widely used k-means to partition data. We set the number of clusters returned by k-means to be the number of predefined activities. In addition, we evaluate the choice of other clustering algorithms in the next section. As for building the hierarchical activity recognizer, we can use the same classification model or different classification models. The two schemes are called homogeneous and heterogeneous schemes. In the study, we evaluate both cases. The naïve Bayes (NB), k nearest neighbor (KNN), support vector machine (SVM), and decision tree (DT) algorithms are adopted as the classification models. For KNN, we choose the nearest neighbor with the Euclidean distance metric. For SVM, the linear kernel is adopted. The four classification models are widely used in many existing studies [6,16,38]. For performance metrics, accuracy, precision, recall, and F1 are used to assess the power of an activity recognizer. F1, the harmonic mean of precision and recall.

5. Results and Analysis

5.1. Hierarchical Activity Recognizer

According to Algorithm 1, we first measure the confusion among activities and then build a clustering-guided hierarchical activity recognition model. Table 2 shows the activity confusion index between activities of UCI-HAR, where the column and row indicate the true activity and confusing labels of the activities. Furthermore, a confusion threshold θ is used to understand the confusion between similar activities. For example, if θ equals 0.03, the confusing activities of go upstairs include walking and go downstairs. If we set θ to be 0.005, the confusing activities of go upstairs include walking, go downstairs, sitting, and lying. Obviously, this makes it easier us not only to quantify the relationships among activities but also to correct the predictions of the top-level classifier.
Figure 2 presents the hierarchical human activity recognizer of UCI-HAR, where we empirically set the value of θ to be 0.01. As shown in Figure 2, the root node is to distinguish the six activities, the task of the internal node “going upstairs” is to distinguish the walking, go upstairs, and go downstairs, and the internal node “standing” aims to classify standing, sitting, and lying. Similarly, we can get the hierarchical activity recognizers for Gyroscope and Accelerometer datasets. After building the framework, the distinct classification models are integrated to train the top-level and second-level classifiers for optimizing a hierarchical activity recognizer.

5.2. Recognition Performance

We present experimental results of both the flat model and the hierarchical model, as shown in Table 3, Table 4 and Table 5 for three datasets, respectively. For the symbol “P-Q” of “the classifier” row, it means we use the classifier P at the top level and use the classifier Q at the second level. For example, NB-KNN indicates that we use NB at the top level, followed by KNN at the second level. To facilitate comparison, Figure 3 presents the comparison results of the four different activity recognition models. As shown in Figure 3, the usage of the accelerometer achieves better performance compared to the usage of the gyroscope. In addition, using both “Acce & Gyro” improves the accuracy. This indicates the priority of the Acce over Gyro in activity recognition and the complementary information shared by the gyroscope and accelerometer data. This is also consistent with previous studies. Exceptionally, we also observed that fusing gyroscope and accelerometer data did not guarantee the improved performance for all the cases, particularly for the use of NB. This is mainly because the redundant information provided by the 3-axis gyroscope and accelerometer destroys the conditional independence assumption. One possible way for further optimization is to conduct feature selection with filter, wrapper, or hybrid methods for filtering out irrelevant and redundant features. Second, we observe that the flat activity recognizers seldom classify a sample from dynamic activity into static activity (or vice versa). However, sometimes the wrong predictions are made in each group, which degraded the discriminant ability of an activity recognizer. This also motivates us to develop a hierarchical framework to recognize activities in a multi-step way. Third, except for the case of using NB at the second level, we observe that the proposed activity recognizer gets enhanced performance in many cases. This occurs because a hierarchical activity recognizer captures the confusion between types of activities and benefits from the finer step to classify confusing activities, which indicates the power of the clustering-guided model in discriminating activities that have similar sensor signals. Particularly, we have done this in a data-driven way instead of relying on expert knowledge to organize the activities of interest in a hierarchical structure. For example, NB-SVM obtains the 91.61% F1, compared with the 81.49% F1 of NB, 81.49% F1 of NB-NB, 83.95% F1 of NB-KNN, and 84.44% F1 of NB-DT in Table 3. On the UCI-HAR dataset, the best accuracy of the flat model is 96.47% achieved by SVM, and SVM-SVM improves it to 96.53%. This may indicate that SVM can better discriminate similar classes in finding an optimal separating hyperplane.
In addition, we present the confusion matrix of the activity recognizers to investigate the performance improvement. Due to the better performance of SVM, we give the results of NB, KNN, SVM, and DT and the comparative results obtained by using SVM at the second level. Figure 4 shows the results on UCI-HAR. The columns indicate the predicted labels and rows represent the actual labels. The used key-value pairs are: {1: walking, 2: go-upstairs, 3: go-downstairs, 4: sitting, 5: standing, 6: lying}. From Figure 4, we observe that the proposed hierarchical activity recognizer generally outperforms the flat model. For example, NB misclassifies 83 go-downstairs test samples into go-upstairs, and NB-SVM only makes 9 errors. Compared with SVM, SVM-SVM better discriminates between go-downstairs and go-upstairs. This indicates the superiority of a hierarchical activity recognizer in handling the activities that have similar sensor readings.

5.3. Evaluation of Different Confusion Thresholds

The confusion threshold θ is a crucial parameter of the hierarchical activity recognizer, which decides the confusing activities of an activity. In this subsection, we experimentally evaluate the effect of varying θ on the performance. Based on our previous work, the candidates of θ include 0.005, 0.01, 0.02, 0.03, 0.05, 0.1, and 0.5. Figure 5 presents the experimental results on UCI-HAR. The X-axis gives the candidate values of θ and the Y-axis shows its accuracy. For Figure 5a, the curves of NB and NB-NB exactly coincide, and so is the case with KNN and KNN-KNN in Figure 5b. From Figure 5, we see that the performance of the hierarchical activity recognizer is closely related to θ and 1% is a reasonable choice that achieves comparable or higher accuracy in most cases. Particularly, a small value of θ tends to get a densely connected graph, and a larger value of θ could lead to sparse connections between the activities at the second level, which decreases the hierarchical model to a flat model. Hence, the accuracy tends to generally decrease with the increase of θ expect the case of KNN-NB. Actually, we could choose the value of θ in a broad range such as between 0.01 and 0.05. This greatly relieves us from parameter tuning. It should be noticed that SVM achieves better performance at the second level.

5.4. Evaluation of Different Number of Clusters

In this subsection, the impact of the different number of clusters on the performance of hierarchical activity recognition models is studied. Figure 6 shows the experimental results on UCI-HAR. The X-axis refers to the number of clusters returned by k-means and the Y-axis is the accuracy obtained. Particularly, the values of the X-axis denote the values relative to the number |L| of predefined activities. For example, −1 means that the number of clusters of k-means equals |L| − 1, 0 indicates that the number of clusters equals |L|, and 1 means that the number of clusters equals |L| + 1. From Figure 6, we observe that setting the number of clusters to be the number of predefined activities generally obtains satisfactory accuracy, although there is no consistent trend between accuracy and |L|. This is most probable because this value corresponds to the true distribution of the dataset. Using the value helps relieve users from selecting the hyperparameter values for model optimization.

5.5. Evaluation of Different Distance Metrics

The distance metric is an important factor that determines the results of a clustering algorithm. We herein explore another two metrics (i.e., cosine and correlation) to calculate the distance between samples and further compare their performance with that of the Euclidean metric. Table 6 presents the experimental results, where the bold text indicates the best results. In addition, Figure 7 gives the corresponding F1. We observe that the use of the three metrics obtains comparable results. For example, in terms of SVM-SVM, the Euclidean, cosine, and correlation metrics obtain the 96.47%, 96.51%, and 96.51% accuracy, respectively. For DT-DT, the results are 86.36%, 86.29%, and 86.29% accuracy for the three metrics. This shows its robustness to the choice of distance metrics. We also observe that the use of the three metrics gets improved accuracy compared with the flat model. This demonstrates the effectiveness of the clustering-guided hierarchical model.

5.6. Evaluation of Different Clustering Algorithms

For the proposed framework, different clustering algorithms, besides k-means, can be integrated into it and function as its first component. In this subsection, we investigate another two commonly used clustering algorithms (i.e., k-medoids and agglomerative clustering), where the former chooses data points as the centers, and the latter builds a hierarchy of clusters. Specifically, according to the previous results, the number of clusters is set as the number of predefined activities. The Euclidean distance metric is used for k-means and k-medoids and applies the Ward’s method to calculate linkage of the agglomerative clustering.
Table 7 presents the corresponding results on UCI-HAR. In the table, the row “agglomerative” refers to the results when agglomerative clustering is used. The bold text indicates the best F1 of each group. In addition, Figure 8 gives the accuracy. From Table 7 and Figure 8, we observe that the hierarchical activity recognizer leads to enhanced recognition rates whichever clustering algorithm is used. We also observe that the use of k-means and agglomerative clustering obtains comparable performance and that the use of k-medoids slightly outperforms its competitors. For example, when NB is used as the top-level classifier, the best F1 of k-medoids, agglomerative clustering, and k-means are 0.9029, 0.8666, and 0.8683, respectively. If SVM is chosen as the top-level classifier, the F1 (0.9657) of k-medoids is higher than the F1 (0.9653) of both k-means and agglomerative clustering. The possible reason is that k-medoids takes data points rather than the average of the points in a cluster as the centers. This is consistent with the defined cluster confusion index and activity confusion index, where we use the majority voting principle to determine the confusing data points.

6. Conclusions

Accurately automating the recognition of human activities plays a prominent role in various applications like elderly care, ambient intelligence, smart homes, and human computer interactions. Training an activity recognition model poses difficulties in data training, modeling, and evaluation. Particularly, in the inter-subject and intra-subject variations, there are activities of interest that have similar sensor signals, which confuses an activity recognizer and result in degraded accuracy. In this study, a data-driven hierarchical activity recognition framework is proposed to quantify the confusion among activities and to build a hierarchical activity recognition model for improving the overall accuracy. Finally, the comparative analysis on the datasets is conducted, and the results show the success ratio of the grouping of the predefined activities in a specific application and its superiority over its competitors. In addition, we experimentally study the role of the two components within the framework from the aspects of sensitivity analysis of the confusion threshold, the choice of clustering algorithms and distance metrics, and the number of used clusters, which indicates its effectiveness.
In instantiating the framework, we can incorporate different clustering algorithms as well as classification models into it, where there are two hyperparameters (i.e., the number of clusters and the confusion threshold) that require us to assign their values. Although experimental results have shown its robustness, it is expected to adaptively set their values based on the data characteristics. This involves further studies for a comprehensive analysis. Second, the hierarchical framework brings larger computation costs, so there is a tradeoff between the recognition accuracy and resource constraints in deploying the model in edge devices.

Author Contributions

Conceptualization, A.W., S.Z. and G.C.; methodology, A.W., S.Z., H.-C.K., G.C. and D.S.R.; validation, A.W., H.-C.K. and D.S.R.; writing—original draft preparation, A.W.; writing—review and editing, A.W., H.-C.K. and D.S.R.; funding acquisition, A.W., S.Z. and G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of China (Grant Nos. 61902068, 61972092 and 62176082), the Featured Innovation Project of the Department of Education of Guangdong Province (Grant No. 2021KTSCX117), the Major Special Projects of Anhui Province (Grant No. 201903A06020026), the Key Research and Development Project of Anhui Province (Grant No. 201904A05020073).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dahmen, J.; Cook, D.J. Indirectly supervised anomaly detection of clinically meaningful health events from smart home data. ACM Trans. Intell. Syst. Technol. 2021, 12, 1–18. [Google Scholar] [CrossRef] [PubMed]
  2. Ghayvat, H.; Awais, M.; Pandya, S.; Ren, H.; Akbarzadeh, S.; Chandra Mukhopadhyay, S.; Chen, W. Smart aging system: Uncovering the hidden wellness parameter for well-being monitoring and anomaly detection. Sensors 2019, 19, 766. [Google Scholar] [CrossRef] [Green Version]
  3. Haque, A.; Milstein, A.; Fei-Fei, L. Illuminating the dark spaces of healthcare with ambient intelligence. Nature 2020, 585, 193–202. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, A.; Zhao, S.; Zheng, C.; Yang, J.; Chen, G.; Chang, C.Y. Activities of daily living recognition with binary environment sensors using deep learning: A comparative study. IEEE Sens. J. 2021, 21, 5423–5433. [Google Scholar] [CrossRef]
  5. Kim, E.; Helal, S.; Nugent, C.; Beattie, M. Analyzing activity recognition uncertainties in smart home environments. ACM Trans. Intell. Syst. Technol. 2015, 6, 52. [Google Scholar] [CrossRef]
  6. Bouchabou, D.; Nguyen, S.M.; Lohr, C.; LeDuc, B.; Kanellos, I. A survey of human activity recognition in smart homes based on IoT sensors algorithms: Taxonomies, challenges, and opportunities with deep learning. Sensors 2021, 21, 6037. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, A.; Chen, G.; Wu, X.; Liu, L.; An, N.; Chang, C.Y. Towards human activity recognition: A hierarchical feature selection framework. Sensors 2018, 18, 3629. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Cho, H.; Yoon, S. Divide and conquer-based 1D CNN human activity recognition using test data sharpening. Sensors 2018, 18, 1055. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Bai, L.; Yao, L.; Wang, X.; Kanhere, S.S.; Guo, B.; Yu, Z. Adversarial multi-view networks for activity recognition. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–22. [Google Scholar] [CrossRef]
  10. Lima, W.S.; Bragança, H.L.; Souto, E.J. NOHAR: NOvelty discrete data stream for human activity recognition based on smartphones with inertial sensors. Expert Syst. Appl. 2021, 166, 114093. [Google Scholar] [CrossRef]
  11. Singh, T.; Vishwakarma, D.K. A deeply coupled ConvNet for human activity recognition using dynamic and RGB images. Neural Comput. Appl. 2021, 33, 469–485. [Google Scholar] [CrossRef]
  12. Chen, G.; Wang, A.; Zhao, S.; Liu, L.; Chang, C.Y. Latent feature learning for activity recognition using simple sensors in smart homes. Multimed. Tools Appl. 2018, 77, 15201–15219. [Google Scholar] [CrossRef]
  13. Liu, Y.; Wang, K.; Li, G.; Lin, L. Semantics-aware adaptive knowledge distillation for sensor-to-vision action recognition. IEEE Trans. Image Process. 2021, 30, 5573–5588. [Google Scholar] [CrossRef] [PubMed]
  14. Snoun, A.; Jlidi, N.; Bouchrika, T.; Jemai, O.; Zaied, M. Towards a deep human activity recognition approach based on video to image transformation with skeleton data. Multimed. Tools Appl. 2021, 80, 29675–29698. [Google Scholar] [CrossRef]
  15. Patterson, D.; Fox, D.; Kautz, H.; Philipose, M. Fine-grained activity recognition by aggregating abstract object usage. In Proceedings of the Ninth IEEE International Symposium Wearable Computers (ISWC), Osaka, Japan, 18–21 October 2005; pp. 44–51. [Google Scholar]
  16. Wang, A.; Chen, G.; Yang, J.; Zhao, S.; Chang, C.Y. A comparative study on human activity recognition using inertial sensors in a smartphone. IEEE Sens. J. 2016, 16, 4566–4578. [Google Scholar] [CrossRef]
  17. Ordóñez, F.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [Green Version]
  18. Akbari, A.; Jafari, R. Personalizing activity recognition models through quantifying different types of uncertainty using wearable sensors. IEEE Trans. Biomed. Eng. 2020, 67, 2530–2541. [Google Scholar] [CrossRef]
  19. Shang, C.; Chang, C.Y.; Chen, G.; Zhao, S.; Chen, H. BIA: Behavior identification algorithm using unsupervised learning based on sensor data for home elderly. IEEE J. Biomed. Health Inform. 2019, 24, 1589–1600. [Google Scholar] [CrossRef]
  20. Xiao, Z.; Xu, X.; Xing, H.; Song, F.; Wang, X.; Zhao, B. A federated learning system with enhanced feature extraction for human activity recognition. Knowl.-Based Syst. 2021, 229, 107338. [Google Scholar] [CrossRef]
  21. Chen, L.; Nugent, C.D.; Wang, H. A knowledge-driven approach to activity recognition in smart homes. IEEE Trans. Knowl. Data Eng. 2012, 24, 961–974. [Google Scholar] [CrossRef]
  22. Fahad, L.G.; Tahir, S.F. Activity recognition and anomaly detection in smart homes. Neurocomputing 2021, 423, 362–372. [Google Scholar] [CrossRef]
  23. Huang, Y.; Yang, X.; Gao, J.; Sang, J.; Xu, C. Knowledge-driven egocentric multimodal activity recognition. ACM Trans. Multimed. Comput. Comm. Appl. 2021, 16, 1–133. [Google Scholar] [CrossRef]
  24. Rawashdeh, M.; Al Zamil, M.; Samarah, S.; Hossain, M.S.; Muhammad, G. A knowledge-driven approach for activity recognition in smart homes based on activity profiling. Future Gener. Comput. Syst. 2020, 107, 924–941. [Google Scholar] [CrossRef]
  25. Wang, L.; Gu, T.; Tao, X.; Lu, J. A hierarchical approach to real-time activity recognition in body sensor networks. Pervasive Mob. Comput. 2012, 8, 115–130. [Google Scholar] [CrossRef]
  26. Xu, L.; Yang, W.; Cao, Y.; Li, Q. Human activity recognition based on random forests. In Proceedings of the 13th International Conference of Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Guilin, China, 29–31 July 2017; pp. 548–553. [Google Scholar]
  27. Ramanujam, E.; Perumal, T.; Padmavathi, S. Human activity recognition with smartphone and wearable sensors using deep learning techniques: A review. IEEE Sens. J. 2021, 12, 13029–13040. [Google Scholar] [CrossRef]
  28. Khan, Z.N.; Ahmad, J. Attention induced multi-head convolutional neural network for human activity recognition. Appl. Soft Comput. 2021, 110, 107671. [Google Scholar] [CrossRef]
  29. Gao, W.; Zhang, L.; Huang, W.; Min, F.; He, J.; Song, A. Deep neural networks for sensor-based human activity recognition using selective kernel convolution. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar]
  30. Ronao, C.; Cho, S. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  31. Azkune, G.; Almeida, A.; López-de-Ipiña, D.; Chen, L. Extending knowledge-driven activity models through data-driven learning techniques. Expert Syst. Appl. 2015, 42, 3115–3128. [Google Scholar] [CrossRef]
  32. Sukor, A.; Zakaria, A.; Rahim, N.; Kamarudin, L.; Setchi, R.; Nishizaki, H. A hybrid approach of knowledge-driven and data-driven reasoning for activity recognition in smart homes. J. Intell. Fuzzy Syst. 2019, 36, 4177–4188. [Google Scholar] [CrossRef]
  33. Chen, J.; Huang, X.; Jiang, H.; Miao, X. Low-cost and device-free human activity recognition based on hierarchical learning model. Sensors 2021, 21, 2359. [Google Scholar] [CrossRef] [PubMed]
  34. Liu, Y.; Nie, L.; Han, L.; Zhang, L.; Rosenblum, D.S. Action2Activity: Recognizing complex activities from sensor data. In Proceedings of the Twenty-fourth International Joint Conference on Artificial Intelligence (IJCAI), Buenos Aires, Argentina, 25–31 July 2015; pp. 1617–1623. [Google Scholar]
  35. Khan, A.; Lee, Y.; Lee, S.Y.; Kim, T. A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 1166–1172. [Google Scholar] [CrossRef] [PubMed]
  36. Wang, A.; Zhao, S.; Zheng, C.; Chen, H.; Liu, L.; Chen, G. HierHAR: Sensor-based data-driven hierarchical human activity recognition. IEEE Sens. J. 2021, 21, 3353–3365. [Google Scholar] [CrossRef]
  37. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A public domain dataset for human activity recognition using smartphones. In Proceedings of the 21st European Symposium Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN), Bruges, Belgium, 24–26 April 2013; pp. 437–442. [Google Scholar]
  38. Wang, A.; Chen, H.; Zheng, C.; Zhao, L.; Liu, J.; Wang, L. Evaluation of random forest for complex human activity recognition using wearable sensors. In Proceedings of the 2020 International Conference on Networking and Network Applications (NaNA), Haikou, China, 11–14 December 2020; pp. 310–315. [Google Scholar]
Figure 1. The Flowchart of the proposed model.
Figure 1. The Flowchart of the proposed model.
Sensors 21 06962 g001
Figure 2. The clustering-guided hierarchical human activity recognizer on UCI-HAR.
Figure 2. The clustering-guided hierarchical human activity recognizer on UCI-HAR.
Sensors 21 06962 g002
Figure 3. Accuracy comparisons on the three datasets. (a) NB; (b) KNN; (c) SVM; (d) DT.
Figure 3. Accuracy comparisons on the three datasets. (a) NB; (b) KNN; (c) SVM; (d) DT.
Sensors 21 06962 g003
Figure 4. Comparison of the confusion matrix on UCI-HAR. (a) NB; (b) NB-SVM; (c) KNN; (d) KNN-SVM; (e) SVM; (f) SVM-SVM; (g) DT; (h) DT-SVM.
Figure 4. Comparison of the confusion matrix on UCI-HAR. (a) NB; (b) NB-SVM; (c) KNN; (d) KNN-SVM; (e) SVM; (f) SVM-SVM; (g) DT; (h) DT-SVM.
Sensors 21 06962 g004aSensors 21 06962 g004b
Figure 5. Accuracy vs. different thresholds. (a) NB; (b) KNN; (c) SVM; (d) DT.
Figure 5. Accuracy vs. different thresholds. (a) NB; (b) KNN; (c) SVM; (d) DT.
Sensors 21 06962 g005
Figure 6. Recognition accuracy vs. different number of clusters. (a) NB; (b) KNN; (c) SVM; (d) DT.
Figure 6. Recognition accuracy vs. different number of clusters. (a) NB; (b) KNN; (c) SVM; (d) DT.
Sensors 21 06962 g006
Figure 7. F1 vs. different distance metrics of k-means. (a) NB; (b) KNN; (c) SVM; (d) DT.
Figure 7. F1 vs. different distance metrics of k-means. (a) NB; (b) KNN; (c) SVM; (d) DT.
Sensors 21 06962 g007
Figure 8. Recognition accuracy vs. different clustering algorithms on UCI-HAR. (a) NB; (b) KNN; (c) SVM; (d) DT.
Figure 8. Recognition accuracy vs. different clustering algorithms on UCI-HAR. (a) NB; (b) KNN; (c) SVM; (d) DT.
Sensors 21 06962 g008
Table 1. The UCI-HAR dataset description.
Table 1. The UCI-HAR dataset description.
Sensing UnitsThe Number of Features
Time-DomainFrequency-DomainTotal
Accelerometer164184348
Gyroscope106105211
Acce & Gyro272289561
Table 2. Class confusion index on UCI-HAR.
Table 2. Class confusion index on UCI-HAR.
WalkingUpstairsDownstairsSittingStandingLying
Walking0629597000
Upstairs0822251000
Downstairs0137849000
Sitting0100123649
Standing000013740
Lying010001641233
Table 3. Recognition performance on the Accelerometer dataset.
Table 3. Recognition performance on the Accelerometer dataset.
ModelNBNB-NBNB-KNNNB-SVMNB-DTKNNKNN-NBKNN-KNNKNN-SVMKNN-DT
Accuracy81.0781.0783.9891.4884.5684.3281.4184.3291.8985.04
Precision80.3880.3883.7291.5084.2984.0380.7084.0391.8884.73
Recall82.6382.6384.1991.7284.5884.5582.8584.5592.1685.08
F181.4981.4983.9591.6184.4484.2981.7684.2992.0284.90
ModelSVMSVM-NBSVM-KNNSVM-SVMSVM-DTDTDT-NBDT-KNNDT-SVMDT-DT
Accuracy91.8981.3484.3291.8984.9784.9381.5784.3291.9685.14
Precision91.8780.6484.0391.8884.6784.6380.8784.0391.9584.82
Recall92.1682.7784.5592.1685.0184.9983.0384.5892.2485.19
F192.0281.6984.2992.0284.8484.8181.9484.3092.0985.01
Table 4. Recognition performance on the Gyroscope dataset.
Table 4. Recognition performance on the Gyroscope dataset.
ModelNBNB-NBNB-KNNNB-SVMNB-DTKNNKNN-NBKNN-KNNKNN-SVMKNN-DT
Accuracy50.8050.8054.0961.5555.3465.8354.2965.8375.3366.75
Precision51.9251.9255.0462.8356.3466.0855.1866.0875.7066.91
Recall55.0055.0061.7671.0462.3266.4160.6766.4177.9769.50
F153.4253.4258.2066.6959.1866.2557.7966.2576.8268.18
ModelSVMSVM-NBSVM-KNNSVM-SVMSVM-DTDTDT-NBDT-KNNDT-SVMDT-DT
Accuracy81.3758.4370.7281.3072.1868.0755.6566.5476.3868.21
Precision81.7359.4371.0481.6672.3968.1756.5766.6876.6168.29
Recall81.6464.1772.3681.6073.4168.2462.0969.4778.6068.34
F181.6961.7171.7081.6372.9068.2059.2068.0577.5968.32
Table 5. Recognition performance on the UCI-HAR dataset.
Table 5. Recognition performance on the UCI-HAR dataset.
ModelNBNB-NBNB-KNNNB-SVMNB-DTKNNKNN-NBKNN-KNNKNN-SVMKNN-DT
Accuracy76.9976.9978.8684.8377.0387.8578.7287.8592.2685.31
Precision76.8876.8878.9085.1377.0887.4478.6387.4492.0884.80
Recall79.2379.2383.4888.6081.0987.9682.8687.9693.0685.98
F178.0478.0481.1286.8379.0387.7080.6987.792.5785.39
ModelSVMSVM-NBSVM-KNNSVM-SVMSVM-DTDTDT-NBDT-KNNDT-SVMDT-DT
Accuracy96.3482.3290.4396.4789.7986.3677.1085.8892.4786.36
Precision96.2682.1690.0196.4289.4085.9976.9585.4092.3885.99
Recall96.5285.5890.9196.6589.8086.3180.7886.8593.1386.31
F196.3983.8490.4696.5389.6086.1578.8286.1292.7586.15
Table 6. Performance of different distance metrics for k-means on UCI-HAR.
Table 6. Performance of different distance metrics for k-means on UCI-HAR.
ModelNB-NBNB-KNNNB-SVMNB-DTKNN-NBKNN-KNNKNN-SVMKNN-DT
MetricsAccF1AccF1AccF1AccF1AccF1AccF1AccF1AccF1
Euclidean76.990.78078.860.81184.830.86877.030.79078.720.80787.850.87792.260.92685.310.854
Cosine76.990.78080.690.82987.720.89778.720.80778.320.80287.850.87794.770.95086.500.864
Correlation76.990.78080.690.82987.720.89778.720.80777.940.79687.850.87794.810.95086.530.865
ModelSVM-NBSVM-KNNSVM-SVMSVM-DTDT-NBDT-KNNDT-SVMDT-DT
MetricsAccF1AccF1AccF1AccF1AccF1AccF1AccF1AccF1
Euclidean82.320.83890.430.90596.470.96589.790.89677.100.78885.880.86192.470.92886.360.862
Cosine80.690.82090.020.89996.510.96688.570.88378.110.79888.190.88294.980.95186.290.861
Correlation80.180.81390.020.89996.510.96688.570.88377.670.79288.190.88294.980.95186.290.861
Table 7. F1 of different clustering algorithms on UCI-HAR.
Table 7. F1 of different clustering algorithms on UCI-HAR.
ModelNBNB-NBNB-KNNNB-SVMNB-DTKNNKNN-NBKNN-KNNKNN-SVMKNN-DT
k-means0.78030.78030.81120.86830.79030.8770.80690.87700.92570.8538
k-medoids0.78030.78030.83500.90290.81220.8770.80190.87700.94950.8640
agglomerative0.78030.78030.79410.86660.77630.8770.8570.87700.93980.8480
ModelSVMSVM-NBSVM-KNNSVM-SVMSVM-DTDTDT-NBDT-KNNDT-SVMDT-DT
k-means0.96390.83840.90450.96530.89600.86150.78820.86120.92750.8615
k-medoids0.96390.82010.89860.96570.88290.86150.79780.88180.95090.8607
agglomerative0.96390.87320.88380.96530.87310.86150.84530.85740.94230.8615
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, A.; Zhao, S.; Keh, H.-C.; Chen, G.; Roy, D.S. Towards a Clustering Guided Hierarchical Framework for Sensor-Based Activity Recognition. Sensors 2021, 21, 6962. https://0-doi-org.brum.beds.ac.uk/10.3390/s21216962

AMA Style

Wang A, Zhao S, Keh H-C, Chen G, Roy DS. Towards a Clustering Guided Hierarchical Framework for Sensor-Based Activity Recognition. Sensors. 2021; 21(21):6962. https://0-doi-org.brum.beds.ac.uk/10.3390/s21216962

Chicago/Turabian Style

Wang, Aiguo, Shenghui Zhao, Huan-Chao Keh, Guilin Chen, and Diptendu Sinha Roy. 2021. "Towards a Clustering Guided Hierarchical Framework for Sensor-Based Activity Recognition" Sensors 21, no. 21: 6962. https://0-doi-org.brum.beds.ac.uk/10.3390/s21216962

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop