Next Article in Journal
A Technique for Improving the Precision of the Direct Measurement of Junction Temperature in Power Light-Emitting Diodes
Next Article in Special Issue
Usability, User Experience, and Acceptance Evaluation of CAPACITY: A Technological Ecosystem for Remote Follow-Up of Frailty
Previous Article in Journal
Panoramic Dental Radiography Image Enhancement Using Multiscale Mathematical Morphology
Previous Article in Special Issue
An Experimental Study on the Validity and Reliability of a Smartphone Application to Acquire Temporal Variables during the Single Sit-to-Stand Test with Older Adults
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network

1
Department of Electrical Engineering, National Taipei University of Technology, Taipei 10608, Taiwan
2
Department of Computer Science and Information Engineering, Chaoyang University of Technology, Taichung 413310, Taiwan
3
Department of Information and Communication Engineering, Chaoyang University of Technology, Taichung 413310, Taiwan
4
Department of Computer Science & Software Engineering, College of Information Technology, United Arab Emirates University, Al Ain P.O. Box 15551, Abu Dhabi, United Arab Emirates
*
Author to whom correspondence should be addressed.
Submission received: 30 March 2021 / Revised: 27 April 2021 / Accepted: 27 April 2021 / Published: 29 April 2021
(This article belongs to the Special Issue Mobile Health Technologies for Ambient Assisted Living and Healthcare)

Abstract

:
Research on the human activity recognition could be utilized for the monitoring of elderly people living alone to reduce the cost of home care. Video sensors can be easily deployed in the different zones of houses to achieve monitoring. The goal of this study is to employ a linear-map convolutional neural network (CNN) to perform action recognition with RGB videos. To reduce the amount of the training data, the posture information is represented by skeleton data extracted from the 300 frames of one film. The two-stream method was applied to increase the accuracy of recognition by using the spatial and motion features of skeleton sequences. The relations of adjacent skeletal joints were employed to build the direct acyclic graph (DAG) matrices, source matrix, and target matrix. Two features were transferred by DAG matrices and expanded as color texture images. The linear-map CNN had a two-dimensional linear map at the beginning of each layer to adjust the number of channels. A two-dimensional CNN was used to recognize the actions. We applied the RGB videos from the action recognition datasets of the NTU RGB+D database, which was established by the Rapid-Rich Object Search Lab, to execute model training and performance evaluation. The experimental results show that the obtained precision, recall, specificity, F1-score, and accuracy were 86.9%, 86.1%, 99.9%, 86.3%, and 99.5%, respectively, in the cross-subject source, and 94.8%, 94.7%, 99.9%, 94.7%, and 99.9%, respectively, in the cross-view source. An important contribution of this work is that by using the skeleton sequences to produce the spatial and motion features and the DAG matrix to enhance the relation of adjacent skeletal joints, the computation speed was faster than the traditional schemes that utilize single frame image convolution. Therefore, this work exhibits the practical potential of real-life action recognition.

1. Introduction

Recently, the lifespans of the world’s population are increasing, and society is gradually aging. According to the report of the United Nations [1], the number of elderly people (over 65) in the world in 2019 was 703 million, and this is estimated to double to 1.5 billion by 2050. From 1990 to 2019, the proportion of the global population over 65 years old increased from 6% to 9%, and the proportion of the elderly population is expected to further increase to 16% by 2050. In Taiwan, the report of the National Development Council indicated that the elderly population with age over 65 will exceed 20% of the national population at 2026 [2].
Taiwan will enter a super-aged society in 2026. This means that the labor manpower will gradually decrease in the future. Thus, the cost of home care for elders will significantly increase. In homecare, the monitoring of elderly people living alone is a major issue. The behaviors of their activities have a high relation with their physical and mental health [3,4]. Therefore, how to use artificial intelligence (AI) techniques to reduce the cost of home care is an important challenge.
The recognition of body activities has two major techniques. One is physical sensors, like accelerometers [5,6], gyroscopes [7], and strain gauges [8], which have the advantage that they can be worn on the body to monitor dangerous activities throughout the day and the disadvantage that few activities can be recognized. Therefore, the physical sensors are typically not used to identify the daily activities. Another technique is the charge-coupled device (CCD) camera [9,10], which has the advantage of being able to recognize many daily activities and the disadvantage that it can only be used to monitor the activities of people in a local area. Thus, it is suitable to be used in a home environment.
Many previous studies have used deep learning techniques to recognize daily activities, including two-stream convolutional neural networks (CNNs), long short-term memory networks (LSTMNs), and three-dimensional CNNs (3D CNNs). For the two-stream CNN, Karpathy et al. used context stream and fovea streams to train a CNN [11]. The two streams proposed by Simonyan et al. were spatial and temporal streams, which represent the static and dynamic frames of each action’s film [12]; however, the spending time for each action was different.
Thus, Wang et al. proposed a time segment network to normalize the spatial and temporal streams [13]. Jiang et al. used two streams as the input combined with CNN and LSTMN [14]. Ji et al. proposed 3D CNN to obtain the features of the spatial and temporal streams [15]. The two-stream methods using image and optical flow to represent the spatial and temporal streams had a better performance for recognizing activities compared with the one-stream methods. However, the weakness is that its doubled data amount requires more time to train the model.
Studies have used skeletal data as the common input feature for human action recognition [16,17,18], where the 3D skeletal data were typically obtained by use of the depth camera. In these studies, the number of recognized actions was less than 20 [17], and the skeletal data had to be processed to extract the features. Machine learning methods were used in these studies. The spatiotemporal information of skeleton sequences was exploited using recurrent neural networks (RNNs) [19,20].
Both the amount of data and recognized actions were less than in the video datasets [16,17,18,19,20]. An RNN tends to overemphasize the temporal information and ignore the spatial information, leading to low accuracy. However, the advantage of methods employing skeletal data is that these requires less training data and training time compared to those using image data. Hou et al. used a CNN to recognize actions with skeletal features [21]. Therefore, an effective method to encode the spatiotemporal information of a skeleton sequence into color texture images that could be recognized by a CNN is a relevant issue.
A directed acyclic graph (DAG) consists of a combination of nodes and edges. Each node points to another node by an edge. These directions do not become a circle graph that will end at the extremities. DAGs are usually used to represent causal relations amongst variables, and they are also used extensively to determine which variables need to be controlled for confounding in order to estimate causal effects [22]. The physical posture of people can be described by the positions of the skeletal joints. The adjacent joints have a causal relation when the body is moving. Thus, we can define the DAG of skeletal joints to explain the relations of physical skeletons.
This study aims to recognize the daily activities with films recorded by CCD cameras. To reduce the large amount of data for model training, we transferred body images to physical postures with an open system, AlphaPose [23]. The posture information is the skeleton sequences captured from the films of actions to build the spatial and motion features. These features all include both the spatial and temporal characteristics of actions. The relations of adjacent joints were used to build direct acyclic graph (DAG) matrices, the source matrix, and the target matrix.
These features are expanded by the DAG matrices as color texture images. The linear-map CNN has a two-dimensional linear map at the beginning of each layer to adjust the number of channels. Then, a two-dimensional CNN is used to recognize the actions. A structure with two streams was used to increase the accuracy of the action recognition. The datasets (NTU RGB+D) used in this study is an open source supported by Rapid-Rich Object Search Lab, National Technological University, Singapore [18].
In our work, a total of 49 actions, including daily actions, medical conditions, and mutual actions, were considered for action recognition. The total number of films was 46,452 for the cross-subject and cross-view sources. Of the cross-subject sources, 32,928 films were used for training, and 13,524 films were used for testing. Of the cross-view sources, 30,968 films were used for training, and 15,484 were used for testing. The experimental results show that the performance of our method was better than those in the previous studies.

2. Methods

Figure 1 shows the flowchart of action recognition in this study, which has three phases. In the feature phase, RGB images were processed by AlphaPose [23] to obtain the coordinate values of the skeletal joints of the subject in an image as a vector. A film contained 300 images that were used to build a posture matrix as the feature. We defined the spatial features and motion features by the coordinate values of the skeletal joints for each film. Spatial features are the position information of skeletons and joints, and motion features are the optical-flow information of skeletons and joints.
Each feature was expanded into two features of source and target by DAGs. In the recognition phase, a 10-layer linear-map CNN was used to recognize the activities. The cross-subject and cross-view evaluations were used to test the performance of this linear-map CNN. In the output phase, the results of the spatial and motion features were fused to show the recognized actions.

2.1. NTU RGB+D Dataset

The datasets of action recognition supported by the Rapid-Rich Object Search Lab, National Technological University, Singapore [24] were used in the study. There were 56,880 files, including 60 action classes. Each file consists of RGB, depth, and skeleton data of human actions. All actions were recorded by three Kinect V2 cameras. The size of the RGB images was 1920 × 1080. There were 40 classes for daily actions, 9 classes for medical conditions, and 11 classes for mutual actions. Forty distinct subjects were invited for this data collection.
The physical activities of only a single person were recognized, and the sample size of 46,452, including the 49 physical activities, was used in this study. To ensure standard evaluations for all the reported results on the benchmark, two types of action classification evaluation (cross-subject evaluation, and cross-view evaluation) were used [24]. In the cross-subject evaluation, the sample sizes for training and testing sets were 40,320 and 16,560, respectively. In the cross-view evaluation, the sample sizes for training and testing sets were 37,920 and 18,960, respectively.

2.2. Spatial and Motion Features

The RGB images were processed by the AlphaPose [23] to obtain the coordinate values of skeleton joints of people in the image, and the format is shown in Equation (1).
pose = x 0 , y 0 , c 0 , , x M , y M , c M M = 17
where xi and yi are the coordinate values of ith joint, ci is the confidence score, and M is the index of the joints. According to the coordinate values of the joints, the spatial and motion variables were defined, as shown in Table 1. n is the index of the frames. The spatial variables are the joint data (vi), and skeleton data (si,j). The motion variables are the motion data of the joints and skeleton (mvi and msi). Thus, there are four features in this study, Fv, Fs, Fmv, and Fms. Table 2 shows the indexes and definition of 18 joints and 17 edges in the body. The 17 edges, ei, are defined as the relations between two adjacent joints, i-1th and ith.

2.3. Directed Acyclic Graph

DAG was used to describe the relations of 18 joints. The nodes of DAG represent the joints, and the flows represent the edges. Each edge has a source joint and a target joint. Thus, two DAG matrices, the source matrix and target matrix, can be defined. If the ith joint is the source point of the jth edge, the element (j, i) of the source matrix is set as 1. Otherwise, the element (j, i) is set as 0.
The target matrix is set as the source matrix. Then, each row of the source and target matrices is normalized to avoid overvalues. To match the size of feature, e(0, 0) = 1 is defined as a virtual edge. The sizes of the source and target matrices are 18∗18. Figure 2a is the source matrix, S, color-none is 0, color-black is 1, and color-original is 0.25. Figure 2b is the target matrix, T.

2.4. Input Features

We used 300 frames for every film. The joint data built the joint feature (Fv), the skeleton data built the skeleton feature (Fs), the joint-motion data built the joint-motion feature (Fmv), and the skeleton-motion data built the skeleton-motion feature (Fms). Figure 3 shows the contents of a feature with x and y values of a data. Thus, the information of the film was reduced to four 600*18 matrices for four features, Fv, Fs, Fmv, and Fms. The Fv was expanded by the DAG matrix into two features, Fvin and Fvout.
F v i n = F v × S T
F v o u t = F v × T T
Fs was expanded by the DAG matrix to Fsin and Fsout; Fmv was expanded to Fmvin and Fmvout; and Fms was expanded to Fmsin and Fmsout. Table 3 shows the contents of the spatial feature (Fspatial) and motion feature (Fmotion). Fsaptial is the combination of Fspatial-joint and Fspatial_skeleton. Fmotion is the combination of Fmotion-joint and Fmotion_skeleton. We used the two features (Fspatial and Fmotion) to evaluate the performance of the linear-map CNN.

2.5. Linear-Map CNN

Figure 4 shows the structure of a 10-layer linear-map CNN. The linear-map was used to adjust the number of channels at the beginning of each layer. Batch normalization (BN) can overcome the disappearance of the learning gradient and, thus, use a larger learning rate. In the CNN, the kernel size is a 9 × 1 matrix, the stride is (1,1), and the padding is (4,0). In the input feature, columns represent the different joints, and rows represent the time sequence of the actions. The relation of the adjacent joints was enhanced by the DAG matrix. Thus, the kernel of convolution is a 9 × 1 matrix. Table 4 shows the detailed information of the linear-map CNN. The output layer has 49 nodes representing the 49 action classes. The optimal method was momentum. The batch number is 32.

2.6. Statistical Analysis

According to our proposed method, a film is considered as true positive (TP) when the classification action is correctly identified; false positive (FP) when the classification action is incorrectly identified; true negative (TN) when the action classification is correctly rejected, and false-negative (FN) when the action classification is incorrectly rejected. Here, the performance of the proposed method was evaluated using these parameters,
P r e c i s i o n % = T P T P + F P × 100 % ,
R e c a l l % = T P T P + F N × 100 % ,
S p e c i f i c i t y % = T N T N + F P × 100 % ,
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l × 100 % ,
A c c u r a c y % = T P + T N T P + T N + F P + F N × 100 % .

3. Results

In this study, the hardware employed was CPU Intel Core i7-8700 and GPU GeForce GTX1080. The operating system was Ubuntu 16.04LTS software, the development system was Anaconda 3 at python 3.7 version, the tool of deep learning was Pytorch 1.10, and the compiler was Jupyter Notebook. We evaluated the performance of DAG with the cross-subject and cross-view sources, and four features (Fspatial_joint, Fspatial-skeleton, Fmotion_joint, and Fmotion_skeleton). At last, we used the two-stream concept, class score fusion for Fspatial and Fmotion, to evaluate the performance of the proposed method with cross-subject and cross-view sources.
Table 5 shows the results without the DAG transfer. The best feature is Fspatial under the cross-subject source, which resulted in an accuracy and F1-score of 99.3% and 82.8%, respectively. There were 10 actions with recall rates below 70%: 0, 4, 9, 10, 11, 16, 28, 29, 43, and 48. The worst feature is Fmotion under the cross-subject source, its accuracy and F1-score are 99.2% and 79.6%, respectively. There are 11 actions with recall rates below 70%: 3, 10, 11, 16, 24, 28, 29, 31, 36, 43, and 45. Table 6 shows the results with the DAG transfer. The best feature was Fspatial under the cross-view source; its accuracy and F1-score were 99.9% and 96.2%, respectively.
Only four actions, 10, 11, 28, and 29, had recall rates below 70%. The worst feature was Fmotion under the cross-subject source, which obtained an accuracy and F1-score of 99.1% and 79.1%, respectively. There were 10 actions with recall rates below 70%: 2, 10, 11, 16, 28, 29, 31, 43, 44, and 45. We found that the DAG transfer could significantly improve the recognition rate of different actions, not only for spatial features but also for motion features. Table 7 shows the results of class score fusion with and without DAG transfer. We found that the performance of DAG transfer used in the cross-view source was better than used in the cross-subject source, with an accuracy of 99.9% vs. 99.5% and F1-score of 94.7% vs. 86.3%. The recall rates for all 49 actions were not below 70%.
We used the two-dimensional joint and skeleton features to perform the training and testing of the linear-map CNN, which could reduce the running time more than those using two- or three-dimensional joint and skeleton images. Table 8 shows the training and testing time with and without DAG transfer. We found that the GPU could process about 30 frames/second (fps) in the training phase, and process about 125 fps in the testing phase. Although the DAG transfer required time to process, the delay time was about 30 min in the training phase. The maximum testing time was 141 s.

4. Discussion

In this study, we used DAG transfer and the two-stream method to improve the accuracy of action recognition. When the input features were transferred with the DAG matrices, the precision, recall, specificity, F1-score, and accuracy were improved by 1.2%, 1.1%, 0.1%, 1.3%, and 0.1%, respectively, for the cross-subject source, and were improved by 9.1%, 7.4%, 0.1%, 7.4%, and 0.5%, respectively, for the cross-view source. In the two-stream method, previous studies have typically used the spatial and temporal, or optical flow features to perform the active score fusion [11,12,13,14]. They also proved that the performance of two streams was better than one stream.
We used joint and skeleton sequences as the spatial motion features that had temporal characteristics. We utilized 300 frames to describe an action. Thus, the spatial feature of one action included the spatial and temporal characters. However, the motion relations of the joints and skeletons were different from one action to another. Therefore, we defined the motion variables, mvi and msi, as shown in Table 1, to establish the motion features.
Our results also show that the precision, recall, specificity, F1-score, and accuracy of the two streams were improved by 0% vs. 6.3%, 0.1% vs. 7.4%, 0% vs. 0.2%, 0.1% vs. 7.1%, and 0.1% vs. 0.4%, over spatial and motion streams with the DAG transfer in the cross-subject source, and improved by 0.5% vs. 3.9%, 0.4% vs. 4.5%, 0% vs. 0%, 0.5% vs. 4.3%, and 0% vs. 0.2% in the cross-view source.
The comparison of our results with the previous studies under the recall rate is shown in Table 9. These studies all used cross-subject and cross-view sources from the NTU RGB+D database to recognize actions and also used three-dimensional characteristics of each posture as the input features [19,25,26,27,28,29,30,31]. Our method had the best recall rates of 86.1% and 94.7% in the cross-subject and cross-view sources.
We analyzed the actions with lower recall rates in the cross-subject and cross-view sources in Table 5, Table 6 and Table 7. The four actions that often had lower recall rates were A10 (reading), A11 (writing), A28 (phone call), and A29 (playing with laptop). Figure 5a is the posture of reading, and Figure 5b is the posture of writing. The subject is standing up, looking down, and holding something. The difference between the two images is only in the gestures of two hands. However, according to the description of the body posture in Table 2, only the right and left wrist joints are marked, which cannot show the gestures of two hands.
The postures of the subject making a phone call (in Figure 5c) and using the laptop (in Figure 5d)) had the same problem. The difference between the two images was also in the gestures of the two hands. These actions were difficult to recognize using spatial features, such as the movement trajectories of the arms, elbows, and wrists. Thus, the results of our method with the DAG transfer and two-stream method in Table 7 show that no action had a lower recall rate in the cross-view source.

5. Conclusions

The large scale of the collected data in the NTU RGB+D database enabled us to apply the posture-driven learning method for action recognition. The posture information represented by the skeleton data was obtained from the 300 frames of the film. The joint and skeleton sequences were used to build spatial and motion features that included the spatial and temporal characteristics of the actions. The relations of the adjacent skeletal joints were used to build the DAG matrices.
The spatial or motion features were expanded by DAG matrices as color texture images. The expanded features all indicated that the relations between adjacent joints were enhanced. Our method effectively reduced the amount of data for training the linear-map CNN, and its performance was superior to the previous schemes using deep learning methods. Notably, since the computation speed can reach around 125 fps in the testing phase with GPU, our scheme could be used to monitor the daily activities of elders in real time in home care applications.

Author Contributions

Conceptualization, T.-H.T. and S.-H.L.; Data curation, J.-H.H.; Investigation, T.-H.T.; Methodology, J.-H.H.; Project administration, T.-H.T.; Software, J.-H.H.; Supervision, Y.-F.H. and M.G.; Validation, Y.-F.H. and M.G.; Writing original draft, S.-H.L.; Writing review and editing, S.-H.L., Y.-F.H., and M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology, Taiwan, under grants MOST 109-2221-E-324-002-MY2 and MOST 109-2221-E-027-97.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. United Nations, Department of Economic Social Affairs. World Population Ageing 2019; United Nations: New York, NY, USA, 2019. [Google Scholar]
  2. Available online: https://www.ndc.gov.tw/Content_List.aspx?n=695E69E28C6AC7F3 (accessed on 28 April 2021).
  3. Nyboe, L.; Lund, H. Low levels of physical activity in patients with severe mental illness. Nord. J. Psychiatry 2013, 67, 43–46. [Google Scholar] [CrossRef]
  4. Oliveira, J.; Ribeiro, F.; Gomes, H. Effects of a home-based cardiac rehabilitation program on the physical activity levels of patients with coronary artery disease. J. Cardiopulm. Rehabil. Prev. 2008, 28, 392–396. [Google Scholar] [CrossRef]
  5. Liu, S.-H.; Cheng, W.-C. Fall detection with the support vector machine during scripted and continuous unscripted activities. Sensors 2012, 12, 12301–12316. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Liu, S.-H.; Chang, Y.-J. Using accelerometers for physical actions recognition by a neural fuzzy network. Telemed. e-Health 2009, 15, 867–876. [Google Scholar] [CrossRef]
  7. Tufek, N.; Yalcin, M.; Altintas, M.; Kalaoglu, F.; Li, Y.; Bahadir, S.K. Human action recognition using deep learning methods on limited sensory data. IEEE Sens. J. 2020, 20, 3101–3112. [Google Scholar] [CrossRef] [Green Version]
  8. Wong, W.Y.; Wong, M.S.; Lo, K.H. Clinical applications of sensors for human posture and movement analysis: A review. Prosthet. Orthot. Int. 2007, 31, 62–75. [Google Scholar] [CrossRef]
  9. Feichtenhofer, C.; Pinz, A.; Zisserman, A. Convolutional two-stream network fusion for video action recognition. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1933–1941. [Google Scholar]
  10. Fernando, B.; Gavves, E.; Oramas, M.J.; Ghodrati, A.; Tuytelaars, T. Modeling video evolution for action recognition. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5378–5387. [Google Scholar]
  11. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Fei, L. Large-scale video classification with convolutional neural networks. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar]
  12. Simonyan, K.; Zisserman, A. Two-stream convolutional networks for action recognition in videos. arXiv Prepr. 2014, arXiv:1406.2199. [Google Scholar]
  13. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Gool, L.V. Temporal segment networks: Towards good practices for deep action recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 20–36. [Google Scholar]
  14. Jiang, Y.; Wu, Z.; Tang, J.; Li, Z.; Xue, X.; Chang, S. Modeling multimodal clues in a hybrid deep learning framework for video classification. IEEE Trans. Multimed. 2018, 20, 3137–3147. [Google Scholar] [CrossRef] [Green Version]
  15. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 221–231. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Chaudhry, R.; Ofli, F.; Kurillo, G.; Bajcsy, R.; Vidal, R. Bioinspired dynamic 3D discriminative skeletal features for human action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 471–478. [Google Scholar]
  17. Vemulapalli, R.; Arrate, F.; Chellappa, R. R3DG features: Relative 3D geometry-based skeletal representations for human action recognition. Comput. Vis. Image Understand. 2016, 152, 155–166. [Google Scholar] [CrossRef]
  18. Wang, P.; Li, W.; Gao, Z.; Zhang, J.; Tang, C.; Ogunbona, P.O. Action recognition from depth maps using deep convolutional neural networks. IEEE Trans. Hum.–Mach. Syst. 2016, 46, 498–509. [Google Scholar] [CrossRef]
  19. Du, Y.; Wang, W.; Wang, L. Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1110–1118. [Google Scholar]
  20. Veeriah, V.; Zhuang, N.; Qi, G.-J. Differential recurrent neural networks for action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4041–4049. [Google Scholar]
  21. Hou, Y.; Li, Z.; Wang, P.; Li, W. Skeleton optical spectra-based action recognition using convolutional neural networks. IEEE Trans. Circ. Syst. Video Technol. 2018, 28, 807–811. [Google Scholar] [CrossRef]
  22. VanderWeele, T.J.; Robins, J.M. Four types of effect modification: A classification based on directed acyclic graphs. Epidemiology 2007, 18, 561–568. [Google Scholar] [CrossRef] [PubMed]
  23. Fang, H.; Xie, S.; Tai, Y.; Lu, C. RMPE: Regional multi-person pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2353–2362. [Google Scholar]
  24. Rapid-Rich Object Search (ROSE) Lab. 2019. Available online: http://rose1.ntu.edu.sg/datasets/actionrecognition.asp (accessed on 28 April 2021).
  25. Vemulapalli, R.; Arrate, F.; Chellappa, R. Human action recognition by representing 3d skeletons as points in a lie group. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 588–595. [Google Scholar]
  26. Shahroudy, A.; Liu, J.; Ng, T.; Wang, G. NTU RGB+D: A large scale dataset for 3d human activity analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1010–1019. [Google Scholar]
  27. Liu, J.; Shahroudy, A.; Xu, D.; Wang, G. Spatial-temporal LSTM with trust gates for 3D human action recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 816–833. [Google Scholar]
  28. Wang, H.; Wang, L. Modeling temporal dynamics and spatial configurations of actions using two-stream recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 499–508. [Google Scholar]
  29. Ke, Q.; Bennamoun, M.; An, S.; Sohel, F.; Boussaid, F. A new representation of skeleton sequences for 3D action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 7444–7452. [Google Scholar]
  31. Si, C.; Jing, Y.; Wang, W.; Wang, L.; Tan, T. Skeleton-based action recognition with spatial reasoning and temporal stack learning. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 106–121. [Google Scholar]
Figure 1. The flowchart of the action recognition in this study.
Figure 1. The flowchart of the action recognition in this study.
Sensors 21 03112 g001
Figure 2. (a) Source matrix, (b) Target matrix. Color-none is 0, color-black is 1, and color-orange is 0.25.
Figure 2. (a) Source matrix, (b) Target matrix. Color-none is 0, color-black is 1, and color-orange is 0.25.
Sensors 21 03112 g002
Figure 3. One feature built by the joint data, skeleton data, joint-motion data, or skeleton-motion data, with the size of 600∗18.
Figure 3. One feature built by the joint data, skeleton data, joint-motion data, or skeleton-motion data, with the size of 600∗18.
Sensors 21 03112 g003
Figure 4. The structure of the 10-layer linear-map CNN.
Figure 4. The structure of the 10-layer linear-map CNN.
Sensors 21 03112 g004
Figure 5. (a) the posture of reading, (b) the posture of writing, (c) the posture of making a phone call, (d) the posture of using the laptop.
Figure 5. (a) the posture of reading, (b) the posture of writing, (c) the posture of making a phone call, (d) the posture of using the laptop.
Sensors 21 03112 g005aSensors 21 03112 g005b
Table 1. The indexes of the joints and relations between every two adjacent joints at the 17 edges.
Table 1. The indexes of the joints and relations between every two adjacent joints at the 17 edges.
Spatial Variables
Joint DataSkeleton Data
v i n = x i n , y i n s i , j n = x i n x j n , y i n y j n
Motion Variables
Joint-Motion DataSkeleton-Motion Data
m v i n = v i n + 1 v i n m s i = s i , i + 1 n + 1 s i , i + 1 n
Table 2. The indexes and positions of the joints and edges for the skeletal data.
Table 2. The indexes and positions of the joints and edges for the skeletal data.
IndexJoint PositionIndexJoint PositionIndexe of Joints and Edges
0Nose9Right knee Sensors 21 03112 i001
1Neck10Right ankle
2Right shoulder11Left femur
3Right elbow12Left knee
4Right wrist13Left ankle
5Left shoulder14Right eye
6Left elbow15Left eye
7Left wrist16Right ear
8Right femur17Left ear
Table 3. The channel contents of the input features.
Table 3. The channel contents of the input features.
Fspatial (1200,18,3)Fspatial_joint (600,18,3)(Fv, Fsin, Fsout)
Fspatial_skeleton (600,18,3)(Fs, Fvin, Fvout)
Fmotion (1200,18,3)Fmotion_joint (600,18,3)(Fmv, Fmsin, Fmsout)
Fmotion_skeleton (600,18,3)(Fms, Fmvin, Fmvout)
Table 4. The parameters of linear-map CNN in each layer.
Table 4. The parameters of linear-map CNN in each layer.
LinearBNAFConv2dBNAF
L1In feature (600,18,3),
Out features (600,18,64)
64ReLu(600,18,64), kernel = (9,1),
stride = (1,1), padding = (4,0)
64ReLu
L2In feature (600,18,192),
Out features (600,18,64)
64ReLu(600,18,64), kernel = (9,1),
stride = (1,1), padding = (4,0)
64ReLu
L3In feature (600,18,192),
Out features (600,18,64)
64ReLu(600,18,64), kernel = (9,1),
stride = (1,1), padding = (4,0)
64ReLu
L4In feature (600,18,192),
Out features (600,18,64)
64ReLu(600,18,64), kernel = (9,1),
stride = (1,1), padding = (4,0)
64ReLu
L5In feature (600,18,192),
Out features (600,18,128)
128ReLu(600,18,128), kernel = (9,1),
stride = (2,1), padding = (4,0)
128ReLu
L6In feature (300,18,384),
Out features (300,18,128)
128ReLu(300,18,128), kernel = (9,1),
stride = (1,1), padding = (4,0)
128ReLu
L7In feature (300,18,384),
Out features (300,18,128)
128ReLu(300,18,128), kernel = (9,1),
stride = (1,1), padding = (4,0)
128ReLu
L8In feature (300,18,384),
Out features (300,18,256)
256ReLu(300,18,256), kernel = (9,1),
stride = (2,1), padding = (4,0)
256ReLu
L9In feature (150,18,768),
Out features (150,18,256)
256ReLu(150,18,256), kernel = (9,1),
stride = (1,1), padding = (4,0)
256ReLu
L10In feature (150,18,768),
Out features (150,18,256)
256ReLu(150,18,256), kernel = (9,1),
stride = (1,1), padding = (4,0)
256ReLu
FlattenInput 512
Output 49
Softmax
Table 5. The results without the DAG transfer.
Table 5. The results without the DAG transfer.
Feature
(Epochs)
Precision
(100%)
Recall
(100%)
Specificity
(100%)
F1-Score
(100%)
Accuracy
(100%)
Actions
(Recall < 70%)
Cross-subjectFspatial
(70)
84.2%82.8%99.7%82.8%99.3%0, 4, 9, 10, 11, 16, 28, 29, 43, 48
Fmotion
(65)
80.8%79.7%99.6%79.6%99.2%3, 10, 11, 16, 24, 28, 29, 31, 36, 43, 45
Cross-viewFspatial
(90)
86.4%82.0%99.7%81.6%99.3%4, 9, 11, 16, 18, 28, 43, 44
Fmotion
(20)
82.9%79.7%99.7%79.7%99.2%4, 9, 10, 11, 28, 29, 31, 32
Table 6. The results with the DAG transfer.
Table 6. The results with the DAG transfer.
Feature
(Epochs)
Precision
(100%)
Recall
(100%)
Specificity
(100%)
F1-Score
(100%)
Accuracy
(100%)
Actions
(Recall < 70%)
Cross-subjectFspatial
(65)
86.9%86.0%99.9%86.2%99.4%10, 11, 28
Fmotion
(35)
80.6%78.7%99.7%79.1%99.1%2, 10, 11, 16, 28, 29, 31, 43, 44, 45
Cross-viewFspatial
(65)
94.3%94.3%99.9%94.2%99.9%10, 11, 28, 29
Fmotion
(65)
90.9%90.2%99.9%90.4%99.7%10, 11, 28, 29
Table 7. The results of class score fusion with and without DAG transfer.
Table 7. The results of class score fusion with and without DAG transfer.
Precision
(100%)
Recall
(100%)
Specificity
(100%)
F1-Score
(100%)
Accuracy
(100%)
Actions
(Recall < 70%)
Cross-subjectNon
GAD
85.7%85.0%99.8%85.0%99.4%10, 11, 16, 28, 29, 31
GAD86.9%86.1%99.9%86.3%99.5%10, 11, 28, 29
Cross-viewNon
GAD
89.8%87.3%99.8%87.3%99.6%5, 10, 12, 29
GAD94.8%94.7%99.9%94.7%99.9%
Table 8. The training and testing time with and without DAG transfer.
Table 8. The training and testing time with and without DAG transfer.
Without DAG
Date SourceFeatureEpochTime: Day: Hr.: Min.:Sec.Fps
TrainCross-subjectSpatial1201:12:41:0030
Motion921:04::04:0030
TrainCross-viewSpatial1201:11:39:0030
Motion921:02:55:0030
TestCross-subjectSpatial 0:00:02:01125
Motion 0:00:01:57125
TestCross-viewSpatial 0:00:02:18125
Motion 0:00:02:19125
With DAG
TrainCross-subjectSpatial1201:13:27:0030
Motion921:04:29:0030
Cross-viewSpatial1201:11:05:0030
Motion921:03:30:0030
TestCross-subjectSpatial 0:00:02:03125
Motion 0:00:02:03125
Cross-viewSpatial 0:00:02:21125
Motion 0:00:02:19125
Table 9. These studies all used cross-subject and cross-view sources in the NTU RGB+D database to recognize the actions, which also used the three-dimensional characteristics of each posture as the input features [19,25,26,27,28,29,30,31]. Our method had the best recall rates in the cross-subject and cross-view sources at 86.1% and 94.7%.
Table 9. These studies all used cross-subject and cross-view sources in the NTU RGB+D database to recognize the actions, which also used the three-dimensional characteristics of each posture as the input features [19,25,26,27,28,29,30,31]. Our method had the best recall rates in the cross-subject and cross-view sources at 86.1% and 94.7%.
MethodsRecall Rate (%)
Cross-SubjectCross-View
Lie Group [25]50.152.8
HBRNN [19]59.164.0
Deep RNN [26]59.2964.09
Deep LSTM [26]60.767.3
Part-aware LSTM [26]62.970.3
ST-LSTM + Trust Gate [27]69.277.7
Two-stream RNN [28]71.379.5
Clips + CNN + MTLN [29]79.684.8
ST-GCN [30]81.588.3
SR-TSL [31]84.892.4
Proposed DAG + linear-map CNN86.194.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tan, T.-H.; Hus, J.-H.; Liu, S.-H.; Huang, Y.-F.; Gochoo, M. Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network. Sensors 2021, 21, 3112. https://0-doi-org.brum.beds.ac.uk/10.3390/s21093112

AMA Style

Tan T-H, Hus J-H, Liu S-H, Huang Y-F, Gochoo M. Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network. Sensors. 2021; 21(9):3112. https://0-doi-org.brum.beds.ac.uk/10.3390/s21093112

Chicago/Turabian Style

Tan, Tan-Hsu, Jin-Hao Hus, Shing-Hong Liu, Yung-Fa Huang, and Munkhjargal Gochoo. 2021. "Using Direct Acyclic Graphs to Enhance Skeleton-Based Action Recognition with a Linear-Map Convolution Neural Network" Sensors 21, no. 9: 3112. https://0-doi-org.brum.beds.ac.uk/10.3390/s21093112

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop