Next Article in Journal
Challenges of Developing a Mobile Game for Children with Down Syndrome to Test Gestural Interface
Previous Article in Journal
Playful Learning with a Location-Based Digital Card Environment: A Promising Tool for Informal, Non-Formal, and Formal Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Task Framework for Action Prediction

School of Computer Science, Shenyang Aerospace University, Shenyang 110136, China
*
Authors to whom correspondence should be addressed.
Submission received: 23 January 2020 / Revised: 27 February 2020 / Accepted: 6 March 2020 / Published: 16 March 2020

Abstract

:
Predicting the categories of actions in partially observed videos is a challenging task in the computer vision field. The temporal progress of an ongoing action is of great importance for action prediction, since actions can present different characteristics at different temporal stages. To this end, we propose a novel multi-task deep forest framework, which treats temporal progress analysis as a relevant task to action prediction and takes advantage of observation ratio labels of incomplete videos during training. The proposed multi-task deep forest is a cascade structure of random forests and multi-task random forests. Unlike the traditional single-task random forests, multi-task random forests are built upon incomplete training videos annotated with action labels as well as temporal progress labels. Meanwhile, incorporating both random forests and multi-task random forests can increase the diversity of classifiers and improve the discriminative power of the multi-task deep forest. Experiments on the UT-Interaction and the BIT-Interaction datasets demonstrate the effectiveness of the proposed multi-task deep forest.

1. Introduction

During the past decade, activity prediction has been attracting increasing interest [1,2,3,4], motivated by many potential applications in video surveillance, auto driving, and elderly care. Action prediction takes partially observed videos as input and makes a decision before an action is fully executed. Similar to the human cognitive processes, action prediction methods establish the relationship between the partial observations and action categories by taking advantage of the prior knowledge. For example, it is difficult for a man who has never watched a fencing competition to infer the activity of a fencing athlete. Nevertheless, after watching several competitions, he can tell that the athlete is about to do fencing when the athlete is unsheathing the sword.
Action prediction is an important task related to action recognition [5,6,7] that classifies an action into a certain category after capturing the complete observations. It is evident that predicting actions with incomplete videos is more challenging compared with recognizing actions with full videos. Particularly, an ongoing action can present different motion patterns with sequentially arrived observations, thus the temporal progress of ongoing actions is of great importance for action prediction. However, temporal progress of actions in an input incomplete video is not available during testing.
To overcome the above problem, this paper treats action prediction and temporal progress analysis as two relevant tasks and incorporates them in a multi-task learning framework. Although temporal progress of actions in testing videos cannot be directly observed, we can benefit from the observation ratios of incomplete training videos. The observation ratio represents the proportion of frames that have been observed currently, and can be calculated through dividing the number of frames in an incomplete video by the total number of frames in the corresponding complete videos. It indicates the temporal progress of a training video and can offer important clues for action prediction.
We present a novel Multi-Task Deep Forest (MTDF) framework inspired by the success of gcForest [8]. As shown in Figure 1, the proposed MTDF is a cascade structure of single-task Random Forests (RF) [9] and Multi-Task Random Forests (MTRF). Each level of the cascade exploits two RFs and two MTRFs to learn mid-level features of an input incomplete video. The mid-level features and the original video feature are concatenated to obtain an augmented representation, which is the input of the next level. The final decision of the action category is made upon the mid-level features of the last level. In the proposed MTDF, two kinds of forests (i.e., RF and MTRF) are employed to enhance the diversity so that the ensemble of multiple forests can learn discriminative mid-level features better. Unlike the traditional single-task RF, our MTRF consists of multiple multi-task decision trees that are built upon samples annotated with both action labels and observation ratio labels indicating the temporal progress of actions. A multi-task decision tree jointly deals with the action prediction task and the temporal progress recognition task. Specifically, in the construction of a multi-task decision tree, one type of labels is randomly selected for node split.
The structure of the paper is as follows. A brief summary of the relevant literature is presented in Section 2. Section 3 gives a fine description of the proposed multi-task deep forest framework and explains how to construct a multi-task random forest in detail. Experimental study of the proposed method is shown in Section 4, and Section 5 summarizes the conclusions and future work.

2. Related Work

In recent years, human action prediction has become a hot topic in computer vision, and the main task is to recognize the action category before it is fully executed. Ryoo [1] proposed the definition of action prediction in 2011. He presented integral bag-of-words (IBOW) and dynamic bag-of-words (DBOW) to deal with the action prediction problem. IBOW analyzes the status of ongoing actions from video streams, while DBOW considers the sequential structures of human action. Cao et al. [2] formulated action prediction task as a posterior-maximization problem by using spatiotemporal features of segments. They applied sparse coding on calculating the likelihood of each temporal stage. Lan et al. [3] proposed a hierarchical representation that describes human movements at different levels and developed a max-margin framework for prediction. Kong et al. [4] proposed a multi-scale discriminative model incorporating local and global temporal dynamics of actions. Wang et al. [10] achieved action prediction by matching a test incomplete video with training complete action videos. Considering the characteristics of action prediction, they improved generalized time warping [11] and presented temporally-weighted generalized time warping method that aims to align an incomplete action sequence at an early stage.
Meanwhile, deep learning methods are widely used to tackle complicated learning tasks in recent years. Ma et al. [12] employed LSTM for action prediction and added ranking losses in LSTM training procedure. Kong et al. [13] proposed deep sequential context networks, which learns from fully-executed actions to supplement the missing effective features of partial videos. Singh et al. [14] presented an integrated detection network to process optical flow images for early action prediction and localization. Wang et al. [15] and Kong et al. [16] predicted actions by using different types of adversarial networks, the former put both partial videos and complete videos into the Generative Adversarial Network (GAN) to get the simulative enhanced features, and the latter used a feature generator to generate representative and discriminative features. Based on 3D skeleton sequences, Liu et al. [17] proposed a scale selection network composed of a dilated convolutional network and an activation sharing scheme. Schydlo et al. [18] applied an encoder–decoder recurrent neural network for predicting multiple and variable-length sequences, and introduced simultaneous prediction for this problem.
Recently, Zhou et al. [8] proposed the multi-Grained Cascade Forest (gcForest), which utilized two random forests [9] and two complete random forests [19] in each level of cascade. In addition, the gcForest has much fewer hyper-parameters than classical deep neural networks. The model complexity benefits from the performance evaluation mechanism, which means it can perform well even when the input data are on a small-scale. Inspired by the gcForest, we design a novel multi-task deep forest framework for action prediction. Due to the observation that actions present different characteristics at different stages, we believe that the observation ratio of videos can provide important cues for action prediction. Accordingly, our multi-task deep forest framework treats action prediction and temporal progress analysis as two relevant tasks, and is trained on incomplete videos annotated with both the action label and the observation ratio label.

3. Materials and Methods

Constructions of Multi-Task Deep Forest and Multi-task Random Forest are detailed in this section. The random forest is introduced in Section 3.2. Moreover, details of generating a class vector is presented in Section 3.3.2.

3.1. Multi-Task Deep Forest

In this section, we introduce the proposed Multi-Task Deep Forest (MTDF), which is a cascade structure extended from gcForest. The general structure of MTDF is illustrated in Figure 1. As shown in Figure 1, MTDF is composed of multiple levels, each of which is an ensemble of four forests, i.e., two multi-task random forests and two random forests [9].
Concretely, we extract dense trajectories from an incomplete video and utilize bag-of-words model to build the original video feature vector, which is then input into the first level of the MTDF. At each level, one forest generates one class vector indicating the probability that this video belongs to each category. Then, as shown in Figure 2, four class vectors constructed by forests of a level are concatenated with the original feature vector as a new feature, which is input into the next level of the MTDF. Finally, the four generated class vectors of the last level are averaged into one class vector, and the action category corresponding to the largest element is treated as the predicted result.
In order to reduce the risk of over-fitting, k-fold cross validation is employed to build each forest. Particularly, we divide the training set into k groups and utilize k 1 groups for constructing a forest in each turn. An average of k class vectors is adopted as the ultimate class vector. Different from traditional random forest, a multi-task random forest is trained on incomplete videos annotated with both action category label and observation ratio label. As shown in Figure 1, during training, features annotated with both action category label and observation ratio label are sent to multi-task random forests in the first level, while the same features annotated with action category label are input into traditional random forests. In the following section, we will describe the construction of multi-task random forest in our framework in detail.

3.2. Random Forest

In our multi-task deep forest framework, two random forests [9] are utilized in each level of cascade. A random forest is an ensemble of decision trees and its principle is to let the trees vote for the most popular class. Suppose that a training set D = { ( x m , y m ) } m = 1 : M containing M samples is input into a random forest. Each sample ( x m , y m ) includes a feature vector x m and an action category label y m . In the construction of a decision tree, randomness is introduced by randomly generating several splitting schemes for a node. When a group of samples arrives at a node, the information gain of each split scheme will be calculated by the action labels. Then, the scheme with the largest information gain is selected as a best scheme and is utilized for splitting the node into a left child node and a right child node. The node will continue to split until the samples that reach the node are of the same class or the node reaches its maximum depth. For a test sample input into a random forest, the sample will traverse from the root node of each tree until it reaches a leaf node. The proportion of samples of each action category in the leaf node will be defined as a class vector υ , and the class vector υ will be regarded as the output of the decision tree. The average of the class vectors generated by all decision trees in the random forest is the final output of the random forest.

3.3. Multi-Task Random Forest

A multi-task random forest is an ensemble classifier composed of multiple multi-task decision trees. Suppose that the training set D = { ( x m , y m A , y m O ) } m = 1 : M contains M segmented videos with known observation ratios. Each sample in D contains a feature vector x m R F × 1 , an action category label y m A { 1 , 2 , , K } , and an observation ratio label y m O { 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 , 0.8 , 0.9 , 1 } , where F is the number of features in a feature vector and K is the number of action categories. In order to reduce the association among different multi-task decision trees and enhance the overall performance of multi-task random forest, we employ the Bootstrap [20] method to generate a subset D from the original training set D for each multi-task decision tree.

3.3.1. Construction of a Multi-task Decision Tree

In the construction of a multi-task decision tree, we keep a node sequence S and a pending node sequence Q. The element in the node sequence S is ( n o d e ,l,f, c 1 , c 2 , α , β , υ ), where n o d e is the node ID, l represents the depth, f indicates the parent node, c 1 c r are the left child node and the right child node, α represents the split feature, β is the split value, and υ is the action category vector. In particular, if the node is an intermediate node, υ is null, and, if the node is a leaf node, c 1 , c 2 , α , β are null. The element in the pending node sequence Q is ( n o d e ,l,f, D * ), where n o d e is the node ID, l represents the depth, f indicates its parent node, and D * is a dataset including all samples arriving at this node.
At first, a root node (1,1,0, D ) is crested and is put into Q, then a variable c o u n t e r is initialized as 1. Then, we take the first node ( n o d e ,l,f, D * ) in Q, and remove it from Q. If l is equal to the preset maximum depth L m a x of multi-task decision tree or samples in D * belong to the same categories on both action category label and observation ratio label, the node is regarded as a lead node; otherwise, the node is an intermediate node.
For an intermediate node ( n o d e ,l,f, D * ) taken from sequence Q, procedures of node split are illustrated in Figure 3. A group of split schemes { ( α g , β g ) } g = 1 : G are generated, where α g is a randomly selected feature candidate and β g represents the split value generated according to values of samples in D * on feature α g . Each scheme ( α g , β g ) divides the dataset D * into two subsets D g L and D g R . The former contains samples whose value on feature α g is smaller than β g , and the latter is composed of samples whose value on feature α g is bigger than β g . Then, the best scheme ( α * , β * ) is chosen from { ( α g , β g ) } g = 1 : G for node split based on information gain maximization. Since each sample in D * has two labels, namely action category and observation ratio, we need to choose one type of label for calculating information gain.
Particularly, if the observation ratio labels of all the samples in D * are consistent, the information gain is calculated according to the action category label. Assuming that the proportion of samples in D * belonging to the k-th action is p k , the information entropy of D * can be defined as
EntAction ( D * ) = k = 1 K p k log 2 p k ,
where K is the number of action categories. For split scheme ( α g , β g ) , the information entropy EntAction ( D g L ) of D g L and the information entropy EntAction ( D g R ) of D g R can be calculated in terms of Equation (1) as well. Accordingly, the information gain of utilizing split scheme ( α g , β g ) is computed as
GainAction ( D * , α g , β g ) = EntAction ( D * ) | D g L | | D * | EntAction ( D g L ) | D g R | | D * | EntAction ( D g R ) ,
where | D | indicates the number of samples in dataset D. Split scheme ( α * , β * ) with the largest information gain is selected for node split according to Equation (3), and it divides the dataset D * into D * L and D * R .
( α * , β * ) = arg max ( α g , β g ) GainAction ( D * , α g , β g ) .
Similarly, if all the samples in D * belong to the same action, the information gain is calculated according to the observation ratio label. Supposing that the proportion of the samples in D * with the j-th observation ratio label is q j , the information entropy is defined as
EntRatio ( D * ) = j = 1 10 q j log 2 q j ,
where the number of discrete observation ratios is 10. For split scheme ( α g , β g ) , the information entropy EntRatio ( D g L ) of D g L and the information entropy EntRatio ( D g R ) of D g R can be calculated in terms of Equation (4). Next, the information gain split scheme ( α g , β g ) is calculated as
GainRatio ( D * , α g , β g ) = EntRatio ( D * ) | D g L | | D * | EntRatio ( D g L ) | D g R | | D * | EntRatio ( D g R ) .
The split scheme ( α * , β * ) with the largest information gain is regarded as the best split scheme and is adopted for node split:
( α * , β * ) = arg max ( α g , β g ) GainRatio ( D * , α g , β g ) .
More frequently, samples in D * belong to different categories in terms of action label or observation ratio label. In this case, one type of label is randomly selected for calculation of information gain. Specifically, a predefined parameter γ [0,1] and a random variable μ [ 0 , 1 ] are employed to control the selection. If μ < γ , an action label is chosen to calculate the information gain GainAction ( D * , α g , β g ) for each scheme ( α g , β g ) , and the best scheme is determined according to Equation (3). Otherwise, an observation ratio label is utilized to compute the information gain GainRatio ( D * , α g , β g ) , and Equation (6) is employed to choose the split scheme.
After dividing D * into D * L and D * R , the ID of the left child node is c o u n t e r + 1 and the ID of the right child node is c o u n t e r + 2 . Then, the current intermediate node is expressed as ( n o d e ,l,f, c o u n t e r + 1 , c o u n t e r + 2 , α * , β * , n u l l ) and is further put into the node sequence S. Meanwhile, the left child node ( c o u n t e r + 1 , l + 1 , n o d e , D * L ) and the right child node ( c o u n t e r + 2 , l + 1 , n o d e , D * R ) are put into the pending node sequence Q. Then, the node counter is updated as c o u n t e r = c o u n t e r + 2 .
For a leaf node ( n o d e ,l,f, D * ) taken from sequence Q, we need to calculate the action category vector υ indicating the possibility that samples arriving at the leaf node belonging to each action category. To this end, we calculate the proportion p k of samples in D * belonging to the k-th action and create a class vector υ = [ p 1 , p 2 , , p K ] for this leaf node. Then, the leaf node is expressed as ( n o d e ,l,f, n u l l , n u l l , n u l l , n u l l , υ ) and put it into the node sequence S.
Finally, if the pending node sequence Q is not empty, we will take the first node in Q again, or the construction of this multi-task decision tree is completed. The main steps of how to construct a multi-task decision tree are summarized in Algorithm 1.

3.3.2. Class Vector of a Multi-Task Random Forest

For a test sample input into a multi-task random forest, the sample will traverse from the root node of each tree until it reaches a leaf node. The class vector υ of the arrived leaf node is regarded as the output of the multi-task decision tree. An average of class vectors output from all the multi-task decision trees forms the final class vector of the multi-task random forest.
Algorithm 1 Construction of a multi-task decision tree
Input: a training set D = { ( x m , y m A , y m O ) } m = 1 : M       // x m i s f e a t u r e v e c t o r , y m A a n d y m O a r e l a b e l s
Output: a node sequence S
 1: create a node sequence S storing settled nodes and a pending node sequence Q containing unsettled nodes;
 2: create the root node as (1,1,0,D’) and put it into Q;
 3:  c o u n t e r 1       // the total number of generated nodes
 4: while Q is not N U L L do
 5:  take out the first node ( n o d e ,l,f, D * ) of Q;
 6:  if l is equal to the preset maximum depth or samples in D * belong to the same categories on both action category label and observation ratio label then
 7:   // the current node is a leaf node
 8:   put the current node ( n o d e , l , f , n u l l , n u l l , n u l l , n u l l , υ ) into S;
 9:  else
 10:   // the current node is an intermediate node need to be split
 11:   randomly select G split schemes { ( α g , β g ) } g = 1 : G , and each scheme ( α g , β g ) splits samples in D * into two subsets D g L and D g R ;
 12:   generate a random value μ [ 0 , 1 ] and compare it with a predefined parameter γ ;
 13:   if μ < γ or samples in D * belong to the same categories on observation ratio label then
 14:    //evaluate each split scheme in terms of action labels of samples in D *
 15:    for g is 1 to G do
 16:     calculate the information gain GainAction ( D * , α g , β g ) according to Equation 2
 17:    end for
 18:    choose the best split scheme ( α * , β * ) according to Equation 3, and the dataset D * is divided into subsets D * L and D * R ;
 19:   else
 20:    //evaluate each split scheme in terms of observation ratio labels of samples in D *
 21:    for g is 1 to G do
 22:     calculate the information gain GainRatio ( D * , α g , β g ) according to Equation 5
 23:    end for
 24:    choose the best split scheme ( α * , β * ) according to Equation 6, and the dataset D * is divided into subsets D * L and D * R ;
 25:   end if
 26:    S ( n o d e , l , f , c o u n t e r + 1 , c o u n t e r + 2 , α * , β * , n u l l )       //put the current node into S
 27:    Q ( c o u n t e r + 1 , l + 1 , n o d e , D * L )       //put the left child node into Q
 28:    Q ( c o u n t e r + 2 , l + 1 , n o d e , D * R )       //put the right child node into Q
 29:  end if
 30:   c o u n t e r c o u n t e r + 2
 31: end while
 32: return S

4. Experiments

We evaluate our method on two datasets, and compare it with other methods. Section 4.1 shows the details of these two datasets and the experimental setup. The results of our MTDF and comparisons are shown in Section 4.2. Additionally, we analyze the influence of several key parameters in Section 4.3.

4.1. Datasets and Experimental Setup

The proposed MTDF approach is evaluated on two datasets: the UT-Interaction dataset including Set 1 (UT-I #1) and Set 2 (UT-I #2) [21], and the BIT-Interaction dataset [22]. UT-I #1 contains six types of human interactions, i.e., shaking hands, hugging, kicking, pointing, punching, and pushing, which were collected on a parking area from real surveillance videos. Each human interaction has 10 videos performed by 10 groups of participants. UT-I #2 shares the same classes of interactions with UT-I #1, but the background movements and camera jitters are larger over UT-I #1. On both UT-I #1 and UT-I #2 datasets, we utilize a 10-fold leave-one-out evaluation strategy.
The BIT-Interaction dataset consists of eight types of human interactions, each of which contains 50 videos. Complex background in this dataset poses great challenges for action prediction. In order to evaluate the proposed method, we take the first 34 videos in each class for training and utilize the last 16 videos in each class for testing.
The dense trajectory features [6] are extracted from videos in both UT dataset and BIT dataset, and the bag-of-words model is employed to encode these dense trajectory features, following the experiment settings in [1,4]. Crucially, in the following experiments, we set γ = 0.5 for all multi-task decision trees, which means there is equal probability of using action category label or observation ratio label to calculate the information gain for splitting.

4.2. Experimental Results

For more visible comparisons, we divide the experimental results into three parts. Comparison between MTDF and gcForest is shown in Section 4.2.1. Meanwhile, comparisons with other methods on UT-dataset and BIT-dataset are presented in Section 4.2.2 and Section 4.2.3, respectively. Due to the randomness of random forest and multi-task random forest, our method is run for 5 times and the average results are reported in this section.

4.2.1. Comparison between MTDF and gcForest

In order to evaluate the effectiveness of the proposed multi-task deep forest, we compare it with the original gcForest [8] on the UT-I #1, UT-I #2, and BIT datasets. The comparison experiments are summarized in Table 1.
On the UT-I #1 dataset, the proposed MTDF has achieved better or equal performance over the original gcForest at all observation ratios. Particularly, the MTDF achieves an improvement of 3.33% when the observation ratio is 0.4. On the UT-I #2 dataset, the proposed MTDF performs better than gcForest in most cases, and makes an improvement of about 4% at observation ratio 0.2. On the BIT dataset, our MTDF has achieved 0.15%-2.19% improvements over gcForest when more than 10% of videos are observed. In summary, it is effective to introduce an observation ratio label as an extra task during training, and our MTDF has shown improvement towards gcForest for action prediction. In particular, utilizing the observation ratio labels in the node splitting enhances the randomness of decision trees, which improves the discriminative power of MTDF.

4.2.2. Results on the UT-I #1 and UT-I #2 Datasets

Our MTDF is compared with DBoW [1], MSSC [2], MTSSVM [4], and MMAPM [23] on the UT-I #1 and UT-I #2 datasets. In addition, KNN-Dynamic and the baseline method adopted in [2] are also employed in comparison. Our experiments utilize the same evaluation strategy as that in [2].
The comparison results on UT-I #1 dataset are shown in Figure 4a. Compared with the other methods, MTDF achieves better performance at r {0.4, 0.5, 0.6, 0.7}. When only the first 40% frames of the testing videos are observed, recognition accuracy of our method is 81.33%. Then, the recognition accuracy of our method reaches 93% when the testing videos have been observed 70% frames. These results are much better than others at the same observation ratios. Our method does not perform well at observation ratio 0.1 because the video is so short that we cannot extract enough effective trajectories.
Figure 4b summarizes the comparison results on UT-I #2 dataset. When more than 40% frames of the testing videos are observed, MTDF is able to achieve higher recognition accuracy than all the compared methods. When only the first 50% frames of the testing videos are observed, our MTDF achieves 83% recognition accuracy, which is much higher than other comparison approaches at the same observation ratio and even higher than MSSC, DBoW, and KNN-Dynamic with full observations.

4.2.3. Results on the BIT Dataset

We also conduct experiments on the BIT-interaction dataset and compare the proposed MTDF with Dynamic BoW [1], MSSC [2], MTSSVM [4], MMAPM [23], DeepSCN [13], and AAPNet [16]. The comparison results on BIT dataset are shown in Figure 5.
Overall, the proposed MTDF outperforms MMAPM [23], MTSSVM [4], MSSC [2], and Dynamic BoW [1] at all observation ratios, and achieves comparable performance with DeepSCN [13] and AAPNet [16]. Different from UT-I #1 and UT-I #2, effective trajectory information can be extracted at the first 10% frames of the videos on the BIT-interaction dataset. Therefore, the proposed MTDF achieves 35.47% prediction accuracy at observation ratio 0.1. For observation ratios 0.2, 0.3,0.4, and 0.5, the prediction performance of MTDF is much better than the other comparison approaches. Particularly, MTDF achieves 63.13% prediction accuracy at observation ratio 0.2, which is 17.96% over AAPNet [16] and 18.3% over DeepSCN [13]. The prediction performance of MTDF does not increase significantly when the observation ratio is more than 0.6. One possible reason is that the key information for action prediction has been captured using the first 60% frames.

4.3. Effects of Parameters

In this section, we talk about several key parameters in the construction of MTDF. To this end, we conduct experiments on the UT-I #1 dataset and analyze the influence of parameters.
In considering the computational complexity of forest construction, we set the number of trees in a forest and limit the maximum depth of each decision tree. Figure 6a demonstrates the accuracy trend with different numbers of decision trees at observation ratios 0.6, 0.7, and 0.8. When the number of trees is less than 100, building more trees makes the accuracy improve continuously. Then, the recognition accuracy tends to be stable when there are more than 100 trees in each forest. In general, the prediction accuracy increases as the number of trees increases, and convergence is at a certain value. Figure 6b shows how the maximum depth of each decision tree affects the action prediction performance. Generally, with the increase of depth, the curves of different observation ratios rise first and then decline slightly and stabilize gradually. One reasonable explanation is that very deep decision trees may overfit the training data and reduce the action prediction performance on testing data.

5. Conclusions

In this paper, we have presented a novel Multi-Task Deep Forest framework for predicting actions from partially observed videos. In the construction of Multi-Task Deep Forest, we have utilized the observation ratio to evaluate the temporal progress of training incomplete videos and treated video temporal progress analysis as another task relevant to action prediction tasks. Experimental results on the UT-Interaction and the BIT-Interaction datasets show the superior performance of Multi-Task Deep Forest over the original single-task gcForest, which demonstrates that incorporating video observation labels during training can improve the action prediction performance. Moreover, the Multi-Task Deep Forest achieved better or comparable performance over the existing action prediction methods on both datasets, which has verified the effectiveness and reasonability of the Multi-Task Deep Forest.

Author Contributions

Investigation, T.Y.; Supervision, C.L., Z.Y., and X.S.; Writing—original draft, T.Y.; Writing—review and editing, C.L. and Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Foundation of Liaoning Educational Committee under Grant No. JYT19029, and in part by the Talent Scientific Start-up Project of Shenyang Aerospace University under Grant No. 18YB39, and in part by the National Natural Science Foundation of China (NSFC) under Grant No. 61602320.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MTDFMulti-Task Deep Forest
MTRFMulti-Task Random Forests
RFRandom Forest

References

  1. Ryoo, M.S. Human activity prediction: Early recognition of ongoing activities from streaming videos. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 7 November 2011; pp. 1036–1043. [Google Scholar]
  2. Yu, C.; Barrett, D.; Barbu, A.; Narayanaswamy, S.; Song, W. Recognize Human Activities from Partially Observed Videos. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2658–2665. [Google Scholar]
  3. Lan, T.; Chen, T.C.; Savarese, S. A Hierarchical Representation for Future Action Prediction. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 689–704. [Google Scholar]
  4. Yu, K.; Kit, D.; Yun, F. A Discriminative Model with Multiple Temporal Scales for Action Prediction. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 596–611. [Google Scholar]
  5. Laptev, I. On Space-Time Interest Points. Int. J. Comput. Vis. 2005, 64, 107–123. [Google Scholar] [CrossRef]
  6. Wang, H.; Schmid, C. Action Recognition with Improved Trajectories. In Proceedings of the IEEE International Conference on Computer Vision, Paris, France, 11 September 2014; pp. 3551–3558. [Google Scholar]
  7. Liu, C.; Wu, X.; Jia, Y. A hierarchical video description for complex activity understanding. Int. J. Comput. Vis. 2016, 118, 240–255. [Google Scholar] [CrossRef]
  8. Zhou, Z.H.; Feng, J. Deep Forest: Towards An Alternative to Deep Neural Networks. In Proceedings of the International Joint Conferences on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 3553–3559. [Google Scholar]
  9. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, H.; Yang, W.; Yuan, C.; Ling, H.; Hu, W. Human activity prediction using temporally-weighted generalized time warping. Neurocomputing 2017, 225, 139–147. [Google Scholar] [CrossRef]
  11. Feng, Z.; Torre, F.D.L. Generalized Time Warping for Multi-modal Alignment of Human Motion. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3551–3558. [Google Scholar]
  12. Ma, S.; Sigal, L.; Sclaroff, S. Learning Activity Progression in LSTMs for Activity Detection and Early Detection. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition, Las Vegas, NV, USA, 26 June 2016; pp. 1942–1950. [Google Scholar]
  13. Kong, Y.; Tao, Z.; Yun, F. Deep Sequential Context Networks for Action Prediction. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1473–1481. [Google Scholar]
  14. Singh, G.; Saha, S.; Sapienza, M.; Torr, P.; Cuzzolin, F. Online Real-time Multiple Spatiotemporal Action Localisation and Prediction. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3637–3646. [Google Scholar]
  15. Wang, D.; Yuan, Y.; Wang, Q. Early Action Prediction with Generative Adversarial Networks. IEEE Access 2019, 7, 35795–35804. [Google Scholar] [CrossRef]
  16. Kong, Y.; Tao, Z.; Fu, Y. Adversarial Action Prediction Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2018. 1-1. Available online: https://0-ieeexplore-ieee-org.brum.beds.ac.uk/abstract/document/8543243 (accessed on 10 March 2020).
  17. Liu, J.; Shahroudy, A.; Gang, W.; Duan, L.Y.; Kot, A.C. SSNet: Scale Selection Network for Online 3D Action Prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 12–18 July 2018; pp. 8349–8358. [Google Scholar]
  18. Schydlo, P.; Rakovic, M.; Jamone, L.; Santos-Victor, J. Anticipation in Human-Robot Cooperation: A Recurrent Neural Network Approach for Multiple Action Sequences Prediction. In Proceedings of the IEEE International Conference on Robotics and Automation, Guangzhou, China, 17–19 November 2018; pp. 1–6. [Google Scholar]
  19. Liu, F.T.; Kai, M.T.; Yu, Y.; Zhou, Z.H. Spectrum of Variable-Random Trees. J. Artif. Intell. Res. 2008, 32, 355–384. [Google Scholar] [CrossRef] [Green Version]
  20. Efron, B.; Tibshirani, R.J. An Introduction to the Boot-Strap. Chapman & Hall 1993, 57, 1–436. [Google Scholar]
  21. Ryoo, M.S.; Aggarwal, J.K. UT-Interaction Dataset, ICPR contest on Semantic Description of Human Activities (SDHA). 2010. Available online: http://cvrc.ece.utexas.edu/SDHA2010/Human_Interaction.html (accessed on 10 March 2020).
  22. Kong, Y.; Jia, Y.; Fu, Y. Learning human interaction by interactive phrases. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012. [Google Scholar]
  23. Kong, Y.; Fu, Y. Max-Margin Action Prediction Machine. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 1844–1858. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed multi-task deep forest.
Figure 1. Overview of the proposed multi-task deep forest.
Information 11 00158 g001
Figure 2. Details of how to concatenate feature vectors at the first level. In this figure, we suppose that there are eight action categories.
Figure 2. Details of how to concatenate feature vectors at the first level. In this figure, we suppose that there are eight action categories.
Information 11 00158 g002
Figure 3. Details of node split for an intermediate node in a multi-task decision tree.
Figure 3. Details of node split for an intermediate node in a multi-task decision tree.
Information 11 00158 g003
Figure 4. Prediction results on the UT-Interaction dataset.
Figure 4. Prediction results on the UT-Interaction dataset.
Information 11 00158 g004
Figure 5. Prediction results on the BIT dataset.
Figure 5. Prediction results on the BIT dataset.
Information 11 00158 g005
Figure 6. Recognition accuracy changes as the parameters are adjusted at different observation ratios r on UT-I #1.
Figure 6. Recognition accuracy changes as the parameters are adjusted at different observation ratios r on UT-I #1.
Information 11 00158 g006
Table 1. Prediction results on UT-I #1, UT-I #2, and BIT dataset, using gcForest [8] and our multi-task deep forest
Table 1. Prediction results on UT-I #1, UT-I #2, and BIT dataset, using gcForest [8] and our multi-task deep forest
Observation ratio0.10.20.30.40.50.60.70.80.91.0
UT-I #1gcForest19.33%33.67%60.00%78.00%86.33%91.67%92.00%91.67%91.00%90.67%
MTDF20.00%33.67%60.33%81.33%86.67%92.33%93.00%92.33%91.67%91.67%
UT-I #2gcForest16.67%32.33%51.33%68.00%83.33%90.67%93.67%93.33%92.00%92.00%
MTDF16.67%36.33%52.67%69.67%83.00%91.67%95.33%94.67%91.67%91.67%
BITgcForest36.72%60.94%77.50%83.44%85.63%85.00%84.69%84.38%84.06%84.06%
MTDF35.47%63.13%79.22%84.38%85.78%85.94%86.09%85.63%85.31%85.31%

Share and Cite

MDPI and ACS Style

Yu, T.; Liu, C.; Yan, Z.; Shi, X. A Multi-Task Framework for Action Prediction. Information 2020, 11, 158. https://0-doi-org.brum.beds.ac.uk/10.3390/info11030158

AMA Style

Yu T, Liu C, Yan Z, Shi X. A Multi-Task Framework for Action Prediction. Information. 2020; 11(3):158. https://0-doi-org.brum.beds.ac.uk/10.3390/info11030158

Chicago/Turabian Style

Yu, Tianyu, Cuiwei Liu, Zhuo Yan, and Xiangbin Shi. 2020. "A Multi-Task Framework for Action Prediction" Information 11, no. 3: 158. https://0-doi-org.brum.beds.ac.uk/10.3390/info11030158

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop