Next Article in Journal
Historical Change and Ecological Risk of Potentially Toxic Elements in the Lake Sediments from North Aral Sea, Central Asia
Next Article in Special Issue
Intelligent Control Based on a Neural Network for Aircraft Landing Gear with a Magnetorheological Damper in Different Landing Scenarios
Previous Article in Journal
Enhancing Friction Pendulum Isolation Systems Using Passive and Semi-Active Dampers
Previous Article in Special Issue
Numerical Estimation of Bonding Force of EPDM Grommet Parts with Hollow Shaft Geometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Next Failure Time Prediction of Escalators via Deep Neural Network with Dynamic Time Warping Preprocessing

1
State Key Laboratory for Manufacturing and Systems Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
School of Traffic and Transportation Engineering, Central South University, Changsha 410075, China
*
Author to whom correspondence should be addressed.
Submission received: 15 July 2020 / Revised: 31 July 2020 / Accepted: 12 August 2020 / Published: 13 August 2020

Abstract

:
The escalator is one of the most popular travel methods in public places, and the safe working of the escalator is significant. Accurately predicting the escalator failure time can provide scientific guidance for maintenance to avoid accidents. However, failure data have features of short length, non-uniform sampling, and random interference, which makes the data modeling difficult. Therefore, a strategy that combines data quality enhancement with deep neural networks is proposed for escalator failure time prediction in this paper. First, a comprehensive selection indicator (CSI) that can describe the stationarity and complexity of time series is established to select inherently excellent failure sequences. According to the CSI, failure sequences with high stationarity and low complexity are selected as the referenced sequences to enhance the quality of other failure sequences by using dynamic time warping preprocessing. Secondly, a deep neural network combining the advantages of a convolutional neural network and long short-term memory is built to train and predict quality-enhanced failure sequences. Finally, the failure-recall record of six escalators used for 6 years is analyzed by using the proposed method as a case study, and the results show that the proposed method can reduce the average prediction error of failure time to less than one month.

1. Introduction

The escalator is one of the most popular transportation vehicles in densely populated areas. According to the data, almost 2 billion trips are taken using elevators and escalators every day. Once escalator failure occurs, it may lead to a terrible accident [1]. In China, about 40 persons died and thousands of people were injured in escalator accidents every year. According to data statistics from Guangzhou Metro, about 67% of total injuries are escalator-related injuries [2]. Thus, the timely prediction of escalator failure can help the manager and maintenance personnel develop maintenance plans to effectively prevent accidents.
The escalator is a typical human–machine–environment system with many factors involved in safety. One of the links and one personal mistake may cause a serious accident. As a research field related to people’s transportation, safety issues have always been the focus of attention [3,4,5]. An escalator, as a complex system, can cause serious casualties in the event of malfunction. Therefore, the safe operation of escalators deserves in-depth study. Failure and risk prediction are critical to ensuring the safe operation of escalators. When the risk level exceeds a certain safety range, an accident will occur [6]. Thus, it is necessary to predict the failure time accurately and timely, and arrange maintenance activities in advance, which can effectively reduce the failure rate and reduce accidents. Failure time series modeling analysis can describe the law of failure occurrence by establishing a corresponding mathematical model of related failure data.
Many scholars have done in-depth research on time series analysis in safety research, such as failure maintenance time based on the Weibull distribution [7], shipyard occupational risk assessment based on multivariate regression and genetic algorithm [8], and the Akaike Information Criterion (AIC) based steel mill alarm mechanism [9]. These methods have achieved good effects in the case of large amounts of data. However, for accident-related failure data, it has special features of short samples, non-uniform sampling, and random interference. Traditional data modeling based on statistical laws or deterministic mathematical models requires more labelled data, while data related to safety accidents have typical characteristics as aforementioned, making data modeling problems much more difficult. Neural network technology, as one of the most popular data analysis methods, can build a model containing complex nonlinear relationships without deliberate attention to the mathematical characteristics of the data itself, and has good generalization ability to map the relationship between input and output quickly and effectively, which has been widely used in many fields.
In recent years, with the development of neural network (NN) technology, neural networks have been used in various fields [10,11,12]. The data modeling by NNs does not deliberately focus on the mathematical characteristics of the data itself but learns to construct the complex relationship between input and output with neurons. Thus, neural networks have good generalization capacity. However, the accuracy of neural networks is sensitive to disturbance in data. In addition, some neural networks only pay attention to the relationship between input and output, ignoring the hidden information of the input themselves, which is unreasonable for time series modeling and prediction. Long short-term memory (LSTM), as one of the recurrent neural networks (RNNs), was proposed to solve the problem of time series prediction [13]. LSTM can analyze and process time series with unknown duration delay. LSTM can improve the memory module of traditional RNNs and avoid the problem where effective historical information cannot be stored for a long time, because of the constant input data [14]. LSTM has also been utilized in time series prediction [15], remaining useful for life prediction [16] and safety analysis [17,18]. However, LSTM does not learn well the characteristics of the data itself for short time series with too few samples. Therefore, this paper considers the short-sequence feature extraction ability of a convolutional neural network (CNN) to extract high-dimensional features, which can help the LSTM unit to understand the data better. In addition, the combination of CNNs and LSTMs in a unified framework has already offered state-of-the-art results in the speech recognition domain [19], health care [20], and power load prediction [21]. In addition, the quality of the data itself makes a big difference to the accuracy of neural network modeling. Meanwhile, considering the short length and random interference in the failure data, it is meaningful to enhance the data quality to improve the prediction accuracy of the neural network before data modeling. Dynamic time warping (DTW) [22], with characteristics of oversampling and similarity matching, is much more suitable for data preprocessing before data modeling, especially for short and interfering failure data. DTW has also been used in many fields [23,24]. For this reason, considering the above-mentioned issues, a new failure time series prediction method for the escalator based on a convolutional long short-term memory neural network combined with dynamic time warping preprocessing (DCLNN) was proposed in this paper.
The remainder of this paper is organized as follows: In Section 2, a time series with high stationarity and low complexity is elected by a comprehensive selection indicator, the main principles of the DCLNN are given, and the diagram of the proposed method for failure time prediction is described. In Section 3, case studies of six escalators are applied with the proposed method and the results are given. Finally, a summary and conclusion are drawn in Section 4.

2. The Proposed DCLNN Method for Failure Time Prediction

The complex working environment and numerous components of the escalator make the failure time data have typical nonlinearity and randomness, which increases the difficulty of data modeling of the escalator failure time series. In this section, an escalator failure time prediction model based on a deep neural network under dynamic time warping preprocessing (DCLNN) will be specifically explained. The DCLNN combined the strategy of DTW pre-processing and CNNLSTM post-learning, and some basic principles about the method are given in the following section.

2.1. The DCLNN Strategy for Failure Time Prediction of Escalator

The accuracy of time series prediction is not only related to the sequence model, but the quality of the time series itself also has a great influence on the prediction result. In this paper, a strategy to combine data quality enhancement with deep neural networks is proposed for time series prediction, which is shown in Figure 1. First, the data sequences are divided into good data and bad data according to a selection indicator; secondly, the bad time series data can be transformed to good data with high quality; finally, the built neural network is used to train and predict the time series.
Based on the principles of DTW and CNNLSTM, a time series prediction method combining DTW pre-processing and CNNLSTM post-learning is proposed in this paper for escalator failure time prediction, which is called DCLNN.

2.2. Comprehensive Selection Indicator before Data Modeling

In general, data sequences with high stationarity and low complexity can be modeled accurately by the behavior of mathematical modeling. The failure time has certain statistical rules, but with various reasons for failure, different escalators have different failure times. Those regular escalators whose failure time series are stationary and have lower complexity can be selected as a reference. For the same batch of equipment, other escalators have similar failure occurrence rules. In this paper, two stationary indicators (SIs) and two complexity indicators (CIs) are combined to form a comprehensive selection indicator to select excellent referenced time series. Here are details about four indicators:
(1) Stationary indicator
(a) Standard deviation
S T D = 1 N i = 1 N ( x i μ ) 2
In statistics, the standard deviation is a measure that is used to quantify the amount of variation or dispersion of a set of data values. A low standard deviation indicates that the time series tends to be more stationary.
(b) p-value
The p-value is the probability value in the statistical significance test, that is, the probability that the hypothesis test occurs. It is generally assumed that if a small probability event occurs, the assumption is true. In this paper, the data sequence is used as the hypothesis test for the stationary data, and the P-value can be calculated as the probability of the unit root test to judge whether the failure time series is stationary or not. A low P-value indicates that the time series tends to be stationary.
P _ v a l u e = P ( x   is   a   non - stationary   sequence |   x   is   a   stationary   sequence )
(2) Complexity indicator
(a) Lempel-ziv
The Lempel-ziv indicator is used to characterize the rate at which new patterns appear in a time series. The higher the Lempel-Ziv complexity of the sequence, the greater the degree to which the signal sequence is near-random. The lower the Lempel-ziv indicator, the lower the complexity of the time series.
C L e m p e l z i v = C N ( N ) lim N C N ( N ) C N ( N ) × log k ( N ) N
C N ( N ) is the Lempel-ziv complexity of the time series, and C L e m p e l z i v is the normalized complexity. Details about the Lempel-ziv algorithm can be found in Ref. [25].
(b) Sample entropy
The Sample Entropy (SampEn) measures the complexity of the time series by measuring the probability of generating a new pattern in the signal. The greater the probability of the new pattern, the greater the complexity of the sequence. The lower the sample entropy indicator, the lower the complexity of the time series.
S a m p E n = ln { A k ( r ) / B m ( r ) }
where A k ( r ) and B m ( r ) are the number of subsets of the original time series that satisfies the requirement within the similar distance r (usually m = 2 ;   k = m + 1 ;   r = 0.1 0.25 ).
(3) Comprehensive selection indicator
In order to describe the inherent characteristics of a data sequence, a comprehensive selection indication (CSI) that can describe both stationarity and complexity is constructed in this part to select an inherently excellent data sequence before data modeling. In addition, CSI can be given by the following equation:
C S I = S T D × P _ v a l u e × C L e m p e l z i v × S a m p E n
CSI can be used to compare the regularity itself between data sequences; the smaller the CSI, the better the data sequence itself and the more accurate the data modeling will be.

2.3. Dynamic Time Warping

The failure time series of escalators are different, and the raw data always have defects (such as disturbances), which are less effective when used directly in the prediction of neural networks. Considering the time delay and local similarity of the failure time series of different escalators, dynamic time warping (DTW) is introduced to pre-process before data modeling [22], and the over-sampling characteristics of data under the space defined by DTW can effectively reduce the impact of the mutated data on the accuracy of the model while increasing the data capacity and finally improve the prediction accuracy of the NN model. An intuitive example is given in Figure 2. The principles of DTW can be further explained as follows:
Given two time series X and Y , whose lengths are, respectively, | X | and | Y | , the warping path can be given,
W = w 1 , w 2 , , w k ,   where   max ( | X | , | Y | ) k | X | + | Y |
w k = ( i , j ) ,   where   i   is   the   coordinate   in   X   and   j   is   the   coordinate   in   Y
The warping path W must start from w 1 = ( 1 , 1 ) and end with w k = ( | X | , | Y | ) to ensure every coordinate in | X | and | Y | would be considered. Besides, the coordinate i and j in w ( i , j ) of the warping path must be monotonically increasing, which can be constrained by the following equation
w k = ( i , j ) , w k + 1 = ( i , j ) , i i i + 1 , j j j + 1
The warping path needs to satisfy continuity and monotonicity. According to this, Equation (8) represents the three directions that are in the next optimal path searching. This process can be shown in Figure 3. DTW always searches forward for the closest distance in three directions: Horizontal, vertical, and diagonal upward, to maintain the continuity and monotonicity.
For the failure time of the escalator, such a warping path in DTW is physically meaningful. No new values will be added to the original data after DTW with the reference time series.
Finally, the wanted warping path is the shortest path in all possible warping paths
D ( i , j ) = D i s t ( i , j ) + min [ D ( i 1 , j ) , D ( i , j 1 ) , D ( i 1 , j 1 ) ]
DTW is a typical optimization problem. It uses a time warping function that satisfies certain conditions to describe the time correspondence between the test template and the reference template, and solves the warping function corresponding to the minimum cumulative distance when the two templates match. In addition, the constraint here is the search direction shown in Figure 3 [26]. DTW always searches forward for its nearest distance from three directions: Horizontal direction, vertical direction, and diagonal upward direction. These three directions are consistent with the basic law of the failure time series: (i) The horizontal direction indicates the same failure time; (ii) the vertical direction indicates sudden failure; (iii) the diagonal line indicates the stable development direction of the failure time. It is this constraint that allows the warping time series to learn the inherent failure development information of ‘good data’ while maintaining the original basic physical meaning of itself. Meanwhile, for short time series prediction, DTW can increase the sample capacity for its oversampling characteristics, especially to increase the sample capacity where the data are abrupt, so that the modeling of the disturbance in the neural network can obtain more attention to reduce the influence of disturbance to the whole data model.

2.4. Convolutional LSTM Neural Network

The significant characteristic of the failure time series in this paper is a few samples with non-uniform sampling. When directly using LSTM for data modeling, the features in time series are not obvious, which may lead to poor accuracy of the modeling and prediction. The deep neural network used in this paper is a combination of CNN and LSTM. The main target of this paper is time series prediction. With the help of the CNN’s feature extraction ability, high-dimensional features can be trained to understand the failure time series more comprehensively. Then, these high-dimensional features can be used for time series prediction in LSTM. Based on the advantages of CNN and LSTM, a convolutional LSTM neural network (CLNN) is proposed to achieve the goal of failure time prediction for an escalator.

2.4.1. Convolutional Neural Network

Convolutional neural networks are a type of deep neural network with the ability to act as feature extractors. Such models are able to learn multiple layers of feature hierarchies automatically (also called ‘representation learning’). A typical CNN network structure is shown in Figure 4. This typical CNN includes an input layer, a convolution layer (Conv), a maximum pooling layer (MP), a fully connected layer (FC), and an output layer.
The structure of the CNN network mainly has two characteristics: Sparse connection and shared weights. As shown in Figure 5, unlikely a fully connected neural network, CNNs use the locally connected mode. The neurons in the m layer are connected to the adjacent neurons in the m-1 layer. The weights of every neuron in the m-1 layer are shared to the neurons in the m layer, which means the weights with the same color in Figure 4 are the same. Due to these advantages, the CNN has the effect of regularization, which can improve the stability and generalization ability of neural networks and avoid over-fitting. At the same time, these two advantages reduce the total amount of weight parameters, which is beneficial to the rapid learning of neural networks. CNNs have good representation learning ability, which can learn excellent features through network training. It is very suitable for a learning process lacking prior knowledge or clear features.
In this paper, the failure time in the failure recall data is a typical time series, so a one-dimensional (1D) convolution operation is used to extract the features of failure time sequences.

2.4.2. Long Short-Term Memory Neural Network

After data sequences have been preprocessed, a long short-term memory neural network (LSTM NN) is applied in this paper for data modeling. To overcome the gradient vanish of traditional RNNs, a LSTM NN was proposed. LSTM is one of the special structures of RNNs that can solve the memory problem of neural networks. It can be applied to process and predict important events with long intervals and delays in time series.
With a special gate and cell in the hidden layer, LSTM can effectively update and deliver critical information in time series. Compared to traditional RNNs, LSTM has stronger capacities of information selection and time series learning, which can solve the problem of long-term dependencies by using remote context information for current prediction tasks. An LSTM is composed of one input layer, one hidden layer, and one output layer. Unlike the traditional NN, the basic unit of the hidden layer is a memory block, and LSTM adds a ‘processor’ in the algorithm to judge whether the information is useful or not, which is called a cell [27].
The typical structure of an LSTM cell is shown in Figure 6. An LSTM cell is configured mainly by three gates: Input gate, forget gate, and output gate. The basic idea of LSTM is that when information enters the network, the information that meets the requirements is left by the input gate, the non-compliant information is discarded by the forget gate, and finally new information is generated by the output gate.
Given x ( t ) is the input time series, the output is denoted as h ( t ) and the memory of the cell is denoted as s ( t ) . Three gates of the LSTM NN can be given as follows:
(1) Forget gate
This unit is decided by considering the current input, the previous output, and the previous memory. It will produce a new output and change its memory. σ determines how much information can be left.
f ( t ) = σ ( W f [ h ( t 1 ) , x ( t ) ] + b f )
(2) Input gate
The input of this unit is the same with the forget gate, but the input gate determines the extent to which new memories should affect old memories. Meanwhile, this unit determines how much new information should be delivered to the next cell. Finally, the cell state is updated through discarding the information that needs to be discarded and adding new information.
i ( t ) = σ ( W i [ h ( t 1 ) , x ( t ) ] + b i )
g ( t ) = tanh ( W C [ h ( t 1 ) , x ( t ) ] + b C )
s ( t ) = f ( t ) s ( t 1 ) + i ( t ) g ( t )
(3) Output gate
Based on the new cell state, some information of the new state should be added to the output and the output information of the cell will finally be determined by deterministic information that needs to be output.
o ( t ) = σ ( W o [ h ( t 1 ) , x ( t ) ] + b o )
h ( t ) = o ( t ) tanh ( s ( t ) )
where σ denotes the sigmoid function defined in the equation and the tanh function is defined in the equation:
σ ( x ) = 1 1 + e x
tanh ( x ) = sinh ( x ) cosh ( x ) = e x e x e x + e x
The learning process of LSTM mainly includes the error backpropagation process and optimization algorithm. The backpropagation through time (BPTT) algorithm is applied in the error backpropagation process of LSTM [28].

2.4.3. CLNN: Combination of CNN and LSTM

The architecture of an CLNN is shown in Figure 7. In addition, the network structure of an CLNN consists of a convolutional layer, a pooling layer, LSTM units, and two fully connected layers.

3. Application for Failure-Recall Data with DCLNN Method

3.1. Description of Failure-Recall Data

The current maintenance activities of an escalator are mainly based on two modes: Periodical maintenance and failure-recall maintenance. The maintenance of different components in periodical maintenance is carried out in different cycles (for example, the four types of maintenance items for the escalator with a cycle of 4, 12, 24, and 52 weeks). The failure-recall maintenance aims at the problems discovered by the operator to perform inspection and repair. The failure-recall record can directly reflect the operational safety of the escalator to a certain extent. The failure-recall records of six escalators during 2012.10–2018.10 are investigated. The statistics of failure-recall reasons is shown in Figure 8. There are five main situations in which the failure is recorded, and the cause of component failure accounts for the most.
The failure of the escalator is related to the failure of the components, the time of use, and the inherent performance degradation of the system. The collection and analysis of relevant data can provide scientific predictions for escalator failure. In addition, predicting the possible time of escalator failure and implementing maintenance in advance can reduce the chance of escalator failure and reduce casualties. The failure-recall record can reflect the failure occurrence rule and safety condition of an escalator. However, data in the failure-record have the characteristics of short samples, non-uniform samplings, and random interference, which brings difficulty in predicting the failure time of an escalator.
In this paper, failure-recall records of six escalators with the same brand are collected. In the records, some details about every failure time are recorded and shown in Appendix A. Then, the failure time is processed into the corresponding failure interval days, and the failure time curves of the escalators are calculated by the accumulation idea. The processed failure time curve is shown in Figure 9.

3.2. Dynamic Time Warping Pre-Processing Before Data Modeling

In this part, failure time datasets of different escalators are distinguished by the abovementioned indicators to find those excellent data sequences with inherent advantages (high stationarity and low complexity) before data modeling. In addition, other data sequences are matched similarly with the selected excellent data using DTW. The above process can be seen as a pre-processing for more accurate data modeling. Table 1 summarizes the stationary indicators and complexity indicators of the original failure time series.
According to the comprehensive judgment from stationarity and complexity, the non-stationarity and complexity of escalator #3 is the smallest. The CSI of Escalator #3 is almost fifty times smaller than those of the others. In order to achieve the purpose of accurate prediction, it is necessary to reduce the influence of the other data disturbances during data modeling. The over-sampling characteristics of DTW can increase the number of time series while weakening the disturbances. Thus, DTW is used to warp other failure time curves with the data of the third escalator as a reference. The specific warping path and warped time series can be found in Figure 10.
The white lines in the left of Figure 10 are the warping path and the grayscale maps show the warping distance between the two time series. Figures on the right show the original time series and warped time series. It is easy to see that the new data after DTW have the same physical meaning as original data for the special warping path. Here, a specific analysis of the warped failure time series of Escalator #1 is given. As can be seen from Figure 10, whether it is the red reference sequence #3 or the blue warped sequence #1, no new failure time is generated on the y-axis, but only extended on the x-axis, where the x-axis represents the order or number of failures. Although some new samples were increased on the x-axis, the occurrence time of failure is the same as before. It can still be regarded as the same failure. This is unambiguous in objective meaning. Meanwhile, it can be seen that due to the oversampling characteristics of DTW, the length of the time series after being warped has been increased. Although the increased samples do not affect the physical meaning of the failure time series, it is possible to increase training samples and increase the network’s understanding of the failure time series, which can further improve the training effect and prediction effect of the network.
In order to show the function of DTW, above-mentioned stationary indicators and complexity indicators are calculated after DTW with escalator #3 as referenced. Table 2 shows the result after DTW pre-processing.
As shown in Table 2 and Table 3, the CSI is reduced to below 0.1 after DTW with the failure time series as the referenced sequence. A specific comparison of CSI before and after DTW preprocessing is summarized in Figure 9 for a clear description. As shown in Figure 11, the CSI of the other five escalators are reduced after DTW preprocessing with Escalator #3 referenced. They become more stationary and lower-complexity after matching with the failure time series of Escalator #3.
Besides, for the convenience of engineering applications, six sigma was applied to the analysis of the CSI of the failure time series after DTW to try to give a possible threshold of the CSI. In addition, the threshold can be used to distinguish whether the data sequences need to be warped before data modeling. The distribution interval of CSI can be given by [ μ 3 σ , μ + 3 σ ] , where μ is the mean value of CSI and σ is the variance. For above-mentioned data sequences after DTW, CSI mean = 0.0204 ,   σ = 0.0064 . Thus, the CSI value can be considered in the interval [0.0012 0.0396], and the upper limit of the interval is used as the judgment threshold considering the fault tolerance. That is, CSI t h r e s h o l d = 0.0396 . For data sequences in this paper, if the CSI is less than CSIthreshold, DTW preprocessing is unnecessary before data modeling, otherwise, it is recommended to be warped with a sequence with a CSI less than CSIthreshold.

3.3. Results of LSTM and DCLNN

It is easy to see that no straightforward rules can be found directly from the failure time curve in Figure 10. The LSTM is used directly to train and predict the data in Figure 10. During the training process, the prediction goal is the failure time of the last day. Among them, about 90% of the time series data are used to train the LSTM NN, and the last 10% of the data are used to test the model and predict the failure time. The specific architecture and hyperparameters of CLNN can be found in Figure 12. Tensorflow was used to build the above neural networks. A related repository was stored in the link (the same name with paper) on github.
Taking into account the time cost and prediction accuracy of network training, the recommended training epochs is around 200. Before data training and testing using a neural network, data normalization using Equation (18) is necessary.
x n ( i ) = x ( i ) x min x max x min
Then, the original failure time series of six escalators are first trained by the CLNN. The training process and prediction result are shown in Figure 13. Training loss and testing loss are also given in Figure 14.
Black lines are the true value of the failure time curve, green lines indicate the training process, and the red point shows the prediction result. All escalators except the third one have poor predictions that can be seen from Figure 13 directly. The results are consistent with the result of CSI. The reason for the poor prediction of other escalators are: 1) The length of the data sequence is short, and the number of data used for CLNN model training and testing are also small; 2) the trend of the third escalator is stable with no obvious disturbance, while there are more or less disturbances in the time series data of the other five escalators. In addition, those disturbances make a big difference on modeling accuracy of the CLNN. Then, DTW is used to warp other failure time curves with the data of the third escalator as a reference. Using new data for training and prediction of the DCLNN, the results are shown in Figure 15. Training loss and testing loss are also given in Figure 16.
The prediction accuracy after pre-processing is significantly improved compared to the direct use of the original data. Here, the root-mean-square error (RMSE) is chosen to assess the prediction accuracy of escalators. The RMSE can measure the difference between two time series. The mathematical formula is:
R M S E = 1 n i = 0 n 1 ( y y ) 2
The comparison results are summarized in Table 3 and Table 4. Table 3 shows the RMSE of the training process between the original data and warping data. Table 4 shows the RMSE of the prediction process between the original data and warping data. Although the RMSE in the training process changes a little, the RMSE has been reduced a lot in the prediction process. After pre-processing by DTW, the accuracy of the DCLNN is improved.
With DCLNN, the RMSE of the escalator can be reduced by an average of 83.96%. From Figure 15, the accuracy of the prediction result can be reduced from a few hundred days to less than one month, in order to compare the accuracy of the proposed DCLNN model for failure time prediction of escalators. For a better explanation about the effectiveness of the DCLNN, the other three classical time series prediction methods (Kalman Filter, Nonlinear Auto Regressive Neural Network, Elman RNN, LSTM) are applied for data analysis of the third escalator. The results of these methods are shown in Figure 17. The results intuitively show huge errors from the true value compared to results from the DCLNN.

3.4. Function of DTW

The use of DTW to warping time series can improve the prediction accuracy and reduce the mean square error. If the prediction effect of two time series is poor when using the neural network for the first time, the result is not obviously improved after using DTW. However, the result will have a significant improvement after DTW preprocessing if one of the time series is relatively good for prediction. Table 5 concludes all RMSE values when using different time series as referenced data for DTW pre-processing and warping data for prediction using DCLNN.
Figure 18a is obtained from each column in Table 5. It is clear to see that the MSE is smallest when choosing the third escalator as the reference time series in DTW preprocessing. Figure 18b shows the RMSE improvement of DCLNN prediction after using different escalator data as the reference time series in DTW pre-processing. Similarly, the RMSE improvement of the third escalator is best compared to others. Meanwhile, the result in Figure 18b shows that almost all MSE of the data sequences with different data sequences as referenced can be improved after DTW preprocessing. However, for the third escalator, the RMSE variation is negative. The main reason is that the data quality of the third escalator is best according to the CSI indicator. After DTW is preprocessed with other data as referenced data, the data quality is not as excellent as the third data itself. Therefore, for data with high quality, it is not recommended to use DTW for data preprocessing, otherwise it will reduce the data quality and increase the prediction error.

4. Conclusions

Failure time, as an important factor for the safe operation of escalators, is used for prediction by the proposed DCLNN neural network in this paper. Due to the short length of the failure time series, and the characteristics of random interference and non-uniform sampling, the failure time prediction is much more difficult. Naturally, higher data quality and a suitable neural network can help solve this problem. Considering the oversampling characteristics and similarity-matching performance of DTW, some inferior data with low stationarity and high complexity can be warped by referring to excellent data while retaining its original physical meaning. Such DTW-based data pre-processing can not only increase the sample numbers of short time series but also reduce the impact of disturbance data on the neural network. After data preprocessing, the combination of CNN and LSTM is used to train and predict the failure time series. With the help of CNN’s feature extraction ability, high-dimensional features can be trained to understand the failure time series more comprehensively. Then, these high-dimensional features can be used for time series prediction in LSTM. Based on the advantages of DTW, CNN, and LSTM, a strategy to combine data quality enhancement with deep neural networks is proposed to achieve the goal of failure time prediction for an escalator. Failure-recall data of an escalator are analyzed using the proposed DCLNN method, and the result shows that the proposed method can effectively reduce the root-mean-square error between the predicted value and the real value in the prediction process. At the same time, compared to some classical time series prediction methods such as the Kalman Filter, Nonlinear Auto Regressive Neural Network, and Elman RNN, the prediction accuracy of the proposed method is obviously improved. Besides, the function of the DTW was further analyzed to show that enhancing data quality can effectively improve the accuracy of prediction. The proposed DCLNN method can reduce the prediction time error of escalator failure to less than one month, which can provide scientific guidance for smarter maintenance planning and economic improvement.
However, due to the limited number of samples, the accuracy of the proposed method needs to be improved for better maintenance guiding, and a more reasonable threshold of CSI indicators needs further study before pre-processing in engineering application. Furthermore, the proposed method is currently applicable to the situation of two or more devices to improve the prediction accuracy. In the future, an in-depth study is necessary for the failure time series prediction problem of a single device under a small number of samples.

Author Contributions

Z.Z. proposed the idea of DCLNN, analyzed the failure data, and wrote the paper; Y.Z. proofread the manuscript; J.C. contributed to the conclusion and guided the manuscript; T.A. and J.X. collected the failure data. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant 2017YFC0805701 and National Natural Science Foundation of China (No. 51775411).

Acknowledgments

The authors would like to sincerely thank all the anonymous reviewers for the valuable comments that greatly helped to improve the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The following table is a summary of the failure-record of six escalators in a shopping mall. In this paper, the location and starting date are used for analysis for failure time prediction.
LocationStarting DateStarting TimeArrival DateArrival TimeRecovery DateRecovery TimeMaintenance Time (Min)
1U2N2012/10/2312:21:002012/10/2312:42:002012/10/2314:00:0078
1U2N2013/1/913:57:002013/1/914:20:002013/1/915:30:0070
1U2N2013/3/213:47:002013/3/215:01:002013/3/219:36:00275
1U2N2013/3/148:29:002013/3/148:56:002013/3/149:43:0047
1U2N2013/3/158:30:002013/3/158:50:002013/3/159:30:0040
1U2N2013/5/1717:43:002013/5/1718:27:002013/5/1719:20:0053
1U2N2013/7/2520:00:002013/7/2520:20:002013/7/2521:00:0040
1U2N2014/3/214:10:002014/3/214:30:002014/3/215:50:0080
1U2N2014/6/2116:35:002014/6/2117:15:002014/6/2118:20:0065
1U2N2014/9/2117:40:012014/9/2118:00:012014/9/2122:15:01255
1U2N2014/12/1116:22:002014/12/1116:37:002014/12/1117:54:0077
1U2N2014/12/228:36:002014/12/2214:00:002014/12/2217:20:00200
1U2N2015/1/78:46:002015/1/79:06:002015/1/79:30:0024
1U2N2015/1/238:34:002015/1/2311:40:002015/1/2311:50:0010
1U2N2015/10/1917:00:002015/10/1917:25:002015/10/1920:00:00155
1U2N2015/12/319:30:002015/12/319:46:002015/12/3111:10:0084
1U2N2016/5/109:34:002016/5/109:50:002016/5/1010:10:0020
1U2N2016/6/618:27:002016/6/618:40:002016/6/619:10:0030
1U2N2018/2/914:23:002018/2/914:41:002018/2/915:10:0029
1U2N2018/10/159:04:002018/10/1510:00:002018/10/1518:02:49483
1U2E2012/9/2813:11:002012/9/2813:21:002012/9/2815:05:00104
1U2E2012/11/515:06:002012/11/515:19:002012/11/517:20:00121
1U2E2012/12/1513:05:002012/12/1513:23:002012/12/1515:00:0097
1U2E2013/3/1020:10:002013/3/1020:25:002013/3/1022:00:0095
1U2E2013/9/3016:40:002013/9/3017:05:002013/9/3023:30:00385
1U2E2013/10/1213:01:002013/10/1213:24:002013/10/1213:50:0026
1U2E2014/11/215:58:012014/11/216:20:012014/11/217:57:0197
1U2E2014/11/1613:40:002014/11/1614:00:002014/11/1615:18:0078
1U2E2014/12/19:24:002014/12/110:00:002014/12/123:50:00830
1U2E2014/12/139:39:002014/12/1310:00:002014/12/1310:30:0030
1U2E2015/5/309:14:002015/5/309:40:002015/5/3010:20:0040
1U2E2015/7/1521:28:002015/7/1522:05:002015/7/1522:20:0015
1U2E2016/6/18:43:002016/6/19:00:002016/6/19:30:0030
1U2E2017/2/2413:34:002017/2/2414:55:002017/2/2416:30:0095
1U2E2017/12/89:23:002017/12/819:42:002017/12/821:50:00128
1U2E2018/9/198:53:002018/9/199:24:002018/9/1911:24:00120
1U2S2012/9/2513:40:002012/9/2514:19:002012/9/272:30:002171
1U2S2012/10/1419:07:002012/10/1419:25:002012/10/1420:30:0065
1U2S2012/11/59:30:002012/11/510:00:002012/11/511:20:0080
1U2S2013/3/2215:30:002013/3/2216:10:002013/3/2220:30:00260
1U2S2013/5/610:00:002013/5/610:30:002013/5/612:30:00120
1U2S2013/5/2714:27:002013/5/2714:40:002013/5/2716:00:0080
1U2S2013/6/279:30:002013/6/2710:00:002013/6/2711:40:00100
1U2S2013/7/816:55:012013/7/817:13:012013/7/818:20:0167
1U2S2013/7/1311:00:002013/7/1311:30:002013/7/1312:45:0075
1U2S2013/8/1716:23:002013/8/1716:44:002013/8/1717:20:0036
1U2S2013/11/2510:00:002013/11/2510:30:002013/11/2511:30:0060
1U2S2014/4/2911:10:002014/4/2911:30:002014/4/2913:00:0090
1U2S2014/6/219:10:002014/6/219:50:002014/6/2111:10:0080
1U2S2014/8/2310:00:002014/8/2310:30:002014/8/2311:30:0060
1U2S2015/6/2611:45:002015/6/2612:20:002015/6/2614:10:00110
1U2S2015/8/291:22:002015/8/291:25:002015/9/24:36:005951
1U2S2016/1/1913:59:002016/1/1914:17:002016/1/1915:50:0093
1U2S2016/5/1314:22:002016/5/1314:45:002016/5/1316:00:0075
1U2S2016/7/2614:07:002016/7/2614:25:002016/7/2614:30:005
1U2S2016/8/2115:01:002016/8/2115:23:002016/8/2115:40:0017
1U2S2017/2/1010:12:002017/2/1010:30:002017/2/1010:50:0020
1U2S2017/2/228:55:002017/2/229:10:002017/2/229:30:0020
1U2S2017/11/229:08:002017/11/2210:02:002017/11/2210:29:4628
1U2S2017/12/819:25:002017/12/819:34:002017/12/819:39:055
1U2S2018/5/1317:01:002018/5/1317:30:002018/5/1318:03:0033
1U2S2018/9/1315:36:002018/9/1316:00:002018/9/1317:00:0060
1U2S2018/10/199:40:002018/10/1910:20:002018/10/1910:32:0012
1U2W2012/8/711:00:002012/8/711:40:002012/8/714:30:00170
1U2W2012/8/1614:10:002012/8/1614:25:002012/8/1620:45:00380
1U2W2012/9/1918:15:002012/9/1918:43:002012/9/1920:30:00107
1U2W2013/1/1821:18:002013/1/1821:39:002013/1/1823:57:00138
1U2W2013/2/2712:05:002013/2/2712:20:002013/2/2713:20:0060
1U2W2013/3/316:03:002013/3/316:33:002013/3/322:20:00347
1U2W2013/3/311:01:002013/3/311:40:002013/3/312:00:0020
1U2W2013/9/618:30:002013/9/618:50:002013/9/622:40:00230
1U2W2013/9/1517:28:002013/9/1517:58:002013/9/1518:30:0032
1U2W2013/10/18:10:002013/10/18:25:002013/10/119:10:00645
1U2W2013/10/1316:40:002013/10/1317:07:002013/10/1319:30:00143
1U2W2014/1/149:33:002014/1/1410:02:002014/1/1410:45:0043
1U2W2014/2/413:33:002014/2/413:59:002014/2/414:30:0031
1U2W2014/10/298:05:002014/10/298:50:002014/10/299:40:0050
1U2W2014/11/178:10:002014/11/178:35:002014/11/1710:00:0085
1U2W2016/12/1122:24:002016/12/1122:50:002016/12/1123:30:0040
1U2W2017/8/239:39:002017/8/2310:00:002017/8/2311:30:0090
2U3N2012/8/3015:10:002012/8/3015:20:002012/8/3018:20:00180
2U3N2012/10/2018:50:002012/10/2019:05:002012/10/2021:10:00125
2U3N2012/11/416:15:002012/11/416:15:002012/11/417:20:0065
2U3N2013/1/1317:45:002013/1/1317:55:002013/1/1319:00:0065
2U3N2013/4/811:30:002013/4/812:00:002013/4/812:40:0040
2U3N2014/5/2815:30:002014/5/2816:00:002014/5/2816:30:0030
2U3N2014/10/258:45:002014/10/259:20:002014/10/2512:10:00170
2U3N2015/1/138:50:012015/1/139:25:012015/1/1311:40:01135
2U3N2015/2/2313:21:002015/2/2313:54:002015/2/2317:00:00186
2U3N2015/10/1215:00:002015/10/1215:23:002015/10/1217:15:00112
2U3N2015/10/3011:50:002015/10/3012:23:002015/10/3014:20:00117
2U3N2015/11/2917:00:012015/11/2917:20:012015/11/2920:18:01178
2U3N2016/10/29:09:002016/10/29:30:002016/10/210:00:0030
2U3N2017/11/119:19:002017/11/119:45:002017/11/1111:30:00105
2U3N2018/1/139:25:002018/1/139:53:002018/1/1310:30:0037
2U3N2018/7/169:15:002018/7/169:38:002018/7/1611:38:00120
2U3S2012/8/320:00:002012/8/320:30:002012/8/322:30:00120
2U3S2012/8/289:20:002012/8/289:40:002012/8/2810:20:0040
2U3S2012/11/29:36:002012/11/210:05:002012/11/211:30:0085
2U3S2013/5/610:00:002013/5/610:30:002013/5/611:30:0060
2U3S2013/12/518:30:012013/12/518:50:012013/12/519:30:0140
2U3S2014/3/319:30:002014/3/3110:00:002014/3/3110:30:0030
2U3S2014/8/2511:00:002014/8/2511:30:002014/8/2512:00:0030
2U3S2014/12/209:08:012014/12/209:17:012014/12/209:28:0111
2U3S2015/7/1318:30:002015/7/1319:12:002015/7/1322:30:00198
2U3S2016/6/219:23:002016/6/219:40:002016/6/2110:10:0030
2U3S2016/7/516:50:002016/7/517:10:002016/7/517:30:0020
2U3S2017/7/210:27:002017/7/210:50:002017/7/211:00:0010
2U3S2018/3/2214:20:002018/3/2220:55:002018/3/2221:16:0021
2U3S2018/5/249:20:002018/5/2420:33:002018/5/2421:08:0035
2U3S2018/10/28:52:002018/10/210:20:002018/10/210:38:0018
2U3S2018/10/1910:35:002018/10/1910:36:002018/10/1911:10:0034

References

  1. Schminke, L.H.; Jeger, V.; Evangelopoulos, D.S.; Zimmerman, H.; Exadaktylos, A.K. Riding the Escalator: How Dangerous is it Really? West. J. Emerg. Med. 2013, 14, 141–145. [Google Scholar] [CrossRef] [Green Version]
  2. Xing, Y.; Dissanayake, S.; Lu, J.; Long, S.; Lou, Y. An analysis of escalator-related injuries in metro stations in China, 2013–2015. Accid. Anal. Prev. 2019, 122, 332–341. [Google Scholar] [CrossRef] [PubMed]
  3. Chi, C.-F.; Chang, T.-C.; Tsou, C.-L. In-depth investigation of escalator riding accidents in heavy capacity MRT stations. Accid. Anal. Prev. 2006, 38, 662–670. [Google Scholar] [CrossRef] [PubMed]
  4. Li, W.; Gong, J.; Yu, P.; Shen, S. Modeling, simulation and analysis of group trampling risks during escalator transfers. Phys. A Stat. Mech. its Appl. 2016, 444, 970–984. [Google Scholar] [CrossRef]
  5. Wang, W.; Li, X.; Pan, Q.-L. Notice of Retraction Risk management based on the escalator overturned accident. In 2013 International Conference on Quality, Reliability, Risk, Maintenance, and Safety Engineering (QR2MSE); IEEE: New York, NY, USA, 2013; Volume 2013, pp. 22–27. [Google Scholar] [CrossRef]
  6. Saleh, J.H.; Marais, K.B.; Favarò, F.M. System safety principles: A multidisciplinary engineering perspective. J. Loss Prev. Process. Ind. 2014, 29, 283–294. [Google Scholar] [CrossRef] [Green Version]
  7. Corman, F.; Kraijema, S.; Godjevac, M.; Lodewijks, G. Optimizing preventive maintenance policy: A data-driven application for a light rail braking system. Proc. Inst. Mech. Eng. Part O J. Risk Reliab. 2017, 231, 534–545. [Google Scholar] [CrossRef]
  8. Tsoukalas, V.; Fragiadakis, N. Prediction of occupational risk in the shipbuilding industry using multivariable linear regression and genetic algorithm analysis. Saf. Sci. 2016, 83, 12–22. [Google Scholar] [CrossRef]
  9. Ebrahim, N.; Gholamhossein, H.; Mehdi, J.; Hossein, F.; Morteza, M. Safety performance evaluation in a steel industry: A short-term time series approach. Saf. Sci. 2018, 110, 285–290. [Google Scholar]
  10. Jiang, H.; Li, X.; Shao, H.; Zhao, K. Intelligent failure diagnosis of rolling bearings using an improved deep recurrent neural network. Meas. Sci. Technol. 2018, 29. [Google Scholar] [CrossRef]
  11. Zhang, C.-Y.; Wei, J.; Jing, H.; Fei, C.-W.; Tang, W. Reliability-Based Low Fatigue Life Analysis of Turbine Blisk with Generalized Regression Extreme Neural Network Method. Materials 2019, 12, 1545. [Google Scholar] [CrossRef] [Green Version]
  12. Liu, X.; Liu, Z.; Liang, Z.; Zhu, S.-P.; Correia, J.A.F.O.; De Jesus, A.M.P. PSO-BP Neural Network-Based Strain Prediction of Wind Turbine Blades. Materials 2019, 12, 1889. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  14. Graves, A. Supervised Sequence Labelling. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin, Germany, 2012. [Google Scholar]
  15. Yang, B.; Sun, S.; Li, J.; Lin, X.; Tian, Y. Traffic flow prediction using LSTM with feature enhancement. Neurocomputing 2019, 332, 320–327. [Google Scholar] [CrossRef]
  16. Guo, L.; Li, N.; Jia, F.; Lei, Y.; Lin, J. A recurrent neural network based health indicator for remaining useful life prediction of bearings. Neurocomputing 2017, 240, 98–109. [Google Scholar] [CrossRef]
  17. Yang, J.; Lee, D.; Kim, J. Accident Diagnosis and Autonomous Control of Safety Functions During the Startup Operation of Nuclear Power Plants Using LSTM. Adv. Intell. Syst. Comput. 2018, 787, 488–499. [Google Scholar] [CrossRef]
  18. Adlen, K.; Abderrezak, M.; Ridha, K.; Mohamed, B. Real-time safety monitoring in the induction motor using deep hierarchic long short-term memory. Int. J. Adv. Manuf. Technol. 2018, 99, 2245–2255. [Google Scholar]
  19. Sainath, T.N.; Vinyals, O.; Senior, A.; Sak, H. Convolutional, Long Short-Term Memory, fully connected Deep Neural Networks. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, QLD, Australia, 19–24 April 2015; pp. 4580–4584. [Google Scholar]
  20. Oh, S.L.; Ng, E.Y.; Tan, R.-S.; Acharya, U.R. Automated diagnosis of arrhythmia using combination of CNN and LSTM techniques with variable length heart beats. Comput. Boil. Med. 2018, 102, 278–287. [Google Scholar] [CrossRef] [PubMed]
  21. Tian, C.; Ma, J.; Zhang, C.; Zhan, P. A Deep Neural Network Model for Short-Term Load Forecast Based on Long Short-Term Memory Network and Convolutional Neural Network. Energies 2018, 11, 3493. [Google Scholar] [CrossRef] [Green Version]
  22. Stainhaouer, G.; Carayannis, G. New parallel implementations for DTW algorithms. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 705–711. [Google Scholar] [CrossRef]
  23. Seokgoo, K.; Nam, H.K.; Joo, H.C. Prediction of remaining useful life by data augmentation technique based on dynamic time warping. Mech. Syst. Signal Process. 2020, 136. [Google Scholar] [CrossRef]
  24. Sharma, S.K.; Phan, H.; Lee, J. An Application Study on Road Surface Monitoring Using DTW Based Image Processing and Ultrasonic Sensors. Appl. Sci. 2020, 10, 4490. [Google Scholar] [CrossRef]
  25. Cui, L.; Li, B.; Ma, J.; Jin, Z. Quantitative trend fault diagnosis of a rolling bearing based on Sparsogram and Lempel-Ziv. Measurement 2018, 128, 410–418. [Google Scholar] [CrossRef]
  26. Salvador, S.; Chan, P.K. Toward accurate dynamic time warping in linear time and space. Intell. Data Anal. 2007, 11, 561–580. [Google Scholar] [CrossRef] [Green Version]
  27. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; pp. 326–366. [Google Scholar]
  28. Bodén, M. A Guide to Recurrent Neural Networks and Back Propagation. Ph.D. Thesis, Halmstad University, Halmstad, Sweden, 10 April 2001. [Google Scholar]
Figure 1. The flowchart of the method proposed in this paper.
Figure 1. The flowchart of the method proposed in this paper.
Applsci 10 05622 g001
Figure 2. Potential searching directions.
Figure 2. Potential searching directions.
Applsci 10 05622 g002
Figure 3. Diagram of dynamic time warping.
Figure 3. Diagram of dynamic time warping.
Applsci 10 05622 g003
Figure 4. Convolutional neural network (CNN) network structure.
Figure 4. Convolutional neural network (CNN) network structure.
Applsci 10 05622 g004
Figure 5. Sparse connection and shared weights in CNN.
Figure 5. Sparse connection and shared weights in CNN.
Applsci 10 05622 g005
Figure 6. A cell of a long short-term memory neural network (LSTM NN).
Figure 6. A cell of a long short-term memory neural network (LSTM NN).
Applsci 10 05622 g006
Figure 7. The architecture of a convolutional LSTM neural network (CLNN).
Figure 7. The architecture of a convolutional LSTM neural network (CLNN).
Applsci 10 05622 g007
Figure 8. The statistics of failure-recall reasons.
Figure 8. The statistics of failure-recall reasons.
Applsci 10 05622 g008
Figure 9. Failure time curve of six escalators.
Figure 9. Failure time curve of six escalators.
Applsci 10 05622 g009
Figure 10. DTW preprocessing results. (a) Warped failure time series of #1. (b) Warped failure time series of #2. (c) Warped failure time series of #4. (d) Warped failure time series of #5. (e) Warped failure time series of #6.
Figure 10. DTW preprocessing results. (a) Warped failure time series of #1. (b) Warped failure time series of #2. (c) Warped failure time series of #4. (d) Warped failure time series of #5. (e) Warped failure time series of #6.
Applsci 10 05622 g010aApplsci 10 05622 g010b
Figure 11. Comparison of comprehensive selection indicator (CSI) before and after DTW pre-processing.
Figure 11. Comparison of comprehensive selection indicator (CSI) before and after DTW pre-processing.
Applsci 10 05622 g011
Figure 12. Specific architecture and hyper-parameters.
Figure 12. Specific architecture and hyper-parameters.
Applsci 10 05622 g012
Figure 13. Training process and prediction result using CLNN directly.
Figure 13. Training process and prediction result using CLNN directly.
Applsci 10 05622 g013aApplsci 10 05622 g013b
Figure 14. Loss of 6 escalators using CLNN for training and prediction. (a) Escalator #1. (b) Escalator #2. (c) Escalator #3. (d) Escalator #4. (e) Escalator #5. (f) Escalator #6.
Figure 14. Loss of 6 escalators using CLNN for training and prediction. (a) Escalator #1. (b) Escalator #2. (c) Escalator #3. (d) Escalator #4. (e) Escalator #5. (f) Escalator #6.
Applsci 10 05622 g014
Figure 15. Training process and prediction using DCLNN.
Figure 15. Training process and prediction using DCLNN.
Applsci 10 05622 g015
Figure 16. Loss of 6 escalators using DCLNN for training and prediction. (a) Escalator #1. (b) Escalator #2. (c) Escalator #3. (d) Escalator #4. (e) Escalator #5. (f) Escalator #6.
Figure 16. Loss of 6 escalators using DCLNN for training and prediction. (a) Escalator #1. (b) Escalator #2. (c) Escalator #3. (d) Escalator #4. (e) Escalator #5. (f) Escalator #6.
Applsci 10 05622 g016
Figure 17. Result comparison. (a) Comparison of different methods. (b) Comparison of RMSE with different methods.
Figure 17. Result comparison. (a) Comparison of different methods. (b) Comparison of RMSE with different methods.
Applsci 10 05622 g017
Figure 18. Statistics of RMSE in prediction process. (a) RMSE of every escalator after DTW. (b) RMSE variation statistics after DTW.
Figure 18. Statistics of RMSE in prediction process. (a) RMSE of every escalator after DTW. (b) RMSE variation statistics after DTW.
Applsci 10 05622 g018
Table 1. Comprehensive selection indicator before dynamic time warping (DTW).
Table 1. Comprehensive selection indicator before dynamic time warping (DTW).
IndicatorE #1E #2E #3E #4E #5E #6
SIstd144.5452134.595680.0760192.6437136.8093111.0329
p-value0.02450.10960.00460.01500.06720.0921
CILempel-ziv0.44720.52090.36160.50000.52090.5209
SampEn0.51081.38630.20760.44181.09860.6931
CSI0.809010.65250.02770.63835.26113.6920
Table 2. Comprehensive selection indicator after DTW.
Table 2. Comprehensive selection indicator after DTW.
IndicatorD-M1D-M2D-M3D-M4D-M5D-M6
SISTD127.73125.3480.79157.41126.12114.32
p-value0.0010.0010.00320.0010.0010.001
CILempel-ziv0.31960.35220.36160.34340.35220.3616
SampEn0.30420.74190.20760.45200.41870.4055
CSI0.01240.03080.01940.02440.01860.0168
Table 3. Root-mean-square error (RMSE) in training process.
Table 3. Root-mean-square error (RMSE) in training process.
Training# Of Escalator#1#2#3#4#5#6
RMSEBefore DTW63.64122.0273.877.69127.7293.1
After DTW119.75117.8373.8150.8123.46107.45
Table 4. RMSE in prediction process.
Table 4. RMSE in prediction process.
Prediction# Of Escalator#1#2#3#4#5#6
RMSEBefore DTW384.87112.783.9397.1225.353.18
After DTW25.1216.813.930.379.37.66
RMSE reduction (%)93.5085.09092.3563.2485.60
Average MSE reduction percentage: 83.96%
Table 5. Statistics of RMSE.
Table 5. Statistics of RMSE.
# 1# 2# 3# 4# 5# 6
RMSE384.87112.783.9397.1225.353.18
#1 referenced384.8718.436.81442.3525.9152.26
#2 referenced225.41112.786.85125.1712.4052.48
#3 referenced25.1216.813.930.379.37.66
#4 referenced117.5632.2610.65397.1211.2343.24
#5 referenced48.4313.785.86241.1725.347.77
#6 referenced119.11117.852.9561.4625.4453.18

Share and Cite

MDPI and ACS Style

Zhou, Z.; Zi, Y.; Xie, J.; Chen, J.; An, T. The Next Failure Time Prediction of Escalators via Deep Neural Network with Dynamic Time Warping Preprocessing. Appl. Sci. 2020, 10, 5622. https://0-doi-org.brum.beds.ac.uk/10.3390/app10165622

AMA Style

Zhou Z, Zi Y, Xie J, Chen J, An T. The Next Failure Time Prediction of Escalators via Deep Neural Network with Dynamic Time Warping Preprocessing. Applied Sciences. 2020; 10(16):5622. https://0-doi-org.brum.beds.ac.uk/10.3390/app10165622

Chicago/Turabian Style

Zhou, Zitong, Yanyang Zi, Jingsong Xie, Jinglong Chen, and Tong An. 2020. "The Next Failure Time Prediction of Escalators via Deep Neural Network with Dynamic Time Warping Preprocessing" Applied Sciences 10, no. 16: 5622. https://0-doi-org.brum.beds.ac.uk/10.3390/app10165622

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop